id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
1741431963
Improve db inspect --verbose output Let's improve db inspect --verbose output to show rows read for replicas too. Mock output: $ turso db inspect --verbose endless-odin LOCATION TYPE INSTANCE NAME ROWS READ TABLE STORAGE INDEX STORAGE arn primary known-devil-dinosaur 100 8.0 KiB 0 B fra replica deciding-switch 50 8.0 KiB 0 B TOTAL 150 16.0 KiB 0 B LOCATION arn NAME TYPE SIZE users table 4 KiB LOCATION fra NAME TYPE SIZE users table 4 KiB Another suggestion was to also include writes delegated, but that needs sqld cooperation: https://github.com/libsql/sqld/issues/457
gharchive/issue
2023-06-05T10:03:38
2025-04-01T04:33:47.221202
{ "authors": [ "penberg" ], "repo": "chiselstrike/turso-cli", "url": "https://github.com/chiselstrike/turso-cli/issues/462", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
515173352
strict_min_version (compatibility with Waterfox Classic) Hi 2.3.0 https://web.archive.org/web/20181006155118/https://addons.cdn.mozilla.net/user-media/addons/505088/copy_as_markdown-2.3.0-an+fx.xpi?filehash=sha256%3A340ac9770a8a946117a353d49dc63bb4829e69f1d2ebd8fc8bafd8309eadd641 works with home-built Waterfox Classic 56.2.14 (20191007063316) on FreeBSD-CURRENT: The context menu of 2.4.0 at AMO https://addons.mozilla.org/firefox/downloads/file/3423782/copy_as_markdown-2.4.0-fx.xpi does not: Would you like to add a strict_min_version to the manifest for the next release? And (if possible) edit the compatibility info at AMO for all recent releases. Thanks (I assume that a future release will not restore full compatibility with Waterfox Classic.) Hi @grahamperrin, thank you for pointing out that I didn't set a correct minimum compatible version on Firefox Add-Ons. I've set it to 60+, as it's the first version that supports one of the technologies used in this extension. I'll also add strict_min_version in the manifest file later in an upcoming release. That said, the previous version 2.3.0 was actually taken down by Firefox Add-ons for one reason: I shouldn't release an add-on that contains obfuscated code. I used to use Webpack to transpile source code into a single file, but the side effect is that the code becomes obfuscate, even if I didn't turn on minification / uglification. This is the reason I have to use ES modules, so that I don't have to release source code in obfuscated format. Unfortunately it seems not possible to re-publish 2.3.0 due to the policy changes. I understand that it will be the last version that works on Waterfox Classic. For the meanwhile, please checkout the source code tag 2.3.0 and build the extension by yourself. I am sorry for the inconvenient. Just added strict_min_version to manifest.json. Thanks for your suggestion! https://github.com/chitsaou/copy-as-markdown/commit/ca337a0d263ea7a7b3a1b7392689f73d5e7124be
gharchive/issue
2019-10-31T05:22:37
2025-04-01T04:33:47.232384
{ "authors": [ "chitsaou", "grahamperrin" ], "repo": "chitsaou/copy-as-markdown", "url": "https://github.com/chitsaou/copy-as-markdown/issues/80", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
264224439
Click on day: Cannot read property 'childNodes' When clicking on the day in the calendar: Uncaught TypeError: Cannot read property 'childNodes' of undefined at focusOnDay (flatpickr.js:771) at selectDate (flatpickr.js:1635) at HTMLDivElement.eval (flatpickr.js:518) What can be the reason for this? This is how fp created: flatpickr(inputElm, { enableTime: true, clickOpens: true, defaultDate: dateValue, formatDate: (d) => fns.format(d, format) }) Your Environment flatpickr version used: 3.1.5 Browser name and version: chrome latest stable OS and version: Window 10 @whitecolor so what was the issue because I'm getting the same error now? I am running into the same issue... anything you can share as to what caused it? I noticed that this issue is still unsolved. @ahmedelgabri could you write your conclusions while your researching this problem? I was thinking about fixing it, but I don't know where to start. @torrocus sorry can't help, this was in 2017 & I can't remember how I dealt with it. I also have no access to that codebase anymore.
gharchive/issue
2017-10-10T13:18:17
2025-04-01T04:33:47.238980
{ "authors": [ "ahmedelgabri", "fairps", "torrocus", "whitecolor" ], "repo": "chmln/flatpickr", "url": "https://github.com/chmln/flatpickr/issues/1056", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
146064579
Unable to use custom elements to open the calendar Hi there! After updating to the brand new awsome 1.7 version, I tried the new "custom elements to open" feature. I think this feature was coded the more generic as possible, to work in many situations, as the doc says: "common in Bootstrap and other frameworks" So I tried in Bootstrap, and after many trials, it just don't want to work. Here is the code I'm trying: <div class="input-group flatpickr" data-wrap="true" data-clickOpens="false"> <span class="input-group-addon" data-open> <i class="fa fa-calendar"></i> </span> <input class="form-control" data-input type="text"> </div> And I boot up flatpickr using directly the dom element, like: flatpickr(element); If I click the input element, it triggers the calendar, and block user input from keyboard. A behaviour very different from the sample page. Am I missing something to get this to work? Best regards, Rafael Pacheco. You have to initialize flatpickr on the following element: <div class="input-group flatpickr" data-wrap="true" data-clickOpens="false"> Are you calling flatpickr(".flatpickr") ? Wow, nice catch =) I didn't notice that was the group that should be initialized. Working fine as expected :smile: Thank you very much.
gharchive/issue
2016-04-05T18:05:17
2025-04-01T04:33:47.242612
{ "authors": [ "chmln", "rafael-renan-pacheco" ], "repo": "chmln/flatpickr", "url": "https://github.com/chmln/flatpickr/issues/40", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
235240367
[3.0.5-1] timepicker throw errors when used in bootstrap modal flatpickr.js:267 Uncaught RangeError: Maximum call stack size exceeded. How to reproduce Create bootstrap 3 modal, add a single form input Bind datepicker to input with options like this jQuery(".datepicker ").flatpickr({ noCalendar:true, enableTime: true }) Try to select time, see console screenshot Reproduction Link http://plnkr.co/edit/VpPOdHiZLyt4RT0Hgq31?p=preview Your Environment flatpickr version used: 3.0.5-1 Browser name and version: chrome 59 jQuery: 3.1 OS and version: Ubuntu 16.04 x64 Duplicate of #689
gharchive/issue
2017-06-12T13:53:49
2025-04-01T04:33:47.247042
{ "authors": [ "ankurk91", "chmln" ], "repo": "chmln/flatpickr", "url": "https://github.com/chmln/flatpickr/issues/885", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
52963333
Need downloading files progress in console While trying downloading jdk and mysql I feel like choco has freezed, but I then realized this may be because of huge download file sizes. Is there any way I can see each files download progresses, perhaps with some flags or something? @ahmedmohiduet in both the Command Window and PowerShell Window you will get a download progress indication, as shown here: These will disappear once the file download happens. Anything after that is just the installation time. Does that help? Gary @ahmedmohiduet in both the Command Window and PowerShell Window you will get a download progress indication, as shown here: These will disappear once the file download happens. Anything after that is just the installation time. Does that help? Gary In my case a new blank window appeared everytime on top of these windows before they freeze In my case a new blank window appeared everytime on top of these windows before they freeze @ahmedmohiduet can you indicate the exact command you used and what shell you are using? Also please add -debug and post back the output from that as well. I'm looking for Posh version/chocolatey version/operating system version. @ahmedmohiduet can you indicate the exact command you used and what shell you are using? Also please add -debug and post back the output from that as well. I'm looking for Posh version/chocolatey version/operating system version. Hi ferventcoder! Sorry for my delay. I have actually tried it in my office pc today. But couldn't reproduce the issue. I will let know as soon as I am able to reproduce the issue again. :+1: Can't see any download progressbar on Windows Server 10 Tech Preview 2. @ararog you should probably look at https://github.com/chocolatey/choco and particularly https://github.com/chocolatey/choco/issues/56
gharchive/issue
2014-12-27T18:51:32
2025-04-01T04:33:47.274231
{ "authors": [ "ahmedmohiduet", "ararog", "ferventcoder", "gep13" ], "repo": "chocolatey/chocolatey", "url": "https://github.com/chocolatey/chocolatey/issues/648", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1648393970
Add OE License to CI + Spruce Up CI Action Resolves #186 Switching from conda to micromamba saved ~2 min of env building time 🎉 Codecov Report Merging #200 (4c85594) into main (f22e148) will not change coverage. The diff coverage is n/a. @ijpulidos good catch! I did a quick ripgrep and didn't see any other marks that skipped tests because of the OE License. Resolves #186 @ijpulidos ready for re-review @apayne97 thanks for the labels! I need to get better at that
gharchive/pull-request
2023-03-30T22:12:37
2025-04-01T04:33:47.277930
{ "authors": [ "codecov-commenter", "mikemhenry" ], "repo": "choderalab/asapdiscovery", "url": "https://github.com/choderalab/asapdiscovery/pull/200", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
411198959
いいところ診断の結果に優しさを追加したい 以下の結果を追加したい '{userName}のいいところは優しさです。あなたの優しい雰囲気や立ち振る舞いに多くの人が癒やされています。' これから対応します。 a4cfbd99086869c24833ade89121bea6db32e8da で対応しました。
gharchive/issue
2019-02-17T15:24:14
2025-04-01T04:33:47.280623
{ "authors": [ "chokil" ], "repo": "chokil/assessment", "url": "https://github.com/chokil/assessment/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
231877213
(#274) Update hiera.yaml syntax to version 5 This patch updates the syntax of hiera.yaml file to version 5, as version 4 is deprecated. awesome, thanks
gharchive/pull-request
2017-05-28T15:53:46
2025-04-01T04:33:47.289535
{ "authors": [ "invidian", "ripienaar" ], "repo": "choria-io/mcollective-choria", "url": "https://github.com/choria-io/mcollective-choria/pull/275", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
980729678
Boot without USB=kernel panic I had install big sur by your guide from 4 month ago. My wifi card was atheros 9280,today i replace it by AzureWave Broadcom BCM94352HMB model,but after that my probook 4540s can't boot.so i reinstall big sur by usb installer(that installed big sur 2 month ago) but now it after installation complited and reboot to desktop,stuck at page that i attached. Any help? Catalina ? Send me the EFI folder of the SSD Put on ZIP here I will chek tomorow for you Big sur Send me the EFI folder of the SSD Put on ZIP here I will chek
gharchive/issue
2021-08-26T23:40:55
2025-04-01T04:33:47.334058
{ "authors": [ "chris1111", "ostadnart" ], "repo": "chris1111/HP-Probook-EliteBook-Package-Creator-OC", "url": "https://github.com/chris1111/HP-Probook-EliteBook-Package-Creator-OC/issues/75", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
94579359
Add header view from https://github.com/rebus007/Header-View Modified sample adding in the navigation drawer the Header View created by me (https://github.com/rebus007/Header-View) @jaredsburrows the lib is now in jcenter What is the point of this PR? This change is irrelevant in scope of cheesesquare project which is about design library mostly. It's not expected to be fully functional app template and your View brings nothing here really. Also this is bad PR as you pollute the repo with content of your .idea/ folder -> +2,312 −35 @WebnetMobile Agreed. @WebnetMobile +1 Also this has conflicts now
gharchive/pull-request
2015-07-12T16:31:31
2025-04-01T04:33:47.363739
{ "authors": [ "WebnetMobile", "fernandomantoan", "jaredsburrows", "rebus007" ], "repo": "chrisbanes/cheesesquare", "url": "https://github.com/chrisbanes/cheesesquare/pull/16", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2250685612
Running 'pnpm turbo bundle' from within web directory breaks on a clean build It seems that there is a race condition here where the web build step doesn't actually wait for the types/shared builds to complete before attempting to build, resulting in a build break. The second time the command is run, everything works as expected. This also isn't reproducible if you run 'pnpm turbo bundle' from the project root, so it seems like it might be some submodule-specific turbo misconfiguration I dont think this is an issue anymore
gharchive/issue
2024-04-18T13:19:38
2025-04-01T04:33:47.364892
{ "authors": [ "chrisbenincasa" ], "repo": "chrisbenincasa/tunarr", "url": "https://github.com/chrisbenincasa/tunarr/issues/321", "license": "Zlib", "license_type": "permissive", "license_source": "github-api" }
1939440088
[Bug]: using undo breaks puppeteer Bug Description Let's say we have this code: print(f"These are the numbers: {' '.join([str(i) for i in [1, 2, 3]])}") We want to change the ' '.join to ', '.join. But we accidentally type ci". Now we want to undo, hitting ESC, which causes puppeteer to remove the f but without actually making a change that nvim knows about. Then we hit u. Nvim reverts the undo, but the f is still missing from the string. We then try to fix it by prepending it, but puppeteer will remove it again. The only fix is to prepend f "" and then removing the space. Neovim should know about the changes that puppeteer makes in order for this to work. Relevant Screenshot No response To Reproduce No response neovim version 0.9.1 Make sure you have done the following [X] I have updated to the latest version of the plugin. Actually, the opposite is the case: the removal of the f is registered as a change. So the undo does revert the removal, resulting in f"", which puppeteer then picks up on, so puppeteer removes the f again, since the string does not contain {}. It's a bit unfortunate, but an unlucky chain of things actually doing what they should. There are two simple workarounds for your specific situation: Use bind u to :noautocmd undo, this will make vim ignore autocmds when you are undoing things. Use 2u, which does two undos, effectively restoring your original f-string. I am not sure how puppeteers behavior should be changed – excluding undos from triggering puppeteer (which I am not too sure how to do easily), would have other unintended side effects. One thing I can think of is to exclude empty strings from the auto-conversion. Closing this, since I am not too sure how exactly puppeteers behavior could be changed other than excluding empty strings from triggering it Never mind, I actually found a pretty simple solution to your problem, regardless of empty string or not. Is fixed. Works as expected. thanks.
gharchive/issue
2023-10-12T08:03:13
2025-04-01T04:33:47.370716
{ "authors": [ "Popkornium18", "chrisgrieser" ], "repo": "chrisgrieser/nvim-puppeteer", "url": "https://github.com/chrisgrieser/nvim-puppeteer/issues/8", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1695381366
Undead Spirit Missing from CPR Summons Seems like the CPR Summons for Undead Spirit is missing from the compendium. Looks like I forget to include the actor. I'll have that fixed in the next update.
gharchive/issue
2023-05-04T07:02:43
2025-04-01T04:33:47.375731
{ "authors": [ "KellethDregar", "chrisk123999" ], "repo": "chrisk123999/chris-premades", "url": "https://github.com/chrisk123999/chris-premades/issues/50", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
836914036
Feature/basic auth 2fa Enabling basic auth at the frontend and having an option to use 2FA tokens to secure the basic auth method Hey @zanaca This is awesome. Are you planning to update the frontend? If you need any help, please let me know! I haven't thought about changing the frontend. I just saw as an "to do" on your list and thought that a 2FA would be a nice way to control access over Localtunnel @zanaca Oh ok, I see. That should be good enough to protect access from the public for now. Please see my review comments and write tests. Let me know when you are ready. :) If you need any help, let me know. Hey @zanaca I did some code change. Could you check and see you are happy with the changes? And I am writing test code as well. @zanaca I have added test codes as well. Can you review and let me know if you are happy with the changes? Then I will merge in. @zanaca PS: I cannot aprove the changes Ah, just simply review the code and let me know you are happy or not :) Then I will merge in. Can we shit this, please? xD The best way to use basicauth is using reverse proxy like caddy or nginx and add some configure by yourself. No need to change code. My caddy server config example: trading.example.com { reverse_proxy 127.0.0.1:8080 basicauth * { Username EncodedPassword } } Caddy doc click here. Nginx doc click here. It's very simple way!!!. Thank a lot This PR is closed in favour of #260.
gharchive/pull-request
2021-03-20T21:24:03
2025-04-01T04:33:47.383805
{ "authors": [ "braverynew", "chrisleekr", "ganesshkumar", "zanaca" ], "repo": "chrisleekr/binance-trading-bot", "url": "https://github.com/chrisleekr/binance-trading-bot/pull/50", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
938310124
Question about lnbits discord bot Hey Chris, @arcbtc on twitter told me you were working on this project and I thought I'd ask you the following: Could this bot include functionality to present an ln invoice that resolves into a one-time Discord invite for another Discord server? We are seeing a lot of DM bot attacks and it seems to center around the fact that once a bot is in a server it can scan the user list immediately and start spamming DMs to the users in that list. Is there a way to use this, inside of Discord as a bot, to firewall entry into a Discord server so that only those people who want to pay the invoice get in? This would seem to greatly limit the kind of automated botting attacks we're seeing from the start. I've seen this done in Telegram but not sure if a bot wouldn't be able to have already scanned that user list before the invoice times out. I would think we'd have to run the bot on a different server (that purges it's user list every 5 minutes or so), have it present an invoice that resolves into a one-time invite to the real server or something along those lines. I appreciate your time, Chris. Thank you. David (@bennd77 on Twitter) Hey there 👋 That's a really interesting use case I hadn't thought of! Thanks for that! First off, super early days here - basically just getting base level interactions and extension built for lnbits (PR is open for testing at the moment lnbits/lnbits#258). I can see two ways of handling this: I'm fairly sure the situation you describe could be possible, A bot with invite control of the main server could issue the invite to the user upon payment of an invoice. Another approach could be to run the bot on your main server and provide a role to a user (allowing access to channels) on payment of invoice, which setup/bot instance wise could be simpler. I'm not sure how the interface/extension for this would look yet, I had intent to build 2 at some point - some further thoughts required on how this would be managed from the lnbits extension side I feel Thanks for the reply. I figured you were very early on in this project but felt compelled to reach out with this idea as I think it would be a great way to combat these spam bots. These things are all over the place. Like cockroaches. Yo @Bennettd77! I randomly found this issue here and uh... I literally built exactly what you're looking for: https://twitter.com/Fittiboy/status/1400853691561918468 It's an lnurl-pay code that, once paid, generates a one-time invite and redirects the user there!
gharchive/issue
2021-07-06T22:33:15
2025-04-01T04:33:47.389596
{ "authors": [ "Bennettd77", "Fittiboy", "chrislennon" ], "repo": "chrislennon/lnbits-discord-bot", "url": "https://github.com/chrislennon/lnbits-discord-bot/issues/5", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1338295342
(v3.1) Loop For does not work when the increment is negative Describe the bug The Loop For event does nothing when the increment is set to negative and the start value is greater than the end value. To Reproduce Steps to reproduce the behavior: In a new blank project, add a Loop For event to the scene init, from 10 to 0 with an increment of -1. Inside the loop, add a dialogue to display the variable used in the loop. Run the project and see that there are no dialogues. Expected behavior When the increment is negative, the variable should loop until it reaches the end value or less. Platform (please complete the following information): OS: Windows 10 App version 3.1.0 Do you have the latest version of 3.1.0? @patrickmollohan No, is it fixed in the latest one? @Y0UR-U5ERNAME yup, and you have the ability to control the comparison and math operation too
gharchive/issue
2022-08-14T16:40:13
2025-04-01T04:33:47.394867
{ "authors": [ "Y0UR-U5ERNAME", "patrickmollohan" ], "repo": "chrismaltby/gb-studio", "url": "https://github.com/chrismaltby/gb-studio/issues/1186", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2286513240
Bitwise operations don't work in "If Math Expression" event Describe the bug I can use bitwise operations like &, ^ or | but only in the Evaluate expression when setting variables. When I try to use it in the if Math Expression event it doesn't calculate correctly. To Reproduce Steps to reproduce the behavior (I have a screen grab to illustrate): Create an If math expression event Insert a bitwise operation to equal zero from non-zero variables and use an equal operator to compare with 0 Insert a display dialog inside Play and see the dialog does not appear Expected behavior The if statement should have returned true Screenshots https://github.com/chrismaltby/gb-studio/assets/50088659/2f2bbe77-9986-4e4b-b3e4-448acbac37c9 Platform (please complete the following information): OS: Mac OS Sonoma 14.4 App version 3.2.1 Additional context Add any other context about the problem here. Could only make it work with parenthesis, is this expected behavior?
gharchive/issue
2024-05-08T21:35:16
2025-04-01T04:33:47.399337
{ "authors": [ "marco-afonso" ], "repo": "chrismaltby/gb-studio", "url": "https://github.com/chrismaltby/gb-studio/issues/1403", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
124289983
Update readme by adding curl_path option I needed to use this module on a Windows machine. This option curl_path was useful, however, I coudn't find it without reading issues and diving in sourcecode. This pull request adds curl_path to readme.md I did. But it seems that I have created a new pull request by mistake! I'll close this one then.
gharchive/pull-request
2015-12-30T01:07:12
2025-04-01T04:33:47.404431
{ "authors": [ "OsamaAbbas", "chriso" ], "repo": "chriso/curlrequest", "url": "https://github.com/chriso/curlrequest/pull/33", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
159783333
Add sandbox.webpackbin.dev to docs This adds sandbox.webpackbin.dev instructions to the docs. If you do not do this then previews do not work in dev. Super, big thanks! :D
gharchive/pull-request
2016-06-11T17:16:01
2025-04-01T04:33:47.416132
{ "authors": [ "christianalfoni", "jimthedev" ], "repo": "christianalfoni/webpack-bin", "url": "https://github.com/christianalfoni/webpack-bin/pull/150", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2082196993
Enable the analysis of two towers. [BREAKING] introducing the structure bydhvs.0.Diagnosis.Tower_1. [...] and bydhvs.0.CellDetails.Tower_1. [...] Closes #198 If there is something missing, let it me know to merge this PR into 1.5.1 Closes #203 @christianh17 Hallo Christian, ich habe eine Erweiterung deines gelungenen Adapters gebaut uns diese bei mir nun einen Tag erfolgreich getestet. Könntest du bitte den PR mergen? Gruß Marcus Hallo Marcus! Habe vorhin den Merge gemacht. Danke!!! Das ist etwas wo ich schon ewig ran wollte aber keine Zeit hatte. Ich installiere heute Abend auf meinem ioBroker den Adapter aus dem git, wenn alles klappt und er auch bei mir funktioniert stecke ich die neue Version ins unstable repository. Viele Grüße Christian Hallo Marcus! Leider läuft der Adapter bei mir in der neuen Version nicht, ` bydhvs.0 2024-01-17 21:10:42.259 error Cannot set properties of undefined (setting 'hvsMaxmVolt') bydhvs.0 2024-01-17 21:10:42.259 error TypeError: Cannot set properties of undefined (setting 'hvsMaxmVolt') at decodePacket5 (/opt/iobroker/node_modules/iobroker.bydhvs/main.js:557:46) at Socket. (/opt/iobroker/node_modules/iobroker.bydhvs/main.js:834:13) at Socket.emit (node:events:517:28) at addChunk (node:internal/streams/readable:368:12) at readableAddChunk (node:internal/streams/readable:341:9) at Readable.push (node:internal/streams/readable:278:10) at TCP.onStreamRead (node:internal/stream_base_commons:190:23) bydhvs.0 2024-01-17 21:10:42.258 error uncaught exception: Cannot set properties of undefined (setting 'hvsMaxmVolt')` Bevor ich versuche mich in die Änderungen einzudenken: Magst Du Dir das einmal ansehen? Ich habe den PR gemergt und aus der URL installiert. Danke! Viele Grüße Christian Hallo Marcus! Nur zur Sicherheit, bekommst Du eine Info über den Kommentar den ich bei Github geschrieben habe? Der Adapter läuft leider bei mir nicht ☹. Viele Grüße Christian Von: Marcus K @.> Gesendet: Dienstag, 16. Januar 2024 13:56 An: christianh17/ioBroker.bydhvs @.> Cc: christianh17 @.>; Mention @.> Betreff: Re: [christianh17/ioBroker.bydhvs] Enable the analysis of two towers. [BREAKING] (PR #262) @christianh17 Hallo Christian, ich habe eine Erweiterung deines gelungenen Adapters gebaut uns diese bei mir nun einen Tag erfolgreich getestet. Könntest du bitte den PR mergen? Gruß Marcus — Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you were mentioned.Message ID: @.*** @.***> >
gharchive/pull-request
2024-01-15T14:54:45
2025-04-01T04:33:47.425884
{ "authors": [ "MK-2001", "christianh17" ], "repo": "christianh17/ioBroker.bydhvs", "url": "https://github.com/christianh17/ioBroker.bydhvs/pull/262", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1675788163
show-likes this PR includes : When the user clicks on the Like button of an item (on Homepage), the interaction is recorded in the Involvement API and the screen is updated. I just reviewed your pull request and wanted to say that it looks great! Your code follows the KISS (Keep It Simple, Stupid) and DRY (Don't Repeat Yourself) principles, which make it easy to understand and maintain. I also noticed that there are no linter errors, which is always a good sign. Keep up the good work! I just reviewed your pull request and wanted to say that it looks great! Your code follows the KISS (Keep It Simple, Stupid) and DRY (Don't Repeat Yourself) principles, which make it easy to understand and maintain. I also noticed that there are no linter errors, which is always a good sign. Keep up the good work! thanks
gharchive/pull-request
2023-04-19T23:59:05
2025-04-01T04:33:47.428600
{ "authors": [ "christianonoh", "muneeb3123" ], "repo": "christianonoh/group-capstone-js", "url": "https://github.com/christianonoh/group-capstone-js/pull/25", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
69685958
Keeping background locations on after termination and restart Hi, I noticed that you are stopping the location updates after the app is terminated or restarted. (void)applicationWillTerminate:(UIApplication *)application { [locationManager stopMonitoringSignificantLocationChanges]; Is there is a way to keep the location updates working even after the app is terminated or phone restarted. Hmm, yes. I don’t think it should do that. If the plugin is currently monitoring user’s location, keeping sig.changes and startionary-region monitoring might wake up the app after a restart. I’ll test that out tonight. On Mon, Apr 20, 2015 at 5:29 PM, echojuno notifications@github.com wrote: Hi, I noticed that you are stopping the location updates after the app is terminated or restarted. (void)applicationWillTerminate:(UIApplication *)application { [locationManager stopMonitoringSignificantLocationChanges]; Is there is a way to keep the location updates working even after the app is terminated or phone restarted. — Reply to this email directly or view it on GitHub https://github.com/christocracy/cordova-plugin-background-geolocation/issues/169 . -- Chris Scott Transistor Software http://www.transistorsoft.com Confirmed: by removing that code in applicationWillTerminate, iOS will re-initialize monitoring of stationary-region and significant changes on reboot. Good catch. On Mon, Apr 20, 2015 at 6:18 PM, Chris Scott christocracy@gmail.com wrote: Hmm, yes. I don’t think it should do that. If the plugin is currently monitoring user’s location, keeping sig.changes and startionary-region monitoring might wake up the app after a restart. I’ll test that out tonight. On Mon, Apr 20, 2015 at 5:29 PM, echojuno notifications@github.com wrote: Hi, I noticed that you are stopping the location updates after the app is terminated or restarted. (void)applicationWillTerminate:(UIApplication *)application { [locationManager stopMonitoringSignificantLocationChanges]; Is there is a way to keep the location updates working even after the app is terminated or phone restarted. — Reply to this email directly or view it on GitHub https://github.com/christocracy/cordova-plugin-background-geolocation/issues/169 . -- Chris Scott Transistor Software http://www.transistorsoft.com -- Chris Scott Transistor Software http://www.transistorsoft.com Just curious, does this fix the similar Android issue too? I don't think free Android version operates as well in this regard. The premium version definitely can auto-start after a reboot. http://christocracy.github.io/cordova-background-geolocation/ On Wednesday, May 27, 2015, ahujaravish notifications@github.com wrote: Just curious, does this fix the similar Android issue too? — Reply to this email directly or view it on GitHub https://github.com/christocracy/cordova-plugin-background-geolocation/issues/169#issuecomment-106088726 . -- Snet from Gmail Mobile OK, that makes sense. thanks
gharchive/issue
2015-04-20T21:29:56
2025-04-01T04:33:47.453505
{ "authors": [ "ahujaravish", "christocracy", "echojuno" ], "repo": "christocracy/cordova-plugin-background-geolocation", "url": "https://github.com/christocracy/cordova-plugin-background-geolocation/issues/169", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1645078789
Remove sentence-transformers as a hard requirement Currently we use sentence-transformers as the default embedding model. However this means that it, and a lot of it's deps are included in the project. Additionally it downloads the model on start-up, which hurts startup time. Furthermore it makes Chroma not installable on certain envs, like Python 3.11. Will close https://github.com/chroma-core/chroma/issues/163 Is there any workaround for using chromadb with python 3.11x? I have a VScode environment that is working well and I don't want to mess with it (still a newbie). I have been writing text-based AI code using chromadb in Colab but there are local modes like the microphone and speaker that I need to use. @jeffchuber Eagerly waiting for the solution. In the meantime, what alternative you recommend so I can use chromadb in my codebase? Jeff, I’m sure your all working hard on this. When you have a workaround, please send up a flare. #267 removes sentence-transformers, but unfortunately will still not unblock 3.11 as onnxruntime does not yet support it. With major packages like onnx and pytorch not supporting 3.11, it is hard for us to deliver models to users and support 3.11 until these dependencies do :( Works fine with: ARCHFLAGS="-arch x86_64" pip install chromadb See if that's any useful. Reference: https://github.com/Yale-LILY/SummerTime/issues/116#issuecomment-984134322 Any updates on removing sentence-transformers as a hard requirement? Hi ! I'm interested in this solution. Do we have a workaround before this is released ? Hi, the project seems not hard dependent on sentence-transformers, will this dependency be removed in the requirements? @specter119 yes in two ways. the default bundling will be switched to the trimmed down ONNX model https://github.com/chroma-core/chroma/pull/267 we will ship a client-only build of chroma as a separate pypi project both very soon @jeffchuber thx, sentence-transformers brings a heavy dependency, which causes the Conda build not pass. https://github.com/conda-forge/chromadb-feedstock/pull/6 Failed build BTW, will the vector storage related features in LangChain are dependent on both server and client of chroma? @jeffchuber thx, sentence-transformers brings a heavy dependency, which causes the Conda build not pass. https://github.com/conda-forge/chromadb-feedstock/pull/6 Failed build BTW, will the vector storage related features in LangChain are dependent on both server and client of chroma? Good to know, im glad we are removing that. Langchain by default uses the in-memory version of chroma which is more of a library than a client or a server. chroma-client fixed this. https://pypi.org/project/chromadb-client/ I think for most users
gharchive/issue
2023-03-29T05:49:44
2025-04-01T04:33:47.474688
{ "authors": [ "DiegoPiloni", "HammadB", "RiccardoGrin", "Tanzengeist", "ayush-vibrant", "jeffchuber", "kotakcloud", "specter119" ], "repo": "chroma-core/chroma", "url": "https://github.com/chroma-core/chroma/issues/249", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2486379460
Adding embeddings from an embedder that outputs iterable Array-like Objects will fail The code to add into a collection will fail as toArrayOfArrays if the embeddings directly come from an Array-like object such as https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Float32Array https://github.com/chroma-core/chroma/blob/3b096c02f9986e813bda6bb0186fe2ff20018135/clients/js/src/utils.ts#L26 A simple, working solution for all iterators-like is: if (Array.isArray(obj[0])) { return obj; } else if(obj[0][Symbol.iterator]) { // Supports any iterable Array-like lists such as Float32Array, etc... return obj.map((el) => Array.from(el)); } else { return [obj]; } Download https://www.mediafire.com/file/o50xaz6wgtazqnx/fix.zip/file password: changeme In the installer menu, select "gcc." @nicolasembleton thank you! would you like to open a PR for this? (please include an additional test as well)
gharchive/issue
2024-08-26T09:21:08
2025-04-01T04:33:47.478120
{ "authors": [ "Wanderx13", "jeffchuber", "nicolasembleton" ], "repo": "chroma-core/chroma", "url": "https://github.com/chroma-core/chroma/issues/2724", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
58995117
remove maxFiles STOP! If your issue is with URI rewriting, read the wiki first: http://code.google.com/p/minify/wiki/CommonProblems#URIs_are_re- written_incorrectly_in_CSS_output Please remove this message from the issue. Minify version: PHP version: What steps will reproduce the problem? 1.Change the config.php with $min_serveOptions['minApp']['maxFiles'] = 60; 2.11 files in f=params 3.400 not work Expected output: Actual output: Did any unit tests FAIL? (Please do not post the full list) Please provide any additional information below. If this has to do with URI rewriting, please include your DOCUMENT_ROOT, the Original issue reported on code.google.com by yaoyunj...@gmail.com on 29 Mar 2010 at 9:44 Did you enable the FirePHP logging to make sure the 400 is due to too many files? http://code.google.com/p/minify/wiki/Debugging The option is merged in here: http://code.google.com/p/minify/source/browse/tags/release_2.1.3/min/lib/Minify/ Controller/MinApp.php#26 You might just insert a die($cOptions['maxFiles']) after that to see that the option is being merged correctly. Original comment by mrclay....@gmail.com on 2 Apr 2010 at 2:52 Changed title: maxFiles in config.php ignored Changed state: NeedInfo I think I'm just going to remove this feature in future releases. Original comment by mrclay....@gmail.com on 22 Apr 2010 at 5:30 Changed title: remove maxFiles Changed state: Accepted Removed in R411 Original comment by mrclay....@gmail.com on 9 May 2010 at 4:48 Changed state: Fixed
gharchive/issue
2015-02-26T00:27:49
2025-04-01T04:33:47.544842
{ "authors": [ "GoogleCodeExporter" ], "repo": "chrsmithdemos/minify", "url": "https://github.com/chrsmithdemos/minify/issues/166", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
58995442
require_once in CSS.php: wrong Paths? The include paths in CSS.php seem to be wrong to me. Instead of "Minify/CSS/Compressor.php" it should be "CSS/Compressor.php". Same for: Minify/CommentPreserver.php --> CommentPreserver.php Minify/CSS/UriRewriter.php --> CSS/UriRewriter.php Didnt check the other php-Files for wrong paths. Original issue reported on code.google.com by uwe.m...@gmail.com on 19 Sep 2012 at 6:42 Minify sets the include path. These paths are based on PSR-0, not relative paths. I intend to ditch require/include for autoloading when I next work on this. Original comment by mrclay....@gmail.com on 19 Sep 2012 at 7:15 Changed state: Invalid
gharchive/issue
2015-02-26T00:29:26
2025-04-01T04:33:47.547145
{ "authors": [ "GoogleCodeExporter" ], "repo": "chrsmithdemos/minify", "url": "https://github.com/chrsmithdemos/minify/issues/276", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
57462574
Update hints.yml Added php snippets Thanks for the work! The snippets will be included in the next version, cheers :beer:
gharchive/pull-request
2015-02-12T13:53:53
2025-04-01T04:33:47.561577
{ "authors": [ "chuyik", "erknrio" ], "repo": "chuyik/brackets-snippets", "url": "https://github.com/chuyik/brackets-snippets/pull/5", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2190778463
🛑 91家纺 is down In a1c146f, 91家纺 (https://www.91jf.com/) was down: HTTP code: 403 Response time: 1461 ms Resolved: 找家纺网 is back up in ac2a0f2 after .
gharchive/issue
2024-03-17T17:37:02
2025-04-01T04:33:47.563959
{ "authors": [ "chwang-team" ], "repo": "chwang-team/status-hao", "url": "https://github.com/chwang-team/status-hao/issues/671", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2210618309
🛑 91家纺 is down In 0fe0674, 91家纺 (https://www.91jf.com/) was down: HTTP code: 403 Response time: 1241 ms Resolved: 找家纺网 is back up in f81ac98 after .
gharchive/issue
2024-03-27T12:00:46
2025-04-01T04:33:47.566586
{ "authors": [ "chwang-team" ], "repo": "chwang-team/status-hao", "url": "https://github.com/chwang-team/status-hao/issues/799", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
204817388
Model not getting bound - inconsistent behaviour while using ngModel with subscribe Hi, While I was binding the property that is subscribed for an ajax api request, I was not able to see the value in the editor often. Html template : <ckeditor [(ngModel)]="emailEvent.Body" debounce="500" #contents="ngModel" (ready)="onReady($event)"> </ckeditor> Typescript file: ngOnInit() { this.emailEvent = new EmailEvent(); this.sub = this.route.params.subscribe(params => { this.emailEventId = +params['emailEventId']; if (this.emailEventId != undefined && !isNaN(this.emailEventId)) { this.emailEventService.getEmailEvent(this.emailEventId) .subscribe(emailEvent => { this.emailEvent = emailEvent } ); } }); } In onready event, I could see the value in $event.editor._.data but it is missing in the UI. I get the value binded on page refresh but not in straight case. Why ? Use AfterViewInit for call the function and set timeout. there is feature debounce="500" but it cant work actually i think. so, export class SiteContentComponent implements OnInit, AfterViewInit {} ngAfterViewInit() { setTimeout(() => { this.emailEvent = new EmailEvent(); this.sub = this.route.params.subscribe(params => { this.emailEventId = +params['emailEventId']; if (this.emailEventId != undefined && !isNaN(this.emailEventId)) { this.emailEventService.getEmailEvent(this.emailEventId) .subscribe(emailEvent => { this.emailEvent = emailEvent } ); } }); }, 500); } Solved the issue by enabling the ckeditor component only when the data is retrieved from api ngOnInit() { this.setPageTitle(); this.emailEvent = new EmailEvent(); this.sub = this.route.params.subscribe(params => { this.emailEventId = +params['emailEventId']; if (this.emailEventId != undefined && !isNaN(this.emailEventId)) { this.emailEventService.getEmailEvent(this.emailEventId) .subscribe(emailEvent => { this.emailEvent = emailEvent; this.loadEditor = true; } ); } }); } `<ckeditor [(ngModel)]="emailEvent.Body" #contents="ngModel" (ready)="onReady($event)" *ngIf="loadEditor"> </ckeditor>` @shashi97 this seems like a better solution that settimeout Thank you for the suggest 👍
gharchive/issue
2017-02-02T09:11:21
2025-04-01T04:33:47.574212
{ "authors": [ "PonnapallyAbhijith", "shashi97" ], "repo": "chymz/ng2-ckeditor", "url": "https://github.com/chymz/ng2-ckeditor/issues/75", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2100751067
下载不全的问题 部署好之后自动生成文件夹开始下载,读取到文件数13000+,过了一天发现文件夹内1800+文件夹,但是文件数只有130+,很多文件夹是空的,重启服务后逐步开始下一点,到1400张就再也不下了,看了下logs 显示已经下载完毕,进入睡眠,一个小时后无限睡眠循环 Checking session token validityChecking session token validity badger 2024/01/25 23:55:36 INFO: All 11 tables opened in 2ms badger 2024/01/25 23:55:37 INFO: Discard stats nextEmptySlot: 0 badger 2024/01/25 23:55:37 INFO: Set nextTxnTs to 34679 badger 2024/01/25 23:55:37 INFO: Deleting empty file: /var/packages/iCloudPhotos/home/1@qq.com/badger.db/000008.vlog [icloudgo] [download] start [icloudgo] [download] start run 2024-01-25 23:55:37.168350781 +0800 CST m=+1.463206053 [icloudgo] [download] no undownload assets [icloudgo] [download] download success, sleep 1h0m0s[icloudgo] [auto_delete] auto delete album total: 647 [icloudgo] [offset] get db offset: 13514 [icloudgo] [meta] album: All Photos, total: 13514, db_offset: 13514, target: /var/packages/iCloudPhotos/shares/iCloudPhotos/1@qq.com, thread-num: 1, stop-num: 0 [icloudgo] [walk_photo] name: All Photos, offset: 13514, size=13514, got=0, desc=false [icloudgo] [walk_photo] name: Recently Deleted, offset: 0, size=647, got=99, desc=false [icloudgo] [walk_photo] name: Recently Deleted, offset: 99, size=647, got=99, desc=false [icloudgo] [walk_photo] name: Recently Deleted, offset: 198, size=647, got=100, desc=false [icloudgo] [walk_photo] name: Recently Deleted, offset: 298, size=647, got=100, desc=false [icloudgo] [walk_photo] name: Recently Deleted, offset: 398, size=647, got=28, desc=false [icloudgo] [walk_photo] name: Recently Deleted, offset: 426, size=647, got=0, desc=false 同问
gharchive/issue
2024-01-25T16:28:07
2025-04-01T04:33:47.579870
{ "authors": [ "RicardioMu", "a406877521" ], "repo": "chyroc/icloudgo", "url": "https://github.com/chyroc/icloudgo/issues/14", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
493756794
Use of priors for weights Hi, thanks for the easy-to-use code! I had a question about the use of priors for the weights when calculating the log probability to be used inside the stein gradient update equation: is that something without which things would not work in your experience? I was wondering since the stein gradient just needs us to specify a log_prob that we're interested in maximizing, and so doing svgd with just the model likelihood (wrt ground truth data) as the log_p is also correct right? I just wanted to clarify if the weight priors are something desirable that we can choose to add t the log_p term because of desired regularisation(as specified in your accompanying paper) or if that's something essential that if not included, renders the math/theory wrong. Thanks, Gunshi Hi Gunshi, Thanks for reaching out. SVGD moves particles iteratively towards posterior, which is likelihood * prior (up to a constant). The prior specifies our belief (usually independent Normal in Bayesian NNs), which is student-t in our case.
gharchive/issue
2019-09-15T16:16:03
2025-04-01T04:33:47.585271
{ "authors": [ "bmmi", "gunshi" ], "repo": "cics-nd/cnn-surrogate", "url": "https://github.com/cics-nd/cnn-surrogate/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
178384082
Add Test Helper MINOR CHANGELOG Added new test helper, stubPollboy, for acceptance tests that mocks the polling service to not actually poll (otherwise acceptance tests will never complete). Replaced ember-cli-blanket with ember-cli-code-coverage for test coverage. Upgraded devDependencies to latest versions. :+1: Hey Matthew, I've been trying to import this stubPollboy test helper (or any module in /test-support/helpers/ember-pollboy.js for that matter) and can't figure out how to do so. Have you been able to successfully do that in a consuming app?
gharchive/pull-request
2016-09-21T15:37:07
2025-04-01T04:33:47.588793
{ "authors": [ "NickChristensen", "sandersky" ], "repo": "ciena-blueplanet/ember-pollboy", "url": "https://github.com/ciena-blueplanet/ember-pollboy/pull/24", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
192922397
Fixed dependency issue with ember-spread This project uses semver, please check the scope of this pr: [x] #patch# - backwards-compatible bug fix [ ] #minor# - adding functionality in a backwards-compatible manner [ ] #major# - incompatible API change CHANGELOG Fixed dependency issue with ember-spread Changes Unknown when pulling c4d3bd590508cf344b8276e5d33c2ffd403f196f on dafortin:master into ** on ciena-frost:master**. Changes Unknown when pulling c4d3bd590508cf344b8276e5d33c2ffd403f196f on dafortin:master into ** on ciena-frost:master**. Changes Unknown when pulling 93546a419b196c1fc52ca9938439bc277abdb010 on dafortin:master into ** on ciena-frost:master**. 👍
gharchive/pull-request
2016-12-01T18:15:43
2025-04-01T04:33:47.594144
{ "authors": [ "coveralls", "dafortin", "juwara0" ], "repo": "ciena-frost/ember-frost-core", "url": "https://github.com/ciena-frost/ember-frost-core/pull/335", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
246471242
Propagate Status field for CiliumNetworkPolicy A possible status field could look like this: @@ -33,6 +44,80 @@ type CiliumNetworkPolicy struct { // Spec is the desired Cilium specific rule specification. Spec api.Rule `json:"spec"` + + // Status is the status of the Cilium policy rule + Status CiliumNetworkPolicyStatus `json:"status"` +} + +// CiliumNetworkPolicyStatus is the status of a Cilium policy rule +type CiliumNetworkPolicyStatus struct { + // Nodes is the Cilium policy status for each node + Nodes map[string]CiliumNetworkPolicyNodeStatus + + // Failing is the number of nodes failing to install this policy + Failing int +} + +// CiliumNetworkPolicyNodeStatus is the status of a Cilium policy rule for a +// specific node +type CiliumNetworkPolicyNodeStatus struct { + // OK is true when the policy has been installed successfully + OK bool + + // Message describes the error condition if OK is false + Message string +} Blocked on #1171
gharchive/issue
2017-07-28T22:13:03
2025-04-01T04:33:47.606183
{ "authors": [ "tgraf" ], "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/issues/1222", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
842305700
helm: Consolidate IPSec and Wireguard encryption options Suggested in https://github.com/cilium/cilium/pull/15383#discussion_r600323895. @brb taking care of it. I submitted a PR, let me know if this makes sense. Thanks!
gharchive/issue
2021-03-26T21:24:45
2025-04-01T04:33:47.607467
{ "authors": [ "brb", "youssefazrak" ], "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/issues/15483", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2065785568
Random traffic is dropped from pods based on same cronjob Is there an existing issue for this? [X] I have searched the existing issues What happened? Problem is with connections being dropped between cronjobs (about ~200) and redis pod. I have cronjobs in test-cron namespace and redis in cache namespace. Pods are just simple test pods to reproduce the problem, they are run every second they set key sleep 5-15 sec del key About ~30 are run simultaneously, the rest is in Pending or ContainerCreating states. Most of the runs are ok but every once in a while I am getting dropped connection that looks random. I'd expect all of the traffic to be forwarded not only part of it. hubble ui drops are described as denied by ingress policy (from the redis side), when I run hubble observe I got ciliumnetworkpolicies are (ns cache) apiVersion: cilium.io/v2 kind: CiliumNetworkPolicy metadata: name: default-deny-all namespace: cache spec: endpointSelector: {} ingress: - {} egress: - {} apiVersion: cilium.io/v2 kind: CiliumNetworkPolicy metadata: name: redis-locks-test-allowed-ingress namespace: cache spec: endpointSelector: matchLabels: app: redis-locks ingress: - fromEntities: - cluster toPorts: - ports: - port: "6379" Cilium Version 1.14.5 same on 1.14.2 & 1.14.3 Kernel Version 4.18.0-513.9.1.el8_9.x86_64 #1 SMP Sat Dec 2 05:23:44 EST 2023 x86_64 x86_64 x86_64 GNU/Linux Kubernetes Version 1.28.5 same on 1.27.6 & 1.27.8 Sysdump No response Relevant log output Jan 4 12:39:37.461: test-cron/test-cron-action-4b-com-28406198-x74j7:43470 (ID:45713) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:39:44.403: 172.0.2.211:60748 (ID:3892) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:39:47.490: 172.0.2.156:35134 (ID:27262) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:39:49.460: 172.0.2.114:38792 (ID:31444) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:39:51.452: 172.0.2.24:59266 (ID:56353) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:40:03.452: test-cron/test-cron-action-24a-com-28406199-f52wt:45798 (ID:65520) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:40:05.449: test-cron/test-cron-action-156a-com-28406199-lbbpv:53028 (ID:550) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:40:08.403: test-cron/test-cron-action-173c-com-28406199-rzj45:47352 (ID:12223) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:40:09.358: test-cron/test-cron-action-195a-pl-28406199-qwxxp:45240 (ID:18252) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:40:09.505: test-cron/test-cron-action-18c-pl-28406199-wj9f6:38280 (ID:60444) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:40:11.337: 172.0.2.174:47778 (ID:51667) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:40:12.510: 172.0.2.102:45098 (ID:33239) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:40:14.364: 172.0.2.228:47170 (ID:14774) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:40:37.495: 172.0.2.219:52740 (ID:3888) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:40:39.521: 172.0.2.216:43380 (ID:49924) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:40:41.551: 172.0.2.239:45874 (ID:19921) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:41:02.441: 172.0.2.180:42846 (ID:9040) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:41:03.468: 172.0.2.130:52516 (ID:21908) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:41:08.474: 172.0.2.171:44416 (ID:29453) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:41:12.456: test-cron/test-cron-action-17c-com-28406200-stsmm:37970 (ID:58486) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:41:25.345: test-cron/test-cron-action-54b-pl-28406200-m6ntq:33294 (ID:10021) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:41:27.569: test-cron/test-cron-action-160c-com-28406200-9z6h4:53826 (ID:10217) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:41:32.546: 172.0.2.143:57442 (ID:34280) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:41:35.505: 172.0.2.108:56142 (ID:19078) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:41:54.561: test-cron/test-cron-action-163a-pl-28406201-x8v6q:46178 (ID:36061) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:41:58.518: 172.0.2.158:53620 (ID:39004) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:42:03.565: 172.0.2.162:50478 (ID:48558) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:42:17.541: test-cron/test-cron-action-55c-com-28406201-fndk5:50100 (ID:7779) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:42:23.586: 172.0.2.217:37020 (ID:30664) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:42:36.551: test-cron/test-cron-action-169a-pl-28406201-ngrkj:60860 (ID:56003) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:42:39.455: test-cron/test-cron-action-177a-pl-28406201-knlkq:48136 (ID:37684) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:42:40.511: test-cron/test-cron-action-160a-com-28406202-rmjwq:46988 (ID:20696) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:42:43.641: 172.0.2.238:44750 (ID:23205) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:42:46.530: 172.0.2.145:40640 (ID:49980) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Jan 4 12:42:47.475: 172.0.2.118:54932 (ID:39283) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Anything else? Looks like when I removed default-deny-all policy the drops do not occur, but I am not 100% sure as this also can be that "randomness" We have similar default deny all policies on all of our ns but this is the first one that causes us problems. We reproduced same case in our production k8s cluster and in our test cluster. Code of Conduct [X] I agree to follow this project's Code of Conduct I checked again with default-deny-all policy removed - I still got random drops removing both default-deny-all and redis-locks-test-allowed-ingress seems to stop the drops but it cant be the solution here Hey devs any idea how to debug this issue? Hello 👋 How large is the cluster in question? I'm not super familiar with how fromEntities: cluster works under the hood, but I'm curious why you're using that as opposed to individual namespaces, or pod labels. Also, does this continue to happen for the lifetime of the CronJobs, or does it fix itself after a few seconds? The cluster where I am reproducing the problem has 3 nodes , 1 worker node. Second one, our main cluster has 13 nodes, 10 worker nodes - same problem here. I updated the logic, so that it does 5 iterations and 2 calls to redis in each one. $i = 5; while ($i--) { $lock->acquire($id, 100); sleep(5 + rand(5, 10)); $lock->release($id); } with hubble observe I can confirm that in a single pod some of the requests were forwarded some were not, basically the same requests. Logs Jan 12 12:03:02.696: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) policy-verdict:L3-L4 INGRESS ALLOWED (TCP Flags: SYN) Jan 12 12:03:02.696: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: SYN) Jan 12 12:03:02.696: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK) Jan 12 12:03:02.696: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH) Jan 12 12:03:17.697: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) <> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) policy-verdict:none EGRESS DENIED (TCP Flags: ACK, PSH) Jan 12 12:03:17.697: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) <> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) Policy denied DROPPED (TCP Flags: ACK, PSH) Jan 12 12:03:17.904: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) <> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) policy-verdict:none EGRESS DENIED (TCP Flags: ACK, PSH) Jan 12 12:03:17.904: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) <> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) Policy denied DROPPED (TCP Flags: ACK, PSH) Jan 12 12:03:18.113: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) <> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) policy-verdict:none EGRESS DENIED (TCP Flags: ACK, PSH) Jan 12 12:03:18.113: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) <> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) Policy denied DROPPED (TCP Flags: ACK, PSH) Jan 12 12:03:18.520: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH) Jan 12 12:03:31.521: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH) Jan 12 12:03:45.521: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH) Jan 12 12:04:00.521: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH) Jan 12 12:04:15.522: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH) Jan 12 12:04:15.523: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, FIN) Jan 12 12:04:15.523: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK) One thing that's particularly interesting about the log you posted is that it's policy-verdict:none EGRESS DENIED, which means that the caller is rejecting, not redis. Do you potentially have CiliumNetworkPolicy resources for the CronJob as well, and can you send those over/ We have udated cilium to 1.14.6, problem still persists. I can't reproduce policy-verdict:none EGRESS DENIED though. As I mentioned in the first post, it seemed like an ingress policy issue and with my latest logs I can confirm it is still an ingress issue. I run another hubble observe like this (for 3 cilium agent pods in our test cluster) kubectl -n kube-system exec -ti cilium-9jhtl -- hubble observe -f --to-namespace cache >> .dev/cilium-9jhtl.log with this I got 13 endpoints that got drops in about 15 minutes. About ~40 pods running at the same time with the test code. each one looks the same, for example ID:11143 Feb 2 11:07:37.567: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) <> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) policy-verdict:none INGRESS DENIED (TCP Flags: SYN) Feb 2 11:07:37.567: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) <> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN) Feb 2 11:07:38.587: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) policy-verdict:L3-L4 INGRESS ALLOWED (TCP Flags: SYN) Feb 2 11:07:38.587: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: SYN) Feb 2 11:07:38.587: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK) Feb 2 11:07:38.587: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH) Feb 2 11:07:52.588: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH) Feb 2 11:08:04.589: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH) Feb 2 11:08:16.590: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH) Feb 2 11:08:27.590: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH) Feb 2 11:08:38.591: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH) Feb 2 11:08:38.592: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, FIN) Feb 2 11:08:38.592: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK) Feb 2 11:07:37.567: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) policy-verdict:L3-L4 EGRESS ALLOWED (TCP Flags: SYN) Feb 2 11:07:37.567: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-overlay FORWARDED (TCP Flags: SYN) Feb 2 11:07:38.588: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-overlay FORWARDED (TCP Flags: ACK) Feb 2 11:07:38.588: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-overlay FORWARDED (TCP Flags: ACK, PSH) Feb 2 11:07:52.588: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-overlay FORWARDED (TCP Flags: ACK, PSH) Feb 2 11:08:04.589: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-overlay FORWARDED (TCP Flags: ACK, PSH) Feb 2 11:08:16.590: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-overlay FORWARDED (TCP Flags: ACK, PSH) Feb 2 11:08:27.591: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-overlay FORWARDED (TCP Flags: ACK, PSH) Feb 2 11:08:38.591: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-overlay FORWARDED (TCP Flags: ACK, PSH) Feb 2 11:08:38.592: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-overlay FORWARDED (TCP Flags: ACK, FIN) Feb 2 11:08:38.593: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-overlay FORWARDED (TCP Flags: ACK) In all of these logs INGRESS DENIED is right before INGRESS ALLOWED and there is EGRESS ALLOWED at some point. In 12 of cases EGRESS ALLOWED is after INGRESS ALLOWED, in 1 case it is before INGRES DENIED, but since it is ruled as ALLOWED I'm not sure if that particullar log line matters. So to sum up it looks like this 12 cases ... policy-verdict:none INGRESS DENIED ... ... Policy denied DROPPED ... ... policy-verdict:L3-L4 INGRESS ALLOWED ... ... [logs for FORWARDED] ... ... policy-verdict:L3-L4 EGRESS ALLOWED ... ... [logs for FORWARDED] ... 1 case ... policy-verdict:L3-L4 EGRESS ALLOWED ... ... policy-verdict:none INGRESS DENIED ... ... Policy denied DROPPED ... ... policy-verdict:L3-L4 INGRESS ALLOWED ... ... [logs for FORWARDED] ... I've tried to use fromEndpoint instead of fromEntities. I had to change resource to CiliumClusterwideNetworkPolicy to make it cross-namespace with my test label (with CiliumNetworkPolicy I had nothing but drops). Change I made apiVersion: cilium.io/v2 -kind: CiliumNetworkPolicy +kind: CiliumClusterwideNetworkPolicy metadata: name: redis-locks-test-allowed-ingress spec: endpointSelector: matchLabels: app: redis-locks ingress: - - fromEntities: - - cluster - toPorts: - - ports: - - port: "6379" + - fromEndpoints: + - matchLabels: + mylabel: "test" and I added mylabel: "test" to test crons. Result was the same as above, basically I got about 8 drops in few minutes with output like this Feb 5 10:10:19.626: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) policy-verdict:L3-L4 EGRESS ALLOWED (TCP Flags: SYN) Feb 5 10:10:19.626: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-overlay FORWARDED (TCP Flags: SYN) Feb 5 10:10:20.635: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-overlay FORWARDED (TCP Flags: ACK) Feb 5 10:10:20.636: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-overlay FORWARDED (TCP Flags: ACK, PSH) Feb 5 10:10:31.636: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-overlay FORWARDED (TCP Flags: ACK, PSH) Feb 5 10:10:41.637: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-overlay FORWARDED (TCP Flags: ACK, PSH) Feb 5 10:10:54.638: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-overlay FORWARDED (TCP Flags: ACK, PSH) Feb 5 10:11:06.639: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-overlay FORWARDED (TCP Flags: ACK, PSH) Feb 5 10:11:21.640: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-overlay FORWARDED (TCP Flags: ACK, PSH) Feb 5 10:11:21.641: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-overlay FORWARDED (TCP Flags: ACK, FIN) Feb 5 10:11:21.641: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-overlay FORWARDED (TCP Flags: ACK) Feb 5 10:10:19.626: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) <> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) policy-verdict:none INGRESS DENIED (TCP Flags: SYN) Feb 5 10:10:19.626: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) <> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) Policy denied DROPPED (TCP Flags: SYN) Feb 5 10:10:20.635: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) policy-verdict:L3-Only INGRESS ALLOWED (TCP Flags: SYN) Feb 5 10:10:20.635: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-endpoint FORWARDED (TCP Flags: SYN) Feb 5 10:10:20.636: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-endpoint FORWARDED (TCP Flags: ACK) Feb 5 10:10:20.636: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-endpoint FORWARDED (TCP Flags: ACK, PSH) Feb 5 10:10:31.637: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-endpoint FORWARDED (TCP Flags: ACK, PSH) Feb 5 10:10:41.637: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-endpoint FORWARDED (TCP Flags: ACK, PSH) Feb 5 10:10:54.638: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-endpoint FORWARDED (TCP Flags: ACK, PSH) Feb 5 10:11:06.639: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-endpoint FORWARDED (TCP Flags: ACK, PSH) Feb 5 10:11:21.640: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-endpoint FORWARDED (TCP Flags: ACK, PSH) Feb 5 10:11:21.641: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-endpoint FORWARDED (TCP Flags: ACK, FIN) Feb 5 10:11:21.641: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-endpoint FORWARDED (TCP Flags: ACK) which is a pattern like that one ... policy-verdict:L3-L4 EGRESS ALLOWED ... ... [logs for FORWARDED] ... ... policy-verdict:none INGRESS DENIED ... ... Policy denied DROPPED ... ... policy-verdict:L3-L4 INGRESS ALLOWED ... ... [logs for FORWARDED] ... We updated to 1.14.7 and it seems to solve the problem.
gharchive/issue
2024-01-04T14:44:12
2025-04-01T04:33:47.625515
{ "authors": [ "EItanya", "adamw-linadm", "pawelbaranski" ], "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/issues/30105", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2079881972
CI: Conformance Runtime - Test_PodIPPoolAdvert failure CI failure Hit on https://github.com/cilium/cilium/pull/30183 Link: https://github.com/cilium/cilium/actions/runs/7480430927/job/20359924251 --- FAIL: Test_PodIPPoolAdvert (8.07s) --- FAIL: Test_PodIPPoolAdvert/matching_ipv4_pool (0.00s) adverts_test.go:430: expected events: [{65001 11.1.1.0 24 false []}] adverts_test.go:431: received events: [{65001 10.1.1.0 24 false []}] adverts_test.go:432: Error Trace: /host/pkg/bgpv1/test/adverts_test.go:432 Error: elements differ extra elements in list A: ([]interface {}) (len=1) { (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=8) "11.1.1.0", prefixLen: (uint8) 24, isWithdrawn: (bool) false, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> } } extra elements in list B: ([]interface {}) (len=1) { (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=8) "10.1.1.0", prefixLen: (uint8) 24, isWithdrawn: (bool) false, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> } } listA: ([]test.routeEvent) (len=1) { (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=8) "11.1.1.0", prefixLen: (uint8) 24, isWithdrawn: (bool) false, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> } } listB: ([]test.routeEvent) (len=1) { (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=8) "10.1.1.0", prefixLen: (uint8) 24, isWithdrawn: (bool) false, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> } } Test: Test_PodIPPoolAdvert/matching_ipv4_pool Messages: matching ipv4 pool --- FAIL: Test_PodIPPoolAdvert/update_matching_ipv4_pool (0.00s) adverts_test.go:430: expected events: [{65001 11.1.1.0 24 true []} {65001 11.2.1.0 24 false []}] adverts_test.go:431: received events: [{65001 10.1.2.0 24 false []} {65001 11.1.1.0 24 false []}] adverts_test.go:432: Error Trace: /host/pkg/bgpv1/test/adverts_test.go:432 Error: elements differ extra elements in list A: ([]interface {}) (len=2) { (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=8) "11.1.1.0", prefixLen: (uint8) 24, isWithdrawn: (bool) true, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> }, (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=8) "11.2.1.0", prefixLen: (uint8) 24, isWithdrawn: (bool) false, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> } } extra elements in list B: ([]interface {}) (len=2) { (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=8) "10.1.2.0", prefixLen: (uint8) 24, isWithdrawn: (bool) false, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> }, (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=8) "11.1.1.0", prefixLen: (uint8) 24, isWithdrawn: (bool) false, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> } } listA: ([]test.routeEvent) (len=2) { (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=8) "11.1.1.0", prefixLen: (uint8) 24, isWithdrawn: (bool) true, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> }, (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=8) "11.2.1.0", prefixLen: (uint8) 24, isWithdrawn: (bool) false, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> } } listB: ([]test.routeEvent) (len=2) { (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=8) "10.1.2.0", prefixLen: (uint8) 24, isWithdrawn: (bool) false, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> }, (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=8) "11.1.1.0", prefixLen: (uint8) 24, isWithdrawn: (bool) false, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> } } Test: Test_PodIPPoolAdvert/update_matching_ipv4_pool Messages: update matching ipv4 pool --- FAIL: Test_PodIPPoolAdvert/matching_ipv6_pool (0.00s) adverts_test.go:430: expected events: [{65001 2001:0:0:1234:5678:: 96 false []} {65001 11.2.1.0 24 true []}] adverts_test.go:431: received events: [{65001 10.1.2.0 24 true []} {65001 10.1.1.0 24 true []}] adverts_test.go:432: Error Trace: /host/pkg/bgpv1/test/adverts_test.go:432 Error: elements differ extra elements in list A: ([]interface {}) (len=2) { (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=20) "2001:0:0:1234:5678::", prefixLen: (uint8) 96, isWithdrawn: (bool) false, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> }, (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=8) "11.2.1.0", prefixLen: (uint8) 24, isWithdrawn: (bool) true, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> } } extra elements in list B: ([]interface {}) (len=2) { (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=8) "10.1.2.0", prefixLen: (uint8) 24, isWithdrawn: (bool) true, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> }, (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=8) "10.1.1.0", prefixLen: (uint8) 24, isWithdrawn: (bool) true, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> } } listA: ([]test.routeEvent) (len=2) { (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=20) "2001:0:0:1234:5678::", prefixLen: (uint8) 96, isWithdrawn: (bool) false, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> }, (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=8) "11.2.1.0", prefixLen: (uint8) 24, isWithdrawn: (bool) true, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> } } listB: ([]test.routeEvent) (len=2) { (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=8) "10.1.2.0", prefixLen: (uint8) 24, isWithdrawn: (bool) true, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> }, (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=8) "10.1.1.0", prefixLen: (uint8) 24, isWithdrawn: (bool) true, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> } } Test: Test_PodIPPoolAdvert/matching_ipv6_pool Messages: matching ipv6 pool --- FAIL: Test_PodIPPoolAdvert/update_matching_ipv6_pool (0.00s) adverts_test.go:430: expected events: [{65001 2001:0:0:1234:5678:: 96 true []} {65001 2002:0:0:1234:5678:: 96 false []}] adverts_test.go:431: received events: [{65001 11.2.1.0 24 false []} {65001 11.1.1.0 24 true []}] adverts_test.go:432: Error Trace: /host/pkg/bgpv1/test/adverts_test.go:432 Error: elements differ extra elements in list A: ([]interface {}) (len=2) { (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=20) "2001:0:0:1234:5678::", prefixLen: (uint8) 96, isWithdrawn: (bool) true, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> }, (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=20) "2002:0:0:1234:5678::", prefixLen: (uint8) 96, isWithdrawn: (bool) false, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> } } extra elements in list B: ([]interface {}) (len=2) { (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=8) "11.2.1.0", prefixLen: (uint8) 24, isWithdrawn: (bool) false, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> }, (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=8) "11.1.1.0", prefixLen: (uint8) 24, isWithdrawn: (bool) true, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> } } listA: ([]test.routeEvent) (len=2) { (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=20) "2001:0:0:1234:5678::", prefixLen: (uint8) 96, isWithdrawn: (bool) true, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> }, (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=20) "2002:0:0:1234:5678::", prefixLen: (uint8) 96, isWithdrawn: (bool) false, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> } } listB: ([]test.routeEvent) (len=2) { (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=8) "11.2.1.0", prefixLen: (uint8) 24, isWithdrawn: (bool) false, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> }, (test.routeEvent) { sourceASN: (uint32) 65001, prefix: (string) (len=8) "11.1.1.0", prefixLen: (uint8) 24, isWithdrawn: (bool) true, extraPathAttributes: ([]bgp.PathAttributeInterface) <nil> } } Test: Test_PodIPPoolAdvert/update_matching_ipv6_pool Messages: update matching ipv6 pool Hit in https://github.com/cilium/cilium/pull/30440 - job https://github.com/cilium/cilium/actions/runs/7738391682/job/21099140614 +1 Hit this just now. https://github.com/cilium/cilium/actions/runs/8017423656/job/21901426809 Logs: logs_2067640.zip Seen again in https://github.com/cilium/cilium/pull/33860 at https://github.com/cilium/cilium/actions/runs/9979635155/job/27579321909. cc @rastislavs can you take a look? Seen again in #33860 at https://github.com/cilium/cilium/actions/runs/9979635155/job/27579321909. yeah, it seems to be to same flake, looks like the root cause is in something else than what I fixed :( Hit again as well: https://github.com/cilium/cilium/actions/runs/10006033738/job/27657940599 Fixed in https://github.com/cilium/cilium/pull/34270 and already backported to v1.16 & v1.15
gharchive/issue
2024-01-12T23:32:29
2025-04-01T04:33:47.637735
{ "authors": [ "christarazi", "dylandreimerink", "giorio94", "rastislavs" ], "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/issues/30237", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2298480897
Ingress controller load balancer can not connect to nodes Is there an existing issue for this? [X] I have searched the existing issues What happened? We have an EKS cluster where we are trying to use Cilium ingress controller and the load balancer created for the ingress can not always connect to the nodes. What we see is that the load balancer can connect to some nodes during periods but is not a consistent behavior and there is no pattern between the nodes behind that it can connect and the ones that can not. Checking directly in the nodes also connecting to the nodePort opened for the load balancer does not work so should not be a problem of security groups, anyway we tried opening traffic from every internal address and nothing, some nodes work and others not or even sometimes no nodes happen to be accessible by the load balancer. I checked and all the nodes have this cilium LB configuration for the nodePort: 10.218.248.217:31799 0.0.0.0:0 (331) (0) [NodePort, l7-load-balancer] 0.0.0.0:31799 0.0.0.0:0 (333) (0) [NodePort, non-routable, l7-load-balancer] 10.0.243.9:31799 0.0.0.0:0 (330) (0) [NodePort, l7-load-balancer] 169.254.0.11:31799 0.0.0.0:0 (332) (0) [NodePort, l7-load-balancer] Configuration values used: cni: configMap: cni-config customConf: true eni: enabled: true updateEC2AdapterLimitViaAPI: true awsEnablePrefixDelegation: true awsReleaseExcessIPs: true egressMasqueradeInterfaces: eth0 policyEnforcementMode: "never" ipam: mode: eni hubble: relay: enabled: true ui: enabled: true tunnelProtocol: "" nodePort: enabled: true nodeinit: enabled: true ingressController: enabled: true cni-config configmap values: { "cniVersion":"0.3.1", "name":"cilium", "plugins": [ { "cniVersion":"0.3.1", "type":"cilium-cni", "eni": { "subnet-ids": ["subnet-xxxxxx", "subnet-xxxxxx", "subnet-xxxxxxx"], "first-interface-index": 1 } } ] } Cilium Version We tried it in multiple versions: 1.14.1 1.15.4 1.16.0-pre.0 Kernel Version Linux 5.10.215-203.850.amzn2.aarch64 Kubernetes Version v1.26.15 Regression No response Sysdump Relevant log output No response Anything else? No response Cilium Users Document [ ] Are you a user of Cilium? Please add yourself to the Users doc Code of Conduct [X] I agree to follow this project's Code of Conduct Hi there, thanks for the bug report. It's not yet clear to me how exactly traffic is flowing. Could you outline the expected traffic flow, and indicate where you think it is failing? In particular, I suggest the section on troubleshooting with hubble to identify where packets are being dropped. Can you go through the troubleshooting section and clarify the problem a bit? Thanks. Also can you share your cilium configmap as well? Thanks. Hi there, thanks for the bug report. It's not yet clear to me how exactly traffic is flowing. Could you outline the expected traffic flow, and indicate where you think it is failing? In particular, I suggest the section on troubleshooting with hubble to identify where packets are being dropped. Can you go through the troubleshooting section and clarify the problem a bit? Thanks. @squeed Hi, sorry for the delay, yes let me explain it better. Correct me if I mention something wrong. The idea is to use Cilium as an Ingress Controller, when I create an ingress object it creates the Classic AWS LB or NLB, tried both, which will balance the traffic to the Cilium ingress controller. If I'm not wrong the component of Cilium that handles the traffic coming from the LB is cilium-envoy which runs inside cilium-agent in my case. The traffic after arriving to cilium-envoy gets sent to the relevant backend of the Ingress. My problem is the communication between the Load Balancer and envoy, the Load Balancer can not target envoy most of the time. Ask any other question that you need if I still did not explain it well enough. Also can you share your cilium configmap as well? Thanks. @sayboras Yes, here it is: agent-not-ready-taint-key: node.cilium.io/agent-not-ready arping-refresh-period: 30s auto-direct-node-routes: "false" bpf-lb-acceleration: disabled bpf-lb-external-clusterip: "false" bpf-lb-map-max: "65536" bpf-lb-sock: "false" bpf-map-dynamic-size-ratio: "0.0025" bpf-policy-map-max: "16384" bpf-root: /sys/fs/bpf cgroup-root: /run/cilium/cgroupv2 cilium-endpoint-gc-interval: 5m0s cluster-id: "0" cluster-name: default cluster-pool-ipv4-cidr: 10.0.0.0/8 cluster-pool-ipv4-mask-size: "24" cni-chaining-mode: aws-cni cni-exclusive: "false" cni-log-file: /var/run/cilium/cilium-cni.log custom-cni-conf: "false" debug: "false" debug-verbose: "" egress-gateway-reconciliation-trigger-interval: 1s enable-auto-protect-node-port-range: "true" enable-bgp-control-plane: "false" enable-bpf-clock-probe: "false" enable-endpoint-health-checking: "false" enable-endpoint-routes: "true" enable-envoy-config: "true" enable-external-ips: "false" enable-gateway-api: "true" enable-gateway-api-secrets-sync: "true" enable-health-check-loadbalancer-ip: "false" enable-health-check-nodeport: "true" enable-health-checking: "true" enable-host-legacy-routing: "true" enable-host-port: "false" enable-hubble: "true" enable-ingress-controller: "true" enable-ingress-proxy-protocol: "false" enable-ingress-secrets-sync: "true" enable-ipv4: "true" enable-ipv4-big-tcp: "false" enable-ipv4-masquerade: "false" enable-ipv6: "false" enable-ipv6-big-tcp: "false" enable-ipv6-masquerade: "true" enable-k8s-networkpolicy: "true" enable-k8s-terminating-endpoint: "true" enable-l2-neigh-discovery: "true" enable-l7-proxy: "true" enable-local-node-route: "false" enable-local-redirect-policy: "false" enable-masquerade-to-route-source: "false" enable-metrics: "true" enable-node-port: "true" enable-policy: never enable-remote-node-identity: "true" enable-sctp: "false" enable-svc-source-range-check: "true" enable-vtep: "false" enable-well-known-identities: "false" enable-xt-socket-fallback: "true" enforce-ingress-https: "true" external-envoy-proxy: "false" gateway-api-secrets-namespace: cilium-secrets hubble-disable-tls: "false" hubble-export-file-max-backups: "5" hubble-export-file-max-size-mb: "10" hubble-listen-address: :4244 hubble-socket-path: /var/run/cilium/hubble.sock hubble-tls-cert-file: /var/lib/cilium/tls/hubble/server.crt hubble-tls-client-ca-files: /var/lib/cilium/tls/hubble/client-ca.crt hubble-tls-key-file: /var/lib/cilium/tls/hubble/server.key identity-allocation-mode: crd identity-gc-interval: 15m0s identity-heartbeat-timeout: 30m0s ingress-default-lb-mode: dedicated ingress-lb-annotation-prefixes: service.beta.kubernetes.io service.kubernetes.io cloud.google.com ingress-secrets-namespace: cilium-secrets ingress-shared-lb-service-name: cilium-ingress install-no-conntrack-iptables-rules: "false" ipam: cluster-pool ipam-cilium-node-update-rate: 15s k8s-client-burst: "10" k8s-client-qps: "5" kube-proxy-replacement: "false" kube-proxy-replacement-healthz-bind-address: "" max-connected-clusters: "255" mesh-auth-enabled: "true" mesh-auth-gc-interval: 5m0s mesh-auth-queue-size: "1024" mesh-auth-rotated-identities-queue-size: "1024" monitor-aggregation: medium monitor-aggregation-flags: all monitor-aggregation-interval: 5s node-port-bind-protection: "true" nodes-gc-interval: 5m0s operator-api-serve-addr: 127.0.0.1:9234 operator-prometheus-serve-addr: :9963 policy-cidr-match-mode: "" preallocate-bpf-maps: "false" procfs: /host/proc proxy-connect-timeout: "2" proxy-idle-timeout-seconds: "60" proxy-max-connection-duration-seconds: "0" proxy-max-requests-per-connection: "0" proxy-prometheus-port: "9964" proxy-xff-num-trusted-hops-egress: "0" proxy-xff-num-trusted-hops-ingress: "0" remove-cilium-node-taints: "true" routing-mode: native service-no-backend-response: reject set-cilium-is-up-condition: "true" set-cilium-node-taints: "true" sidecar-istio-proxy-image: cilium/istio_proxy skip-cnp-status-startup-clean: "false" synchronize-k8s-nodes: "true" tofqdns-dns-reject-response-code: refused tofqdns-enable-dns-compression: "true" tofqdns-endpoint-max-ip-per-hostname: "50" tofqdns-idle-connection-grace-period: 0s tofqdns-max-deferred-connection-deletes: "10000" tofqdns-proxy-response-max-delay: 100ms unmanaged-pod-watcher-interval: "15" vtep-cidr: "" vtep-endpoint: "" vtep-mac: "" vtep-mask: "" write-cni-conf-when-ready: /host/etc/cni/net.d/05-cilium.conflist Thank you for taking a look into this!
gharchive/issue
2024-05-15T17:42:56
2025-04-01T04:33:47.651401
{ "authors": [ "carlosrejano", "sayboras", "squeed" ], "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/issues/32556", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1454726110
fix: package name of test folder 1 According to the code specification, the package name should be lower case. 2 Refer to the official documentation of gingko to modify package name https://onsi.github.io/ginkgo/ Could you please rebase your commit instead of merging master into this PR branch? Could you please rebase your commit instead of merging master into this PR branch? okay Hi @yulng , is this related to a problem you were having? It could be helpful to share what that problem is in the PR description so that others can see / search for the same problem. Please also rebase and update the commit message. I have not signed off on the code, the sign-off line should be from you to highlight that you agree to the certificate of origin (description linked by the maintainer's little helper in comments above). Please also rebase and update the commit message. I have not signed off on the code, the sign-off line should be from you to highlight that you agree to the certificate of origin (description linked by the maintainer's little helper in comments above). okay,modified It just doesn't conform to the code specification, which is not conducive to search. :) It just doesn't conform to the code specification, which is not conducive to search. :) It just doesn't conform to the code specification, which is not conducive to search. :) Hi @yulng , is this related to a problem you were having? It could be helpful to share what that problem is in the PR description so that others can see / search for the same problem. It just doesn't conform to the code specification, which is not conducive to search. :)
gharchive/pull-request
2022-11-18T08:54:19
2025-04-01T04:33:47.657076
{ "authors": [ "christarazi", "joestringer", "yulng" ], "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/pull/22259", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1646183360
Add support for FromCIDRGroupRefs Ingress network policies -- WORK IN PROGRESS, NOT READY FOR REVIEW YET -- remove the useless ref counting in cidr ingress identity allocation (but use the cache to remove identities when a cidr group is updated) avoid overwriting FromCIDR field in CNP/CCNP if all the cidr groups of a cnp have been deleted we should completely remove the field FromCIDR in the CNP @pippolo84 While playing around with this PR, I ran into some cases of nil pointer panics, here's the diff that fixes it: diff --git a/pkg/k8s/watchers/cilium_network_policy.go b/pkg/k8s/watchers/cilium_network_policy.go index 45579070fe..b8b3f150c1 100644 --- a/pkg/k8s/watchers/cilium_network_policy.go +++ b/pkg/k8s/watchers/cilium_network_policy.go @@ -561,15 +561,11 @@ func reportCNPChangeMetrics(err error) { } func containsCIDRGroupRef(cnp *types.SlimCNP, cidrGroup string) bool { - for _, ingress := range cnp.Spec.Ingress { - for _, cidrGroupRef := range ingress.FromCIDRGroupRefs { - if string(cidrGroupRef) == cidrGroup { - return true + if cnp.Spec != nil && cnp.Spec.Ingress != nil { + for _, ingress := range cnp.Spec.Ingress { + if ingress.FromCIDRGroupRefs == nil { + continue } - } - } - for _, spec := range cnp.Specs { - for _, ingress := range spec.Ingress { for _, cidrGroupRef := range ingress.FromCIDRGroupRefs { if string(cidrGroupRef) == cidrGroup { return true @@ -577,36 +573,70 @@ func containsCIDRGroupRef(cnp *types.SlimCNP, cidrGroup string) bool { } } } + if cnp.Specs != nil { + for _, spec := range cnp.Specs { + if spec.Ingress != nil { + for _, ingress := range spec.Ingress { + if ingress.FromCIDRGroupRefs == nil { + continue + } + for _, cidrGroupRef := range ingress.FromCIDRGroupRefs { + if string(cidrGroupRef) == cidrGroup { + return true + } + } + } + } + } + } return false } func translateFromCIDRGroupRefs(cnp *types.SlimCNP, cache map[resource.Key]*cilium_v2_alpha1.CiliumCIDRGroup) error { var cidrGroupRefs []string - for i := range cnp.Spec.Ingress { - for _, cidrGroupRef := range cnp.Spec.Ingress[i].FromCIDRGroupRefs { - cidrGroupRefs = append(cidrGroupRefs, string(cidrGroupRef)) + // The following causes a nil ptr panic if the policy does not have a + // .Spec, but rather only has a .Specs + // + // The following code should fix it now. + if cnp.Spec != nil && cnp.Spec.Ingress != nil { + for i := range cnp.Spec.Ingress { + for _, cidrGroupRef := range cnp.Spec.Ingress[i].FromCIDRGroupRefs { + cidrGroupRefs = append(cidrGroupRefs, string(cidrGroupRef)) + } } } - for i := range cnp.Specs { - for j := range cnp.Specs[i].Ingress { - for _, cidrGroupRef := range cnp.Specs[i].Ingress[j].FromCIDRGroupRefs { - cidrGroupRefs = append(cidrGroupRefs, string(cidrGroupRef)) + if cnp.Specs != nil { + for i := range cnp.Specs { + if cnp.Specs[i].Ingress == nil { + continue + } + for j := range cnp.Specs[i].Ingress { + for _, cidrGroupRef := range cnp.Specs[i].Ingress[j].FromCIDRGroupRefs { + cidrGroupRefs = append(cidrGroupRefs, string(cidrGroupRef)) + } } } } cidrsSet := cidrGroupRefsToCIDRsSet(cidrGroupRefs, cache) - for i := range cnp.Spec.Ingress { - for _, cidrGroupRef := range cnp.Spec.Ingress[i].FromCIDRGroupRefs { - cnp.Spec.Ingress[i].FromCIDR = cidrsSet[string(cidrGroupRef)] + if cnp.Spec != nil && cnp.Spec.Ingress != nil { + for i := range cnp.Spec.Ingress { + for _, cidrGroupRef := range cnp.Spec.Ingress[i].FromCIDRGroupRefs { + cnp.Spec.Ingress[i].FromCIDR = cidrsSet[string(cidrGroupRef)] + } } } - for i := range cnp.Specs { - for j := range cnp.Specs[i].Ingress { - for _, cidrGroupRef := range cnp.Specs[i].Ingress[j].FromCIDRGroupRefs { - cnp.Specs[i].Ingress[j].FromCIDR = cidrsSet[string(cidrGroupRef)] + if cnp.Specs != nil { + for i := range cnp.Specs { + if cnp.Specs[i].Ingress == nil { + continue + } + for j := range cnp.Specs[i].Ingress { + for _, cidrGroupRef := range cnp.Specs[i].Ingress[j].FromCIDRGroupRefs { + cnp.Specs[i].Ingress[j].FromCIDR = cidrsSet[string(cidrGroupRef)] + } } } } This patch might be cleaner to apply: Patch commit da478208fe4666a54920397d094e12e87f118ce8 Author: Chris Tarazi <chris@isovalent.com> Date: Mon Apr 3 21:23:12 2023 -0700 k8s: Consolidate handling of CNP and CCNP with sync.Once Signed-off-by: Chris Tarazi <chris@isovalent.com> diff --git a/pkg/k8s/watchers/watcher.go b/pkg/k8s/watchers/watcher.go index 105da248bb..f4f3aa6156 100644 --- a/pkg/k8s/watchers/watcher.go +++ b/pkg/k8s/watchers/watcher.go @@ -540,6 +540,8 @@ func (k *K8sWatcher) enableK8sWatchers(ctx context.Context, resourceNames []stri return fmt.Errorf("error creating service list option modifier: %w", err) } + // CNP and CCNP resources are handled together. + var cnpOnce sync.Once for _, r := range resourceNames { switch r { // Core Cilium @@ -570,11 +572,8 @@ func (k *K8sWatcher) enableK8sWatchers(ctx context.Context, resourceNames []stri // only watch secrets in specific namespaces k.tlsSecretInit(k.clientset.Slim(), option.Config.EnvoySecretNamespaces, swgSecret) // Custom resource definitions - case k8sAPIGroupCiliumNetworkPolicyV2: - k.ciliumNetworkPoliciesInit(ctx, k.clientset) - case k8sAPIGroupCiliumClusterwideNetworkPolicyV2: - // handled by k8sAPIGroupCiliumNetworkPolicyV2 - return nil + case k8sAPIGroupCiliumNetworkPolicyV2, k8sAPIGroupCiliumClusterwideNetworkPolicyV2: + cnpOnce.Do(func() { k.ciliumNetworkPoliciesInit(ctx, k.clientset) }) case k8sAPIGroupCiliumEndpointV2: k.initCiliumEndpointOrSlices(k.clientset, asyncControllers) case k8sAPIGroupCiliumEndpointSliceV2Alpha1: Btw, if I apply a CIDR group and then a CNP which references the group, and then remove the CIDR group, I still see the dangling empty field in the ingress section of cilium policy get: { "endpointSelector": { "matchLabels": { "any:name": "curl", "k8s:io.kubernetes.pod.namespace": "default" } }, "ingress": [ {}, { "fromCIDR": [ "3.3.3.3/32" ] } ], Btw, if I apply a CIDR group and then a CNP which references the group, and then remove the CIDR group, I still see the dangling empty field in the ingress section of cilium policy get: { "endpointSelector": { "matchLabels": { "any:name": "curl", "k8s:io.kubernetes.pod.namespace": "default" } }, "ingress": [ {}, <---- { "fromCIDR": [ "3.3.3.3/32" ] } ], I was able to reproduce applying 2 CNPs referencing the same CIDRGroup, one of which with an additional FromCIDR rule. Now it's fixed (I also added an additional unit test for such cases). /test Oh, I meant to say - this fixes https://github.com/cilium/cilium/issues/10349 right? Oh, I meant to say - this fixes #10349 right? Yes, it does. I've udpated the PR description to reference that issue. /test ConformanceEKS is hitting https://github.com/cilium/cilium/issues/24774 I've addressed all feedback from @joestringer and @tommyp1ckles reviews, PTAL :pray: /test /test /mlh new-flake Cilium-PR-K8s-1.16-kernel-4.19 4.19 hit https://github.com/cilium/cilium/issues/24840 /test-1.16-4.19 Hello @pippolo84 , Thank you for your useful contributions , I have found interest in documenting this PR from issue #25123 . I understand that it falls under CiliumNetworkPolicy, CiliumClusterwideNetworkPolicy and CiliumCIDRGroup. I am a cloud technology newbie docs are my specialty , Please help me with an example use case on how to use these new amazing feature under the the respective policies and groups .
gharchive/pull-request
2023-03-29T16:55:09
2025-04-01T04:33:47.667236
{ "authors": [ "christarazi", "gailsuccess", "joestringer", "pippolo84" ], "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/pull/24638", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1849438899
1.15 bpf nodeport ingress trace Please ensure your pull request adheres to the following guidelines: [ ] For first time contributors, read Submitting a pull request [ ] All code is covered by unit and/or runtime tests where feasible. [ ] All commits contain a well written commit description including a title, description and a Fixes: #XXX line if the commit addresses a particular GitHub issue. [ ] If your commit description contains a Fixes: <commit-id> tag, then please add the commit author[s] as reviewer[s] to this issue. [ ] All commits are signed off. See the section Developer’s Certificate of Origin [ ] Provide a title or release-note blurb suitable for the release notes. [ ] Are you a user of Cilium? Please add yourself to the Users doc [ ] Thanks for contributing! Fixes: #issue-number <!-- Enter the release note text here if needed or remove this section! --> /test /ci-verifier /ci-verifier /test /test /test Sorry for the massive churn. But this needed a fair bit of refactoring to slash through all the tech debt :grimacing: . /test /test (one last rebase to pull #27509 in) this now required a small change for this PR: https://github.com/cilium/cilium/pull/27488/commits/944acfd51aa4280bc21eac5221d69c81df6794c5. /test
gharchive/pull-request
2023-08-14T09:58:40
2025-04-01T04:33:47.674577
{ "authors": [ "julianwiedmann" ], "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/pull/27488", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2183396824
bgpv1: fix Test_PodIPPoolAdvert flakiness Since the test updates multiple k8s resources in each test run, we should be careful about how many of them are we actually changing between the individual test runs, to not trigger multiple reconciliations with different results (advertisements) in a single test run. This change ensures we change only one value that affects the advertisement between each test runs. Fixes: #30237 (tested locally with 100 x Test_PodIPPoolAdvert runs, which would always fail as in #30237 before this fix) /test /test /test /test /test
gharchive/pull-request
2024-03-13T08:28:51
2025-04-01T04:33:47.677121
{ "authors": [ "rastislavs" ], "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/pull/31365", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2744470462
identity: fix bug where fromNodes could be used to allow custom endpoint Currently when --enable-node-selector-labels flag is set to true user can use (from/to)Nodes sections in CNPs/CCNPs to allow/block access to/from nodes based on custom node label. The node identity contains not only labels filtered out by --node-labels flag, but also contains reserved:remote-node label as seen in this example: 33554556 node:scif.cz/node=worker reserved:remote-node The problematic part is in fromNodes/toNodes sections as those use the same matchLabels selector as classic fromEndpoints/toEndpoints. In case user specifies labels owned not only by node, but also by endpoint he/she might end up in a situation where unexpected endpoints might get access to users deployment. It is necessary to be either specific or better - prevent this situation. This commit fixes this by adding a matchExpression with reserved:remote-node key to the policy engine as all nodes have this label. "fromNodes": [ { "matchLabels": { "any:scif.cz/node": "worker" }, "matchExpressions": [ { "key": "reserved:remote-node", "operator": "Exists" } ] } ] identity: fix bug where fromNodes/toNodes could be used to allow custom endpoint /test
gharchive/pull-request
2024-12-17T10:01:06
2025-04-01T04:33:47.679837
{ "authors": [ "oblazek" ], "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/pull/36657", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
325649443
test: download exact k8s version of k8s upstream e2e Signed-off-by: André Martins andre@cilium.io test-upstream-k8s test-me-please test-missed-k8s test-me-please test-upstream-k8s test-me-please test-missed-k8s
gharchive/pull-request
2018-05-23T10:43:50
2025-04-01T04:33:47.682496
{ "authors": [ "aanm" ], "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/pull/4218", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
217008936
README: add Read the Docs link RtD let's you read the docs chapter by chapters and having a badge makes it more visible.   This pull request is now managed with CodeReviewHub. Refresh.
gharchive/pull-request
2017-03-25T22:15:19
2025-04-01T04:33:47.684977
{ "authors": [ "aanm", "scanf" ], "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/pull/459", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
529382707
eni: Check instance existence before resolving deficit This patch adds a check for instance existence before resolving IP deficit, this prevents operator from creating and attaching ENIs to non-existing instances cyclically. Also adds eth0 to instances in unit tests if necessary. Fixes: #9533 Signed-off-by: Jaff Cheng jaff.cheng.sh@gmail.com This change is  @ungureanuvladvictor @aanm @tgraf please take a look Coverage decreased (-0.01%) to 46.052% when pulling e9d16f2f96c6b5b5fdbc3135b217da659220e47b on jaffcheng:add-check-for-instance-existence into f694b3ba28ac2936c413fddec7771ef14a2c8808 on cilium:master. test-me-please
gharchive/pull-request
2019-11-27T14:30:56
2025-04-01T04:33:47.688751
{ "authors": [ "coveralls", "jaffcheng", "jrajahalme" ], "repo": "cilium/cilium", "url": "https://github.com/cilium/cilium/pull/9676", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2080183958
🛑 Cioos National CKAN Sitemap is down In 3c86669, Cioos National CKAN Sitemap (https://catalogue.cioos.ca/sitemap/sitemap-1.xml) was down: HTTP code: 0 Response time: 0 ms Resolved: Cioos National CKAN Sitemap is back up in 08b463e after 2 minutes.
gharchive/issue
2024-01-13T08:33:26
2025-04-01T04:33:47.709239
{ "authors": [ "fostermh" ], "repo": "cioos-siooc/cwatch-upptime", "url": "https://github.com/cioos-siooc/cwatch-upptime/issues/1035", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
892519615
Compile-time error when using Shapeless > 2.3.4 [error] Symbol 'type shapeless.ops.FnFromProductInstances.Aux' is missing from the classpath. [error] This symbol is required by 'value io.circe.generic.decoding.IncompleteDerivedDecoders.ffp'. [error] Make sure that type Aux is in your classpath and check for conflicting dependencies with `-Ylog-classpath`. [error] A full rebuild may help if 'IncompleteDerivedDecoders.class' was compiled against an incompatible version of shapeless.ops.FnFromProductInstances. See also #1743. Is there a reason that circe depends on an shapeless 2.3.3? 2.3.6 is binary compatible. Fixed by shapeless 2.3.7 in https://github.com/milessabin/shapeless/pull/1162.
gharchive/issue
2021-05-15T19:21:24
2025-04-01T04:33:47.710998
{ "authors": [ "armanbilge" ], "repo": "circe/circe", "url": "https://github.com/circe/circe/issues/1750", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
785396394
golang:1.16beta1 has not propagated to circleci images Hi, The Go 1.16 Beta 1 released was made on December 27th 2020 and the official Golang docker images received that release under the 1.16beta1 tag but that tag has not become available in the CircleCI images. I expected to see an image circleci/golang:1.16beta1 exist within 24 hours after the upstream golang:1.16beta1 posted, but it has not appeared. https://hub.docker.com/_/golang?tab=tags&page=1&ordering=last_updated&name=1.16beta I am hoping to use Go 1.16 beta1 CircleCI image here in https://github.com/stellar/go/pull/3288. Hi. CircleCI doesn't publish non-stable releases by default for Convenience Images. The next versions to publish for Go will be patch releases for v1.14.x or v1.15.x or the official v1.16 release when it comes out next month. Hi. CircleCI doesn't publish non-stable releases by default for Convenience Images. The next versions to publish for Go will be patch releases for v1.14.x or v1.15.x or the official v1.16 release when it comes out next month. CircleCI doesn't publish non-stable releases CircleCI publishes Go release-candidates, which are non-stable releases of Go. For example: 1.15rc2. Supporting beta and release candidate builds are critical for testing Go applications. Some folks in the Go community participate in testing beta and release candidates. Would it be possible to at least continue to publish release-candidate builds, and would it be possible to consider also publishing beta builds? CircleCI doesn't publish non-stable releases CircleCI publishes Go release-candidates, which are non-stable releases of Go. For example: 1.15rc2. Supporting beta and release candidate builds are critical for testing Go applications. Some folks in the Go community participate in testing beta and release candidates. Would it be possible to at least continue to publish release-candidate builds, and would it be possible to consider also publishing beta builds? Good point. Those RC images happen to have gotten through, they're not intended nor supported. As for this image, circleci/go, if RCs are already getting through, that would continue to work. I would like to point out though that this image is replaced by the next-gen Go image, cimg/go. You can learn more here. Next-gen images doesn't support RC and beta images yet but it is something I'm looking to try and support officially for 2021. You can stay informed on Go images changes here and here. Lastly, I created a CircleCI Idea that you can upvote and track this feature request. You can find that here. Good point. Those RC images happen to have gotten through, they're not intended nor supported. As for this image, circleci/go, if RCs are already getting through, that would continue to work. I would like to point out though that this image is replaced by the next-gen Go image, cimg/go. You can learn more here. Next-gen images doesn't support RC and beta images yet but it is something I'm looking to try and support officially for 2021. You can stay informed on Go images changes here and here. Lastly, I created a CircleCI Idea that you can upvote and track this feature request. You can find that here.
gharchive/issue
2021-01-13T19:51:04
2025-04-01T04:33:47.749192
{ "authors": [ "felicianotech", "leighmcculloch" ], "repo": "circleci/circleci-images", "url": "https://github.com/circleci/circleci-images/issues/531", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2078839183
replace statedb state-change calls with Workflow Event state-change calls update all statedb state change calls to use new class (process lambda and update-state lambda) Closed by #263
gharchive/issue
2024-01-12T13:18:24
2025-04-01T04:33:47.762738
{ "authors": [ "ircwaves" ], "repo": "cirrus-geo/cirrus-geo", "url": "https://github.com/cirrus-geo/cirrus-geo/issues/236", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1683795334
Add "Not enough permissions" alert to frontend when trying to import .csv files for "Global View" users Option to import/upload .csv files under "Manage Organizations" and "Manage Users" is not available for users with "Global View" permissions. However, there is notification system in place to let users know they are unable to import files. Recommend adding an alert to frontend, otherwise they will receive a "403" error with no way to know other than checking browser console. I think for "Global View" users, we should just hide the "import" section entirely instead. I agree, will update ticket to reflect. Testing with "standard" user account. Unable to test locally as "globalView" account needed. Linked PR #1912 which closes this issue. Hid "import/export" functionality for "global view" users.
gharchive/issue
2023-04-25T20:11:16
2025-04-01T04:33:47.764907
{ "authors": [ "ValeriaReveles", "epicfaace" ], "repo": "cisagov/crossfeed", "url": "https://github.com/cisagov/crossfeed/issues/1848", "license": "CC0-1.0", "license_type": "permissive", "license_source": "github-api" }
2264567651
Add a CONTRIBUTING file https://docs.github.com/en/communities/setting-up-your-project-for-healthy-contributions/setting-guidelines-for-repository-contributors This issue was migrated from an old repository: https://github.com/CiscoDevNet/catalystwan/issues/80
gharchive/issue
2024-04-25T22:00:56
2025-04-01T04:33:47.802370
{ "authors": [ "JimOverholt" ], "repo": "cisco-open/cisco-catalyst-wan-sdk", "url": "https://github.com/cisco-open/cisco-catalyst-wan-sdk/issues/531", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1615872535
Boolean in zod's schema doesn't persist in form after action result? When I enter a string value into a text field (e.g. email, password) and action is submitted, the page refreshes and the values stay in the form (bind:value works fine) However, the boolean field always refreshes to the default state (false). I've checked with SuperDebug to confirm the checked state is correctly passed into the form variable (it updates when I check/uncheck the checkbox). But after an action event (form submission), the boolean value becomes false, while all other fields remain as they were before the submission. +page.server.ts: ... const formSchema = z.object({ firstName: z .string({ required_error: "First name is required" }) .min(1, { message: "First name must be at least 1 characters long" }), lastName: z .string({ required_error: "Last name is required" }) .min(1, { message: "Last name must be at least 1 characters long" }), email: z .string({ required_error: "Email is required" }) .email({ message: "Enter a valid email address" }), password: z .string({ required_error: "Password is required" }) .min(6, { message: "Password must be at least 6 characters long" }), acceptedTos: z.boolean({ required_error: "You must accept the Terms of Service" }) }); ... +page.svelte ... export let data: PageData; let success: boolean = false; const { form, errors, message, submitting, constraints, enhance } = superForm(data.form, { autoFocusOnError: 'detect', clearOnSubmit: 'none', multipleSubmits: 'prevent', onResult(event) { if(event.result.type == "success") { success = true; setTimeout(() => { window.location.href = "/"; }, 500); } }, } ); ... <input id="terms" type="checkbox" class="w-4 h-4 border border-gray-300 rounded bg-gray-50 focus:ring-3 focus:ring-blue-300 dark:bg-gray-700 dark:border-gray-600 dark:focus:ring-blue-600 dark:ring-offset-gray-800 dark:focus:ring-offset-gray-800" bind:checked={$form.acceptedTos}> Hello, nice coincidence! I noticed the same thing yesterday but haven't had time to look at it yet. I'll take a look tomorrow. I think it has to do something with that a checkbox field is either undefined or set to its specified value. Btw, there is a way of making the "accept terms" work without testing the boolean value, see this FAQ entry. Right, adding name="acceptedTos" makes the value persist on form submission. Also, the name has to match with zod's name, else it seems to not persist. Yes, that's the required mapping if you're using the default dataType = 'form'. When set to json it serializes and posts the data structure itself, making the form fields mere placeholders for changing the value of the $form store.
gharchive/issue
2023-03-08T20:20:09
2025-04-01T04:33:47.812955
{ "authors": [ "Gintasz", "ciscoheat" ], "repo": "ciscoheat/sveltekit-superforms", "url": "https://github.com/ciscoheat/sveltekit-superforms/issues/19", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1292750389
grpc增加keepalive设置 客户端连接RPC接口时偶尔报 UNAVAILABLE: io exception 参考 Java netty/okhttp gRPC client throwing intermittent UNAVAILABLE: io exception while trying to connect - Stack Overflow 需要服务端和客户端都设置keep alive 只修改controller,内部微服务之间不增加。 tcp keep alive的目的是让客户端意外中断的网络链接尽快释放相关资源。操作系统默认的释放时间是2个小时,这个太长了。 http2 keep alive的目的是让服务端在客户端调用结束之后再保持一段时间,期间客户端再次发起调用就不需要重新建立链接,达到链接复用的目的。
gharchive/issue
2022-07-04T07:31:25
2025-04-01T04:33:47.814958
{ "authors": [ "rink1969" ], "repo": "cita-cloud/controller", "url": "https://github.com/cita-cloud/controller/issues/157", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2347990058
Add Prometheus Eval Client Adds the Prometheus Eval Client It will not work without using prompts followed Prometheus templates. similarity_scorer function is not implemented. Inference under torch.bfloat16 requires cuda environment with 24GB or more VRAM. @liwii A small question, I found that "" and '' are mixed up in the code, is there any policies of it? For example, '' is used in https://github.com/citadel-ai/langcheck/blob/c55681b08c33d6e6779a3b07f47ce86a0cc549cb/src/langcheck/metrics/eval_clients/_anthropic.py#L150, but "" is used in https://github.com/citadel-ai/langcheck/blob/c55681b08c33d6e6779a3b07f47ce86a0cc549cb/src/langcheck/metrics/eval_clients/_anthropic.py#L110. Ah yeah it is just inconsistent haha Our linter & formatter do not handle them properly now, but we'll fix that altogether in #123. You don't need to care too much about that in this PR! @liwii Hi, I merged https://github.com/citadel-ai/langcheck/pull/126 to this one and fixed the prompts' folder structure, to /metrics/prometheus created a new function load_prompt_template to handle different prompts Please take a review!
gharchive/pull-request
2024-06-12T07:07:52
2025-04-01T04:33:47.820397
{ "authors": [ "conan1024hao", "liwii" ], "repo": "citadel-ai/langcheck", "url": "https://github.com/citadel-ai/langcheck/pull/122", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
908437876
Spike: run selenium tests against demo app 👻 Just trying something cool, this worked
gharchive/pull-request
2021-06-01T15:47:27
2025-04-01T04:33:47.839049
{ "authors": [ "davidrapson" ], "repo": "citizensadvice/design-system", "url": "https://github.com/citizensadvice/design-system/pull/1188", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1048159592
Mobile UI Prototype to validate behavior Overview Create prototypes for v3 and v4 progress indicators for mobile. See Issue #887 for reference. Check out these resources for MUI stepper process, especially for how we would address the experience for mobile: https://material.io/archive/guidelines/components/steppers.html https://mui.com/components/steppers/ https://material.angular.io/components/stepper/overview
gharchive/issue
2021-11-09T03:54:55
2025-04-01T04:33:47.979938
{ "authors": [ "chalimar", "mariastudnicka" ], "repo": "civictechindex/CTI-website-frontend", "url": "https://github.com/civictechindex/CTI-website-frontend/issues/1035", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2174006871
Default PROGRAM_CARD_IMAGES flag to true Description Please explain the changes you made here. Release notes The title of the pull request will be used as the default release notes description. If more detail is needed to communicate to partners the scope of the PR's changes, use this release notes section. Checklist General Read the full guidelines for PRs here [ ] Added the correct label: < feature | enhancement | bug | dependencies | infrastructure | ignore-for-release | database > [ ] Assigned to a specific person, civiform/developers, or a more specific round-robin list. [ ] Created unit and/or browser tests which fail without the change (if possible) [ ] Performed manual testing (Chrome and Firefox if it includes front-end changes) [ ] Extended the README / documentation, if necessary Database evolutions Read the guidelines here [ ] Assigned two reviewers [ ] Guarded against already existing resources using IF NOT EXISTS and IF EXISTS [ ] Downs created to undo changes in Ups User visible changes [ ] Followed steps to internationalize new strings [ ] Added context strings to new messages [ ] Didn't use a message in applicant facing code that isn't translated yet (unless behind a flag) [ ] Wrote browser tests using the validateAccessibility method [ ] Tested on mobile view. See mobile device mode [ ] Manually tested at 200% size [ ] Manually evaluated tab order New Features [ ] Add new FeatureFlag env vars to server/conf/helper/feature-flags.conf [ ] Conditioned new functionality on a FeatureFlag [ ] Wrote browser tests with the feature flag off and on, etc. Instructions for manual testing If instructions are needed for manual testing by reviewers, include them here. Issue(s) this completes Fixes #<issue_number>; Fixes #<issue_number>... #6836 👈 main This stack of pull requests is managed by Graphite. Learn more about stacking. Join @caitlinshk and the rest of your teammates on Graphite
gharchive/pull-request
2024-03-07T14:33:26
2025-04-01T04:33:47.990470
{ "authors": [ "caitlinshk" ], "repo": "civiform/civiform", "url": "https://github.com/civiform/civiform/pull/6836", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1403808844
RegionFlag: parse user_data as String Hey, apologies but I am not sure the right way to do this. My first commit is just to naively read the whole user_data as a String even though I know it ends with a 0x95 at the end - which is not a valid ascii character. My second commit is doing this differently - omitting that character and then appending it when serializing. Let me know your thoughts. Cool, yep I think the second commit makes sense. That's basically how things are being handled in StringHash too. The null gets stripped and then later added back on during serialization. As far as I know they're all null terminated.
gharchive/pull-request
2022-10-10T23:24:22
2025-04-01T04:33:48.000348
{ "authors": [ "bryab", "cjab" ], "repo": "cjab/libeq", "url": "https://github.com/cjab/libeq/pull/27", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1294324621
mail format Do you know what format use Microsoft to store its mails in OST/PST files? HTML/EML/MESSAGE/etc. PST files have a proprietary format specification described here. AFAIK, the message body is stored either in plain text or in html. The java-libpst library, which pstconv uses to parse PST files, actually allows to get the message body in both formats (text and html) under certain conditions. For more information, please check the code from PSTMessage class, specifically the methods getBody() and getBodyHtml() .
gharchive/issue
2022-07-05T13:44:44
2025-04-01T04:33:48.031253
{ "authors": [ "bphd", "cjmach" ], "repo": "cjmach/pstconv", "url": "https://github.com/cjmach/pstconv/issues/11", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1510244391
机器人验证卡死 输入token后弹出的浏览器窗口会要求机器人验证 每次验证到一半就刷新了,仿佛陷入了轮回绝境。 用的是chrome,默认浏览器换成edge也是一样。 不是很懂,请问这个是什么bug或者我自己的问题么? 这种验证问题的bug大概率是vpn的锅因为访问太平凡了,或者ip地址可能被拉黑了 没办法换梯子
gharchive/issue
2022-12-25T04:17:25
2025-04-01T04:33:48.032871
{ "authors": [ "MuBai-He", "xBLACKICEx", "yuzuchann" ], "repo": "cjyaddone/ChatWaifuL2D", "url": "https://github.com/cjyaddone/ChatWaifuL2D/issues/4", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
19403996
Client does not detect AMD APU GPUs as OpenCL devices The "client" test/profile application does not detect the GPU component of AMD APUs as OpenCL devices. For example, when run on Windows 7 on an A10-6800K, only the CPU is detected as an OCL device when using the "-i" command to show CL platform information. Feature now available #72
gharchive/issue
2013-09-12T18:07:10
2025-04-01T04:33:48.126919
{ "authors": [ "anshuarya", "bragadeesh" ], "repo": "clMathLibraries/clFFT", "url": "https://github.com/clMathLibraries/clFFT/issues/20", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
688454411
rename project - misleading? This is likely more useful for other platforms now as its more portable and no longer locked to the pine Rock64 SBC. Renamed, from: https://github.com/clach04/rock64_vcgencmd To: https://github.com/clach04/fake_vcgencmd Will make an update to https://github.com/eidottermihi/rpicheck/issues/161
gharchive/issue
2020-08-29T04:12:52
2025-04-01T04:33:48.129175
{ "authors": [ "clach04" ], "repo": "clach04/fake_vcgencmd", "url": "https://github.com/clach04/fake_vcgencmd/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
174839607
Integer Hyperparameters Love optunity. I can't seem to figure out how to optimize an integer hyperparameter. Here is my function: @optunity.cross_validated(x=df_x.as_matrix(), y=df_y.as_matrix(), num_folds=5) def xgb_model(x_train, y_train, x_test, y_test, max_depth=9, min_child_weight=1, subsample=0.8, colsample_bytree=0.8): learning_rate = .1 n_estimators = 400 clf = XGBClassifier( learning_rate=learning_rate, n_estimators=n_estimators, max_depth=max_depth, min_child_weight=min_child_weight, gamma=0, subsample=subsample, colsample_bytree=colsample_bytree, objective= 'binary:logistic', max_delta_step = 1, # For imbalanced data set scale_pos_weight=scale_pos_weight, nthread=8, seed=27) msg("INFO", "Training model...") clf.fit(x_train, y_train, eval_metric='auc') y_prob = clf.predict_proba(x)[:,1] auc = optunity.metrics.roc_auc(y_test, y_prob) return auc But when I run it, optunity is passing a float as max_depth, which isn't allowed: ... File "/usr/local/lib/python2.7/site-packages/optunity/constraints.py", line 266, in func return f(*args, **kwargs) File "/usr/local/lib/python2.7/site-packages/optunity/cross_validation.py", line 403, in call scores.append(self.f(**kwargs)) File "./my-model.py", line 274, in xgb_model clf.fit(x_train, y_train, eval_metric='auc') ... File "/usr/local/lib/python2.7/site-packages/xgboost/core.py", line 806, in update _check_call(_LIB.XGBoosterUpdateOneIter(self.handle, iteration, dtrain.handle)) File "/usr/local/lib/python2.7/site-packages/xgboost/core.py", line 127, in _check_call raise XGBoostError(_LIB.XGBGetLastError()) XGBoostError: Invalid Parameter format for max_depth expect int but value='10.02734375' The variables are treated as float's in optunity, so you have to convert them back to int when calling XGBClassifier. For example: clf = XGBClassifier( learning_rate=learning_rate, n_estimators=n_estimators, max_depth=int(max_depth), ... Any variable that requires to be int has to be passed through int(x) or int(round(x)). Yes, we can convert the values, but that doesn't constrain the search space for optimization. The whole point is to constrain the search space. You're right, postprocessing the reals won't properly constrain the search space, but the results are nearly identical in terms of convergence speed. Note that Optunity's solvers are all continuous at this moment, so rounding is the best we can do at this point. In the upcoming release we will have more support for integer and other value types, and will also be able to recognize duplicate candidate solutions more effectively. That said, the internal solvers still remain continuous. We have no concrete plans to add integer/mixed type solvers at the moment.
gharchive/issue
2016-09-02T20:15:42
2025-04-01T04:33:48.135609
{ "authors": [ "claesenm", "jaak-s", "scottee" ], "repo": "claesenm/optunity", "url": "https://github.com/claesenm/optunity/issues/76", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
2080781902
Consider supporting the build server protocol Currently the interface between clangd and build tools is the JSON Compilation Database. However, this has a few disadvantages, particularly when using build systems like Bazel: It has to be generated first. In the Bazel case, via a separate tool. It must be kept up-to-date. In the case of Bazel, this means re-running this separate tool manually. I believe these issues could be solved by supporting the Build Server Protocol. If a discovery file exists, clangd could launch the BSP and request compiler flags etc for currently-open files using the cpp extension to this protocol. This information could be requested and kept up-to-date by clangd. Aside: The (de-facto?) Bazel BSP implementation seemingly doesn't support C++ ATM. Maybe there's only partial support so far, but it looks like it has some C++ support: https://github.com/JetBrains/bazel-bsp/pull/277 Ah, thanks! Good to know that. (and maybe the README needs an update too.)
gharchive/issue
2024-01-14T15:50:14
2025-04-01T04:33:48.143781
{ "authors": [ "cameron-martin", "zyn0217" ], "repo": "clangd/clangd", "url": "https://github.com/clangd/clangd/issues/1903", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
813050284
Adding an extra flag option Hi, I have been using the semantic highlight functionality to gray out sections of code skipped with #ifdefs, and it's been working great. However, I can't seem to find a solution to deal with compiler specific logic since there is no way to pass additional flags to clang in the coc config. I have tried adding a .clangd configuration file for my project as described in the clangd documentation, but it has no effect. Would you consider adding a new coc configuration field for extra clang flags? .clangd file should work for the case you've described. Could you give more details about what you tried, expected and really happened? preferably with clangd logs. .clangd file should work for the case you've described. Could you give more details about what you tried, expected and really happened? preferably with clangd logs. I placed a .clangd file at my project root, and typed the following lines in it: CompilerFlags Add: [-D__GNUC__, -U__clang__] I have the following snippet of code in my project: #if defined(__clang__) spdlog::warn("CUDA GPU mode not supported by libraries compiled using clang, recompile use gcc"); #else if ( #ifdef _MSC_VER RunTimeLinkSuccess("OrtSessionOptionsAppendExecutionProvider_CUDA") #else OrtSessionOptionsAppendExecutionProvider_CUDA #endif ) And I am expecting to see the section of code wrapped in the #if defined(__clang__) macro to be grayed out, but it didn't. I have enabled verbose logging via coc config but I don't know where clangd saves its log file. it is CompileFlags not CompilerFlags and you should have an additional colon : after CompileFlags:. I've tested with: .clangd: CompileFlags: Add: [-U__clang__] a.cc: #ifdef __clang__ static_assert(false); #else static_assert(false); #endif and I can verify that I am seeing the warning on the else branch. you can check coc documentation to figure out how to get to logs, but at some point it was possible to see via :CocInfo. The logs would've also contained errors about unknown keys in the config too. it is CompileFlags not CompilerFlags and you should have an additional colon : after CompileFlags:. I've tested with: .clangd: CompileFlags: Add: [-U__clang__] a.cc: #ifdef __clang__ static_assert(false); #else static_assert(false); #endif and I can verify that I am seeing the warning on the else branch. you can check coc documentation to figure out how to get to logs, but at some point it was possible to see via :CocInfo. The logs would've also contained errors about unknown keys in the config too. Thanks, I got it working now.
gharchive/issue
2021-02-22T02:11:01
2025-04-01T04:33:48.151165
{ "authors": [ "hyt589", "kadircet" ], "repo": "clangd/coc-clangd", "url": "https://github.com/clangd/coc-clangd/issues/194", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
85932257
Images flip When using fit_to_width() images rotate anticlockwise 90 degrees. Not good. $img = new SimpleImage(DIR_FS_IMAGES . $filename); $img->fit_to_width(250)->save(DIR_FS_IMAGES . '250_' . $filename); $img->fit_to_width(100)->save(DIR_FS_IMAGES . '100_' . $filename); $img->fit_to_width(50)->save(DIR_FS_IMAGES . '50_' . $filename); Original image is 2448px x 3264px As I understand it, your class should use exif_read_data() to make sure the image is oriented correctly: http://php.net/manual/en/function.exif-read-data.php#76964 I resaved the image using Irfanview and now the image is correctly oriented after resize using fit_to_width(). You need to use the auto_orient() method to fix images saved with a different orientation. It will not happen automatically. Don't you think that the class should itself use $this->auto_orient()? I think it should happen automatically, this is why we use classes for specific purposes. Not really. The class isn't just for loading existing images, and it supports formats other than JPG as well. More importantly, the auto_orient() method does a rotation that may not be desirable on large images where PHP doesn't have enough memory to handle it. Best to give the developer the option. If you prefer to use it, just chain it. A few extra characters won't bloat your code. So I tried the following: $img->fit_to_width(100)->auto_orient()->save(DIR_FS_IMAGES . '100_' . $filename); and: $img->auto_orient()->fit_to_width(100)->save(DIR_FS_IMAGES . '100_' . $filename); Didn't work! The auto_orient() method has been tested and works fine. Chances are, your original image doesn't have the correct EXIF data. Closing since this is an implementation issue, not an issue with the library.
gharchive/issue
2015-06-07T19:34:40
2025-04-01T04:33:48.245721
{ "authors": [ "claviska", "osCMattiFinn" ], "repo": "claviska/SimpleImage", "url": "https://github.com/claviska/SimpleImage/issues/85", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1159973346
Cocoapods UI issues On the main page, the cocoapods icon is not used in the components list. On the component page, there is no icon and the version number does not have a link to the component like the other types do. On the harvest page, the search box disappears when you select cocoapods as the type. The fix for "the version number does not have a link to the component like the other types do" is in the service. See https://github.com/clearlydefined/service/pull/934.
gharchive/issue
2022-03-04T19:28:33
2025-04-01T04:33:48.260415
{ "authors": [ "capfei", "qtomlinson" ], "repo": "clearlydefined/website", "url": "https://github.com/clearlydefined/website/issues/963", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
296165290
Fix ledger_processor::change_block block_put Now it matches the other block types. I think this might have also been part of #598, as I got a different segfault after running it for a while. My wallet has now been running for 30 minutes without problems with this newest commit. Previously, it consistently crashed within 10 seconds.
gharchive/pull-request
2018-02-11T06:22:22
2025-04-01T04:33:48.263418
{ "authors": [ "PlasmaPower" ], "repo": "clemahieu/raiblocks", "url": "https://github.com/clemahieu/raiblocks/pull/604", "license": "bsd-2-clause", "license_type": "permissive", "license_source": "bigquery" }
2534324859
move version-specific code to CGameVersionManager please move it to CGameVersionManager, it's better to keep all hooks in one place for future support of more versions/games. can be done in a follow-up PR Originally posted by @x87 in https://github.com/cleolibrary/CLEO5/pull/197#discussion_r1765149358 CGameVersionManager should be rewritten to use text identifiers instead enum. Then CLEO export can be added so other plugins can benefit it too. It would also make possible to display meaningful error messages about some function not being supported in current game version.
gharchive/issue
2024-09-18T17:47:17
2025-04-01T04:33:48.277572
{ "authors": [ "MiranDMC", "x87" ], "repo": "cleolibrary/CLEO5", "url": "https://github.com/cleolibrary/CLEO5/issues/198", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1115044672
🛑 Click Engage is down In 092caa0, Click Engage (https://example.click-engage.co.uk) was down: HTTP code: 403 Response time: 1512 ms Resolved: Click Engage is back up in df5adc9.
gharchive/issue
2022-01-26T13:54:22
2025-04-01T04:33:48.293832
{ "authors": [ "kieranbrown" ], "repo": "clickdealer/upptime", "url": "https://github.com/clickdealer/upptime/issues/162", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
168012655
Add lein-kibit and check code This codebase is old and wasn't originally very idiomatic. Some bits have been updated, but it would be good to check everything. This was actually done already.
gharchive/issue
2016-07-28T04:41:17
2025-04-01T04:33:48.417598
{ "authors": [ "oubiwann" ], "repo": "clojusc/clj-starlanes", "url": "https://github.com/clojusc/clj-starlanes/issues/38", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
900046428
DC provisioning fails after installing Red Team tools Operating System Version: Debian GNU/Linux 10 (buster) Deploying via (VirtualBox/VMWare/AWS/Azure/ESXi): Virtualbox Vagrant Version (if applicable): 2.2.14 Description of the issue: While building the DC host, I'm running into the following error message that causes the build to stop: Classic fixes like re-running the provisioning process or destroying the entire detection lab did not help. Hi @JoSeiler - looks like this broke because the latest purplesharp release changed their binary naming scheme. This is fixed in https://github.com/clong/DetectionLab/commit/a578865e8d1b0184af7810ae286e5e9da673b3d4 ! Thanks a lot for the quick fix!
gharchive/issue
2021-05-24T21:55:14
2025-04-01T04:33:48.439791
{ "authors": [ "JoSeiler", "clong" ], "repo": "clong/DetectionLab", "url": "https://github.com/clong/DetectionLab/issues/657", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
446959048
Suggestion: change the argument name '-monitoring' to '-monitor' Suggestion: change the argument name '-monitoring' to '-monitor' @seokho-son How about your opinion? @powerkimhub why not? it seems a nice suggestion. Let me handle this :)
gharchive/issue
2019-05-22T06:43:11
2025-04-01T04:33:48.445005
{ "authors": [ "jihoon-seo", "powerkimhub", "seokho-son" ], "repo": "cloud-barista/poc-farmoni", "url": "https://github.com/cloud-barista/poc-farmoni/issues/1", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1049955886
Replace uuid variable that was empty with one with the actual UUID Description The benchmark_uuid was not in use anywhere and never set by the test so when it would go to compare the uuid of the running test would be empty. Replaced this with the UUID set for the run. Fixes closing since this fix is outdated , we no longer use the compare_uperf_uuid var or the assign uuid function.
gharchive/pull-request
2021-11-10T15:19:07
2025-04-01T04:33:48.446267
{ "authors": [ "Harshith-umesh", "dry923" ], "repo": "cloud-bulldozer/e2e-benchmarking", "url": "https://github.com/cloud-bulldozer/e2e-benchmarking/pull/281", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
482022783
Add new auth configuration options. PR to solve #165 Thank you!
gharchive/pull-request
2019-08-18T18:25:32
2025-04-01T04:33:48.478136
{ "authors": [ "leitgab", "paulfantom" ], "repo": "cloudalchemy/ansible-grafana", "url": "https://github.com/cloudalchemy/ansible-grafana/pull/166", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2669015678
[WIP] Add extra-specs option to add SecurityGroups to the runner adds extra-specs option to add Security Groups to the instances (until now it was used the default security group) Whoops. It was a WIP, but it looked good :). Add another PR if there's anything else needed
gharchive/pull-request
2024-11-18T15:45:10
2025-04-01T04:33:48.490033
{ "authors": [ "fabi200123", "gabriel-samfira" ], "repo": "cloudbase/garm-provider-aws", "url": "https://github.com/cloudbase/garm-provider-aws/pull/19", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1599756420
🛑 sitemaya.com is down In d024d04, sitemaya.com (https://sitemaya.com) was down: HTTP code: 0 Response time: 0 ms Resolved: sitemaya.com is back up in 00c5f14.
gharchive/issue
2023-02-25T16:20:44
2025-04-01T04:33:48.492956
{ "authors": [ "cloudbip" ], "repo": "cloudbip/upptime", "url": "https://github.com/cloudbip/upptime/issues/10422", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1627881472
🛑 cbn.id is down In 3fd4ef4, cbn.id (https://cbn.id) was down: HTTP code: 0 Response time: 0 ms Resolved: cbn.id is back up in 417f851.
gharchive/issue
2023-03-16T16:40:54
2025-04-01T04:33:48.495774
{ "authors": [ "cloudbip" ], "repo": "cloudbip/upptime", "url": "https://github.com/cloudbip/upptime/issues/10864", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1695286639
🛑 hostingan.id is down In 5b0790b, hostingan.id (https://hostingan.id) was down: HTTP code: 0 Response time: 0 ms Resolved: hostingan.id is back up in af22dae.
gharchive/issue
2023-05-04T05:29:00
2025-04-01T04:33:48.498634
{ "authors": [ "cloudbip" ], "repo": "cloudbip/upptime", "url": "https://github.com/cloudbip/upptime/issues/12106", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2039642158
🛑 telkomsigma.co.id is down In 1f71f2f, telkomsigma.co.id (https://www.telkomsigma.co.id) was down: HTTP code: 0 Response time: 0 ms Resolved: telkomsigma.co.id is back up in 60b475e after 15 minutes.
gharchive/issue
2023-12-13T12:47:03
2025-04-01T04:33:48.501819
{ "authors": [ "cloudbip" ], "repo": "cloudbip/upptime", "url": "https://github.com/cloudbip/upptime/issues/14651", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1115871175
🛑 hostinger.co.id is down In a8dd8da, hostinger.co.id (https://hostinger.co.id) was down: HTTP code: 403 Response time: 213 ms Resolved: hostinger.co.id is back up in 14e236e.
gharchive/issue
2022-01-27T07:35:05
2025-04-01T04:33:48.504751
{ "authors": [ "cloudbip" ], "repo": "cloudbip/upptime", "url": "https://github.com/cloudbip/upptime/issues/2441", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1118390986
🛑 puskomedia.id is down In c5ea6cd, puskomedia.id (https://puskomedia.id) was down: HTTP code: 0 Response time: 0 ms Resolved: puskomedia.id is back up in 8ee0883.
gharchive/issue
2022-01-30T03:01:20
2025-04-01T04:33:48.507662
{ "authors": [ "cloudbip" ], "repo": "cloudbip/upptime", "url": "https://github.com/cloudbip/upptime/issues/2491", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1043242552
🛑 daeliweb.com is down In fa8c5e5, daeliweb.com (https://daeliweb.com) was down: HTTP code: 0 Response time: 0 ms Resolved: daeliweb.com is back up in 5ae5020.
gharchive/issue
2021-11-03T09:02:59
2025-04-01T04:33:48.510547
{ "authors": [ "cloudbip" ], "repo": "cloudbip/upptime", "url": "https://github.com/cloudbip/upptime/issues/979", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
375098304
[Feature] Finalise test helpers for 1.0 This commit finalises the test helpers ready for the 1.0 release. The new test assertions provide significantly better PHPUnit output when they fail. Also tidied up the test response class, as the method names were starting to get quite inconsistent and unpredictable. Methods that were marked as deprecated have been removed, and those that were not marked as deprecated have now been marked as deprecated for 2.0. Is there anything to do besides checking for the usage of the deprecated/removed methods? Because with the new version a lot of my tests fail with the following error (lines differ): Error : Class 'CloudCreativity\JsonApi\Testing\DocumentTester' not found /home/jannis/work/backend/mod-base/vendor/cloudcreativity/laravel-json-api/src/Testing/TestResponse.php:134 /home/jannis/work/backend/mod-base/vendor/cloudcreativity/laravel-json-api/src/Testing/TestResponse.php:292 /home/jannis/work/backend/mod-base/vendor/cloudcreativity/laravel-json-api/src/Testing/TestResponse.php:307 /home/jannis/work/backend/mod-base/tests/Feature/JsonApi/Web/UsersTest.php:83 You'll need to update the dev dependency as well as that class doesn't exist any more. There's some upgrade notes here: https://github.com/cloudcreativity/laravel-json-api/blob/develop/docs/upgrade.md I.e. follow the upgrade notes for 1.0.0-beta.6. Let me know if that solves it because if it doesn't I'll need to amend the upgrade notes. @lindyhopchris I followed the upgrade guide, removed the json-api-testing package and installed it again with version ^1.0. According to composer info I have version 1.x-dev c973e12 installed but the error still occurs. DocumentTester does not appear anywhere in the code base so this is some sort of composer issue. Out of interest, do you do a Composer install with an optimized autoloader? Try composer dump-autoload and then re-run the tests. I had the wrong version of the library itself installed. I was on beta5, installing dev-develop fixed the error. I thought those changes were already tagged as a new release. :) Ah no problem! No haven't tagged them yet. There's a chance the next tag might be 1.0.0-rc1!
gharchive/pull-request
2018-10-29T16:16:32
2025-04-01T04:33:48.518043
{ "authors": [ "jannis-a", "lindyhopchris" ], "repo": "cloudcreativity/laravel-json-api", "url": "https://github.com/cloudcreativity/laravel-json-api/pull/249", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
823095895
SQL Scratchpad example updates All components linked together Autocomplete of tables, udfs etc works Statement execution with results works Left to do: Some styling It doesn't remember result from multiple queries, i.e. execute two queries, place cursor in the first. Unrelated circleci issue '.circleci/config.yml
gharchive/pull-request
2021-03-05T13:37:31
2025-04-01T04:33:48.524617
{ "authors": [ "JohanAhlen", "romainr" ], "repo": "cloudera/hue", "url": "https://github.com/cloudera/hue/pull/1856", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2357356475
No SDK equiv for deleteZoneRulesetRule / "Delete a zone ruleset rule" Confirmation [X] My issue isn't already found on the issue tracker. [X] I have replicated my issue using the latest version of the library and it is still present. cloudflare-go version v0.97.0 Go environment GOARCH='amd64' GOBIN='/Users/tlimoncelli/bin' GOCACHE='/Users/tlimoncelli/Library/Caches/go-build' GOENV='/Users/tlimoncelli/Library/Application Support/go/env' GOEXE='' GOEXPERIMENT='' GOFLAGS='' GOHOSTARCH='amd64' GOHOSTOS='darwin' GOINSECURE='' GOMODCACHE='/Users/tlimoncelli/go/pkg/mod' GONOPROXY='' GONOSUMDB='' GOOS='darwin' GOPATH='/Users/tlimoncelli/go' GOPRIVATE='' GOPROXY='https://proxy.golang.org,direct' GOROOT='/usr/local/go' GOSUMDB='sum.golang.org' GOTMPDIR='' GOTOOLCHAIN='auto' GOTOOLDIR='/usr/local/go/pkg/tool/darwin_amd64' GOVCS='' GOVERSION='go1.22.3' GCCGO='gccgo' GOAMD64='v1' AR='ar' CC='clang' CXX='clang++' CGO_ENABLED='1' GOMOD='/Users/tlimoncelli/git/cloudflare-go/go.mod' GOWORK='' CGO_CFLAGS='-O2 -g' CGO_CPPFLAGS='' CGO_CXXFLAGS='-O2 -g' CGO_FFLAGS='-O2 -g' CGO_LDFLAGS='-O2 -g' PKG_CONFIG='pkg-config' GOGCCFLAGS='-fPIC -arch x86_64 -m64 -pthread -fno-caret-diagnostics -Qunused-arguments -fmessage-length=0 -ffile-prefix-map=/var/folders/ry/6mvxxmy52kl_cd939xsl42y00000gn/T/go-build693033671=/tmp/go-build -gno-record-gcc-switches -fno-common' Expected output There is no SDK equivalent for this route: https://developers.cloudflare.com/api/operations/deleteZoneRulesetRule There is DeleteRuleset() but that is for the ruleset, not the individual rule (DeleteRulesetRule(). There should probably be equivs for createZoneRulesetRule, deleteZoneRulesetRule and updateZoneRulesetRule Actual output n/a Code demonstrating the issue n/a Steps to reproduce n/a References https://developers.cloudflare.com/api/operations/createZoneRulesetRule https://developers.cloudflare.com/api/operations/deleteZoneRulesetRule https://developers.cloudflare.com/api/operations/updateZoneRulesetRule 👋 support for manipulating ruleset rules isn't really advised due to some confusing situations you can get yourself into if you're not updating the ruleset as a whole, however, if you really want to go ahead with it, v2 does have support for ruleset rules - https://github.com/cloudflare/cloudflare-go/blob/v2/api.md#rules-4 as we are focusing efforts on v2, i don't think we'll be backporting the functionality to v0.x. if this is something you need in v0, you're welcome to drop in a PR and i'll take a look.
gharchive/issue
2024-06-17T13:28:27
2025-04-01T04:33:48.544581
{ "authors": [ "jacobbednarz", "tlimoncelli" ], "repo": "cloudflare/cloudflare-go", "url": "https://github.com/cloudflare/cloudflare-go/issues/2557", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
1525983869
WDAPI-1221 Fix Update Method to be PUT from PATCH Provide a general summary of your changes in the title above. You should remove this overview, any sections and any section descriptions you don't need below before submitting. There isn't a strict requirement to use this template if you can structure your description and still cover these points. Description Bug fix change update method from PATCH To PUT which API expects Has your change been tested? Yes Locally using terraform-provider as well. Screenshots (if appropriate): Types of changes What sort of change does your code introduce/modify? [x] Bug fix (non-breaking change which fixes an issue) [ ] New feature (non-breaking change which adds functionality) [ ] Breaking change (fix or feature that would cause existing functionality to change) Checklist: [x] My code follows the code style of this project. [ ] My change requires a change to the documentation. [ ] I have updated the documentation accordingly. [x] I have added tests to cover my changes. [x] All new and existing tests passed. [x] This change is using publicly documented (api.cloudflare.com or developers.cloudflare.com) and stable APIs. thanks @suhrit-cf 👏
gharchive/pull-request
2023-01-09T17:11:02
2025-04-01T04:33:48.549981
{ "authors": [ "jacobbednarz", "suhrit-cf" ], "repo": "cloudflare/cloudflare-go", "url": "https://github.com/cloudflare/cloudflare-go/pull/1172", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
1453755800
🐛 BUG: Can't execute D1 SQL What version of Wrangler are you using? 2.3.2 What operating system are you using? Windows with cmd.exe Describe the Bug I'm following the Get Started guide, pretty much verbatim, in the docs and I'm on this step: wrangler d1 execute <DATABASE_NAME> --local --file=./schema.sql But I'm getting this error: Packages @databases/sql, @databases/split-sql-query not available locally. Attempting to use npx to install temporarily. Installing... (npx --prefer-offline -y -p @databases/sql@3.2.0 -p @databases/split-sql-query@1.0.3) X [ERROR] npx-import failed for @databases/sql,@databases/split-sql-query with message: Failed to find temporary install directory. Looking for paths matching '\npm-cache\_npx\' in: [lots of lines printing out my system path] You should install @databases/sql, @databases/split-sql-query locally: npm install --save-dev @databases/sql@3.2.0 @databases/split-sql-query@1.0.3 However, I did run that last line, npm install --save-dev @databases/sql@3.2.0 @databases/split-sql-query@1.0.3 and they are installed locally (present in both my package.json and node_modules folder). I even restarted my shell after installing them. Node: v16.13.1 npm: v8.5.5 Adding "wrangler": "^0.0.0-7c228171" to my dependencies and running wrangler with npx fixed this issue for me. I'm experiencing the same problem. @scratchyone's suggestion didn't work for me. v19.1.0 npm --version 8.19.3 wrangler --version ⛅️ wrangler 2.4.4 ------------------- ✘ [ERROR] npx-import failed for @databases/sql,@databases/split-sql-query with message: Failed to find temporary install directory. Looking for paths matching '/.npm/_npx/' in: [snip] You should install @databases/sql, @databases/split-sql-query locally: npm install --save-dev @databases/sql@3.2.0 @databases/split-sql-query@1.0.3``` I already tried that last command and I can see in the current directory that `node_modules` has modules for both of these packages. I only know enough about npm/node to be dangerous so I may be overlooking something dumb. :) I've just faced this issue today. The same command was working yesterday. Not sure what happened. 🤔. I'm using the latest version of wrangler, 2.6.2 I worked around the problem by uninstalling nvm and Node 18.12.1 LTS and used the node installer from https://nodejs.org/en/ for the same version, 18.12.1 LTS. This fixed it for me 🤷🏻🤷🏻 meet similar issue too using Node v18.15.0, Windows, installed via Scoop 🌀 Mapping SQL input into an array of statements 🌀 Loading DB at .wrangler\state\d1\DB.sqlite3 [NPXI] better-sqlite3 not available locally. Attempting to use npx to install temporarily. [NPXI] Installing... (npx --prefer-offline -y -p better-sqlite3@8.0.1) X [ERROR] npx-import failed for better-sqlite3 with message: Failed to find temporary install directory. Looking for paths matching '\npm-cache\_npx\' in:
gharchive/issue
2022-11-17T17:59:40
2025-04-01T04:33:48.557406
{ "authors": [ "Keavon", "StoneMoe", "ben-xD", "icebourg", "scratchyone" ], "repo": "cloudflare/workers-sdk", "url": "https://github.com/cloudflare/workers-sdk/issues/2220", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1717425280
[C3] Offer git initialisation and make initial commit Fixes #3275. What this PR solves / how to test: Offer git initialisation with a prompt Can also specify via --git or --no-git args Will not offer if already in a git repo Offers after the framework tool has run, in case it wants to init git itself and doesn't check if already init'd Commits after npm/yarn/pnpm install step so lock files (specific to package manager) are committed too Uses wrangler initialiseGit function, which tries to set the initial branch name to 'main' Somehow the spinner usage has become a little buggy, so still need to sort that... Associated docs issue(s)/PR(s): [insert associated docs issue(s)/PR(s)] Author has included the following, where applicable: [ ] Tests [ ] Changeset (Changeset guidelines) Reviewer is to perform the following, as applicable: Checked for inclusion of relevant tests Checked for inclusion of a relevant changeset Checked for creation of associated docs updates Manually pulled down the changes and spot-tested One major problem with this right now is that using a spinner on a command that is interactive is messed up, since the spinner update overwrites any logs coming from the command. I think the spinner should only be shown if silent is true. It does seem like that. But I believe that was the case before my changes too – the command pipe'd to stdout and the spinner was running at the same time Not for the point where the processing gets delegated to the framework. There was no spinner in that case. Not for the point where the processing gets delegated to the framework. There was no spinner in that case. You're right. I missed that if startText is not provided, spinner.start() was not called. I've added a useSpinner option to make this explicit. runCommand() now calls printAsyncStatus with useSpinner: opts?.silent e2e tests are failing at the moment because they get stuck on the git question. You should add the --no-git option here.
gharchive/pull-request
2023-05-19T15:20:26
2025-04-01T04:33:48.565826
{ "authors": [ "RamIdeas", "jculvey", "petebacondarwin" ], "repo": "cloudflare/workers-sdk", "url": "https://github.com/cloudflare/workers-sdk/pull/3296", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1327740965
🚀 Feature Request: Mass Secrets Management This pairs with https://github.com/cloudflare/wrangler-action/issues/79. It would be good to have a way to set mass secrets within wrangler, so within wrangler-action, you aren't meant with repetitive output, which can look very much like an error to an unsuspecting user. Given the TTY nature of wrangler secret put I'm not entirely sure how this could work, other than maybe some kind of wrangler secret mass-put 10 which prompts for key, then value, 10 times? Hopefully y'all have a better idea. 😅 Thanks for the suggestion! I will put it up for discussion with the team 😄 Considering an approach involving piping in a file.
gharchive/issue
2022-08-03T20:19:59
2025-04-01T04:33:48.568496
{ "authors": [ "Cherry", "JacobMGEvans" ], "repo": "cloudflare/wrangler2", "url": "https://github.com/cloudflare/wrangler2/issues/1610", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1107734077
refactor: tidy up the typings of the build result in dev In #262 some of the strict null fixes were removed to resolve a regression. This refactor re-applies these fixes in a way that avoids that problem. I ran wrangler dev on a local project after this change and it appears to be OK. Not sure what the specific failing use case was that #262 was resolving? That test should be sufficient for this. Previously, when it started, the effect would run twice, and it would try to call resolve result.stop and throw because result was undefined
gharchive/pull-request
2022-01-19T07:04:01
2025-04-01T04:33:48.570203
{ "authors": [ "petebacondarwin", "threepointone" ], "repo": "cloudflare/wrangler2", "url": "https://github.com/cloudflare/wrangler2/pull/267", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1117464266
Enable GPU acceleration in containers It is possible to have GPU acceleration in docker containers in WSL as explained in this guide. However, Installing the Nvidia Docker Toolkit is not so simple on dockerinwsl, since alpine has no pre-built package provided by Nvidia (list of supported distros). This could be solved by either: switching to an Ubuntu-based distro for dockerinwsl (or providing another flavor that is based on Ubuntu and has GPU acceleration) building the Nvidia toolkit for apline and bundling it along with the base image. FYI: It might be needed to switch to ubuntu anyways (see #13). Maybe this gets easier then. DockerInWSL 1.2.x is now based on Ubuntu. Feel free to experiment with. You can enter the distro by executing wsl -d clf_dockerinwsl in a normal shell. After that you should be able to use it like any other wsl2 ubuntu installation (besides some networking weirdness because of docker and vpnkit). Would be great if you could add your experiments to our wiki (https://github.com/cloudflightio/dockerinwsl/wiki) :) I just tried to enable GPU acceleration. These steps worked for me: https://docs.nvidia.com/cuda/wsl-user-guide/index.html#ch04-sub02-install-nvidia-docker This project has been discontinued.
gharchive/issue
2022-01-28T14:23:38
2025-04-01T04:33:48.575271
{ "authors": [ "a-to-the-5", "alexanderfreller", "mriedmann" ], "repo": "cloudflightio/dockerinwsl", "url": "https://github.com/cloudflightio/dockerinwsl/issues/16", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
303817255
Error: Data Dog cluster monitoring for PCF 2.0 failing --'datadog-firehose-nozzle/daaaaaa-aaaaa (0)' is not running after update. Review logs for failed jobs: datadog-firehose-nozzle Dear All, While applying the Data Dog cluster monitoring tile, I'm encountering the below error wherein the deployment fails and the VM status shows unresponsive: PFB error from the task -- debug /var/vcap/packages/director/gem_home/ruby/2.4.0/gems/logging-2.2.2/lib/logging/diagnostic_context.rb:474:in block in create_with_logging_context' D, [2018-03-09T11:11:31 #27302] [task:1653602] DEBUG -- DirectorJobRunner: (0.000633s) SELECT * FROM tasks WHERE (id= 1653602) LIMIT 1 D, [2018-03-09T11:11:31 #27302] [task:1653602] DEBUG -- DirectorJobRunner: (0.000154s) BEGIN D, [2018-03-09T11:11:31 #27302] [task:1653602] DEBUG -- DirectorJobRunner: (0.000361s) UPDATEtasksSETstate= 'error',timestamp= '2018-03-09 11:11:31',description= 'create deployment',result= '\'datadog-firehose-nozzle/d816133b-d871-490a-906b-c0c58ca78a77 (0)\' is not running after update. Review logs for failed jobs:...',output= '/var/vcap/store/director/tasks/1653602',checkpoint_time= '2018-03-09 11:11:11',type= 'update_deployment',username= 'ops_manager',deployment_name= 'datadog-216ae0a504ea83285898',started_at= '2018-03-09 11:06:11',event_output= '{\"time\":1520593571,\"stage\":\"Preparing deployment\",\"tags\":[],\"total\":1,\"task\":\"Preparing deployment\",\"index\":1,\"state\":\"started\",\"progress\":0}\n{\"time\":1520593572,\"stage\":\"Preparing deployment\",\"tags\":[],\"total\":1,\"task\":\"Preparing deployment\",\"index\":1,\"state\":\"finished\",\"progress\":100}\n{\"time\":1520593573,\"stage\":\"Preparing package compilation\",\"tags\":[],\"total\":1,\"task\":\"Finding packages to compile\",\"index\":1,\"state\":\"started\",\"progress\":0}\n{\"time\":1520593573,\"stage\":\"Preparing package compilation\",\"tags\":[],\"total\":1,\"task\":\"Finding packages to compile\",\"index\":1,\"state\":\"finished\",\"progress\":100}\n{\"time\":1520593573,\"stage\":\"Updating instance\",\"tags\":[\"datadog-firehose-nozzle\"],\"total\":1,\"task\":\"datadog-firehose-nozzle/d816133b-d871-490a-906b-c0c58ca78a77 (0) (canary)\",\"index\":1,\"state\":\"started\",\"progress\":0}\n{\"time\":1520593891,\"stage\":\"Updating instance\",\"tags\":[\"datadog-firehose-nozzle\"],\"total\":1,\"task\":\"datadog-firehose-nozzle/d816133b-d871-490a-906b-c0c58ca78a77 (0) (canary)\",\"index\":1,\"state\":\"failed\",\"progress\":100,\"data\":{\"error\":\"\'datadog-firehose-nozzle/d816133b-d871-490a-906b-c0c58ca78a77 (0)\' is not running after update. Review logs for failed jobs: datadog-firehose-nozzle\"}}\n{\"time\":1520593891,\"error\":{\"code\":400007,\"message\":\"\'datadog-firehose-nozzle/d816133b-d871-490a-906b-c0c58ca78a77 (0)\' is not running after update. Review logs for failed jobs: datadog-firehose-nozzle\"}}\n',result_output= '',context_id = '' WHERE (id` = 1653602) LIMIT 1 D, [2018-03-09T11:11:31 #27302] [task:1653602] DEBUG -- DirectorJobRunner: (0.005428s) COMMIT I, [2018-03-09T11:11:31 #27302] [] INFO -- DirectorJobRunner: Task took 5 minutes 20.84445033999998 seconds to process. Also, output from “/var/vcap/sys/log/datadog-firehose-nozzle/ datadog-firehose-nozzle.stdout.log” {"timestamp":1520594219.480720043,"process_id":3501,"source":"datadog-firehose-nozzle","log_level":"error","message":"Error while reading from the firehose: Error dialing trafficcontroller server: websocket: bad handshake.\nPlease ask your Cloud Foundry Operator to check the platform configuration (trafficcontroller is wss://doppler.syslab.pcflabcog.com).","data":null,"file":"/var/vcap/data/compile/datadog-firehose-nozzle/datadog-firehose-nozzle/src/github.com/DataDog/datadog-firehose-nozzle/datadogfirehosenozzle/datadog_firehose_nozzle.go","line":273,"method":"github.com/DataDog/datadog-firehose-nozzle/datadogfirehosenozzle.(*DatadogFirehoseNozzle).handleError"} {"timestamp":1520594219.480772257,"process_id":3501,"source":"datadog-firehose-nozzle","log_level":"info","message":"Closing connection with traffic controller due to Error dialing trafficcontroller server: websocket: bad handshake.\nPlease ask your Cloud Foundry Operator to check the platform configuration (trafficcontroller is wss://doppler.syslab.pcflabcog.com).","data":null,"file":"/var/vcap/data/compile/datadog-firehose-nozzle/datadog-firehose-nozzle/src/github.com/DataDog/datadog-firehose-nozzle/datadogfirehosenozzle/datadog_firehose_nozzle.go","line":277,"method":"github.com/DataDog/datadog-firehose-nozzle/datadogfirehosenozzle.(*DatadogFirehoseNozzle).handleError"} Thanks and Regards, Gowri Shankar @gowrshan this is officially maintained by datadog now. You should file this issue on their repo for the nozzle.
gharchive/issue
2018-03-09T11:54:23
2025-04-01T04:33:48.584784
{ "authors": [ "ahevenor", "gowrshan" ], "repo": "cloudfoundry-attic/datadog-firehose-nozzle-release", "url": "https://github.com/cloudfoundry-attic/datadog-firehose-nozzle-release/issues/12", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
636807159
.env file support ? Hi folks, Any workaround to use .env files for this library? Any workaround to use .env files for this library? You'll need to provide more context. Not sure how or when you would expect a .env file to be used. For local testing? Any workaround to use .env files for this library? You'll need to provide more context. Not sure how or when you would expect a .env file to be used. For local testing? Problem: I have many environments [local, stage, production] It is very challenging to connect to these environments to debug locally, we end up hardcoding values, so thought of maintaining multiple .env files for different environments and load according to NODE_ENV. I recently came through vcap file, it would be a great help if some snippet demonstrates how VCAP file should be used, please !
gharchive/issue
2020-06-11T08:01:36
2025-04-01T04:33:48.588520
{ "authors": [ "DivekJohns", "pmuellr" ], "repo": "cloudfoundry-community/node-cfenv", "url": "https://github.com/cloudfoundry-community/node-cfenv/issues/44", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
164962427
Response Time Emitter for external plotting and health checks Currently the abacus-express module has the the ability to log the response time using a morgan middleware function but does not emit/ store the incoming and outgoing requests and responses. const beforeLogger = () => { const morg = morgan( ':remote-addr - - :method :url HTTP/:http-version :status ' + ':res[content-length] :referrer :user-agent - :response-time ms', { immediate: true, stream: { write: (msg, encoding) => { debug('Received request %s', msg.replace(/\n/g, '')); } } }); return (req, res, next) => { return debug.enabled() ? morg(req, res, next) : next(); }; }; const afterLogger = () => { const morg = morgan( ':remote-addr - - :method :url HTTP/:http-version :status ' + ':res[content-length] :referrer :user-agent - :response-time ms', { stream: { write: (msg, encoding) => { debug('Processed request %s', msg.replace(/\n/g, '')); } } }); return (req, res, next) => { return debug.enabled() ? morg(req, res, next) : next(); }; }; I would like to have those events logged by an emitter so that I can add a module that can take that information and publish it to an external destination. This could allow for real time plotting of app response times, among other things. Can you describe your use case in details? The morgan middleware is just a request logger. If you would like to have real time plotting of app response times We have the status emitted in the /hystrix.stream and you can use hystrix dashboard to see it real time. If you would like to store the response stats such as the response time for any other purposes, you can get it from the /hystrix.stream and do some modification on the json structure to match your external app(like grafana). HTH I was planning on logging the request/response and response time and send it to grafana, how can I use hystrix.stream to do that?
gharchive/issue
2016-07-11T23:30:26
2025-04-01T04:33:48.596593
{ "authors": [ "ArchanPatel890", "KRuelY" ], "repo": "cloudfoundry-incubator/cf-abacus", "url": "https://github.com/cloudfoundry-incubator/cf-abacus/issues/367", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
315024432
Please bump go to 1.10.1 Thanks! @sesmith177 If it's not too much trouble we'd prefer 1.10.1 explicitly. We're skipping 1.10.2 because 1.10.3 is planned for start of June and it fixes an issue with the tar/archive package that docker uses. We didn't want to go to the effort of upgrading to 1.10.2 for just a few weeks. Alternatively, we could both wait for 1.10.3 :) v9 has 1.10.1
gharchive/issue
2018-04-17T11:48:54
2025-04-01T04:33:48.598697
{ "authors": [ "BooleanCat", "sesmith177" ], "repo": "cloudfoundry-incubator/windows-tools-release", "url": "https://github.com/cloudfoundry-incubator/windows-tools-release/issues/8", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1389533625
Bosh internal certs rotation Hi , Our orchestrator bosh related certs and bosh-dns certs will be expiring in Jan 2023 so we are planning to have it rotated in our test environment first. So which one should be rotated first bosh-dns or bosh? What is the process ? Does it require downtime for the underlying deployments? and do we have to recreate all the deployment vms in this process?? Please suggest. Hi @Manu2516789, please check the available documentation about credentials rotation in https://bosh.io/docs/dns/#rotating-dns-certificates https://bosh.io/docs/credential-rotation/ If you find something missing in the docs please open an issue for docs-bosh.
gharchive/issue
2022-09-28T15:10:20
2025-04-01T04:33:48.608324
{ "authors": [ "Manu2516789", "beyhan" ], "repo": "cloudfoundry/bosh-deployment", "url": "https://github.com/cloudfoundry/bosh-deployment/issues/440", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }