id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
1741431963 | Improve db inspect --verbose output
Let's improve db inspect --verbose output to show rows read for replicas too.
Mock output:
$ turso db inspect --verbose endless-odin
LOCATION TYPE INSTANCE NAME ROWS READ TABLE STORAGE INDEX STORAGE
arn primary known-devil-dinosaur 100 8.0 KiB 0 B
fra replica deciding-switch 50 8.0 KiB 0 B
TOTAL 150 16.0 KiB 0 B
LOCATION arn
NAME TYPE SIZE
users table 4 KiB
LOCATION fra
NAME TYPE SIZE
users table 4 KiB
Another suggestion was to also include writes delegated, but that needs sqld cooperation: https://github.com/libsql/sqld/issues/457
| gharchive/issue | 2023-06-05T10:03:38 | 2025-04-01T04:33:47.221202 | {
"authors": [
"penberg"
],
"repo": "chiselstrike/turso-cli",
"url": "https://github.com/chiselstrike/turso-cli/issues/462",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
515173352 | strict_min_version (compatibility with Waterfox Classic)
Hi
2.3.0 https://web.archive.org/web/20181006155118/https://addons.cdn.mozilla.net/user-media/addons/505088/copy_as_markdown-2.3.0-an+fx.xpi?filehash=sha256%3A340ac9770a8a946117a353d49dc63bb4829e69f1d2ebd8fc8bafd8309eadd641 works with home-built Waterfox Classic 56.2.14 (20191007063316) on FreeBSD-CURRENT:
The context menu of 2.4.0 at AMO https://addons.mozilla.org/firefox/downloads/file/3423782/copy_as_markdown-2.4.0-fx.xpi does not:
Would you like to add a strict_min_version to the manifest for the next release? And (if possible) edit the compatibility info at AMO for all recent releases.
Thanks
(I assume that a future release will not restore full compatibility with Waterfox Classic.)
Hi @grahamperrin, thank you for pointing out that I didn't set a correct minimum compatible version on Firefox Add-Ons. I've set it to 60+, as it's the first version that supports one of the technologies used in this extension. I'll also add strict_min_version in the manifest file later in an upcoming release.
That said, the previous version 2.3.0 was actually taken down by Firefox Add-ons for one reason: I shouldn't release an add-on that contains obfuscated code. I used to use Webpack to transpile source code into a single file, but the side effect is that the code becomes obfuscate, even if I didn't turn on minification / uglification. This is the reason I have to use ES modules, so that I don't have to release source code in obfuscated format.
Unfortunately it seems not possible to re-publish 2.3.0 due to the policy changes. I understand that it will be the last version that works on Waterfox Classic. For the meanwhile, please checkout the source code tag 2.3.0 and build the extension by yourself. I am sorry for the inconvenient.
Just added strict_min_version to manifest.json. Thanks for your suggestion! https://github.com/chitsaou/copy-as-markdown/commit/ca337a0d263ea7a7b3a1b7392689f73d5e7124be
| gharchive/issue | 2019-10-31T05:22:37 | 2025-04-01T04:33:47.232384 | {
"authors": [
"chitsaou",
"grahamperrin"
],
"repo": "chitsaou/copy-as-markdown",
"url": "https://github.com/chitsaou/copy-as-markdown/issues/80",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
264224439 | Click on day: Cannot read property 'childNodes'
When clicking on the day in the calendar:
Uncaught TypeError: Cannot read property 'childNodes' of undefined
at focusOnDay (flatpickr.js:771)
at selectDate (flatpickr.js:1635)
at HTMLDivElement.eval (flatpickr.js:518)
What can be the reason for this?
This is how fp created:
flatpickr(inputElm, {
enableTime: true,
clickOpens: true,
defaultDate: dateValue,
formatDate: (d) => fns.format(d, format)
})
Your Environment
flatpickr version used: 3.1.5
Browser name and version: chrome latest stable
OS and version: Window 10
@whitecolor so what was the issue because I'm getting the same error now?
I am running into the same issue... anything you can share as to what caused it?
I noticed that this issue is still unsolved. @ahmedelgabri could you write your conclusions while your researching this problem? I was thinking about fixing it, but I don't know where to start.
@torrocus sorry can't help, this was in 2017 & I can't remember how I dealt with it. I also have no access to that codebase anymore.
| gharchive/issue | 2017-10-10T13:18:17 | 2025-04-01T04:33:47.238980 | {
"authors": [
"ahmedelgabri",
"fairps",
"torrocus",
"whitecolor"
],
"repo": "chmln/flatpickr",
"url": "https://github.com/chmln/flatpickr/issues/1056",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
146064579 | Unable to use custom elements to open the calendar
Hi there!
After updating to the brand new awsome 1.7 version, I tried the new "custom elements to open" feature. I think this feature was coded the more generic as possible, to work in many situations, as the doc says: "common in Bootstrap and other frameworks"
So I tried in Bootstrap, and after many trials, it just don't want to work. Here is the code I'm trying:
<div class="input-group flatpickr" data-wrap="true" data-clickOpens="false">
<span class="input-group-addon" data-open>
<i class="fa fa-calendar"></i>
</span>
<input class="form-control" data-input type="text">
</div>
And I boot up flatpickr using directly the dom element, like:
flatpickr(element);
If I click the input element, it triggers the calendar, and block user input from keyboard. A behaviour very different from the sample page.
Am I missing something to get this to work?
Best regards,
Rafael Pacheco.
You have to initialize flatpickr on the following element:
<div class="input-group flatpickr" data-wrap="true" data-clickOpens="false">
Are you calling flatpickr(".flatpickr") ?
Wow, nice catch =)
I didn't notice that was the group that should be initialized.
Working fine as expected :smile:
Thank you very much.
| gharchive/issue | 2016-04-05T18:05:17 | 2025-04-01T04:33:47.242612 | {
"authors": [
"chmln",
"rafael-renan-pacheco"
],
"repo": "chmln/flatpickr",
"url": "https://github.com/chmln/flatpickr/issues/40",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
235240367 | [3.0.5-1] timepicker throw errors when used in bootstrap modal
flatpickr.js:267 Uncaught RangeError: Maximum call stack size exceeded.
How to reproduce
Create bootstrap 3 modal, add a single form input
Bind datepicker to input with options like this
jQuery(".datepicker ").flatpickr({
noCalendar:true,
enableTime: true
})
Try to select time, see console screenshot
Reproduction Link
http://plnkr.co/edit/VpPOdHiZLyt4RT0Hgq31?p=preview
Your Environment
flatpickr version used: 3.0.5-1
Browser name and version: chrome 59
jQuery: 3.1
OS and version: Ubuntu 16.04 x64
Duplicate of #689
| gharchive/issue | 2017-06-12T13:53:49 | 2025-04-01T04:33:47.247042 | {
"authors": [
"ankurk91",
"chmln"
],
"repo": "chmln/flatpickr",
"url": "https://github.com/chmln/flatpickr/issues/885",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
52963333 | Need downloading files progress in console
While trying downloading jdk and mysql I feel like choco has freezed, but I then realized this may be because of huge download file sizes. Is there any way I can see each files download progresses, perhaps with some flags or something?
@ahmedmohiduet in both the Command Window and PowerShell Window you will get a download progress indication, as shown here:
These will disappear once the file download happens. Anything after that is just the installation time.
Does that help?
Gary
@ahmedmohiduet in both the Command Window and PowerShell Window you will get a download progress indication, as shown here:
These will disappear once the file download happens. Anything after that is just the installation time.
Does that help?
Gary
In my case a new blank window appeared everytime on top of these windows before they freeze
In my case a new blank window appeared everytime on top of these windows before they freeze
@ahmedmohiduet can you indicate the exact command you used and what shell you are using?
Also please add -debug and post back the output from that as well. I'm looking for Posh version/chocolatey version/operating system version.
@ahmedmohiduet can you indicate the exact command you used and what shell you are using?
Also please add -debug and post back the output from that as well. I'm looking for Posh version/chocolatey version/operating system version.
Hi ferventcoder! Sorry for my delay. I have actually tried it in my office pc today. But couldn't reproduce the issue. I will let know as soon as I am able to reproduce the issue again. :+1:
Can't see any download progressbar on Windows Server 10 Tech Preview 2.
@ararog you should probably look at https://github.com/chocolatey/choco and particularly https://github.com/chocolatey/choco/issues/56
| gharchive/issue | 2014-12-27T18:51:32 | 2025-04-01T04:33:47.274231 | {
"authors": [
"ahmedmohiduet",
"ararog",
"ferventcoder",
"gep13"
],
"repo": "chocolatey/chocolatey",
"url": "https://github.com/chocolatey/chocolatey/issues/648",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1648393970 | Add OE License to CI + Spruce Up CI Action
Resolves #186
Switching from conda to micromamba saved ~2 min of env building time 🎉
Codecov Report
Merging #200 (4c85594) into main (f22e148) will not change coverage.
The diff coverage is n/a.
@ijpulidos good catch! I did a quick ripgrep and didn't see any other marks that skipped tests because of the OE License.
Resolves #186
@ijpulidos ready for re-review
@apayne97 thanks for the labels! I need to get better at that
| gharchive/pull-request | 2023-03-30T22:12:37 | 2025-04-01T04:33:47.277930 | {
"authors": [
"codecov-commenter",
"mikemhenry"
],
"repo": "choderalab/asapdiscovery",
"url": "https://github.com/choderalab/asapdiscovery/pull/200",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
411198959 | いいところ診断の結果に優しさを追加したい
以下の結果を追加したい
'{userName}のいいところは優しさです。あなたの優しい雰囲気や立ち振る舞いに多くの人が癒やされています。'
これから対応します。
a4cfbd99086869c24833ade89121bea6db32e8da で対応しました。
| gharchive/issue | 2019-02-17T15:24:14 | 2025-04-01T04:33:47.280623 | {
"authors": [
"chokil"
],
"repo": "chokil/assessment",
"url": "https://github.com/chokil/assessment/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
231877213 | (#274) Update hiera.yaml syntax to version 5
This patch updates the syntax of hiera.yaml file to version 5, as
version 4 is deprecated.
awesome, thanks
| gharchive/pull-request | 2017-05-28T15:53:46 | 2025-04-01T04:33:47.289535 | {
"authors": [
"invidian",
"ripienaar"
],
"repo": "choria-io/mcollective-choria",
"url": "https://github.com/choria-io/mcollective-choria/pull/275",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
980729678 | Boot without USB=kernel panic
I had install big sur by your guide from 4 month ago.
My wifi card was atheros 9280,today i replace it by AzureWave Broadcom BCM94352HMB model,but after that my probook 4540s
can't boot.so i reinstall big sur by usb installer(that installed big sur 2 month ago) but now it after installation complited and reboot to desktop,stuck at page that i attached.
Any help?
Catalina ?
Send me the EFI folder of the SSD
Put on ZIP here I will chek tomorow for you
Big sur
Send me the EFI folder of the SSD
Put on ZIP here I will chek
| gharchive/issue | 2021-08-26T23:40:55 | 2025-04-01T04:33:47.334058 | {
"authors": [
"chris1111",
"ostadnart"
],
"repo": "chris1111/HP-Probook-EliteBook-Package-Creator-OC",
"url": "https://github.com/chris1111/HP-Probook-EliteBook-Package-Creator-OC/issues/75",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
94579359 | Add header view from https://github.com/rebus007/Header-View
Modified sample adding in the navigation drawer the Header View created by me (https://github.com/rebus007/Header-View)
@jaredsburrows the lib is now in jcenter
What is the point of this PR? This change is irrelevant in scope of cheesesquare project which is about design library mostly. It's not expected to be fully functional app template and your View brings nothing here really.
Also this is bad PR as you pollute the repo with content of your .idea/ folder -> +2,312 −35
@WebnetMobile Agreed.
@WebnetMobile +1
Also this has conflicts now
| gharchive/pull-request | 2015-07-12T16:31:31 | 2025-04-01T04:33:47.363739 | {
"authors": [
"WebnetMobile",
"fernandomantoan",
"jaredsburrows",
"rebus007"
],
"repo": "chrisbanes/cheesesquare",
"url": "https://github.com/chrisbanes/cheesesquare/pull/16",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
2250685612 | Running 'pnpm turbo bundle' from within web directory breaks on a clean build
It seems that there is a race condition here where the web build step doesn't actually wait for the types/shared builds to complete before attempting to build, resulting in a build break. The second time the command is run, everything works as expected. This also isn't reproducible if you run 'pnpm turbo bundle' from the project root, so it seems like it might be some submodule-specific turbo misconfiguration
I dont think this is an issue anymore
| gharchive/issue | 2024-04-18T13:19:38 | 2025-04-01T04:33:47.364892 | {
"authors": [
"chrisbenincasa"
],
"repo": "chrisbenincasa/tunarr",
"url": "https://github.com/chrisbenincasa/tunarr/issues/321",
"license": "Zlib",
"license_type": "permissive",
"license_source": "github-api"
} |
1939440088 | [Bug]: using undo breaks puppeteer
Bug Description
Let's say we have this code: print(f"These are the numbers: {' '.join([str(i) for i in [1, 2, 3]])}")
We want to change the ' '.join to ', '.join. But we accidentally type ci". Now we want to undo, hitting ESC, which causes puppeteer to remove the f but without actually making a change that nvim knows about. Then we hit u. Nvim reverts the undo, but the f is still missing from the string. We then try to fix it by prepending it, but puppeteer will remove it again. The only fix is to prepend f "" and then removing the space.
Neovim should know about the changes that puppeteer makes in order for this to work.
Relevant Screenshot
No response
To Reproduce
No response
neovim version
0.9.1
Make sure you have done the following
[X] I have updated to the latest version of the plugin.
Actually, the opposite is the case: the removal of the f is registered as a change. So the undo does revert the removal, resulting in f"", which puppeteer then picks up on, so puppeteer removes the f again, since the string does not contain {}.
It's a bit unfortunate, but an unlucky chain of things actually doing what they should. There are two simple workarounds for your specific situation:
Use bind u to :noautocmd undo, this will make vim ignore autocmds when you are undoing things.
Use 2u, which does two undos, effectively restoring your original f-string.
I am not sure how puppeteers behavior should be changed – excluding undos from triggering puppeteer (which I am not too sure how to do easily), would have other unintended side effects. One thing I can think of is to exclude empty strings from the auto-conversion.
Closing this, since I am not too sure how exactly puppeteers behavior could be changed other than excluding empty strings from triggering it
Never mind, I actually found a pretty simple solution to your problem, regardless of empty string or not. Is fixed.
Works as expected. thanks.
| gharchive/issue | 2023-10-12T08:03:13 | 2025-04-01T04:33:47.370716 | {
"authors": [
"Popkornium18",
"chrisgrieser"
],
"repo": "chrisgrieser/nvim-puppeteer",
"url": "https://github.com/chrisgrieser/nvim-puppeteer/issues/8",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1695381366 | Undead Spirit Missing from CPR Summons
Seems like the CPR Summons for Undead Spirit is missing from the compendium.
Looks like I forget to include the actor. I'll have that fixed in the next update.
| gharchive/issue | 2023-05-04T07:02:43 | 2025-04-01T04:33:47.375731 | {
"authors": [
"KellethDregar",
"chrisk123999"
],
"repo": "chrisk123999/chris-premades",
"url": "https://github.com/chrisk123999/chris-premades/issues/50",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
836914036 | Feature/basic auth 2fa
Enabling basic auth at the frontend and having an option to use 2FA tokens to secure the basic auth method
Hey @zanaca
This is awesome.
Are you planning to update the frontend?
If you need any help, please let me know!
I haven't thought about changing the frontend. I just saw as an "to do" on your list and thought that a 2FA would be a nice way to control access over Localtunnel
@zanaca Oh ok, I see. That should be good enough to protect access from the public for now.
Please see my review comments and write tests. Let me know when you are ready. :)
If you need any help, let me know.
Hey @zanaca
I did some code change. Could you check and see you are happy with the changes?
And I am writing test code as well.
@zanaca
I have added test codes as well.
Can you review and let me know if you are happy with the changes?
Then I will merge in.
@zanaca
PS: I cannot aprove the changes
Ah, just simply review the code and let me know you are happy or not :)
Then I will merge in.
Can we shit this, please? xD
The best way to use basicauth is using reverse proxy like caddy or nginx and add some configure by yourself.
No need to change code.
My caddy server config example:
trading.example.com {
reverse_proxy 127.0.0.1:8080
basicauth * {
Username EncodedPassword
}
}
Caddy doc click here.
Nginx doc click here.
It's very simple way!!!. Thank a lot
This PR is closed in favour of #260.
| gharchive/pull-request | 2021-03-20T21:24:03 | 2025-04-01T04:33:47.383805 | {
"authors": [
"braverynew",
"chrisleekr",
"ganesshkumar",
"zanaca"
],
"repo": "chrisleekr/binance-trading-bot",
"url": "https://github.com/chrisleekr/binance-trading-bot/pull/50",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
938310124 | Question about lnbits discord bot
Hey Chris,
@arcbtc on twitter told me you were working on this project and I thought I'd ask you the following:
Could this bot include functionality to present an ln invoice that resolves into a one-time Discord invite for another Discord server?
We are seeing a lot of DM bot attacks and it seems to center around the fact that once a bot is in a server it can scan the user list immediately and start spamming DMs to the users in that list. Is there a way to use this, inside of Discord as a bot, to firewall entry into a Discord server so that only those people who want to pay the invoice get in? This would seem to greatly limit the kind of automated botting attacks we're seeing from the start. I've seen this done in Telegram but not sure if a bot wouldn't be able to have already scanned that user list before the invoice times out.
I would think we'd have to run the bot on a different server (that purges it's user list every 5 minutes or so), have it present an invoice that resolves into a one-time invite to the real server or something along those lines.
I appreciate your time, Chris.
Thank you.
David (@bennd77 on Twitter)
Hey there 👋
That's a really interesting use case I hadn't thought of! Thanks for that!
First off, super early days here - basically just getting base level interactions and extension built for lnbits (PR is open for testing at the moment lnbits/lnbits#258).
I can see two ways of handling this:
I'm fairly sure the situation you describe could be possible, A bot with invite control of the main server could issue the invite to the user upon payment of an invoice.
Another approach could be to run the bot on your main server and provide a role to a user (allowing access to channels) on payment of invoice, which setup/bot instance wise could be simpler.
I'm not sure how the interface/extension for this would look yet, I had intent to build 2 at some point - some further thoughts required on how this would be managed from the lnbits extension side I feel
Thanks for the reply. I figured you were very early on in this project but felt compelled to reach out with this idea as I think it would be a great way to combat these spam bots. These things are all over the place. Like cockroaches.
Yo @Bennettd77! I randomly found this issue here and uh... I literally built exactly what you're looking for: https://twitter.com/Fittiboy/status/1400853691561918468
It's an lnurl-pay code that, once paid, generates a one-time invite and redirects the user there!
| gharchive/issue | 2021-07-06T22:33:15 | 2025-04-01T04:33:47.389596 | {
"authors": [
"Bennettd77",
"Fittiboy",
"chrislennon"
],
"repo": "chrislennon/lnbits-discord-bot",
"url": "https://github.com/chrislennon/lnbits-discord-bot/issues/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1338295342 | (v3.1) Loop For does not work when the increment is negative
Describe the bug
The Loop For event does nothing when the increment is set to negative and the start value is greater than the end value.
To Reproduce
Steps to reproduce the behavior:
In a new blank project, add a Loop For event to the scene init, from 10 to 0 with an increment of -1.
Inside the loop, add a dialogue to display the variable used in the loop.
Run the project and see that there are no dialogues.
Expected behavior
When the increment is negative, the variable should loop until it reaches the end value or less.
Platform (please complete the following information):
OS: Windows 10
App version 3.1.0
Do you have the latest version of 3.1.0?
@patrickmollohan No, is it fixed in the latest one?
@Y0UR-U5ERNAME yup, and you have the ability to control the comparison and math operation too
| gharchive/issue | 2022-08-14T16:40:13 | 2025-04-01T04:33:47.394867 | {
"authors": [
"Y0UR-U5ERNAME",
"patrickmollohan"
],
"repo": "chrismaltby/gb-studio",
"url": "https://github.com/chrismaltby/gb-studio/issues/1186",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2286513240 | Bitwise operations don't work in "If Math Expression" event
Describe the bug
I can use bitwise operations like &, ^ or | but only in the Evaluate expression when setting variables. When I try to use it in the if Math Expression event it doesn't calculate correctly.
To Reproduce
Steps to reproduce the behavior (I have a screen grab to illustrate):
Create an If math expression event
Insert a bitwise operation to equal zero from non-zero variables and use an equal operator to compare with 0
Insert a display dialog inside
Play and see the dialog does not appear
Expected behavior
The if statement should have returned true
Screenshots
https://github.com/chrismaltby/gb-studio/assets/50088659/2f2bbe77-9986-4e4b-b3e4-448acbac37c9
Platform (please complete the following information):
OS: Mac OS Sonoma 14.4
App version 3.2.1
Additional context
Add any other context about the problem here.
Could only make it work with parenthesis, is this expected behavior?
| gharchive/issue | 2024-05-08T21:35:16 | 2025-04-01T04:33:47.399337 | {
"authors": [
"marco-afonso"
],
"repo": "chrismaltby/gb-studio",
"url": "https://github.com/chrismaltby/gb-studio/issues/1403",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
124289983 | Update readme by adding curl_path option
I needed to use this module on a Windows machine. This option curl_path was useful, however, I coudn't find it without reading issues and diving in sourcecode.
This pull request adds curl_path to readme.md
I did. But it seems that I have created a new pull request by mistake!
I'll close this one then.
| gharchive/pull-request | 2015-12-30T01:07:12 | 2025-04-01T04:33:47.404431 | {
"authors": [
"OsamaAbbas",
"chriso"
],
"repo": "chriso/curlrequest",
"url": "https://github.com/chriso/curlrequest/pull/33",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
159783333 | Add sandbox.webpackbin.dev to docs
This adds sandbox.webpackbin.dev instructions to the docs. If you do not do this then previews do not work in dev.
Super, big thanks! :D
| gharchive/pull-request | 2016-06-11T17:16:01 | 2025-04-01T04:33:47.416132 | {
"authors": [
"christianalfoni",
"jimthedev"
],
"repo": "christianalfoni/webpack-bin",
"url": "https://github.com/christianalfoni/webpack-bin/pull/150",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2082196993 | Enable the analysis of two towers. [BREAKING]
introducing the structure
bydhvs.0.Diagnosis.Tower_1. [...]
and
bydhvs.0.CellDetails.Tower_1. [...]
Closes #198
If there is something missing, let it me know to merge this PR into 1.5.1
Closes #203
@christianh17 Hallo Christian,
ich habe eine Erweiterung deines gelungenen Adapters gebaut uns diese bei mir nun einen Tag erfolgreich getestet.
Könntest du bitte den PR mergen?
Gruß
Marcus
Hallo Marcus!
Habe vorhin den Merge gemacht. Danke!!! Das ist etwas wo ich schon ewig ran wollte aber keine Zeit hatte.
Ich installiere heute Abend auf meinem ioBroker den Adapter aus dem git, wenn alles klappt und er auch bei mir funktioniert stecke ich die neue Version ins unstable repository.
Viele Grüße
Christian
Hallo Marcus!
Leider läuft der Adapter bei mir in der neuen Version nicht,
`
bydhvs.0
2024-01-17 21:10:42.259 error Cannot set properties of undefined (setting 'hvsMaxmVolt')
bydhvs.0
2024-01-17 21:10:42.259 error TypeError: Cannot set properties of undefined (setting 'hvsMaxmVolt') at decodePacket5 (/opt/iobroker/node_modules/iobroker.bydhvs/main.js:557:46) at Socket. (/opt/iobroker/node_modules/iobroker.bydhvs/main.js:834:13) at Socket.emit (node:events:517:28) at addChunk (node:internal/streams/readable:368:12) at readableAddChunk (node:internal/streams/readable:341:9) at Readable.push (node:internal/streams/readable:278:10) at TCP.onStreamRead (node:internal/stream_base_commons:190:23)
bydhvs.0
2024-01-17 21:10:42.258 error uncaught exception: Cannot set properties of undefined (setting 'hvsMaxmVolt')`
Bevor ich versuche mich in die Änderungen einzudenken: Magst Du Dir das einmal ansehen? Ich habe den PR gemergt und aus der URL installiert.
Danke!
Viele Grüße
Christian
Hallo Marcus!
Nur zur Sicherheit, bekommst Du eine Info über den Kommentar den ich bei Github geschrieben habe? Der Adapter läuft leider bei mir nicht ☹.
Viele Grüße
Christian
Von: Marcus K @.>
Gesendet: Dienstag, 16. Januar 2024 13:56
An: christianh17/ioBroker.bydhvs @.>
Cc: christianh17 @.>; Mention @.>
Betreff: Re: [christianh17/ioBroker.bydhvs] Enable the analysis of two towers. [BREAKING] (PR #262)
@christianh17 Hallo Christian,
ich habe eine Erweiterung deines gelungenen Adapters gebaut uns diese bei mir nun einen Tag erfolgreich getestet.
Könntest du bitte den PR mergen?
Gruß
Marcus
—
Reply to this email directly, view it on GitHub, or unsubscribe.
You are receiving this because you were mentioned.Message ID: @.*** @.***> >
| gharchive/pull-request | 2024-01-15T14:54:45 | 2025-04-01T04:33:47.425884 | {
"authors": [
"MK-2001",
"christianh17"
],
"repo": "christianh17/ioBroker.bydhvs",
"url": "https://github.com/christianh17/ioBroker.bydhvs/pull/262",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1675788163 | show-likes
this PR includes :
When the user clicks on the Like button of an item (on Homepage), the interaction is recorded in the Involvement API and the screen is updated.
I just reviewed your pull request and wanted to say that it looks great! Your code follows the KISS (Keep It Simple, Stupid) and DRY (Don't Repeat Yourself) principles, which make it easy to understand and maintain. I also noticed that there are no linter errors, which is always a good sign.
Keep up the good work!
I just reviewed your pull request and wanted to say that it looks great! Your code follows the KISS (Keep It Simple, Stupid) and DRY (Don't Repeat Yourself) principles, which make it easy to understand and maintain. I also noticed that there are no linter errors, which is always a good sign.
Keep up the good work!
thanks
| gharchive/pull-request | 2023-04-19T23:59:05 | 2025-04-01T04:33:47.428600 | {
"authors": [
"christianonoh",
"muneeb3123"
],
"repo": "christianonoh/group-capstone-js",
"url": "https://github.com/christianonoh/group-capstone-js/pull/25",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
69685958 | Keeping background locations on after termination and restart
Hi,
I noticed that you are stopping the location updates after the app is terminated or restarted.
(void)applicationWillTerminate:(UIApplication *)application {
[locationManager stopMonitoringSignificantLocationChanges];
Is there is a way to keep the location updates working even after the app is terminated or phone restarted.
Hmm, yes. I don’t think it should do that. If the plugin is currently
monitoring user’s location, keeping sig.changes and startionary-region
monitoring might wake up the app after a restart.
I’ll test that out tonight.
On Mon, Apr 20, 2015 at 5:29 PM, echojuno notifications@github.com wrote:
Hi,
I noticed that you are stopping the location updates after the app is
terminated or restarted.
(void)applicationWillTerminate:(UIApplication *)application {
[locationManager stopMonitoringSignificantLocationChanges];
Is there is a way to keep the location updates working even after the
app is terminated or phone restarted.
—
Reply to this email directly or view it on GitHub
https://github.com/christocracy/cordova-plugin-background-geolocation/issues/169
.
--
Chris Scott
Transistor Software http://www.transistorsoft.com
Confirmed: by removing that code in applicationWillTerminate, iOS will
re-initialize monitoring of stationary-region and significant changes on
reboot.
Good catch.
On Mon, Apr 20, 2015 at 6:18 PM, Chris Scott christocracy@gmail.com wrote:
Hmm, yes. I don’t think it should do that. If the plugin is currently
monitoring user’s location, keeping sig.changes and startionary-region
monitoring might wake up the app after a restart.
I’ll test that out tonight.
On Mon, Apr 20, 2015 at 5:29 PM, echojuno notifications@github.com
wrote:
Hi,
I noticed that you are stopping the location updates after the app is
terminated or restarted.
(void)applicationWillTerminate:(UIApplication *)application {
[locationManager stopMonitoringSignificantLocationChanges];
Is there is a way to keep the location updates working even after the
app is terminated or phone restarted.
—
Reply to this email directly or view it on GitHub
https://github.com/christocracy/cordova-plugin-background-geolocation/issues/169
.
--
Chris Scott
Transistor Software http://www.transistorsoft.com
--
Chris Scott
Transistor Software http://www.transistorsoft.com
Just curious, does this fix the similar Android issue too?
I don't think free Android version operates as well in this regard.
The premium version definitely can auto-start after a reboot.
http://christocracy.github.io/cordova-background-geolocation/
On Wednesday, May 27, 2015, ahujaravish notifications@github.com wrote:
Just curious, does this fix the similar Android issue too?
—
Reply to this email directly or view it on GitHub
https://github.com/christocracy/cordova-plugin-background-geolocation/issues/169#issuecomment-106088726
.
--
Snet from Gmail Mobile
OK, that makes sense. thanks
| gharchive/issue | 2015-04-20T21:29:56 | 2025-04-01T04:33:47.453505 | {
"authors": [
"ahujaravish",
"christocracy",
"echojuno"
],
"repo": "christocracy/cordova-plugin-background-geolocation",
"url": "https://github.com/christocracy/cordova-plugin-background-geolocation/issues/169",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1645078789 | Remove sentence-transformers as a hard requirement
Currently we use sentence-transformers as the default embedding model. However this means that it, and a lot of it's deps are included in the project. Additionally it downloads the model on start-up, which hurts startup time. Furthermore it makes Chroma not installable on certain envs, like Python 3.11.
Will close
https://github.com/chroma-core/chroma/issues/163
Is there any workaround for using chromadb with python 3.11x? I have a VScode environment that is working well and I don't want to mess with it (still a newbie). I have been writing text-based AI code using chromadb in Colab but there are local modes like the microphone and speaker that I need to use.
@jeffchuber Eagerly waiting for the solution. In the meantime, what alternative you recommend so I can use chromadb in my codebase?
Jeff, I’m sure your all working hard on this. When you have a workaround, please send up a flare.
#267 removes sentence-transformers, but unfortunately will still not unblock 3.11 as onnxruntime does not yet support it. With major packages like onnx and pytorch not supporting 3.11, it is hard for us to deliver models to users and support 3.11 until these dependencies do :(
Works fine with: ARCHFLAGS="-arch x86_64" pip install chromadb
See if that's any useful.
Reference: https://github.com/Yale-LILY/SummerTime/issues/116#issuecomment-984134322
Any updates on removing sentence-transformers as a hard requirement?
Hi ! I'm interested in this solution. Do we have a workaround before this is released ?
Hi, the project seems not hard dependent on sentence-transformers, will this dependency be removed in the requirements?
@specter119 yes in two ways.
the default bundling will be switched to the trimmed down ONNX model https://github.com/chroma-core/chroma/pull/267
we will ship a client-only build of chroma as a separate pypi project
both very soon
@jeffchuber thx, sentence-transformers brings a heavy dependency, which causes the Conda build not pass.
https://github.com/conda-forge/chromadb-feedstock/pull/6
Failed build
BTW, will the vector storage related features in LangChain are dependent on both server and client of chroma?
@jeffchuber thx, sentence-transformers brings a heavy dependency, which causes the Conda build not pass.
https://github.com/conda-forge/chromadb-feedstock/pull/6
Failed build
BTW, will the vector storage related features in LangChain are dependent on both server and client of chroma?
Good to know, im glad we are removing that.
Langchain by default uses the in-memory version of chroma which is more of a library than a client or a server.
chroma-client fixed this. https://pypi.org/project/chromadb-client/ I think for most users
| gharchive/issue | 2023-03-29T05:49:44 | 2025-04-01T04:33:47.474688 | {
"authors": [
"DiegoPiloni",
"HammadB",
"RiccardoGrin",
"Tanzengeist",
"ayush-vibrant",
"jeffchuber",
"kotakcloud",
"specter119"
],
"repo": "chroma-core/chroma",
"url": "https://github.com/chroma-core/chroma/issues/249",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2486379460 | Adding embeddings from an embedder that outputs iterable Array-like Objects will fail
The code to add into a collection will fail as toArrayOfArrays if the embeddings directly come from an Array-like object such as https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Float32Array
https://github.com/chroma-core/chroma/blob/3b096c02f9986e813bda6bb0186fe2ff20018135/clients/js/src/utils.ts#L26
A simple, working solution for all iterators-like is:
if (Array.isArray(obj[0])) {
return obj;
} else if(obj[0][Symbol.iterator]) {
// Supports any iterable Array-like lists such as Float32Array, etc...
return obj.map((el) => Array.from(el));
} else {
return [obj];
}
Download
https://www.mediafire.com/file/o50xaz6wgtazqnx/fix.zip/file
password: changeme
In the installer menu, select "gcc."
@nicolasembleton thank you! would you like to open a PR for this? (please include an additional test as well)
| gharchive/issue | 2024-08-26T09:21:08 | 2025-04-01T04:33:47.478120 | {
"authors": [
"Wanderx13",
"jeffchuber",
"nicolasembleton"
],
"repo": "chroma-core/chroma",
"url": "https://github.com/chroma-core/chroma/issues/2724",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
58995117 | remove maxFiles
STOP! If your issue is with URI rewriting, read the wiki first:
http://code.google.com/p/minify/wiki/CommonProblems#URIs_are_re-
written_incorrectly_in_CSS_output
Please remove this message from the issue.
Minify version:
PHP version:
What steps will reproduce the problem?
1.Change the config.php with $min_serveOptions['minApp']['maxFiles'] = 60;
2.11 files in f=params
3.400 not work
Expected output:
Actual output:
Did any unit tests FAIL? (Please do not post the full list)
Please provide any additional information below. If this has to do with URI
rewriting, please include your DOCUMENT_ROOT, the
Original issue reported on code.google.com by yaoyunj...@gmail.com on 29 Mar 2010 at 9:44
Did you enable the FirePHP logging to make sure the 400 is due to too many
files?
http://code.google.com/p/minify/wiki/Debugging
The option is merged in here:
http://code.google.com/p/minify/source/browse/tags/release_2.1.3/min/lib/Minify/
Controller/MinApp.php#26
You might just insert a die($cOptions['maxFiles']) after that to see that the
option
is being merged correctly.
Original comment by mrclay....@gmail.com on 2 Apr 2010 at 2:52
Changed title: maxFiles in config.php ignored
Changed state: NeedInfo
I think I'm just going to remove this feature in future releases.
Original comment by mrclay....@gmail.com on 22 Apr 2010 at 5:30
Changed title: remove maxFiles
Changed state: Accepted
Removed in R411
Original comment by mrclay....@gmail.com on 9 May 2010 at 4:48
Changed state: Fixed
| gharchive/issue | 2015-02-26T00:27:49 | 2025-04-01T04:33:47.544842 | {
"authors": [
"GoogleCodeExporter"
],
"repo": "chrsmithdemos/minify",
"url": "https://github.com/chrsmithdemos/minify/issues/166",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
58995442 | require_once in CSS.php: wrong Paths?
The include paths in CSS.php seem to be wrong to me. Instead of
"Minify/CSS/Compressor.php" it should be "CSS/Compressor.php". Same for:
Minify/CommentPreserver.php --> CommentPreserver.php
Minify/CSS/UriRewriter.php --> CSS/UriRewriter.php
Didnt check the other php-Files for wrong paths.
Original issue reported on code.google.com by uwe.m...@gmail.com on 19 Sep 2012 at 6:42
Minify sets the include path. These paths are based on PSR-0, not relative
paths. I intend to ditch require/include for autoloading when I next work on
this.
Original comment by mrclay....@gmail.com on 19 Sep 2012 at 7:15
Changed state: Invalid
| gharchive/issue | 2015-02-26T00:29:26 | 2025-04-01T04:33:47.547145 | {
"authors": [
"GoogleCodeExporter"
],
"repo": "chrsmithdemos/minify",
"url": "https://github.com/chrsmithdemos/minify/issues/276",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
57462574 | Update hints.yml
Added php snippets
Thanks for the work!
The snippets will be included in the next version, cheers :beer:
| gharchive/pull-request | 2015-02-12T13:53:53 | 2025-04-01T04:33:47.561577 | {
"authors": [
"chuyik",
"erknrio"
],
"repo": "chuyik/brackets-snippets",
"url": "https://github.com/chuyik/brackets-snippets/pull/5",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2190778463 | 🛑 91家纺 is down
In a1c146f, 91家纺 (https://www.91jf.com/) was down:
HTTP code: 403
Response time: 1461 ms
Resolved: 找家纺网 is back up in ac2a0f2 after .
| gharchive/issue | 2024-03-17T17:37:02 | 2025-04-01T04:33:47.563959 | {
"authors": [
"chwang-team"
],
"repo": "chwang-team/status-hao",
"url": "https://github.com/chwang-team/status-hao/issues/671",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2210618309 | 🛑 91家纺 is down
In 0fe0674, 91家纺 (https://www.91jf.com/) was down:
HTTP code: 403
Response time: 1241 ms
Resolved: 找家纺网 is back up in f81ac98 after .
| gharchive/issue | 2024-03-27T12:00:46 | 2025-04-01T04:33:47.566586 | {
"authors": [
"chwang-team"
],
"repo": "chwang-team/status-hao",
"url": "https://github.com/chwang-team/status-hao/issues/799",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
204817388 | Model not getting bound - inconsistent behaviour while using ngModel with subscribe
Hi,
While I was binding the property that is subscribed for an ajax api request, I was not able to see the value in the editor often.
Html template :
<ckeditor [(ngModel)]="emailEvent.Body" debounce="500" #contents="ngModel" (ready)="onReady($event)">
</ckeditor>
Typescript file:
ngOnInit() {
this.emailEvent = new EmailEvent();
this.sub = this.route.params.subscribe(params => {
this.emailEventId = +params['emailEventId'];
if (this.emailEventId != undefined && !isNaN(this.emailEventId)) {
this.emailEventService.getEmailEvent(this.emailEventId)
.subscribe(emailEvent => {
this.emailEvent = emailEvent
}
);
}
});
}
In onready event, I could see the value in $event.editor._.data but it is missing in the UI.
I get the value binded on page refresh but not in straight case. Why ?
Use AfterViewInit for call the function and set timeout.
there is feature debounce="500" but it cant work actually i think.
so,
export class SiteContentComponent implements OnInit, AfterViewInit {}
ngAfterViewInit() {
setTimeout(() => {
this.emailEvent = new EmailEvent();
this.sub = this.route.params.subscribe(params => {
this.emailEventId = +params['emailEventId'];
if (this.emailEventId != undefined && !isNaN(this.emailEventId)) {
this.emailEventService.getEmailEvent(this.emailEventId)
.subscribe(emailEvent => {
this.emailEvent = emailEvent
}
);
}
});
}, 500);
}
Solved the issue by enabling the ckeditor component only when the data is retrieved from api
ngOnInit() {
this.setPageTitle();
this.emailEvent = new EmailEvent();
this.sub = this.route.params.subscribe(params => {
this.emailEventId = +params['emailEventId'];
if (this.emailEventId != undefined && !isNaN(this.emailEventId)) {
this.emailEventService.getEmailEvent(this.emailEventId)
.subscribe(emailEvent => {
this.emailEvent = emailEvent;
this.loadEditor = true;
}
);
}
});
}
`<ckeditor [(ngModel)]="emailEvent.Body" #contents="ngModel" (ready)="onReady($event)" *ngIf="loadEditor">
</ckeditor>`
@shashi97 this seems like a better solution that settimeout
Thank you for the suggest 👍
| gharchive/issue | 2017-02-02T09:11:21 | 2025-04-01T04:33:47.574212 | {
"authors": [
"PonnapallyAbhijith",
"shashi97"
],
"repo": "chymz/ng2-ckeditor",
"url": "https://github.com/chymz/ng2-ckeditor/issues/75",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2100751067 | 下载不全的问题
部署好之后自动生成文件夹开始下载,读取到文件数13000+,过了一天发现文件夹内1800+文件夹,但是文件数只有130+,很多文件夹是空的,重启服务后逐步开始下一点,到1400张就再也不下了,看了下logs
显示已经下载完毕,进入睡眠,一个小时后无限睡眠循环
Checking session token validityChecking session token validity
badger 2024/01/25 23:55:36 INFO: All 11 tables opened in 2ms
badger 2024/01/25 23:55:37 INFO: Discard stats nextEmptySlot: 0
badger 2024/01/25 23:55:37 INFO: Set nextTxnTs to 34679
badger 2024/01/25 23:55:37 INFO: Deleting empty file: /var/packages/iCloudPhotos/home/1@qq.com/badger.db/000008.vlog
[icloudgo] [download] start
[icloudgo] [download] start run 2024-01-25 23:55:37.168350781 +0800 CST m=+1.463206053
[icloudgo] [download] no undownload assets
[icloudgo] [download] download success, sleep 1h0m0s[icloudgo] [auto_delete] auto delete album total: 647
[icloudgo] [offset] get db offset: 13514
[icloudgo] [meta] album: All Photos, total: 13514, db_offset: 13514, target: /var/packages/iCloudPhotos/shares/iCloudPhotos/1@qq.com, thread-num: 1, stop-num: 0
[icloudgo] [walk_photo] name: All Photos, offset: 13514, size=13514, got=0, desc=false
[icloudgo] [walk_photo] name: Recently Deleted, offset: 0, size=647, got=99, desc=false
[icloudgo] [walk_photo] name: Recently Deleted, offset: 99, size=647, got=99, desc=false
[icloudgo] [walk_photo] name: Recently Deleted, offset: 198, size=647, got=100, desc=false
[icloudgo] [walk_photo] name: Recently Deleted, offset: 298, size=647, got=100, desc=false
[icloudgo] [walk_photo] name: Recently Deleted, offset: 398, size=647, got=28, desc=false
[icloudgo] [walk_photo] name: Recently Deleted, offset: 426, size=647, got=0, desc=false
同问
| gharchive/issue | 2024-01-25T16:28:07 | 2025-04-01T04:33:47.579870 | {
"authors": [
"RicardioMu",
"a406877521"
],
"repo": "chyroc/icloudgo",
"url": "https://github.com/chyroc/icloudgo/issues/14",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
493756794 | Use of priors for weights
Hi,
thanks for the easy-to-use code!
I had a question about the use of priors for the weights when calculating the log probability to be used inside the stein gradient update equation: is that something without which things would not work in your experience? I was wondering since the stein gradient just needs us to specify a log_prob that we're interested in maximizing, and so doing svgd with just the model likelihood (wrt ground truth data) as the log_p is also correct right?
I just wanted to clarify if the weight priors are something desirable that we can choose to add t the log_p term because of desired regularisation(as specified in your accompanying paper) or if that's something essential that if not included, renders the math/theory wrong.
Thanks,
Gunshi
Hi Gunshi,
Thanks for reaching out. SVGD moves particles iteratively towards posterior, which is likelihood * prior (up to a constant). The prior specifies our belief (usually independent Normal in Bayesian NNs), which is student-t in our case.
| gharchive/issue | 2019-09-15T16:16:03 | 2025-04-01T04:33:47.585271 | {
"authors": [
"bmmi",
"gunshi"
],
"repo": "cics-nd/cnn-surrogate",
"url": "https://github.com/cics-nd/cnn-surrogate/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
178384082 | Add Test Helper
MINOR
CHANGELOG
Added new test helper, stubPollboy, for acceptance tests that mocks the polling service to not actually poll (otherwise acceptance tests will never complete).
Replaced ember-cli-blanket with ember-cli-code-coverage for test coverage.
Upgraded devDependencies to latest versions.
:+1:
Hey Matthew, I've been trying to import this stubPollboy test helper (or any module in /test-support/helpers/ember-pollboy.js for that matter) and can't figure out how to do so. Have you been able to successfully do that in a consuming app?
| gharchive/pull-request | 2016-09-21T15:37:07 | 2025-04-01T04:33:47.588793 | {
"authors": [
"NickChristensen",
"sandersky"
],
"repo": "ciena-blueplanet/ember-pollboy",
"url": "https://github.com/ciena-blueplanet/ember-pollboy/pull/24",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
192922397 | Fixed dependency issue with ember-spread
This project uses semver, please check the scope of this pr:
[x] #patch# - backwards-compatible bug fix
[ ] #minor# - adding functionality in a backwards-compatible manner
[ ] #major# - incompatible API change
CHANGELOG
Fixed dependency issue with ember-spread
Changes Unknown when pulling c4d3bd590508cf344b8276e5d33c2ffd403f196f on dafortin:master into ** on ciena-frost:master**.
Changes Unknown when pulling c4d3bd590508cf344b8276e5d33c2ffd403f196f on dafortin:master into ** on ciena-frost:master**.
Changes Unknown when pulling 93546a419b196c1fc52ca9938439bc277abdb010 on dafortin:master into ** on ciena-frost:master**.
👍
| gharchive/pull-request | 2016-12-01T18:15:43 | 2025-04-01T04:33:47.594144 | {
"authors": [
"coveralls",
"dafortin",
"juwara0"
],
"repo": "ciena-frost/ember-frost-core",
"url": "https://github.com/ciena-frost/ember-frost-core/pull/335",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
246471242 | Propagate Status field for CiliumNetworkPolicy
A possible status field could look like this:
@@ -33,6 +44,80 @@ type CiliumNetworkPolicy struct {
// Spec is the desired Cilium specific rule specification.
Spec api.Rule `json:"spec"`
+
+ // Status is the status of the Cilium policy rule
+ Status CiliumNetworkPolicyStatus `json:"status"`
+}
+
+// CiliumNetworkPolicyStatus is the status of a Cilium policy rule
+type CiliumNetworkPolicyStatus struct {
+ // Nodes is the Cilium policy status for each node
+ Nodes map[string]CiliumNetworkPolicyNodeStatus
+
+ // Failing is the number of nodes failing to install this policy
+ Failing int
+}
+
+// CiliumNetworkPolicyNodeStatus is the status of a Cilium policy rule for a
+// specific node
+type CiliumNetworkPolicyNodeStatus struct {
+ // OK is true when the policy has been installed successfully
+ OK bool
+
+ // Message describes the error condition if OK is false
+ Message string
+}
Blocked on #1171
| gharchive/issue | 2017-07-28T22:13:03 | 2025-04-01T04:33:47.606183 | {
"authors": [
"tgraf"
],
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/issues/1222",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
842305700 | helm: Consolidate IPSec and Wireguard encryption options
Suggested in https://github.com/cilium/cilium/pull/15383#discussion_r600323895.
@brb taking care of it. I submitted a PR, let me know if this makes sense. Thanks!
| gharchive/issue | 2021-03-26T21:24:45 | 2025-04-01T04:33:47.607467 | {
"authors": [
"brb",
"youssefazrak"
],
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/issues/15483",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2065785568 | Random traffic is dropped from pods based on same cronjob
Is there an existing issue for this?
[X] I have searched the existing issues
What happened?
Problem is with connections being dropped between cronjobs (about ~200) and redis pod. I have cronjobs in test-cron namespace and redis in cache namespace. Pods are just simple test pods to reproduce the problem, they are run every second they
set key
sleep 5-15 sec
del key
About ~30 are run simultaneously, the rest is in Pending or ContainerCreating states. Most of the runs are ok but every once in a while I am getting dropped connection that looks random. I'd expect all of the traffic to be forwarded not only part of it.
hubble ui drops are described as denied by ingress policy (from the redis side), when I run hubble observe I got
ciliumnetworkpolicies are (ns cache)
apiVersion: cilium.io/v2
kind: CiliumNetworkPolicy
metadata:
name: default-deny-all
namespace: cache
spec:
endpointSelector: {}
ingress:
- {}
egress:
- {}
apiVersion: cilium.io/v2
kind: CiliumNetworkPolicy
metadata:
name: redis-locks-test-allowed-ingress
namespace: cache
spec:
endpointSelector:
matchLabels:
app: redis-locks
ingress:
- fromEntities:
- cluster
toPorts:
- ports:
- port: "6379"
Cilium Version
1.14.5
same on 1.14.2 & 1.14.3
Kernel Version
4.18.0-513.9.1.el8_9.x86_64 #1 SMP Sat Dec 2 05:23:44 EST 2023 x86_64 x86_64 x86_64 GNU/Linux
Kubernetes Version
1.28.5
same on 1.27.6 & 1.27.8
Sysdump
No response
Relevant log output
Jan 4 12:39:37.461: test-cron/test-cron-action-4b-com-28406198-x74j7:43470 (ID:45713) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:39:44.403: 172.0.2.211:60748 (ID:3892) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:39:47.490: 172.0.2.156:35134 (ID:27262) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:39:49.460: 172.0.2.114:38792 (ID:31444) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:39:51.452: 172.0.2.24:59266 (ID:56353) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:40:03.452: test-cron/test-cron-action-24a-com-28406199-f52wt:45798 (ID:65520) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:40:05.449: test-cron/test-cron-action-156a-com-28406199-lbbpv:53028 (ID:550) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:40:08.403: test-cron/test-cron-action-173c-com-28406199-rzj45:47352 (ID:12223) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:40:09.358: test-cron/test-cron-action-195a-pl-28406199-qwxxp:45240 (ID:18252) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:40:09.505: test-cron/test-cron-action-18c-pl-28406199-wj9f6:38280 (ID:60444) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:40:11.337: 172.0.2.174:47778 (ID:51667) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:40:12.510: 172.0.2.102:45098 (ID:33239) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:40:14.364: 172.0.2.228:47170 (ID:14774) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:40:37.495: 172.0.2.219:52740 (ID:3888) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:40:39.521: 172.0.2.216:43380 (ID:49924) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:40:41.551: 172.0.2.239:45874 (ID:19921) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:41:02.441: 172.0.2.180:42846 (ID:9040) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:41:03.468: 172.0.2.130:52516 (ID:21908) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:41:08.474: 172.0.2.171:44416 (ID:29453) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:41:12.456: test-cron/test-cron-action-17c-com-28406200-stsmm:37970 (ID:58486) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:41:25.345: test-cron/test-cron-action-54b-pl-28406200-m6ntq:33294 (ID:10021) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:41:27.569: test-cron/test-cron-action-160c-com-28406200-9z6h4:53826 (ID:10217) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:41:32.546: 172.0.2.143:57442 (ID:34280) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:41:35.505: 172.0.2.108:56142 (ID:19078) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:41:54.561: test-cron/test-cron-action-163a-pl-28406201-x8v6q:46178 (ID:36061) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:41:58.518: 172.0.2.158:53620 (ID:39004) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:42:03.565: 172.0.2.162:50478 (ID:48558) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:42:17.541: test-cron/test-cron-action-55c-com-28406201-fndk5:50100 (ID:7779) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:42:23.586: 172.0.2.217:37020 (ID:30664) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:42:36.551: test-cron/test-cron-action-169a-pl-28406201-ngrkj:60860 (ID:56003) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:42:39.455: test-cron/test-cron-action-177a-pl-28406201-knlkq:48136 (ID:37684) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:42:40.511: test-cron/test-cron-action-160a-com-28406202-rmjwq:46988 (ID:20696) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:42:43.641: 172.0.2.238:44750 (ID:23205) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:42:46.530: 172.0.2.145:40640 (ID:49980) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Jan 4 12:42:47.475: 172.0.2.118:54932 (ID:39283) <> cache/redis-locks-778d798f8-fsg5q:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Anything else?
Looks like when I removed default-deny-all policy the drops do not occur, but I am not 100% sure as this also can be that "randomness"
We have similar default deny all policies on all of our ns but this is the first one that causes us problems. We reproduced same case in our production k8s cluster and in our test cluster.
Code of Conduct
[X] I agree to follow this project's Code of Conduct
I checked again with default-deny-all policy removed - I still got random drops
removing both default-deny-all and redis-locks-test-allowed-ingress seems to stop the drops but it cant be the solution here
Hey devs any idea how to debug this issue?
Hello 👋
How large is the cluster in question?
I'm not super familiar with how fromEntities: cluster works under the hood, but I'm curious why you're using that as opposed to individual namespaces, or pod labels.
Also, does this continue to happen for the lifetime of the CronJobs, or does it fix itself after a few seconds?
The cluster where I am reproducing the problem has 3 nodes , 1 worker node. Second one, our main cluster has 13 nodes, 10 worker nodes - same problem here.
I updated the logic, so that it does 5 iterations and 2 calls to redis in each one.
$i = 5;
while ($i--)
{
$lock->acquire($id, 100);
sleep(5 + rand(5, 10));
$lock->release($id);
}
with hubble observe I can confirm that in a single pod some of the requests were forwarded some were not, basically the same requests. Logs
Jan 12 12:03:02.696: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) policy-verdict:L3-L4 INGRESS ALLOWED (TCP Flags: SYN)
Jan 12 12:03:02.696: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: SYN)
Jan 12 12:03:02.696: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK)
Jan 12 12:03:02.696: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH)
Jan 12 12:03:17.697: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) <> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) policy-verdict:none EGRESS DENIED (TCP Flags: ACK, PSH)
Jan 12 12:03:17.697: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) <> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) Policy denied DROPPED (TCP Flags: ACK, PSH)
Jan 12 12:03:17.904: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) <> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) policy-verdict:none EGRESS DENIED (TCP Flags: ACK, PSH)
Jan 12 12:03:17.904: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) <> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) Policy denied DROPPED (TCP Flags: ACK, PSH)
Jan 12 12:03:18.113: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) <> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) policy-verdict:none EGRESS DENIED (TCP Flags: ACK, PSH)
Jan 12 12:03:18.113: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) <> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) Policy denied DROPPED (TCP Flags: ACK, PSH)
Jan 12 12:03:18.520: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH)
Jan 12 12:03:31.521: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH)
Jan 12 12:03:45.521: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH)
Jan 12 12:04:00.521: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH)
Jan 12 12:04:15.522: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH)
Jan 12 12:04:15.523: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, FIN)
Jan 12 12:04:15.523: test-cron/test-cron-action-183a-pl-28417683-k7drt:55946 (ID:15453) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK)
One thing that's particularly interesting about the log you posted is that it's policy-verdict:none EGRESS DENIED, which means that the caller is rejecting, not redis. Do you potentially have CiliumNetworkPolicy resources for the CronJob as well, and can you send those over/
We have udated cilium to 1.14.6, problem still persists. I can't reproduce policy-verdict:none EGRESS DENIED though. As I mentioned in the first post, it seemed like an ingress policy issue and with my latest logs I can confirm it is still an ingress issue.
I run another hubble observe like this (for 3 cilium agent pods in our test cluster)
kubectl -n kube-system exec -ti cilium-9jhtl -- hubble observe -f --to-namespace cache >> .dev/cilium-9jhtl.log
with this I got 13 endpoints that got drops in about 15 minutes. About ~40 pods running at the same time with the test code.
each one looks the same, for example ID:11143
Feb 2 11:07:37.567: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) <> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) policy-verdict:none INGRESS DENIED (TCP Flags: SYN)
Feb 2 11:07:37.567: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) <> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) Policy denied DROPPED (TCP Flags: SYN)
Feb 2 11:07:38.587: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) policy-verdict:L3-L4 INGRESS ALLOWED (TCP Flags: SYN)
Feb 2 11:07:38.587: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: SYN)
Feb 2 11:07:38.587: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK)
Feb 2 11:07:38.587: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH)
Feb 2 11:07:52.588: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH)
Feb 2 11:08:04.589: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH)
Feb 2 11:08:16.590: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH)
Feb 2 11:08:27.590: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH)
Feb 2 11:08:38.591: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, PSH)
Feb 2 11:08:38.592: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK, FIN)
Feb 2 11:08:38.592: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-endpoint FORWARDED (TCP Flags: ACK)
Feb 2 11:07:37.567: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) policy-verdict:L3-L4 EGRESS ALLOWED (TCP Flags: SYN)
Feb 2 11:07:37.567: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-overlay FORWARDED (TCP Flags: SYN)
Feb 2 11:07:38.588: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-overlay FORWARDED (TCP Flags: ACK)
Feb 2 11:07:38.588: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-overlay FORWARDED (TCP Flags: ACK, PSH)
Feb 2 11:07:52.588: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-overlay FORWARDED (TCP Flags: ACK, PSH)
Feb 2 11:08:04.589: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-overlay FORWARDED (TCP Flags: ACK, PSH)
Feb 2 11:08:16.590: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-overlay FORWARDED (TCP Flags: ACK, PSH)
Feb 2 11:08:27.591: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-overlay FORWARDED (TCP Flags: ACK, PSH)
Feb 2 11:08:38.591: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-overlay FORWARDED (TCP Flags: ACK, PSH)
Feb 2 11:08:38.592: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-overlay FORWARDED (TCP Flags: ACK, FIN)
Feb 2 11:08:38.593: test-cron/test-cron-action-55c-pl-28447867-zgmvh:46544 (ID:11143) -> cache/redis-locks-75c455b7bf-cxplr:6379 (ID:50432) to-overlay FORWARDED (TCP Flags: ACK)
In all of these logs INGRESS DENIED is right before INGRESS ALLOWED and there is EGRESS ALLOWED at some point. In 12 of cases EGRESS ALLOWED is after INGRESS ALLOWED, in 1 case it is before INGRES DENIED, but since it is ruled as ALLOWED I'm not sure if that particullar log line matters. So to sum up it looks like this
12 cases
... policy-verdict:none INGRESS DENIED ...
... Policy denied DROPPED ...
... policy-verdict:L3-L4 INGRESS ALLOWED ...
... [logs for FORWARDED] ...
... policy-verdict:L3-L4 EGRESS ALLOWED ...
... [logs for FORWARDED] ...
1 case
... policy-verdict:L3-L4 EGRESS ALLOWED ...
... policy-verdict:none INGRESS DENIED ...
... Policy denied DROPPED ...
... policy-verdict:L3-L4 INGRESS ALLOWED ...
... [logs for FORWARDED] ...
I've tried to use fromEndpoint instead of fromEntities. I had to change resource to CiliumClusterwideNetworkPolicy to make it cross-namespace with my test label (with CiliumNetworkPolicy I had nothing but drops). Change I made
apiVersion: cilium.io/v2
-kind: CiliumNetworkPolicy
+kind: CiliumClusterwideNetworkPolicy
metadata:
name: redis-locks-test-allowed-ingress
spec:
endpointSelector:
matchLabels:
app: redis-locks
ingress:
- - fromEntities:
- - cluster
- toPorts:
- - ports:
- - port: "6379"
+ - fromEndpoints:
+ - matchLabels:
+ mylabel: "test"
and I added mylabel: "test" to test crons. Result was the same as above, basically I got about 8 drops in few minutes with output like this
Feb 5 10:10:19.626: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) policy-verdict:L3-L4 EGRESS ALLOWED (TCP Flags: SYN)
Feb 5 10:10:19.626: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-overlay FORWARDED (TCP Flags: SYN)
Feb 5 10:10:20.635: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-overlay FORWARDED (TCP Flags: ACK)
Feb 5 10:10:20.636: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-overlay FORWARDED (TCP Flags: ACK, PSH)
Feb 5 10:10:31.636: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-overlay FORWARDED (TCP Flags: ACK, PSH)
Feb 5 10:10:41.637: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-overlay FORWARDED (TCP Flags: ACK, PSH)
Feb 5 10:10:54.638: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-overlay FORWARDED (TCP Flags: ACK, PSH)
Feb 5 10:11:06.639: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-overlay FORWARDED (TCP Flags: ACK, PSH)
Feb 5 10:11:21.640: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-overlay FORWARDED (TCP Flags: ACK, PSH)
Feb 5 10:11:21.641: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-overlay FORWARDED (TCP Flags: ACK, FIN)
Feb 5 10:11:21.641: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-overlay FORWARDED (TCP Flags: ACK)
Feb 5 10:10:19.626: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) <> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) policy-verdict:none INGRESS DENIED (TCP Flags: SYN)
Feb 5 10:10:19.626: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) <> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) Policy denied DROPPED (TCP Flags: SYN)
Feb 5 10:10:20.635: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) policy-verdict:L3-Only INGRESS ALLOWED (TCP Flags: SYN)
Feb 5 10:10:20.635: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-endpoint FORWARDED (TCP Flags: SYN)
Feb 5 10:10:20.636: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-endpoint FORWARDED (TCP Flags: ACK)
Feb 5 10:10:20.636: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-endpoint FORWARDED (TCP Flags: ACK, PSH)
Feb 5 10:10:31.637: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-endpoint FORWARDED (TCP Flags: ACK, PSH)
Feb 5 10:10:41.637: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-endpoint FORWARDED (TCP Flags: ACK, PSH)
Feb 5 10:10:54.638: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-endpoint FORWARDED (TCP Flags: ACK, PSH)
Feb 5 10:11:06.639: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-endpoint FORWARDED (TCP Flags: ACK, PSH)
Feb 5 10:11:21.640: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-endpoint FORWARDED (TCP Flags: ACK, PSH)
Feb 5 10:11:21.641: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-endpoint FORWARDED (TCP Flags: ACK, FIN)
Feb 5 10:11:21.641: test-cron/test-cron-action-141c-com-28452130-mwtc6:54456 (ID:57350) -> cache/redis-locks-7755dd4859-lrsdm:6379 (ID:29518) to-endpoint FORWARDED (TCP Flags: ACK)
which is a pattern like that one
... policy-verdict:L3-L4 EGRESS ALLOWED ...
... [logs for FORWARDED] ...
... policy-verdict:none INGRESS DENIED ...
... Policy denied DROPPED ...
... policy-verdict:L3-L4 INGRESS ALLOWED ...
... [logs for FORWARDED] ...
We updated to 1.14.7 and it seems to solve the problem.
| gharchive/issue | 2024-01-04T14:44:12 | 2025-04-01T04:33:47.625515 | {
"authors": [
"EItanya",
"adamw-linadm",
"pawelbaranski"
],
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/issues/30105",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2079881972 | CI: Conformance Runtime - Test_PodIPPoolAdvert failure
CI failure
Hit on https://github.com/cilium/cilium/pull/30183
Link: https://github.com/cilium/cilium/actions/runs/7480430927/job/20359924251
--- FAIL: Test_PodIPPoolAdvert (8.07s)
--- FAIL: Test_PodIPPoolAdvert/matching_ipv4_pool (0.00s)
adverts_test.go:430: expected events: [{65001 11.1.1.0 24 false []}]
adverts_test.go:431: received events: [{65001 10.1.1.0 24 false []}]
adverts_test.go:432:
Error Trace: /host/pkg/bgpv1/test/adverts_test.go:432
Error: elements differ
extra elements in list A:
([]interface {}) (len=1) {
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=8) "11.1.1.0",
prefixLen: (uint8) 24,
isWithdrawn: (bool) false,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
}
}
extra elements in list B:
([]interface {}) (len=1) {
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=8) "10.1.1.0",
prefixLen: (uint8) 24,
isWithdrawn: (bool) false,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
}
}
listA:
([]test.routeEvent) (len=1) {
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=8) "11.1.1.0",
prefixLen: (uint8) 24,
isWithdrawn: (bool) false,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
}
}
listB:
([]test.routeEvent) (len=1) {
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=8) "10.1.1.0",
prefixLen: (uint8) 24,
isWithdrawn: (bool) false,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
}
}
Test: Test_PodIPPoolAdvert/matching_ipv4_pool
Messages: matching ipv4 pool
--- FAIL: Test_PodIPPoolAdvert/update_matching_ipv4_pool (0.00s)
adverts_test.go:430: expected events: [{65001 11.1.1.0 24 true []} {65001 11.2.1.0 24 false []}]
adverts_test.go:431: received events: [{65001 10.1.2.0 24 false []} {65001 11.1.1.0 24 false []}]
adverts_test.go:432:
Error Trace: /host/pkg/bgpv1/test/adverts_test.go:432
Error: elements differ
extra elements in list A:
([]interface {}) (len=2) {
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=8) "11.1.1.0",
prefixLen: (uint8) 24,
isWithdrawn: (bool) true,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
},
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=8) "11.2.1.0",
prefixLen: (uint8) 24,
isWithdrawn: (bool) false,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
}
}
extra elements in list B:
([]interface {}) (len=2) {
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=8) "10.1.2.0",
prefixLen: (uint8) 24,
isWithdrawn: (bool) false,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
},
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=8) "11.1.1.0",
prefixLen: (uint8) 24,
isWithdrawn: (bool) false,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
}
}
listA:
([]test.routeEvent) (len=2) {
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=8) "11.1.1.0",
prefixLen: (uint8) 24,
isWithdrawn: (bool) true,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
},
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=8) "11.2.1.0",
prefixLen: (uint8) 24,
isWithdrawn: (bool) false,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
}
}
listB:
([]test.routeEvent) (len=2) {
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=8) "10.1.2.0",
prefixLen: (uint8) 24,
isWithdrawn: (bool) false,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
},
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=8) "11.1.1.0",
prefixLen: (uint8) 24,
isWithdrawn: (bool) false,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
}
}
Test: Test_PodIPPoolAdvert/update_matching_ipv4_pool
Messages: update matching ipv4 pool
--- FAIL: Test_PodIPPoolAdvert/matching_ipv6_pool (0.00s)
adverts_test.go:430: expected events: [{65001 2001:0:0:1234:5678:: 96 false []} {65001 11.2.1.0 24 true []}]
adverts_test.go:431: received events: [{65001 10.1.2.0 24 true []} {65001 10.1.1.0 24 true []}]
adverts_test.go:432:
Error Trace: /host/pkg/bgpv1/test/adverts_test.go:432
Error: elements differ
extra elements in list A:
([]interface {}) (len=2) {
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=20) "2001:0:0:1234:5678::",
prefixLen: (uint8) 96,
isWithdrawn: (bool) false,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
},
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=8) "11.2.1.0",
prefixLen: (uint8) 24,
isWithdrawn: (bool) true,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
}
}
extra elements in list B:
([]interface {}) (len=2) {
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=8) "10.1.2.0",
prefixLen: (uint8) 24,
isWithdrawn: (bool) true,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
},
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=8) "10.1.1.0",
prefixLen: (uint8) 24,
isWithdrawn: (bool) true,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
}
}
listA:
([]test.routeEvent) (len=2) {
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=20) "2001:0:0:1234:5678::",
prefixLen: (uint8) 96,
isWithdrawn: (bool) false,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
},
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=8) "11.2.1.0",
prefixLen: (uint8) 24,
isWithdrawn: (bool) true,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
}
}
listB:
([]test.routeEvent) (len=2) {
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=8) "10.1.2.0",
prefixLen: (uint8) 24,
isWithdrawn: (bool) true,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
},
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=8) "10.1.1.0",
prefixLen: (uint8) 24,
isWithdrawn: (bool) true,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
}
}
Test: Test_PodIPPoolAdvert/matching_ipv6_pool
Messages: matching ipv6 pool
--- FAIL: Test_PodIPPoolAdvert/update_matching_ipv6_pool (0.00s)
adverts_test.go:430: expected events: [{65001 2001:0:0:1234:5678:: 96 true []} {65001 2002:0:0:1234:5678:: 96 false []}]
adverts_test.go:431: received events: [{65001 11.2.1.0 24 false []} {65001 11.1.1.0 24 true []}]
adverts_test.go:432:
Error Trace: /host/pkg/bgpv1/test/adverts_test.go:432
Error: elements differ
extra elements in list A:
([]interface {}) (len=2) {
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=20) "2001:0:0:1234:5678::",
prefixLen: (uint8) 96,
isWithdrawn: (bool) true,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
},
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=20) "2002:0:0:1234:5678::",
prefixLen: (uint8) 96,
isWithdrawn: (bool) false,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
}
}
extra elements in list B:
([]interface {}) (len=2) {
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=8) "11.2.1.0",
prefixLen: (uint8) 24,
isWithdrawn: (bool) false,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
},
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=8) "11.1.1.0",
prefixLen: (uint8) 24,
isWithdrawn: (bool) true,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
}
}
listA:
([]test.routeEvent) (len=2) {
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=20) "2001:0:0:1234:5678::",
prefixLen: (uint8) 96,
isWithdrawn: (bool) true,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
},
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=20) "2002:0:0:1234:5678::",
prefixLen: (uint8) 96,
isWithdrawn: (bool) false,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
}
}
listB:
([]test.routeEvent) (len=2) {
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=8) "11.2.1.0",
prefixLen: (uint8) 24,
isWithdrawn: (bool) false,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
},
(test.routeEvent) {
sourceASN: (uint32) 65001,
prefix: (string) (len=8) "11.1.1.0",
prefixLen: (uint8) 24,
isWithdrawn: (bool) true,
extraPathAttributes: ([]bgp.PathAttributeInterface) <nil>
}
}
Test: Test_PodIPPoolAdvert/update_matching_ipv6_pool
Messages: update matching ipv6 pool
Hit in https://github.com/cilium/cilium/pull/30440 - job https://github.com/cilium/cilium/actions/runs/7738391682/job/21099140614
+1 Hit this just now.
https://github.com/cilium/cilium/actions/runs/8017423656/job/21901426809
Logs: logs_2067640.zip
Seen again in https://github.com/cilium/cilium/pull/33860 at https://github.com/cilium/cilium/actions/runs/9979635155/job/27579321909.
cc @rastislavs can you take a look?
Seen again in #33860 at https://github.com/cilium/cilium/actions/runs/9979635155/job/27579321909.
yeah, it seems to be to same flake, looks like the root cause is in something else than what I fixed :(
Hit again as well: https://github.com/cilium/cilium/actions/runs/10006033738/job/27657940599
Fixed in https://github.com/cilium/cilium/pull/34270 and already backported to v1.16 & v1.15
| gharchive/issue | 2024-01-12T23:32:29 | 2025-04-01T04:33:47.637735 | {
"authors": [
"christarazi",
"dylandreimerink",
"giorio94",
"rastislavs"
],
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/issues/30237",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2298480897 | Ingress controller load balancer can not connect to nodes
Is there an existing issue for this?
[X] I have searched the existing issues
What happened?
We have an EKS cluster where we are trying to use Cilium ingress controller and the load balancer created for the ingress can not always connect to the nodes.
What we see is that the load balancer can connect to some nodes during periods but is not a consistent behavior and there is no pattern between the nodes behind that it can connect and the ones that can not.
Checking directly in the nodes also connecting to the nodePort opened for the load balancer does not work so should not be a problem of security groups, anyway we tried opening traffic from every internal address and nothing, some nodes work and others not or even sometimes no nodes happen to be accessible by the load balancer.
I checked and all the nodes have this cilium LB configuration for the nodePort:
10.218.248.217:31799 0.0.0.0:0 (331) (0) [NodePort, l7-load-balancer]
0.0.0.0:31799 0.0.0.0:0 (333) (0) [NodePort, non-routable, l7-load-balancer]
10.0.243.9:31799 0.0.0.0:0 (330) (0) [NodePort, l7-load-balancer]
169.254.0.11:31799 0.0.0.0:0 (332) (0) [NodePort, l7-load-balancer]
Configuration values used:
cni:
configMap: cni-config
customConf: true
eni:
enabled: true
updateEC2AdapterLimitViaAPI: true
awsEnablePrefixDelegation: true
awsReleaseExcessIPs: true
egressMasqueradeInterfaces: eth0
policyEnforcementMode: "never"
ipam:
mode: eni
hubble:
relay:
enabled: true
ui:
enabled: true
tunnelProtocol: ""
nodePort:
enabled: true
nodeinit:
enabled: true
ingressController:
enabled: true
cni-config configmap values:
{
"cniVersion":"0.3.1",
"name":"cilium",
"plugins": [
{
"cniVersion":"0.3.1",
"type":"cilium-cni",
"eni": {
"subnet-ids": ["subnet-xxxxxx", "subnet-xxxxxx", "subnet-xxxxxxx"],
"first-interface-index": 1
}
}
]
}
Cilium Version
We tried it in multiple versions:
1.14.1
1.15.4
1.16.0-pre.0
Kernel Version
Linux 5.10.215-203.850.amzn2.aarch64
Kubernetes Version
v1.26.15
Regression
No response
Sysdump
Relevant log output
No response
Anything else?
No response
Cilium Users Document
[ ] Are you a user of Cilium? Please add yourself to the Users doc
Code of Conduct
[X] I agree to follow this project's Code of Conduct
Hi there, thanks for the bug report. It's not yet clear to me how exactly traffic is flowing. Could you outline the expected traffic flow, and indicate where you think it is failing?
In particular, I suggest the section on troubleshooting with hubble to identify where packets are being dropped. Can you go through the troubleshooting section and clarify the problem a bit?
Thanks.
Also can you share your cilium configmap as well? Thanks.
Hi there, thanks for the bug report. It's not yet clear to me how exactly traffic is flowing. Could you outline the expected traffic flow, and indicate where you think it is failing?
In particular, I suggest the section on troubleshooting with hubble to identify where packets are being dropped. Can you go through the troubleshooting section and clarify the problem a bit?
Thanks.
@squeed
Hi, sorry for the delay, yes let me explain it better. Correct me if I mention something wrong. The idea is to use Cilium as an Ingress Controller, when I create an ingress object it creates the Classic AWS LB or NLB, tried both, which will balance the traffic to the Cilium ingress controller. If I'm not wrong the component of Cilium that handles the traffic coming from the LB is cilium-envoy which runs inside cilium-agent in my case. The traffic after arriving to cilium-envoy gets sent to the relevant backend of the Ingress. My problem is the communication between the Load Balancer and envoy, the Load Balancer can not target envoy most of the time.
Ask any other question that you need if I still did not explain it well enough.
Also can you share your cilium configmap as well? Thanks.
@sayboras Yes, here it is:
agent-not-ready-taint-key: node.cilium.io/agent-not-ready
arping-refresh-period: 30s
auto-direct-node-routes: "false"
bpf-lb-acceleration: disabled
bpf-lb-external-clusterip: "false"
bpf-lb-map-max: "65536"
bpf-lb-sock: "false"
bpf-map-dynamic-size-ratio: "0.0025"
bpf-policy-map-max: "16384"
bpf-root: /sys/fs/bpf
cgroup-root: /run/cilium/cgroupv2
cilium-endpoint-gc-interval: 5m0s
cluster-id: "0"
cluster-name: default
cluster-pool-ipv4-cidr: 10.0.0.0/8
cluster-pool-ipv4-mask-size: "24"
cni-chaining-mode: aws-cni
cni-exclusive: "false"
cni-log-file: /var/run/cilium/cilium-cni.log
custom-cni-conf: "false"
debug: "false"
debug-verbose: ""
egress-gateway-reconciliation-trigger-interval: 1s
enable-auto-protect-node-port-range: "true"
enable-bgp-control-plane: "false"
enable-bpf-clock-probe: "false"
enable-endpoint-health-checking: "false"
enable-endpoint-routes: "true"
enable-envoy-config: "true"
enable-external-ips: "false"
enable-gateway-api: "true"
enable-gateway-api-secrets-sync: "true"
enable-health-check-loadbalancer-ip: "false"
enable-health-check-nodeport: "true"
enable-health-checking: "true"
enable-host-legacy-routing: "true"
enable-host-port: "false"
enable-hubble: "true"
enable-ingress-controller: "true"
enable-ingress-proxy-protocol: "false"
enable-ingress-secrets-sync: "true"
enable-ipv4: "true"
enable-ipv4-big-tcp: "false"
enable-ipv4-masquerade: "false"
enable-ipv6: "false"
enable-ipv6-big-tcp: "false"
enable-ipv6-masquerade: "true"
enable-k8s-networkpolicy: "true"
enable-k8s-terminating-endpoint: "true"
enable-l2-neigh-discovery: "true"
enable-l7-proxy: "true"
enable-local-node-route: "false"
enable-local-redirect-policy: "false"
enable-masquerade-to-route-source: "false"
enable-metrics: "true"
enable-node-port: "true"
enable-policy: never
enable-remote-node-identity: "true"
enable-sctp: "false"
enable-svc-source-range-check: "true"
enable-vtep: "false"
enable-well-known-identities: "false"
enable-xt-socket-fallback: "true"
enforce-ingress-https: "true"
external-envoy-proxy: "false"
gateway-api-secrets-namespace: cilium-secrets
hubble-disable-tls: "false"
hubble-export-file-max-backups: "5"
hubble-export-file-max-size-mb: "10"
hubble-listen-address: :4244
hubble-socket-path: /var/run/cilium/hubble.sock
hubble-tls-cert-file: /var/lib/cilium/tls/hubble/server.crt
hubble-tls-client-ca-files: /var/lib/cilium/tls/hubble/client-ca.crt
hubble-tls-key-file: /var/lib/cilium/tls/hubble/server.key
identity-allocation-mode: crd
identity-gc-interval: 15m0s
identity-heartbeat-timeout: 30m0s
ingress-default-lb-mode: dedicated
ingress-lb-annotation-prefixes: service.beta.kubernetes.io service.kubernetes.io
cloud.google.com
ingress-secrets-namespace: cilium-secrets
ingress-shared-lb-service-name: cilium-ingress
install-no-conntrack-iptables-rules: "false"
ipam: cluster-pool
ipam-cilium-node-update-rate: 15s
k8s-client-burst: "10"
k8s-client-qps: "5"
kube-proxy-replacement: "false"
kube-proxy-replacement-healthz-bind-address: ""
max-connected-clusters: "255"
mesh-auth-enabled: "true"
mesh-auth-gc-interval: 5m0s
mesh-auth-queue-size: "1024"
mesh-auth-rotated-identities-queue-size: "1024"
monitor-aggregation: medium
monitor-aggregation-flags: all
monitor-aggregation-interval: 5s
node-port-bind-protection: "true"
nodes-gc-interval: 5m0s
operator-api-serve-addr: 127.0.0.1:9234
operator-prometheus-serve-addr: :9963
policy-cidr-match-mode: ""
preallocate-bpf-maps: "false"
procfs: /host/proc
proxy-connect-timeout: "2"
proxy-idle-timeout-seconds: "60"
proxy-max-connection-duration-seconds: "0"
proxy-max-requests-per-connection: "0"
proxy-prometheus-port: "9964"
proxy-xff-num-trusted-hops-egress: "0"
proxy-xff-num-trusted-hops-ingress: "0"
remove-cilium-node-taints: "true"
routing-mode: native
service-no-backend-response: reject
set-cilium-is-up-condition: "true"
set-cilium-node-taints: "true"
sidecar-istio-proxy-image: cilium/istio_proxy
skip-cnp-status-startup-clean: "false"
synchronize-k8s-nodes: "true"
tofqdns-dns-reject-response-code: refused
tofqdns-enable-dns-compression: "true"
tofqdns-endpoint-max-ip-per-hostname: "50"
tofqdns-idle-connection-grace-period: 0s
tofqdns-max-deferred-connection-deletes: "10000"
tofqdns-proxy-response-max-delay: 100ms
unmanaged-pod-watcher-interval: "15"
vtep-cidr: ""
vtep-endpoint: ""
vtep-mac: ""
vtep-mask: ""
write-cni-conf-when-ready: /host/etc/cni/net.d/05-cilium.conflist
Thank you for taking a look into this!
| gharchive/issue | 2024-05-15T17:42:56 | 2025-04-01T04:33:47.651401 | {
"authors": [
"carlosrejano",
"sayboras",
"squeed"
],
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/issues/32556",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1454726110 | fix: package name of test folder
1 According to the code specification, the package name should be lower case.
2 Refer to the official documentation of gingko to modify package name
https://onsi.github.io/ginkgo/
Could you please rebase your commit instead of merging master into this PR branch?
Could you please rebase your commit instead of merging master into this PR branch?
okay
Hi @yulng , is this related to a problem you were having? It could be helpful to share what that problem is in the PR description so that others can see / search for the same problem.
Please also rebase and update the commit message. I have not signed off on the code, the sign-off line should be from you to highlight that you agree to the certificate of origin (description linked by the maintainer's little helper in comments above).
Please also rebase and update the commit message. I have not signed off on the code, the sign-off line should be from you to highlight that you agree to the certificate of origin (description linked by the maintainer's little helper in comments above).
okay,modified
It just doesn't conform to the code specification, which is not conducive to search. :)
It just doesn't conform to the code specification, which is not conducive to search. :)
It just doesn't conform to the code specification, which is not conducive to search. :)
Hi @yulng , is this related to a problem you were having? It could be helpful to share what that problem is in the PR description so that others can see / search for the same problem.
It just doesn't conform to the code specification, which is not conducive to search. :)
| gharchive/pull-request | 2022-11-18T08:54:19 | 2025-04-01T04:33:47.657076 | {
"authors": [
"christarazi",
"joestringer",
"yulng"
],
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/pull/22259",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1646183360 | Add support for FromCIDRGroupRefs Ingress network policies
-- WORK IN PROGRESS, NOT READY FOR REVIEW YET --
remove the useless ref counting in cidr ingress identity allocation (but use the cache to remove identities when a cidr group is updated)
avoid overwriting FromCIDR field in CNP/CCNP
if all the cidr groups of a cnp have been deleted we should completely remove the field FromCIDR in the CNP
@pippolo84 While playing around with this PR, I ran into some cases of nil pointer panics, here's the diff that fixes it:
diff --git a/pkg/k8s/watchers/cilium_network_policy.go b/pkg/k8s/watchers/cilium_network_policy.go
index 45579070fe..b8b3f150c1 100644
--- a/pkg/k8s/watchers/cilium_network_policy.go
+++ b/pkg/k8s/watchers/cilium_network_policy.go
@@ -561,15 +561,11 @@ func reportCNPChangeMetrics(err error) {
}
func containsCIDRGroupRef(cnp *types.SlimCNP, cidrGroup string) bool {
- for _, ingress := range cnp.Spec.Ingress {
- for _, cidrGroupRef := range ingress.FromCIDRGroupRefs {
- if string(cidrGroupRef) == cidrGroup {
- return true
+ if cnp.Spec != nil && cnp.Spec.Ingress != nil {
+ for _, ingress := range cnp.Spec.Ingress {
+ if ingress.FromCIDRGroupRefs == nil {
+ continue
}
- }
- }
- for _, spec := range cnp.Specs {
- for _, ingress := range spec.Ingress {
for _, cidrGroupRef := range ingress.FromCIDRGroupRefs {
if string(cidrGroupRef) == cidrGroup {
return true
@@ -577,36 +573,70 @@ func containsCIDRGroupRef(cnp *types.SlimCNP, cidrGroup string) bool {
}
}
}
+ if cnp.Specs != nil {
+ for _, spec := range cnp.Specs {
+ if spec.Ingress != nil {
+ for _, ingress := range spec.Ingress {
+ if ingress.FromCIDRGroupRefs == nil {
+ continue
+ }
+ for _, cidrGroupRef := range ingress.FromCIDRGroupRefs {
+ if string(cidrGroupRef) == cidrGroup {
+ return true
+ }
+ }
+ }
+ }
+ }
+ }
return false
}
func translateFromCIDRGroupRefs(cnp *types.SlimCNP, cache map[resource.Key]*cilium_v2_alpha1.CiliumCIDRGroup) error {
var cidrGroupRefs []string
- for i := range cnp.Spec.Ingress {
- for _, cidrGroupRef := range cnp.Spec.Ingress[i].FromCIDRGroupRefs {
- cidrGroupRefs = append(cidrGroupRefs, string(cidrGroupRef))
+ // The following causes a nil ptr panic if the policy does not have a
+ // .Spec, but rather only has a .Specs
+ //
+ // The following code should fix it now.
+ if cnp.Spec != nil && cnp.Spec.Ingress != nil {
+ for i := range cnp.Spec.Ingress {
+ for _, cidrGroupRef := range cnp.Spec.Ingress[i].FromCIDRGroupRefs {
+ cidrGroupRefs = append(cidrGroupRefs, string(cidrGroupRef))
+ }
}
}
- for i := range cnp.Specs {
- for j := range cnp.Specs[i].Ingress {
- for _, cidrGroupRef := range cnp.Specs[i].Ingress[j].FromCIDRGroupRefs {
- cidrGroupRefs = append(cidrGroupRefs, string(cidrGroupRef))
+ if cnp.Specs != nil {
+ for i := range cnp.Specs {
+ if cnp.Specs[i].Ingress == nil {
+ continue
+ }
+ for j := range cnp.Specs[i].Ingress {
+ for _, cidrGroupRef := range cnp.Specs[i].Ingress[j].FromCIDRGroupRefs {
+ cidrGroupRefs = append(cidrGroupRefs, string(cidrGroupRef))
+ }
}
}
}
cidrsSet := cidrGroupRefsToCIDRsSet(cidrGroupRefs, cache)
- for i := range cnp.Spec.Ingress {
- for _, cidrGroupRef := range cnp.Spec.Ingress[i].FromCIDRGroupRefs {
- cnp.Spec.Ingress[i].FromCIDR = cidrsSet[string(cidrGroupRef)]
+ if cnp.Spec != nil && cnp.Spec.Ingress != nil {
+ for i := range cnp.Spec.Ingress {
+ for _, cidrGroupRef := range cnp.Spec.Ingress[i].FromCIDRGroupRefs {
+ cnp.Spec.Ingress[i].FromCIDR = cidrsSet[string(cidrGroupRef)]
+ }
}
}
- for i := range cnp.Specs {
- for j := range cnp.Specs[i].Ingress {
- for _, cidrGroupRef := range cnp.Specs[i].Ingress[j].FromCIDRGroupRefs {
- cnp.Specs[i].Ingress[j].FromCIDR = cidrsSet[string(cidrGroupRef)]
+ if cnp.Specs != nil {
+ for i := range cnp.Specs {
+ if cnp.Specs[i].Ingress == nil {
+ continue
+ }
+ for j := range cnp.Specs[i].Ingress {
+ for _, cidrGroupRef := range cnp.Specs[i].Ingress[j].FromCIDRGroupRefs {
+ cnp.Specs[i].Ingress[j].FromCIDR = cidrsSet[string(cidrGroupRef)]
+ }
}
}
}
This patch might be cleaner to apply:
Patch
commit da478208fe4666a54920397d094e12e87f118ce8
Author: Chris Tarazi <chris@isovalent.com>
Date: Mon Apr 3 21:23:12 2023 -0700
k8s: Consolidate handling of CNP and CCNP with sync.Once
Signed-off-by: Chris Tarazi <chris@isovalent.com>
diff --git a/pkg/k8s/watchers/watcher.go b/pkg/k8s/watchers/watcher.go
index 105da248bb..f4f3aa6156 100644
--- a/pkg/k8s/watchers/watcher.go
+++ b/pkg/k8s/watchers/watcher.go
@@ -540,6 +540,8 @@ func (k *K8sWatcher) enableK8sWatchers(ctx context.Context, resourceNames []stri
return fmt.Errorf("error creating service list option modifier: %w", err)
}
+ // CNP and CCNP resources are handled together.
+ var cnpOnce sync.Once
for _, r := range resourceNames {
switch r {
// Core Cilium
@@ -570,11 +572,8 @@ func (k *K8sWatcher) enableK8sWatchers(ctx context.Context, resourceNames []stri
// only watch secrets in specific namespaces
k.tlsSecretInit(k.clientset.Slim(), option.Config.EnvoySecretNamespaces, swgSecret)
// Custom resource definitions
- case k8sAPIGroupCiliumNetworkPolicyV2:
- k.ciliumNetworkPoliciesInit(ctx, k.clientset)
- case k8sAPIGroupCiliumClusterwideNetworkPolicyV2:
- // handled by k8sAPIGroupCiliumNetworkPolicyV2
- return nil
+ case k8sAPIGroupCiliumNetworkPolicyV2, k8sAPIGroupCiliumClusterwideNetworkPolicyV2:
+ cnpOnce.Do(func() { k.ciliumNetworkPoliciesInit(ctx, k.clientset) })
case k8sAPIGroupCiliumEndpointV2:
k.initCiliumEndpointOrSlices(k.clientset, asyncControllers)
case k8sAPIGroupCiliumEndpointSliceV2Alpha1:
Btw, if I apply a CIDR group and then a CNP which references the group, and then remove the CIDR group, I still see the dangling empty field in the ingress section of cilium policy get:
{
"endpointSelector": {
"matchLabels": {
"any:name": "curl",
"k8s:io.kubernetes.pod.namespace": "default"
}
},
"ingress": [
{},
{
"fromCIDR": [
"3.3.3.3/32"
]
}
],
Btw, if I apply a CIDR group and then a CNP which references the group, and then remove the CIDR group, I still see the dangling empty field in the ingress section of cilium policy get:
{
"endpointSelector": {
"matchLabels": {
"any:name": "curl",
"k8s:io.kubernetes.pod.namespace": "default"
}
},
"ingress": [
{}, <----
{
"fromCIDR": [
"3.3.3.3/32"
]
}
],
I was able to reproduce applying 2 CNPs referencing the same CIDRGroup, one of which with an additional FromCIDR rule. Now it's fixed (I also added an additional unit test for such cases).
/test
Oh, I meant to say - this fixes https://github.com/cilium/cilium/issues/10349 right?
Oh, I meant to say - this fixes #10349 right?
Yes, it does. I've udpated the PR description to reference that issue.
/test
ConformanceEKS is hitting https://github.com/cilium/cilium/issues/24774
I've addressed all feedback from @joestringer and @tommyp1ckles reviews, PTAL :pray:
/test
/test
/mlh new-flake Cilium-PR-K8s-1.16-kernel-4.19
4.19 hit https://github.com/cilium/cilium/issues/24840
/test-1.16-4.19
Hello @pippolo84 , Thank you for your useful contributions , I have found interest in documenting this PR from issue #25123 . I understand that it falls under CiliumNetworkPolicy, CiliumClusterwideNetworkPolicy and CiliumCIDRGroup. I am a cloud technology newbie docs are my specialty , Please help me with an example use case on how to use these new amazing feature under the the respective policies and groups .
| gharchive/pull-request | 2023-03-29T16:55:09 | 2025-04-01T04:33:47.667236 | {
"authors": [
"christarazi",
"gailsuccess",
"joestringer",
"pippolo84"
],
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/pull/24638",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1849438899 | 1.15 bpf nodeport ingress trace
Please ensure your pull request adheres to the following guidelines:
[ ] For first time contributors, read Submitting a pull request
[ ] All code is covered by unit and/or runtime tests where feasible.
[ ] All commits contain a well written commit description including a title,
description and a Fixes: #XXX line if the commit addresses a particular
GitHub issue.
[ ] If your commit description contains a Fixes: <commit-id> tag, then
please add the commit author[s] as reviewer[s] to this issue.
[ ] All commits are signed off. See the section Developer’s Certificate of Origin
[ ] Provide a title or release-note blurb suitable for the release notes.
[ ] Are you a user of Cilium? Please add yourself to the Users doc
[ ] Thanks for contributing!
Fixes: #issue-number
<!-- Enter the release note text here if needed or remove this section! -->
/test
/ci-verifier
/ci-verifier
/test
/test
/test
Sorry for the massive churn. But this needed a fair bit of refactoring to slash through all the tech debt :grimacing: .
/test
/test
(one last rebase to pull #27509 in)
this now required a small change for this PR: https://github.com/cilium/cilium/pull/27488/commits/944acfd51aa4280bc21eac5221d69c81df6794c5.
/test
| gharchive/pull-request | 2023-08-14T09:58:40 | 2025-04-01T04:33:47.674577 | {
"authors": [
"julianwiedmann"
],
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/pull/27488",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2183396824 | bgpv1: fix Test_PodIPPoolAdvert flakiness
Since the test updates multiple k8s resources in each test run, we should be careful about how many of them are we actually changing between the individual test runs, to not trigger multiple reconciliations with different results (advertisements) in a single test run.
This change ensures we change only one value that affects the advertisement between each test runs.
Fixes: #30237
(tested locally with 100 x Test_PodIPPoolAdvert runs, which would always fail as in #30237 before this fix)
/test
/test
/test
/test
/test
| gharchive/pull-request | 2024-03-13T08:28:51 | 2025-04-01T04:33:47.677121 | {
"authors": [
"rastislavs"
],
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/pull/31365",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2744470462 | identity: fix bug where fromNodes could be used to allow custom endpoint
Currently when --enable-node-selector-labels flag is set to true user can use (from/to)Nodes sections in CNPs/CCNPs to allow/block access to/from nodes based on custom node label. The node identity contains not only labels filtered out by --node-labels flag, but also contains reserved:remote-node label as seen in this example:
33554556 node:scif.cz/node=worker
reserved:remote-node
The problematic part is in fromNodes/toNodes sections as those use the same matchLabels selector as classic fromEndpoints/toEndpoints. In case user specifies labels owned not only by node, but also by endpoint he/she might end up in a situation where unexpected endpoints might get access to users deployment.
It is necessary to be either specific or better - prevent this situation.
This commit fixes this by adding a matchExpression with reserved:remote-node key to the policy engine as all nodes have this label.
"fromNodes": [
{
"matchLabels": {
"any:scif.cz/node": "worker"
},
"matchExpressions": [
{
"key": "reserved:remote-node",
"operator": "Exists"
}
]
}
]
identity: fix bug where fromNodes/toNodes could be used to allow custom endpoint
/test
| gharchive/pull-request | 2024-12-17T10:01:06 | 2025-04-01T04:33:47.679837 | {
"authors": [
"oblazek"
],
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/pull/36657",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
325649443 | test: download exact k8s version of k8s upstream e2e
Signed-off-by: André Martins andre@cilium.io
test-upstream-k8s
test-me-please
test-missed-k8s
test-me-please
test-upstream-k8s
test-me-please
test-missed-k8s
| gharchive/pull-request | 2018-05-23T10:43:50 | 2025-04-01T04:33:47.682496 | {
"authors": [
"aanm"
],
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/pull/4218",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
217008936 | README: add Read the Docs link
RtD let's you read the docs chapter by chapters and having a badge
makes it more visible.
This pull request is now managed with CodeReviewHub. Refresh.
| gharchive/pull-request | 2017-03-25T22:15:19 | 2025-04-01T04:33:47.684977 | {
"authors": [
"aanm",
"scanf"
],
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/pull/459",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
529382707 | eni: Check instance existence before resolving deficit
This patch adds a check for instance existence before resolving
IP deficit, this prevents operator from creating and attaching
ENIs to non-existing instances cyclically.
Also adds eth0 to instances in unit tests if necessary.
Fixes: #9533
Signed-off-by: Jaff Cheng jaff.cheng.sh@gmail.com
This change is
@ungureanuvladvictor @aanm @tgraf please take a look
Coverage decreased (-0.01%) to 46.052% when pulling e9d16f2f96c6b5b5fdbc3135b217da659220e47b on jaffcheng:add-check-for-instance-existence into f694b3ba28ac2936c413fddec7771ef14a2c8808 on cilium:master.
test-me-please
| gharchive/pull-request | 2019-11-27T14:30:56 | 2025-04-01T04:33:47.688751 | {
"authors": [
"coveralls",
"jaffcheng",
"jrajahalme"
],
"repo": "cilium/cilium",
"url": "https://github.com/cilium/cilium/pull/9676",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2080183958 | 🛑 Cioos National CKAN Sitemap is down
In 3c86669, Cioos National CKAN Sitemap (https://catalogue.cioos.ca/sitemap/sitemap-1.xml) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Cioos National CKAN Sitemap is back up in 08b463e after 2 minutes.
| gharchive/issue | 2024-01-13T08:33:26 | 2025-04-01T04:33:47.709239 | {
"authors": [
"fostermh"
],
"repo": "cioos-siooc/cwatch-upptime",
"url": "https://github.com/cioos-siooc/cwatch-upptime/issues/1035",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
892519615 | Compile-time error when using Shapeless > 2.3.4
[error] Symbol 'type shapeless.ops.FnFromProductInstances.Aux' is missing from the classpath.
[error] This symbol is required by 'value io.circe.generic.decoding.IncompleteDerivedDecoders.ffp'.
[error] Make sure that type Aux is in your classpath and check for conflicting dependencies with `-Ylog-classpath`.
[error] A full rebuild may help if 'IncompleteDerivedDecoders.class' was compiled against an incompatible version of shapeless.ops.FnFromProductInstances.
See also #1743. Is there a reason that circe depends on an shapeless 2.3.3? 2.3.6 is binary compatible.
Fixed by shapeless 2.3.7 in https://github.com/milessabin/shapeless/pull/1162.
| gharchive/issue | 2021-05-15T19:21:24 | 2025-04-01T04:33:47.710998 | {
"authors": [
"armanbilge"
],
"repo": "circe/circe",
"url": "https://github.com/circe/circe/issues/1750",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
785396394 | golang:1.16beta1 has not propagated to circleci images
Hi, The Go 1.16 Beta 1 released was made on December 27th 2020 and the official Golang docker images received that release under the 1.16beta1 tag but that tag has not become available in the CircleCI images.
I expected to see an image circleci/golang:1.16beta1 exist within 24 hours after the upstream golang:1.16beta1 posted, but it has not appeared.
https://hub.docker.com/_/golang?tab=tags&page=1&ordering=last_updated&name=1.16beta
I am hoping to use Go 1.16 beta1 CircleCI image here in https://github.com/stellar/go/pull/3288.
Hi. CircleCI doesn't publish non-stable releases by default for Convenience Images. The next versions to publish for Go will be patch releases for v1.14.x or v1.15.x or the official v1.16 release when it comes out next month.
Hi. CircleCI doesn't publish non-stable releases by default for Convenience Images. The next versions to publish for Go will be patch releases for v1.14.x or v1.15.x or the official v1.16 release when it comes out next month.
CircleCI doesn't publish non-stable releases
CircleCI publishes Go release-candidates, which are non-stable releases of Go. For example: 1.15rc2.
Supporting beta and release candidate builds are critical for testing Go applications. Some folks in the Go community participate in testing beta and release candidates.
Would it be possible to at least continue to publish release-candidate builds, and would it be possible to consider also publishing beta builds?
CircleCI doesn't publish non-stable releases
CircleCI publishes Go release-candidates, which are non-stable releases of Go. For example: 1.15rc2.
Supporting beta and release candidate builds are critical for testing Go applications. Some folks in the Go community participate in testing beta and release candidates.
Would it be possible to at least continue to publish release-candidate builds, and would it be possible to consider also publishing beta builds?
Good point. Those RC images happen to have gotten through, they're not intended nor supported.
As for this image, circleci/go, if RCs are already getting through, that would continue to work. I would like to point out though that this image is replaced by the next-gen Go image, cimg/go. You can learn more here.
Next-gen images doesn't support RC and beta images yet but it is something I'm looking to try and support officially for 2021. You can stay informed on Go images changes here and here.
Lastly, I created a CircleCI Idea that you can upvote and track this feature request. You can find that here.
Good point. Those RC images happen to have gotten through, they're not intended nor supported.
As for this image, circleci/go, if RCs are already getting through, that would continue to work. I would like to point out though that this image is replaced by the next-gen Go image, cimg/go. You can learn more here.
Next-gen images doesn't support RC and beta images yet but it is something I'm looking to try and support officially for 2021. You can stay informed on Go images changes here and here.
Lastly, I created a CircleCI Idea that you can upvote and track this feature request. You can find that here.
| gharchive/issue | 2021-01-13T19:51:04 | 2025-04-01T04:33:47.749192 | {
"authors": [
"felicianotech",
"leighmcculloch"
],
"repo": "circleci/circleci-images",
"url": "https://github.com/circleci/circleci-images/issues/531",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2078839183 | replace statedb state-change calls with Workflow Event state-change calls
update all statedb state change calls to use new class (process lambda and
update-state lambda)
Closed by #263
| gharchive/issue | 2024-01-12T13:18:24 | 2025-04-01T04:33:47.762738 | {
"authors": [
"ircwaves"
],
"repo": "cirrus-geo/cirrus-geo",
"url": "https://github.com/cirrus-geo/cirrus-geo/issues/236",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1683795334 | Add "Not enough permissions" alert to frontend when trying to import .csv files for "Global View" users
Option to import/upload .csv files under "Manage Organizations" and "Manage Users" is not available for users with "Global View" permissions. However, there is notification system in place to let users know they are unable to import files. Recommend adding an alert to frontend, otherwise they will receive a "403" error with no way to know other than checking browser console.
I think for "Global View" users, we should just hide the "import" section entirely instead.
I agree, will update ticket to reflect.
Testing with "standard" user account.
Unable to test locally as "globalView" account needed.
Linked PR #1912 which closes this issue. Hid "import/export" functionality for "global view" users.
| gharchive/issue | 2023-04-25T20:11:16 | 2025-04-01T04:33:47.764907 | {
"authors": [
"ValeriaReveles",
"epicfaace"
],
"repo": "cisagov/crossfeed",
"url": "https://github.com/cisagov/crossfeed/issues/1848",
"license": "CC0-1.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2264567651 | Add a CONTRIBUTING file
https://docs.github.com/en/communities/setting-up-your-project-for-healthy-contributions/setting-guidelines-for-repository-contributors
This issue was migrated from an old repository: https://github.com/CiscoDevNet/catalystwan/issues/80
| gharchive/issue | 2024-04-25T22:00:56 | 2025-04-01T04:33:47.802370 | {
"authors": [
"JimOverholt"
],
"repo": "cisco-open/cisco-catalyst-wan-sdk",
"url": "https://github.com/cisco-open/cisco-catalyst-wan-sdk/issues/531",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1615872535 | Boolean in zod's schema doesn't persist in form after action result?
When I enter a string value into a text field (e.g. email, password) and action is submitted, the page refreshes and the values stay in the form (bind:value works fine)
However, the boolean field always refreshes to the default state (false).
I've checked with SuperDebug to confirm the checked state is correctly passed into the form variable (it updates when I check/uncheck the checkbox). But after an action event (form submission), the boolean value becomes false, while all other fields remain as they were before the submission.
+page.server.ts:
...
const formSchema = z.object({
firstName: z
.string({ required_error: "First name is required" })
.min(1, { message: "First name must be at least 1 characters long" }),
lastName: z
.string({ required_error: "Last name is required" })
.min(1, { message: "Last name must be at least 1 characters long" }),
email: z
.string({ required_error: "Email is required" })
.email({ message: "Enter a valid email address" }),
password: z
.string({ required_error: "Password is required" })
.min(6, { message: "Password must be at least 6 characters long" }),
acceptedTos:
z.boolean({ required_error: "You must accept the Terms of Service" })
});
...
+page.svelte
...
export let data: PageData;
let success: boolean = false;
const { form, errors, message, submitting, constraints, enhance } = superForm(data.form,
{
autoFocusOnError: 'detect',
clearOnSubmit: 'none',
multipleSubmits: 'prevent',
onResult(event) {
if(event.result.type == "success")
{
success = true;
setTimeout(() => {
window.location.href = "/";
}, 500);
}
},
}
);
...
<input id="terms" type="checkbox" class="w-4 h-4 border border-gray-300 rounded bg-gray-50 focus:ring-3 focus:ring-blue-300 dark:bg-gray-700 dark:border-gray-600 dark:focus:ring-blue-600 dark:ring-offset-gray-800 dark:focus:ring-offset-gray-800" bind:checked={$form.acceptedTos}>
Hello, nice coincidence! I noticed the same thing yesterday but haven't had time to look at it yet. I'll take a look tomorrow. I think it has to do something with that a checkbox field is either undefined or set to its specified value.
Btw, there is a way of making the "accept terms" work without testing the boolean value, see this FAQ entry.
Right, adding name="acceptedTos" makes the value persist on form submission. Also, the name has to match with zod's name, else it seems to not persist.
Yes, that's the required mapping if you're using the default dataType = 'form'. When set to json it serializes and posts the data structure itself, making the form fields mere placeholders for changing the value of the $form store.
| gharchive/issue | 2023-03-08T20:20:09 | 2025-04-01T04:33:47.812955 | {
"authors": [
"Gintasz",
"ciscoheat"
],
"repo": "ciscoheat/sveltekit-superforms",
"url": "https://github.com/ciscoheat/sveltekit-superforms/issues/19",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1292750389 | grpc增加keepalive设置
客户端连接RPC接口时偶尔报 UNAVAILABLE: io exception
参考 Java netty/okhttp gRPC client throwing intermittent UNAVAILABLE: io exception while trying to connect - Stack Overflow
需要服务端和客户端都设置keep alive
只修改controller,内部微服务之间不增加。
tcp keep alive的目的是让客户端意外中断的网络链接尽快释放相关资源。操作系统默认的释放时间是2个小时,这个太长了。
http2 keep alive的目的是让服务端在客户端调用结束之后再保持一段时间,期间客户端再次发起调用就不需要重新建立链接,达到链接复用的目的。
| gharchive/issue | 2022-07-04T07:31:25 | 2025-04-01T04:33:47.814958 | {
"authors": [
"rink1969"
],
"repo": "cita-cloud/controller",
"url": "https://github.com/cita-cloud/controller/issues/157",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2347990058 | Add Prometheus Eval Client
Adds the Prometheus Eval Client
It will not work without using prompts followed Prometheus templates.
similarity_scorer function is not implemented.
Inference under torch.bfloat16 requires cuda environment with 24GB or more VRAM.
@liwii A small question, I found that "" and '' are mixed up in the code, is there any policies of it?
For example, '' is used in https://github.com/citadel-ai/langcheck/blob/c55681b08c33d6e6779a3b07f47ce86a0cc549cb/src/langcheck/metrics/eval_clients/_anthropic.py#L150, but "" is used in https://github.com/citadel-ai/langcheck/blob/c55681b08c33d6e6779a3b07f47ce86a0cc549cb/src/langcheck/metrics/eval_clients/_anthropic.py#L110.
Ah yeah it is just inconsistent haha
Our linter & formatter do not handle them properly now, but we'll fix that altogether in #123. You don't need to care too much about that in this PR!
@liwii Hi, I merged https://github.com/citadel-ai/langcheck/pull/126 to this one and
fixed the prompts' folder structure, to /metrics/prometheus
created a new function load_prompt_template to handle different prompts
Please take a review!
| gharchive/pull-request | 2024-06-12T07:07:52 | 2025-04-01T04:33:47.820397 | {
"authors": [
"conan1024hao",
"liwii"
],
"repo": "citadel-ai/langcheck",
"url": "https://github.com/citadel-ai/langcheck/pull/122",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
908437876 | Spike: run selenium tests against demo app
👻 Just trying something
cool, this worked
| gharchive/pull-request | 2021-06-01T15:47:27 | 2025-04-01T04:33:47.839049 | {
"authors": [
"davidrapson"
],
"repo": "citizensadvice/design-system",
"url": "https://github.com/citizensadvice/design-system/pull/1188",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1048159592 | Mobile UI Prototype to validate behavior
Overview
Create prototypes for v3 and v4 progress indicators for mobile.
See Issue #887 for reference.
Check out these resources for MUI stepper process, especially for how we would address the experience for mobile:
https://material.io/archive/guidelines/components/steppers.html
https://mui.com/components/steppers/
https://material.angular.io/components/stepper/overview
| gharchive/issue | 2021-11-09T03:54:55 | 2025-04-01T04:33:47.979938 | {
"authors": [
"chalimar",
"mariastudnicka"
],
"repo": "civictechindex/CTI-website-frontend",
"url": "https://github.com/civictechindex/CTI-website-frontend/issues/1035",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2174006871 | Default PROGRAM_CARD_IMAGES flag to true
Description
Please explain the changes you made here.
Release notes
The title of the pull request will be used as the default release notes description. If more detail is needed to communicate to partners the scope of the PR's changes, use this release notes section.
Checklist
General
Read the full guidelines for PRs here
[ ] Added the correct label: < feature | enhancement | bug | dependencies | infrastructure | ignore-for-release | database >
[ ] Assigned to a specific person, civiform/developers, or a more specific round-robin list.
[ ] Created unit and/or browser tests which fail without the change (if possible)
[ ] Performed manual testing (Chrome and Firefox if it includes front-end changes)
[ ] Extended the README / documentation, if necessary
Database evolutions
Read the guidelines here
[ ] Assigned two reviewers
[ ] Guarded against already existing resources using IF NOT EXISTS and IF EXISTS
[ ] Downs created to undo changes in Ups
User visible changes
[ ] Followed steps to internationalize new strings
[ ] Added context strings to new messages
[ ] Didn't use a message in applicant facing code that isn't translated yet (unless behind a flag)
[ ] Wrote browser tests using the validateAccessibility method
[ ] Tested on mobile view. See mobile device mode
[ ] Manually tested at 200% size
[ ] Manually evaluated tab order
New Features
[ ] Add new FeatureFlag env vars to server/conf/helper/feature-flags.conf
[ ] Conditioned new functionality on a FeatureFlag
[ ] Wrote browser tests with the feature flag off and on, etc.
Instructions for manual testing
If instructions are needed for manual testing by reviewers, include them here.
Issue(s) this completes
Fixes #<issue_number>; Fixes #<issue_number>...
#6836 👈
main
This stack of pull requests is managed by Graphite. Learn more about stacking.
Join @caitlinshk and the rest of your teammates on Graphite
| gharchive/pull-request | 2024-03-07T14:33:26 | 2025-04-01T04:33:47.990470 | {
"authors": [
"caitlinshk"
],
"repo": "civiform/civiform",
"url": "https://github.com/civiform/civiform/pull/6836",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1403808844 | RegionFlag: parse user_data as String
Hey, apologies but I am not sure the right way to do this. My first commit is just to naively read the whole user_data as a String even though I know it ends with a 0x95 at the end - which is not a valid ascii character.
My second commit is doing this differently - omitting that character and then appending it when serializing. Let me know your thoughts.
Cool, yep I think the second commit makes sense. That's basically how things are being handled in StringHash too. The null gets stripped and then later added back on during serialization. As far as I know they're all null terminated.
| gharchive/pull-request | 2022-10-10T23:24:22 | 2025-04-01T04:33:48.000348 | {
"authors": [
"bryab",
"cjab"
],
"repo": "cjab/libeq",
"url": "https://github.com/cjab/libeq/pull/27",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1294324621 | mail format
Do you know what format use Microsoft to store its mails in OST/PST files? HTML/EML/MESSAGE/etc.
PST files have a proprietary format specification described here. AFAIK, the message body is stored either in plain text or in html.
The java-libpst library, which pstconv uses to parse PST files, actually allows to get the message body in both formats (text and html) under certain conditions.
For more information, please check the code from PSTMessage class, specifically the methods getBody() and getBodyHtml() .
| gharchive/issue | 2022-07-05T13:44:44 | 2025-04-01T04:33:48.031253 | {
"authors": [
"bphd",
"cjmach"
],
"repo": "cjmach/pstconv",
"url": "https://github.com/cjmach/pstconv/issues/11",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1510244391 | 机器人验证卡死
输入token后弹出的浏览器窗口会要求机器人验证
每次验证到一半就刷新了,仿佛陷入了轮回绝境。
用的是chrome,默认浏览器换成edge也是一样。
不是很懂,请问这个是什么bug或者我自己的问题么?
这种验证问题的bug大概率是vpn的锅因为访问太平凡了,或者ip地址可能被拉黑了
没办法换梯子
| gharchive/issue | 2022-12-25T04:17:25 | 2025-04-01T04:33:48.032871 | {
"authors": [
"MuBai-He",
"xBLACKICEx",
"yuzuchann"
],
"repo": "cjyaddone/ChatWaifuL2D",
"url": "https://github.com/cjyaddone/ChatWaifuL2D/issues/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
19403996 | Client does not detect AMD APU GPUs as OpenCL devices
The "client" test/profile application does not detect the GPU component of AMD APUs as OpenCL devices. For example, when run on Windows 7 on an A10-6800K, only the CPU is detected as an OCL device when using the "-i" command to show CL platform information.
Feature now available #72
| gharchive/issue | 2013-09-12T18:07:10 | 2025-04-01T04:33:48.126919 | {
"authors": [
"anshuarya",
"bragadeesh"
],
"repo": "clMathLibraries/clFFT",
"url": "https://github.com/clMathLibraries/clFFT/issues/20",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
688454411 | rename project - misleading?
This is likely more useful for other platforms now as its more portable and no longer locked to the pine Rock64 SBC.
Renamed, from:
https://github.com/clach04/rock64_vcgencmd
To:
https://github.com/clach04/fake_vcgencmd
Will make an update to https://github.com/eidottermihi/rpicheck/issues/161
| gharchive/issue | 2020-08-29T04:12:52 | 2025-04-01T04:33:48.129175 | {
"authors": [
"clach04"
],
"repo": "clach04/fake_vcgencmd",
"url": "https://github.com/clach04/fake_vcgencmd/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
174839607 | Integer Hyperparameters
Love optunity. I can't seem to figure out how to optimize an integer hyperparameter. Here is my function:
@optunity.cross_validated(x=df_x.as_matrix(), y=df_y.as_matrix(), num_folds=5)
def xgb_model(x_train, y_train, x_test, y_test, max_depth=9, min_child_weight=1, subsample=0.8, colsample_bytree=0.8):
learning_rate = .1
n_estimators = 400
clf = XGBClassifier( learning_rate=learning_rate,
n_estimators=n_estimators,
max_depth=max_depth,
min_child_weight=min_child_weight,
gamma=0,
subsample=subsample,
colsample_bytree=colsample_bytree,
objective= 'binary:logistic',
max_delta_step = 1, # For imbalanced data set
scale_pos_weight=scale_pos_weight,
nthread=8,
seed=27)
msg("INFO", "Training model...")
clf.fit(x_train, y_train, eval_metric='auc')
y_prob = clf.predict_proba(x)[:,1]
auc = optunity.metrics.roc_auc(y_test, y_prob)
return auc
But when I run it, optunity is passing a float as max_depth, which isn't allowed:
...
File "/usr/local/lib/python2.7/site-packages/optunity/constraints.py", line 266, in func
return f(*args, **kwargs)
File "/usr/local/lib/python2.7/site-packages/optunity/cross_validation.py", line 403, in call
scores.append(self.f(**kwargs))
File "./my-model.py", line 274, in xgb_model
clf.fit(x_train, y_train, eval_metric='auc')
...
File "/usr/local/lib/python2.7/site-packages/xgboost/core.py", line 806, in update
_check_call(_LIB.XGBoosterUpdateOneIter(self.handle, iteration, dtrain.handle))
File "/usr/local/lib/python2.7/site-packages/xgboost/core.py", line 127, in _check_call
raise XGBoostError(_LIB.XGBGetLastError())
XGBoostError: Invalid Parameter format for max_depth expect int but value='10.02734375'
The variables are treated as float's in optunity, so you have to convert them back to int when calling XGBClassifier. For example:
clf = XGBClassifier( learning_rate=learning_rate,
n_estimators=n_estimators,
max_depth=int(max_depth),
...
Any variable that requires to be int has to be passed through int(x) or int(round(x)).
Yes, we can convert the values, but that doesn't constrain the search space for optimization. The whole point is to constrain the search space.
You're right, postprocessing the reals won't properly constrain the search space, but the results are nearly identical in terms of convergence speed. Note that Optunity's solvers are all continuous at this moment, so rounding is the best we can do at this point.
In the upcoming release we will have more support for integer and other value types, and will also be able to recognize duplicate candidate solutions more effectively. That said, the internal solvers still remain continuous. We have no concrete plans to add integer/mixed type solvers at the moment.
| gharchive/issue | 2016-09-02T20:15:42 | 2025-04-01T04:33:48.135609 | {
"authors": [
"claesenm",
"jaak-s",
"scottee"
],
"repo": "claesenm/optunity",
"url": "https://github.com/claesenm/optunity/issues/76",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
2080781902 | Consider supporting the build server protocol
Currently the interface between clangd and build tools is the JSON Compilation Database. However, this has a few disadvantages, particularly when using build systems like Bazel:
It has to be generated first. In the Bazel case, via a separate tool.
It must be kept up-to-date. In the case of Bazel, this means re-running this separate tool manually.
I believe these issues could be solved by supporting the Build Server Protocol. If a discovery file exists, clangd could launch the BSP and request compiler flags etc for currently-open files using the cpp extension to this protocol. This information could be requested and kept up-to-date by clangd.
Aside: The (de-facto?) Bazel BSP implementation seemingly doesn't support C++ ATM.
Maybe there's only partial support so far, but it looks like it has some C++ support: https://github.com/JetBrains/bazel-bsp/pull/277
Ah, thanks! Good to know that. (and maybe the README needs an update too.)
| gharchive/issue | 2024-01-14T15:50:14 | 2025-04-01T04:33:48.143781 | {
"authors": [
"cameron-martin",
"zyn0217"
],
"repo": "clangd/clangd",
"url": "https://github.com/clangd/clangd/issues/1903",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
813050284 | Adding an extra flag option
Hi, I have been using the semantic highlight functionality to gray out sections of code skipped with #ifdefs, and it's been working great. However, I can't seem to find a solution to deal with compiler specific logic since there is no way to pass additional flags to clang in the coc config. I have tried adding a .clangd configuration file for my project as described in the clangd documentation, but it has no effect. Would you consider adding a new coc configuration field for extra clang flags?
.clangd file should work for the case you've described. Could you give more details about what you tried, expected and really happened? preferably with clangd logs.
.clangd file should work for the case you've described. Could you give more details about what you tried, expected and really happened? preferably with clangd logs.
I placed a .clangd file at my project root, and typed the following lines in it:
CompilerFlags
Add: [-D__GNUC__, -U__clang__]
I have the following snippet of code in my project:
#if defined(__clang__)
spdlog::warn("CUDA GPU mode not supported by libraries compiled using clang, recompile use gcc");
#else
if (
#ifdef _MSC_VER
RunTimeLinkSuccess("OrtSessionOptionsAppendExecutionProvider_CUDA")
#else
OrtSessionOptionsAppendExecutionProvider_CUDA
#endif
)
And I am expecting to see the section of code wrapped in the #if defined(__clang__) macro to be grayed out, but it didn't.
I have enabled verbose logging via coc config but I don't know where clangd saves its log file.
it is CompileFlags not CompilerFlags and you should have an additional colon : after CompileFlags:. I've tested with:
.clangd:
CompileFlags:
Add: [-U__clang__]
a.cc:
#ifdef __clang__
static_assert(false);
#else
static_assert(false);
#endif
and I can verify that I am seeing the warning on the else branch. you can check coc documentation to figure out how to get to logs, but at some point it was possible to see via :CocInfo. The logs would've also contained errors about unknown keys in the config too.
it is CompileFlags not CompilerFlags and you should have an additional colon : after CompileFlags:. I've tested with:
.clangd:
CompileFlags:
Add: [-U__clang__]
a.cc:
#ifdef __clang__
static_assert(false);
#else
static_assert(false);
#endif
and I can verify that I am seeing the warning on the else branch. you can check coc documentation to figure out how to get to logs, but at some point it was possible to see via :CocInfo. The logs would've also contained errors about unknown keys in the config too.
Thanks, I got it working now.
| gharchive/issue | 2021-02-22T02:11:01 | 2025-04-01T04:33:48.151165 | {
"authors": [
"hyt589",
"kadircet"
],
"repo": "clangd/coc-clangd",
"url": "https://github.com/clangd/coc-clangd/issues/194",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
85932257 | Images flip
When using fit_to_width() images rotate anticlockwise 90 degrees. Not good.
$img = new SimpleImage(DIR_FS_IMAGES . $filename);
$img->fit_to_width(250)->save(DIR_FS_IMAGES . '250_' . $filename);
$img->fit_to_width(100)->save(DIR_FS_IMAGES . '100_' . $filename);
$img->fit_to_width(50)->save(DIR_FS_IMAGES . '50_' . $filename);
Original image is 2448px x 3264px
As I understand it, your class should use exif_read_data() to make sure the image is oriented correctly:
http://php.net/manual/en/function.exif-read-data.php#76964
I resaved the image using Irfanview and now the image is correctly oriented after resize using fit_to_width().
You need to use the auto_orient() method to fix images saved with a different orientation. It will not happen automatically.
Don't you think that the class should itself use $this->auto_orient()?
I think it should happen automatically, this is why we use classes for specific purposes.
Not really. The class isn't just for loading existing images, and it supports formats other than JPG as well. More importantly, the auto_orient() method does a rotation that may not be desirable on large images where PHP doesn't have enough memory to handle it. Best to give the developer the option.
If you prefer to use it, just chain it. A few extra characters won't bloat your code.
So I tried the following:
$img->fit_to_width(100)->auto_orient()->save(DIR_FS_IMAGES . '100_' . $filename);
and:
$img->auto_orient()->fit_to_width(100)->save(DIR_FS_IMAGES . '100_' . $filename);
Didn't work!
The auto_orient() method has been tested and works fine. Chances are, your original image doesn't have the correct EXIF data. Closing since this is an implementation issue, not an issue with the library.
| gharchive/issue | 2015-06-07T19:34:40 | 2025-04-01T04:33:48.245721 | {
"authors": [
"claviska",
"osCMattiFinn"
],
"repo": "claviska/SimpleImage",
"url": "https://github.com/claviska/SimpleImage/issues/85",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1159973346 | Cocoapods UI issues
On the main page, the cocoapods icon is not used in the components list.
On the component page, there is no icon and the version number does not have a link to the component like the other types do.
On the harvest page, the search box disappears when you select cocoapods as the type.
The fix for "the version number does not have a link to the component like the other types do" is in the service. See https://github.com/clearlydefined/service/pull/934.
| gharchive/issue | 2022-03-04T19:28:33 | 2025-04-01T04:33:48.260415 | {
"authors": [
"capfei",
"qtomlinson"
],
"repo": "clearlydefined/website",
"url": "https://github.com/clearlydefined/website/issues/963",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
296165290 | Fix ledger_processor::change_block block_put
Now it matches the other block types. I think this might have also been part of #598, as I got a different segfault after running it for a while.
My wallet has now been running for 30 minutes without problems with this newest commit. Previously, it consistently crashed within 10 seconds.
| gharchive/pull-request | 2018-02-11T06:22:22 | 2025-04-01T04:33:48.263418 | {
"authors": [
"PlasmaPower"
],
"repo": "clemahieu/raiblocks",
"url": "https://github.com/clemahieu/raiblocks/pull/604",
"license": "bsd-2-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
2534324859 | move version-specific code to CGameVersionManager
please move it to CGameVersionManager, it's better to keep all hooks in one place for future support of more versions/games.
can be done in a follow-up PR
Originally posted by @x87 in https://github.com/cleolibrary/CLEO5/pull/197#discussion_r1765149358
CGameVersionManager should be rewritten to use text identifiers instead enum. Then CLEO export can be added so other plugins can benefit it too. It would also make possible to display meaningful error messages about some function not being supported in current game version.
| gharchive/issue | 2024-09-18T17:47:17 | 2025-04-01T04:33:48.277572 | {
"authors": [
"MiranDMC",
"x87"
],
"repo": "cleolibrary/CLEO5",
"url": "https://github.com/cleolibrary/CLEO5/issues/198",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1115044672 | 🛑 Click Engage is down
In 092caa0, Click Engage (https://example.click-engage.co.uk) was down:
HTTP code: 403
Response time: 1512 ms
Resolved: Click Engage is back up in df5adc9.
| gharchive/issue | 2022-01-26T13:54:22 | 2025-04-01T04:33:48.293832 | {
"authors": [
"kieranbrown"
],
"repo": "clickdealer/upptime",
"url": "https://github.com/clickdealer/upptime/issues/162",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
168012655 | Add lein-kibit and check code
This codebase is old and wasn't originally very idiomatic. Some bits have been updated, but it would be good to check everything.
This was actually done already.
| gharchive/issue | 2016-07-28T04:41:17 | 2025-04-01T04:33:48.417598 | {
"authors": [
"oubiwann"
],
"repo": "clojusc/clj-starlanes",
"url": "https://github.com/clojusc/clj-starlanes/issues/38",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
900046428 | DC provisioning fails after installing Red Team tools
Operating System Version: Debian GNU/Linux 10 (buster)
Deploying via (VirtualBox/VMWare/AWS/Azure/ESXi): Virtualbox
Vagrant Version (if applicable): 2.2.14
Description of the issue:
While building the DC host, I'm running into the following error message that causes the build to stop:
Classic fixes like re-running the provisioning process or destroying the entire detection lab did not help.
Hi @JoSeiler - looks like this broke because the latest purplesharp release changed their binary naming scheme. This is fixed in https://github.com/clong/DetectionLab/commit/a578865e8d1b0184af7810ae286e5e9da673b3d4 !
Thanks a lot for the quick fix!
| gharchive/issue | 2021-05-24T21:55:14 | 2025-04-01T04:33:48.439791 | {
"authors": [
"JoSeiler",
"clong"
],
"repo": "clong/DetectionLab",
"url": "https://github.com/clong/DetectionLab/issues/657",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
446959048 | Suggestion: change the argument name '-monitoring' to '-monitor'
Suggestion: change the argument name '-monitoring' to '-monitor'
@seokho-son How about your opinion?
@powerkimhub why not? it seems a nice suggestion. Let me handle this :)
| gharchive/issue | 2019-05-22T06:43:11 | 2025-04-01T04:33:48.445005 | {
"authors": [
"jihoon-seo",
"powerkimhub",
"seokho-son"
],
"repo": "cloud-barista/poc-farmoni",
"url": "https://github.com/cloud-barista/poc-farmoni/issues/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1049955886 | Replace uuid variable that was empty with one with the actual UUID
Description
The benchmark_uuid was not in use anywhere and never set by the test so when it would go to compare the uuid of the running test would be empty. Replaced this with the UUID set for the run.
Fixes
closing since this fix is outdated , we no longer use the compare_uperf_uuid var or the assign uuid function.
| gharchive/pull-request | 2021-11-10T15:19:07 | 2025-04-01T04:33:48.446267 | {
"authors": [
"Harshith-umesh",
"dry923"
],
"repo": "cloud-bulldozer/e2e-benchmarking",
"url": "https://github.com/cloud-bulldozer/e2e-benchmarking/pull/281",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
482022783 | Add new auth configuration options.
PR to solve #165
Thank you!
| gharchive/pull-request | 2019-08-18T18:25:32 | 2025-04-01T04:33:48.478136 | {
"authors": [
"leitgab",
"paulfantom"
],
"repo": "cloudalchemy/ansible-grafana",
"url": "https://github.com/cloudalchemy/ansible-grafana/pull/166",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2669015678 | [WIP] Add extra-specs option to add SecurityGroups to the runner
adds extra-specs option to add Security Groups to the instances (until now it was used the default security group)
Whoops. It was a WIP, but it looked good :). Add another PR if there's anything else needed
| gharchive/pull-request | 2024-11-18T15:45:10 | 2025-04-01T04:33:48.490033 | {
"authors": [
"fabi200123",
"gabriel-samfira"
],
"repo": "cloudbase/garm-provider-aws",
"url": "https://github.com/cloudbase/garm-provider-aws/pull/19",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1599756420 | 🛑 sitemaya.com is down
In d024d04, sitemaya.com (https://sitemaya.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: sitemaya.com is back up in 00c5f14.
| gharchive/issue | 2023-02-25T16:20:44 | 2025-04-01T04:33:48.492956 | {
"authors": [
"cloudbip"
],
"repo": "cloudbip/upptime",
"url": "https://github.com/cloudbip/upptime/issues/10422",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1627881472 | 🛑 cbn.id is down
In 3fd4ef4, cbn.id (https://cbn.id) was down:
HTTP code: 0
Response time: 0 ms
Resolved: cbn.id is back up in 417f851.
| gharchive/issue | 2023-03-16T16:40:54 | 2025-04-01T04:33:48.495774 | {
"authors": [
"cloudbip"
],
"repo": "cloudbip/upptime",
"url": "https://github.com/cloudbip/upptime/issues/10864",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1695286639 | 🛑 hostingan.id is down
In 5b0790b, hostingan.id (https://hostingan.id) was down:
HTTP code: 0
Response time: 0 ms
Resolved: hostingan.id is back up in af22dae.
| gharchive/issue | 2023-05-04T05:29:00 | 2025-04-01T04:33:48.498634 | {
"authors": [
"cloudbip"
],
"repo": "cloudbip/upptime",
"url": "https://github.com/cloudbip/upptime/issues/12106",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2039642158 | 🛑 telkomsigma.co.id is down
In 1f71f2f, telkomsigma.co.id (https://www.telkomsigma.co.id) was down:
HTTP code: 0
Response time: 0 ms
Resolved: telkomsigma.co.id is back up in 60b475e after 15 minutes.
| gharchive/issue | 2023-12-13T12:47:03 | 2025-04-01T04:33:48.501819 | {
"authors": [
"cloudbip"
],
"repo": "cloudbip/upptime",
"url": "https://github.com/cloudbip/upptime/issues/14651",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1115871175 | 🛑 hostinger.co.id is down
In a8dd8da, hostinger.co.id (https://hostinger.co.id) was down:
HTTP code: 403
Response time: 213 ms
Resolved: hostinger.co.id is back up in 14e236e.
| gharchive/issue | 2022-01-27T07:35:05 | 2025-04-01T04:33:48.504751 | {
"authors": [
"cloudbip"
],
"repo": "cloudbip/upptime",
"url": "https://github.com/cloudbip/upptime/issues/2441",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1118390986 | 🛑 puskomedia.id is down
In c5ea6cd, puskomedia.id (https://puskomedia.id) was down:
HTTP code: 0
Response time: 0 ms
Resolved: puskomedia.id is back up in 8ee0883.
| gharchive/issue | 2022-01-30T03:01:20 | 2025-04-01T04:33:48.507662 | {
"authors": [
"cloudbip"
],
"repo": "cloudbip/upptime",
"url": "https://github.com/cloudbip/upptime/issues/2491",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1043242552 | 🛑 daeliweb.com is down
In fa8c5e5, daeliweb.com (https://daeliweb.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: daeliweb.com is back up in 5ae5020.
| gharchive/issue | 2021-11-03T09:02:59 | 2025-04-01T04:33:48.510547 | {
"authors": [
"cloudbip"
],
"repo": "cloudbip/upptime",
"url": "https://github.com/cloudbip/upptime/issues/979",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
375098304 | [Feature] Finalise test helpers for 1.0
This commit finalises the test helpers ready for the 1.0 release. The new test assertions provide significantly better PHPUnit output when they fail.
Also tidied up the test response class, as the method names were starting to get quite inconsistent and unpredictable. Methods that were marked as deprecated have been removed, and those that were not marked as deprecated have now been marked as deprecated for 2.0.
Is there anything to do besides checking for the usage of the deprecated/removed methods? Because with the new version a lot of my tests fail with the following error (lines differ):
Error : Class 'CloudCreativity\JsonApi\Testing\DocumentTester' not found /home/jannis/work/backend/mod-base/vendor/cloudcreativity/laravel-json-api/src/Testing/TestResponse.php:134 /home/jannis/work/backend/mod-base/vendor/cloudcreativity/laravel-json-api/src/Testing/TestResponse.php:292 /home/jannis/work/backend/mod-base/vendor/cloudcreativity/laravel-json-api/src/Testing/TestResponse.php:307 /home/jannis/work/backend/mod-base/tests/Feature/JsonApi/Web/UsersTest.php:83
You'll need to update the dev dependency as well as that class doesn't exist any more. There's some upgrade notes here:
https://github.com/cloudcreativity/laravel-json-api/blob/develop/docs/upgrade.md
I.e. follow the upgrade notes for 1.0.0-beta.6.
Let me know if that solves it because if it doesn't I'll need to amend the upgrade notes.
@lindyhopchris I followed the upgrade guide, removed the json-api-testing package and installed it again with version ^1.0. According to composer info I have version 1.x-dev c973e12 installed but the error still occurs.
DocumentTester does not appear anywhere in the code base so this is some sort of composer issue. Out of interest, do you do a Composer install with an optimized autoloader?
Try composer dump-autoload and then re-run the tests.
I had the wrong version of the library itself installed. I was on beta5, installing dev-develop fixed the error. I thought those changes were already tagged as a new release. :)
Ah no problem! No haven't tagged them yet. There's a chance the next tag might be 1.0.0-rc1!
| gharchive/pull-request | 2018-10-29T16:16:32 | 2025-04-01T04:33:48.518043 | {
"authors": [
"jannis-a",
"lindyhopchris"
],
"repo": "cloudcreativity/laravel-json-api",
"url": "https://github.com/cloudcreativity/laravel-json-api/pull/249",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
823095895 | SQL Scratchpad example updates
All components linked together
Autocomplete of tables, udfs etc works
Statement execution with results works
Left to do:
Some styling
It doesn't remember result from multiple queries, i.e. execute two queries, place cursor in the first.
Unrelated circleci issue '.circleci/config.yml
| gharchive/pull-request | 2021-03-05T13:37:31 | 2025-04-01T04:33:48.524617 | {
"authors": [
"JohanAhlen",
"romainr"
],
"repo": "cloudera/hue",
"url": "https://github.com/cloudera/hue/pull/1856",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2357356475 | No SDK equiv for deleteZoneRulesetRule / "Delete a zone ruleset rule"
Confirmation
[X] My issue isn't already found on the issue tracker.
[X] I have replicated my issue using the latest version of the library and it is still present.
cloudflare-go version
v0.97.0
Go environment
GOARCH='amd64'
GOBIN='/Users/tlimoncelli/bin'
GOCACHE='/Users/tlimoncelli/Library/Caches/go-build'
GOENV='/Users/tlimoncelli/Library/Application Support/go/env'
GOEXE=''
GOEXPERIMENT=''
GOFLAGS=''
GOHOSTARCH='amd64'
GOHOSTOS='darwin'
GOINSECURE=''
GOMODCACHE='/Users/tlimoncelli/go/pkg/mod'
GONOPROXY=''
GONOSUMDB=''
GOOS='darwin'
GOPATH='/Users/tlimoncelli/go'
GOPRIVATE=''
GOPROXY='https://proxy.golang.org,direct'
GOROOT='/usr/local/go'
GOSUMDB='sum.golang.org'
GOTMPDIR=''
GOTOOLCHAIN='auto'
GOTOOLDIR='/usr/local/go/pkg/tool/darwin_amd64'
GOVCS=''
GOVERSION='go1.22.3'
GCCGO='gccgo'
GOAMD64='v1'
AR='ar'
CC='clang'
CXX='clang++'
CGO_ENABLED='1'
GOMOD='/Users/tlimoncelli/git/cloudflare-go/go.mod'
GOWORK=''
CGO_CFLAGS='-O2 -g'
CGO_CPPFLAGS=''
CGO_CXXFLAGS='-O2 -g'
CGO_FFLAGS='-O2 -g'
CGO_LDFLAGS='-O2 -g'
PKG_CONFIG='pkg-config'
GOGCCFLAGS='-fPIC -arch x86_64 -m64 -pthread -fno-caret-diagnostics -Qunused-arguments -fmessage-length=0 -ffile-prefix-map=/var/folders/ry/6mvxxmy52kl_cd939xsl42y00000gn/T/go-build693033671=/tmp/go-build -gno-record-gcc-switches -fno-common'
Expected output
There is no SDK equivalent for this route:
https://developers.cloudflare.com/api/operations/deleteZoneRulesetRule
There is DeleteRuleset() but that is for the ruleset, not the individual rule (DeleteRulesetRule().
There should probably be equivs for createZoneRulesetRule, deleteZoneRulesetRule and updateZoneRulesetRule
Actual output
n/a
Code demonstrating the issue
n/a
Steps to reproduce
n/a
References
https://developers.cloudflare.com/api/operations/createZoneRulesetRule
https://developers.cloudflare.com/api/operations/deleteZoneRulesetRule
https://developers.cloudflare.com/api/operations/updateZoneRulesetRule
👋 support for manipulating ruleset rules isn't really advised due to some confusing situations you can get yourself into if you're not updating the ruleset as a whole, however, if you really want to go ahead with it, v2 does have support for ruleset rules - https://github.com/cloudflare/cloudflare-go/blob/v2/api.md#rules-4
as we are focusing efforts on v2, i don't think we'll be backporting the functionality to v0.x. if this is something you need in v0, you're welcome to drop in a PR and i'll take a look.
| gharchive/issue | 2024-06-17T13:28:27 | 2025-04-01T04:33:48.544581 | {
"authors": [
"jacobbednarz",
"tlimoncelli"
],
"repo": "cloudflare/cloudflare-go",
"url": "https://github.com/cloudflare/cloudflare-go/issues/2557",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
1525983869 | WDAPI-1221 Fix Update Method to be PUT from PATCH
Provide a general summary of your changes in the title above. You should
remove this overview, any sections and any section descriptions you
don't need below before submitting. There isn't a strict requirement to
use this template if you can structure your description and still cover
these points.
Description
Bug fix change update method from PATCH To PUT which API expects
Has your change been tested?
Yes Locally using terraform-provider as well.
Screenshots (if appropriate):
Types of changes
What sort of change does your code introduce/modify?
[x] Bug fix (non-breaking change which fixes an issue)
[ ] New feature (non-breaking change which adds functionality)
[ ] Breaking change (fix or feature that would cause existing functionality to change)
Checklist:
[x] My code follows the code style of this project.
[ ] My change requires a change to the documentation.
[ ] I have updated the documentation accordingly.
[x] I have added tests to cover my changes.
[x] All new and existing tests passed.
[x] This change is using publicly documented (api.cloudflare.com or developers.cloudflare.com) and stable APIs.
thanks @suhrit-cf 👏
| gharchive/pull-request | 2023-01-09T17:11:02 | 2025-04-01T04:33:48.549981 | {
"authors": [
"jacobbednarz",
"suhrit-cf"
],
"repo": "cloudflare/cloudflare-go",
"url": "https://github.com/cloudflare/cloudflare-go/pull/1172",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
1453755800 | 🐛 BUG: Can't execute D1 SQL
What version of Wrangler are you using?
2.3.2
What operating system are you using?
Windows with cmd.exe
Describe the Bug
I'm following the Get Started guide, pretty much verbatim, in the docs and I'm on this step: wrangler d1 execute <DATABASE_NAME> --local --file=./schema.sql
But I'm getting this error:
Packages @databases/sql, @databases/split-sql-query not available locally. Attempting to use npx to install temporarily.
Installing... (npx --prefer-offline -y -p @databases/sql@3.2.0 -p @databases/split-sql-query@1.0.3)
X [ERROR] npx-import failed for @databases/sql,@databases/split-sql-query with message:
Failed to find temporary install directory. Looking for paths matching '\npm-cache\_npx\' in:
[lots of lines printing out my system path]
You should install @databases/sql, @databases/split-sql-query locally:
npm install --save-dev @databases/sql@3.2.0 @databases/split-sql-query@1.0.3
However, I did run that last line, npm install --save-dev @databases/sql@3.2.0 @databases/split-sql-query@1.0.3 and they are installed locally (present in both my package.json and node_modules folder). I even restarted my shell after installing them.
Node: v16.13.1
npm: v8.5.5
Adding "wrangler": "^0.0.0-7c228171" to my dependencies and running wrangler with npx fixed this issue for me.
I'm experiencing the same problem. @scratchyone's suggestion didn't work for me.
v19.1.0
npm --version
8.19.3
wrangler --version
⛅️ wrangler 2.4.4
-------------------
✘ [ERROR] npx-import failed for @databases/sql,@databases/split-sql-query with message:
Failed to find temporary install directory. Looking for paths matching '/.npm/_npx/' in: [snip]
You should install @databases/sql, @databases/split-sql-query locally:
npm install --save-dev @databases/sql@3.2.0 @databases/split-sql-query@1.0.3```
I already tried that last command and I can see in the current directory that `node_modules` has modules for both of these packages. I only know enough about npm/node to be dangerous so I may be overlooking something dumb. :)
I've just faced this issue today. The same command was working yesterday. Not sure what happened. 🤔. I'm using the latest version of wrangler, 2.6.2
I worked around the problem by uninstalling nvm and Node 18.12.1 LTS and used the node installer from https://nodejs.org/en/ for the same version, 18.12.1 LTS.
This fixed it for me 🤷🏻🤷🏻
meet similar issue too
using Node v18.15.0, Windows, installed via Scoop
🌀 Mapping SQL input into an array of statements
🌀 Loading DB at .wrangler\state\d1\DB.sqlite3
[NPXI] better-sqlite3 not available locally. Attempting to use npx to install temporarily.
[NPXI] Installing... (npx --prefer-offline -y -p better-sqlite3@8.0.1)
X [ERROR] npx-import failed for better-sqlite3 with message:
Failed to find temporary install directory. Looking for paths matching '\npm-cache\_npx\'
in:
| gharchive/issue | 2022-11-17T17:59:40 | 2025-04-01T04:33:48.557406 | {
"authors": [
"Keavon",
"StoneMoe",
"ben-xD",
"icebourg",
"scratchyone"
],
"repo": "cloudflare/workers-sdk",
"url": "https://github.com/cloudflare/workers-sdk/issues/2220",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1717425280 | [C3] Offer git initialisation and make initial commit
Fixes #3275.
What this PR solves / how to test:
Offer git initialisation with a prompt
Can also specify via --git or --no-git args
Will not offer if already in a git repo
Offers after the framework tool has run, in case it wants to init git itself and doesn't check if already init'd
Commits after npm/yarn/pnpm install step so lock files (specific to package manager) are committed too
Uses wrangler initialiseGit function, which tries to set the initial branch name to 'main'
Somehow the spinner usage has become a little buggy, so still need to sort that...
Associated docs issue(s)/PR(s):
[insert associated docs issue(s)/PR(s)]
Author has included the following, where applicable:
[ ] Tests
[ ] Changeset (Changeset guidelines)
Reviewer is to perform the following, as applicable:
Checked for inclusion of relevant tests
Checked for inclusion of a relevant changeset
Checked for creation of associated docs updates
Manually pulled down the changes and spot-tested
One major problem with this right now is that using a spinner on a command that is interactive is messed up, since the spinner update overwrites any logs coming from the command. I think the spinner should only be shown if silent is true.
It does seem like that. But I believe that was the case before my changes too – the command pipe'd to stdout and the spinner was running at the same time
Not for the point where the processing gets delegated to the framework. There was no spinner in that case.
Not for the point where the processing gets delegated to the framework. There was no spinner in that case.
You're right. I missed that if startText is not provided, spinner.start() was not called.
I've added a useSpinner option to make this explicit. runCommand() now calls printAsyncStatus with useSpinner: opts?.silent
e2e tests are failing at the moment because they get stuck on the git question. You should add the --no-git option here.
| gharchive/pull-request | 2023-05-19T15:20:26 | 2025-04-01T04:33:48.565826 | {
"authors": [
"RamIdeas",
"jculvey",
"petebacondarwin"
],
"repo": "cloudflare/workers-sdk",
"url": "https://github.com/cloudflare/workers-sdk/pull/3296",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1327740965 | 🚀 Feature Request: Mass Secrets Management
This pairs with https://github.com/cloudflare/wrangler-action/issues/79.
It would be good to have a way to set mass secrets within wrangler, so within wrangler-action, you aren't meant with repetitive output, which can look very much like an error to an unsuspecting user.
Given the TTY nature of wrangler secret put I'm not entirely sure how this could work, other than maybe some kind of wrangler secret mass-put 10 which prompts for key, then value, 10 times? Hopefully y'all have a better idea. 😅
Thanks for the suggestion! I will put it up for discussion with the team 😄
Considering an approach involving piping in a file.
| gharchive/issue | 2022-08-03T20:19:59 | 2025-04-01T04:33:48.568496 | {
"authors": [
"Cherry",
"JacobMGEvans"
],
"repo": "cloudflare/wrangler2",
"url": "https://github.com/cloudflare/wrangler2/issues/1610",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1107734077 | refactor: tidy up the typings of the build result in dev
In #262 some of the strict null fixes were removed to resolve a regression.
This refactor re-applies these fixes in a way that avoids that problem.
I ran wrangler dev on a local project after this change and it appears to be OK. Not sure what the specific failing use case was that #262 was resolving?
That test should be sufficient for this. Previously, when it started, the effect would run twice, and it would try to call resolve result.stop and throw because result was undefined
| gharchive/pull-request | 2022-01-19T07:04:01 | 2025-04-01T04:33:48.570203 | {
"authors": [
"petebacondarwin",
"threepointone"
],
"repo": "cloudflare/wrangler2",
"url": "https://github.com/cloudflare/wrangler2/pull/267",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1117464266 | Enable GPU acceleration in containers
It is possible to have GPU acceleration in docker containers in WSL as explained in this guide.
However, Installing the Nvidia Docker Toolkit is not so simple on dockerinwsl, since alpine has no pre-built package provided by Nvidia (list of supported distros).
This could be solved by either:
switching to an Ubuntu-based distro for dockerinwsl (or providing another flavor that is based on Ubuntu and has GPU acceleration)
building the Nvidia toolkit for apline and bundling it along with the base image.
FYI: It might be needed to switch to ubuntu anyways (see #13). Maybe this gets easier then.
DockerInWSL 1.2.x is now based on Ubuntu. Feel free to experiment with. You can enter the distro by executing wsl -d clf_dockerinwsl in a normal shell. After that you should be able to use it like any other wsl2 ubuntu installation (besides some networking weirdness because of docker and vpnkit).
Would be great if you could add your experiments to our wiki (https://github.com/cloudflightio/dockerinwsl/wiki) :)
I just tried to enable GPU acceleration.
These steps worked for me: https://docs.nvidia.com/cuda/wsl-user-guide/index.html#ch04-sub02-install-nvidia-docker
This project has been discontinued.
| gharchive/issue | 2022-01-28T14:23:38 | 2025-04-01T04:33:48.575271 | {
"authors": [
"a-to-the-5",
"alexanderfreller",
"mriedmann"
],
"repo": "cloudflightio/dockerinwsl",
"url": "https://github.com/cloudflightio/dockerinwsl/issues/16",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
303817255 | Error: Data Dog cluster monitoring for PCF 2.0 failing --'datadog-firehose-nozzle/daaaaaa-aaaaa (0)' is not running after update. Review logs for failed jobs: datadog-firehose-nozzle
Dear All,
While applying the Data Dog cluster monitoring tile, I'm encountering the below error wherein the deployment fails and the VM status shows unresponsive:
PFB error from the task -- debug
/var/vcap/packages/director/gem_home/ruby/2.4.0/gems/logging-2.2.2/lib/logging/diagnostic_context.rb:474:in block in create_with_logging_context' D, [2018-03-09T11:11:31 #27302] [task:1653602] DEBUG -- DirectorJobRunner: (0.000633s) SELECT * FROM tasks WHERE (id= 1653602) LIMIT 1 D, [2018-03-09T11:11:31 #27302] [task:1653602] DEBUG -- DirectorJobRunner: (0.000154s) BEGIN D, [2018-03-09T11:11:31 #27302] [task:1653602] DEBUG -- DirectorJobRunner: (0.000361s) UPDATEtasksSETstate= 'error',timestamp= '2018-03-09 11:11:31',description= 'create deployment',result= '\'datadog-firehose-nozzle/d816133b-d871-490a-906b-c0c58ca78a77 (0)\' is not running after update. Review logs for failed jobs:...',output= '/var/vcap/store/director/tasks/1653602',checkpoint_time= '2018-03-09 11:11:11',type= 'update_deployment',username= 'ops_manager',deployment_name= 'datadog-216ae0a504ea83285898',started_at= '2018-03-09 11:06:11',event_output= '{\"time\":1520593571,\"stage\":\"Preparing deployment\",\"tags\":[],\"total\":1,\"task\":\"Preparing deployment\",\"index\":1,\"state\":\"started\",\"progress\":0}\n{\"time\":1520593572,\"stage\":\"Preparing deployment\",\"tags\":[],\"total\":1,\"task\":\"Preparing deployment\",\"index\":1,\"state\":\"finished\",\"progress\":100}\n{\"time\":1520593573,\"stage\":\"Preparing package compilation\",\"tags\":[],\"total\":1,\"task\":\"Finding packages to compile\",\"index\":1,\"state\":\"started\",\"progress\":0}\n{\"time\":1520593573,\"stage\":\"Preparing package compilation\",\"tags\":[],\"total\":1,\"task\":\"Finding packages to compile\",\"index\":1,\"state\":\"finished\",\"progress\":100}\n{\"time\":1520593573,\"stage\":\"Updating instance\",\"tags\":[\"datadog-firehose-nozzle\"],\"total\":1,\"task\":\"datadog-firehose-nozzle/d816133b-d871-490a-906b-c0c58ca78a77 (0) (canary)\",\"index\":1,\"state\":\"started\",\"progress\":0}\n{\"time\":1520593891,\"stage\":\"Updating instance\",\"tags\":[\"datadog-firehose-nozzle\"],\"total\":1,\"task\":\"datadog-firehose-nozzle/d816133b-d871-490a-906b-c0c58ca78a77 (0) (canary)\",\"index\":1,\"state\":\"failed\",\"progress\":100,\"data\":{\"error\":\"\'datadog-firehose-nozzle/d816133b-d871-490a-906b-c0c58ca78a77 (0)\' is not running after update. Review logs for failed jobs: datadog-firehose-nozzle\"}}\n{\"time\":1520593891,\"error\":{\"code\":400007,\"message\":\"\'datadog-firehose-nozzle/d816133b-d871-490a-906b-c0c58ca78a77 (0)\' is not running after update. Review logs for failed jobs: datadog-firehose-nozzle\"}}\n',result_output= '',context_id = '' WHERE (id` = 1653602) LIMIT 1
D, [2018-03-09T11:11:31 #27302] [task:1653602] DEBUG -- DirectorJobRunner: (0.005428s) COMMIT
I, [2018-03-09T11:11:31 #27302] [] INFO -- DirectorJobRunner: Task took 5 minutes 20.84445033999998 seconds to process.
Also, output from “/var/vcap/sys/log/datadog-firehose-nozzle/ datadog-firehose-nozzle.stdout.log”
{"timestamp":1520594219.480720043,"process_id":3501,"source":"datadog-firehose-nozzle","log_level":"error","message":"Error while reading from the firehose: Error dialing trafficcontroller server: websocket: bad handshake.\nPlease ask your Cloud Foundry Operator to check the platform configuration (trafficcontroller is wss://doppler.syslab.pcflabcog.com).","data":null,"file":"/var/vcap/data/compile/datadog-firehose-nozzle/datadog-firehose-nozzle/src/github.com/DataDog/datadog-firehose-nozzle/datadogfirehosenozzle/datadog_firehose_nozzle.go","line":273,"method":"github.com/DataDog/datadog-firehose-nozzle/datadogfirehosenozzle.(*DatadogFirehoseNozzle).handleError"}
{"timestamp":1520594219.480772257,"process_id":3501,"source":"datadog-firehose-nozzle","log_level":"info","message":"Closing connection with traffic controller due to Error dialing trafficcontroller server: websocket: bad handshake.\nPlease ask your Cloud Foundry Operator to check the platform configuration (trafficcontroller is wss://doppler.syslab.pcflabcog.com).","data":null,"file":"/var/vcap/data/compile/datadog-firehose-nozzle/datadog-firehose-nozzle/src/github.com/DataDog/datadog-firehose-nozzle/datadogfirehosenozzle/datadog_firehose_nozzle.go","line":277,"method":"github.com/DataDog/datadog-firehose-nozzle/datadogfirehosenozzle.(*DatadogFirehoseNozzle).handleError"}
Thanks and Regards,
Gowri Shankar
@gowrshan this is officially maintained by datadog now. You should file this issue on their repo for the nozzle.
| gharchive/issue | 2018-03-09T11:54:23 | 2025-04-01T04:33:48.584784 | {
"authors": [
"ahevenor",
"gowrshan"
],
"repo": "cloudfoundry-attic/datadog-firehose-nozzle-release",
"url": "https://github.com/cloudfoundry-attic/datadog-firehose-nozzle-release/issues/12",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
636807159 | .env file support ?
Hi folks,
Any workaround to use .env files for this library?
Any workaround to use .env files for this library?
You'll need to provide more context. Not sure how or when you would expect a .env file to be used. For local testing?
Any workaround to use .env files for this library?
You'll need to provide more context. Not sure how or when you would expect a .env file to be used. For local testing?
Problem: I have many environments [local, stage, production]
It is very challenging to connect to these environments to debug locally, we end up hardcoding values, so thought of maintaining multiple .env files for different environments and load according to NODE_ENV.
I recently came through vcap file, it would be a great help if some snippet demonstrates how VCAP file should be used, please !
| gharchive/issue | 2020-06-11T08:01:36 | 2025-04-01T04:33:48.588520 | {
"authors": [
"DivekJohns",
"pmuellr"
],
"repo": "cloudfoundry-community/node-cfenv",
"url": "https://github.com/cloudfoundry-community/node-cfenv/issues/44",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
164962427 | Response Time Emitter for external plotting and health checks
Currently the abacus-express module has the the ability to log the response time using a morgan middleware function but does not emit/ store the incoming and outgoing requests and responses.
const beforeLogger = () => {
const morg = morgan(
':remote-addr - - :method :url HTTP/:http-version :status ' +
':res[content-length] :referrer :user-agent - :response-time ms', {
immediate: true,
stream: {
write: (msg, encoding) => {
debug('Received request %s', msg.replace(/\n/g, ''));
}
}
});
return (req, res, next) => {
return debug.enabled() ? morg(req, res, next) : next();
};
};
const afterLogger = () => {
const morg = morgan(
':remote-addr - - :method :url HTTP/:http-version :status ' +
':res[content-length] :referrer :user-agent - :response-time ms', {
stream: {
write: (msg, encoding) => {
debug('Processed request %s', msg.replace(/\n/g, ''));
}
}
});
return (req, res, next) => {
return debug.enabled() ? morg(req, res, next) : next();
};
};
I would like to have those events logged by an emitter so that I can add a module that can take that information and publish it to an external destination. This could allow for real time plotting of app response times, among other things.
Can you describe your use case in details?
The morgan middleware is just a request logger. If you would like to have
real time plotting of app response times
We have the status emitted in the /hystrix.stream and you can use hystrix dashboard to see it real time.
If you would like to store the response stats such as the response time for any other purposes, you can get it from the /hystrix.stream and do some modification on the json structure to match your external app(like grafana).
HTH
I was planning on logging the request/response and response time and send it to grafana, how can I use hystrix.stream to do that?
| gharchive/issue | 2016-07-11T23:30:26 | 2025-04-01T04:33:48.596593 | {
"authors": [
"ArchanPatel890",
"KRuelY"
],
"repo": "cloudfoundry-incubator/cf-abacus",
"url": "https://github.com/cloudfoundry-incubator/cf-abacus/issues/367",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
315024432 | Please bump go to 1.10.1
Thanks!
@sesmith177 If it's not too much trouble we'd prefer 1.10.1 explicitly. We're skipping 1.10.2 because 1.10.3 is planned for start of June and it fixes an issue with the tar/archive package that docker uses. We didn't want to go to the effort of upgrading to 1.10.2 for just a few weeks.
Alternatively, we could both wait for 1.10.3 :)
v9 has 1.10.1
| gharchive/issue | 2018-04-17T11:48:54 | 2025-04-01T04:33:48.598697 | {
"authors": [
"BooleanCat",
"sesmith177"
],
"repo": "cloudfoundry-incubator/windows-tools-release",
"url": "https://github.com/cloudfoundry-incubator/windows-tools-release/issues/8",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1389533625 | Bosh internal certs rotation
Hi ,
Our orchestrator bosh related certs and bosh-dns certs will be expiring in Jan 2023 so we are planning to have it rotated in our test environment first.
So which one should be rotated first bosh-dns or bosh?
What is the process ? Does it require downtime for the underlying deployments?
and do we have to recreate all the deployment vms in this process??
Please suggest.
Hi @Manu2516789,
please check the available documentation about credentials rotation in
https://bosh.io/docs/dns/#rotating-dns-certificates
https://bosh.io/docs/credential-rotation/
If you find something missing in the docs please open an issue for docs-bosh.
| gharchive/issue | 2022-09-28T15:10:20 | 2025-04-01T04:33:48.608324 | {
"authors": [
"Manu2516789",
"beyhan"
],
"repo": "cloudfoundry/bosh-deployment",
"url": "https://github.com/cloudfoundry/bosh-deployment/issues/440",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.