id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
1894249721
🛑 ISTAT is down In 44ad9ab, ISTAT (https://esploradati.istat.it/SDMXWS/rest/dataflow/all/all/latest) was down: HTTP code: 0 Response time: 0 ms Resolved: ISTAT is back up in e19f9c2 after 12 minutes.
gharchive/issue
2023-09-13T10:39:02
2025-04-01T04:35:10.502289
{ "authors": [ "charphi" ], "repo": "nbbrd/sdmx-upptime", "url": "https://github.com/nbbrd/sdmx-upptime/issues/15348", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1287228266
🛑 ESTAT is down In 50bb6ad, ESTAT (http://ec.europa.eu/eurostat/SDMX/diss-web/rest/dataflow/ESTAT/all/latest/) was down: HTTP code: 0 Response time: 0 ms Resolved: ESTAT is back up in 3748630.
gharchive/issue
2022-06-28T11:53:09
2025-04-01T04:35:10.504555
{ "authors": [ "charphi" ], "repo": "nbbrd/sdmx-upptime", "url": "https://github.com/nbbrd/sdmx-upptime/issues/2136", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1326701676
🛑 SPC is down In 93d01b8, SPC (https://stats-nsi-stable.pacificdata.org/rest/dataflow/all/all/latest) was down: HTTP code: 502 Response time: 1167 ms Resolved: SPC is back up in 275465a.
gharchive/issue
2022-08-03T05:33:58
2025-04-01T04:35:10.506838
{ "authors": [ "charphi" ], "repo": "nbbrd/sdmx-upptime", "url": "https://github.com/nbbrd/sdmx-upptime/issues/2589", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2246787276
🛑 ESTAT is down In 8f38e2f, ESTAT (https://ec.europa.eu/eurostat/api/dissemination/sdmx/2.1/dataflow/ESTAT/all/latest) was down: HTTP code: 0 Response time: 0 ms Resolved: ESTAT is back up in bb2b362 after 11 minutes.
gharchive/issue
2024-04-16T19:56:18
2025-04-01T04:35:10.509286
{ "authors": [ "charphi" ], "repo": "nbbrd/sdmx-upptime", "url": "https://github.com/nbbrd/sdmx-upptime/issues/28971", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2317175685
🛑 ESTAT is down In cfb1104, ESTAT (https://ec.europa.eu/eurostat/api/dissemination/sdmx/2.1/dataflow/ESTAT/all/latest) was down: HTTP code: 0 Response time: 0 ms Resolved: ESTAT is back up in c3fd620 after 5 minutes.
gharchive/issue
2024-05-25T18:52:39
2025-04-01T04:35:10.511771
{ "authors": [ "charphi" ], "repo": "nbbrd/sdmx-upptime", "url": "https://github.com/nbbrd/sdmx-upptime/issues/31556", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2324743711
🛑 IMF is down In a096c07, IMF (http://dataservices.imf.org/REST/SDMX_XML.svc/Dataflow) was down: HTTP code: 0 Response time: 0 ms Resolved: IMF is back up in bf2be2d after 7 minutes.
gharchive/issue
2024-05-30T05:34:15
2025-04-01T04:35:10.514091
{ "authors": [ "charphi" ], "repo": "nbbrd/sdmx-upptime", "url": "https://github.com/nbbrd/sdmx-upptime/issues/31825", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2368841630
🛑 ESTAT is down In 48bd474, ESTAT (https://ec.europa.eu/eurostat/api/dissemination/sdmx/2.1/dataflow/ESTAT/all/latest) was down: HTTP code: 0 Response time: 0 ms Resolved: ESTAT is back up in 4e751c8 after 6 minutes.
gharchive/issue
2024-06-23T19:29:16
2025-04-01T04:35:10.516436
{ "authors": [ "charphi" ], "repo": "nbbrd/sdmx-upptime", "url": "https://github.com/nbbrd/sdmx-upptime/issues/33401", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2414171757
🛑 NBB is down In 860f788, NBB (https://stat.nbb.be/restsdmx/sdmx.ashx/GetDataStructure/ALL) was down: HTTP code: 429 Response time: 833 ms Resolved: NBB is back up in 2bf15cc after 3 hours, 7 minutes.
gharchive/issue
2024-07-17T17:37:13
2025-04-01T04:35:10.518739
{ "authors": [ "charphi" ], "repo": "nbbrd/sdmx-upptime", "url": "https://github.com/nbbrd/sdmx-upptime/issues/34833", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2502483831
🛑 ISTAT is down In c3effdf, ISTAT (https://esploradati.istat.it/SDMXWS/rest/dataflow/all/all/latest) was down: HTTP code: 0 Response time: 0 ms Resolved: ISTAT is back up in 1f00a04 after 15 minutes.
gharchive/issue
2024-09-03T10:18:02
2025-04-01T04:35:10.521027
{ "authors": [ "charphi" ], "repo": "nbbrd/sdmx-upptime", "url": "https://github.com/nbbrd/sdmx-upptime/issues/37482", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2541126633
🛑 NBB is down In f4a455b, NBB (https://stat.nbb.be/restsdmx/sdmx.ashx/GetDataStructure/ALL) was down: HTTP code: 429 Response time: 665 ms Resolved: NBB is back up in 93c153c after 11 minutes.
gharchive/issue
2024-09-22T15:17:59
2025-04-01T04:35:10.523472
{ "authors": [ "charphi" ], "repo": "nbbrd/sdmx-upptime", "url": "https://github.com/nbbrd/sdmx-upptime/issues/38651", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2549067822
🛑 ISTAT is down In e7fee0d, ISTAT (https://esploradati.istat.it/SDMXWS/rest/dataflow/all/all/latest) was down: HTTP code: 0 Response time: 0 ms Resolved: ISTAT is back up in 52c21be after 11 minutes.
gharchive/issue
2024-09-25T21:57:55
2025-04-01T04:35:10.525731
{ "authors": [ "charphi" ], "repo": "nbbrd/sdmx-upptime", "url": "https://github.com/nbbrd/sdmx-upptime/issues/38836", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2594192734
🛑 OECD is down In c5d1bca, OECD (https://sdmx.oecd.org/public/rest/dataflow/all/all/latest) was down: HTTP code: 0 Response time: 0 ms Resolved: OECD is back up in c26885f after 7 minutes.
gharchive/issue
2024-10-17T09:43:57
2025-04-01T04:35:10.528001
{ "authors": [ "charphi" ], "repo": "nbbrd/sdmx-upptime", "url": "https://github.com/nbbrd/sdmx-upptime/issues/40208", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1532311494
🛑 INSEE is down In c13deaf, INSEE (https://bdm.insee.fr/series/sdmx/dataflow/all/all/latest/) was down: HTTP code: 0 Response time: 0 ms Resolved: INSEE is back up in 0d76356.
gharchive/issue
2023-01-13T13:31:22
2025-04-01T04:35:10.530266
{ "authors": [ "charphi" ], "repo": "nbbrd/sdmx-upptime", "url": "https://github.com/nbbrd/sdmx-upptime/issues/5215", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1592523807
🛑 ESTAT is down In 58f3353, ESTAT (https://ec.europa.eu/eurostat/api/dissemination/sdmx/2.1/dataflow/ESTAT/all/latest) was down: HTTP code: 0 Response time: 0 ms Resolved: ESTAT is back up in 70de40a.
gharchive/issue
2023-02-20T22:36:13
2025-04-01T04:35:10.532572
{ "authors": [ "charphi" ], "repo": "nbbrd/sdmx-upptime", "url": "https://github.com/nbbrd/sdmx-upptime/issues/5955", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1630559191
🛑 ESTAT is down In 749627e, ESTAT (https://ec.europa.eu/eurostat/api/dissemination/sdmx/2.1/dataflow/ESTAT/all/latest) was down: HTTP code: 0 Response time: 0 ms Resolved: ESTAT is back up in 105f25d.
gharchive/issue
2023-03-18T20:00:23
2025-04-01T04:35:10.535211
{ "authors": [ "charphi" ], "repo": "nbbrd/sdmx-upptime", "url": "https://github.com/nbbrd/sdmx-upptime/issues/6918", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1676101560
🛑 ESTAT is down In ac70859, ESTAT (https://ec.europa.eu/eurostat/api/dissemination/sdmx/2.1/dataflow/ESTAT/all/latest) was down: HTTP code: 0 Response time: 0 ms Resolved: ESTAT is back up in b31a394.
gharchive/issue
2023-04-20T06:46:27
2025-04-01T04:35:10.537512
{ "authors": [ "charphi" ], "repo": "nbbrd/sdmx-upptime", "url": "https://github.com/nbbrd/sdmx-upptime/issues/8217", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2432499365
🛑 Radarr (Kids/Nordic) is down In aaead40, Radarr (Kids/Nordic) ($SECRET_SITE019) was down: HTTP code: 0 Response time: 0 ms Resolved: Radarr (Kids/Nordic) is back up in 20ab091 after 42 minutes.
gharchive/issue
2024-07-26T15:42:50
2025-04-01T04:35:10.540653
{ "authors": [ "nboxdev" ], "repo": "nboxdev/upptime", "url": "https://github.com/nboxdev/upptime/issues/236", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2152960388
Exercise 4 final submission This is the only pull request that needs to be approved. Thanks! So I'm not totally sure how to do it on my end, but the only one that actually needs to be approved is 'New (and final) exercise 4 upload). Everything else is old. Just kidding, now a couple things need to be approved: ''New (and final) exercise 4 upload' and 'completed exercise 5'. Thanks! Looks great!
gharchive/pull-request
2024-02-25T22:01:31
2025-04-01T04:35:10.607433
{ "authors": [ "chesterharvey", "msanford7615" ], "repo": "ncsg/ursp688y_sp2024", "url": "https://github.com/ncsg/ursp688y_sp2024/pull/72", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
418782183
Valid initrd fix boot issue on SD820 This change fix a boot issue: [ 0.849690] Initramfs unpacking failed: junk in compressed archive [ 0.913369] ufs_qcom_phy_qmp_14nm 627000.phy: invalid resource [ 0.916028] qcom-pcie 600000.pcie: Failed to get supply 'vddpe-3v3': -517 [ 1.924849] qcom-pcie 608000.pcie: Phy link never came up [ 1.926456] qcom-pcie 608000.pcie: cannot initialize host [ 2.173672] ufshcd-qcom 624000.ufshc: ufshcd_print_pwr_info:[RX, TX]: gear=[1, 1], lane[1, 1], pwr[SLOWAUTO_MODE, SLOWAUTO_MODE], rate = 0 [ 2.235323] dwc3 7600000.dwc3: Failed to get clk 'ref': -2 [ 2.237718] dwc3 6a00000.dwc3: Failed to get clk 'ref': -2 [ 2.242360] i2c_qup 75b5000.i2c: [ 2.242360] tx channel not available [ 2.246507] i2c_qup 75b6000.i2c: [ 2.246507] tx channel not available [ 2.253187] i2c_qup 7577000.i2c: [ 2.253187] tx channel not available [ 2.416921] ufshcd-qcom 624000.ufshc: ufshcd_print_pwr_info:[RX, TX]: gear=[3, 3], lane[1, 1], pwr[FAST MODE, FAST MODE], rate = 2 [ 2.530560] dwc3 7600000.dwc3: Failed to get clk 'ref': -2 [ 2.538541] dwc3 6a00000.dwc3: Failed to get clk 'ref': -2 [ 2.574903] Kernel panic - not syncing: VFS: Unable to mount root fs on unknown-block(259,2) [ 2.574928] CPU: 0 PID: 1 Comm: swapper/0 Not tainted 5.0.0-rc1 #1 [ 2.582393] Hardware name: Qualcomm Technologies, Inc. DB820c (DT) [ 2.588382] Call trace: [ 2.594552] dump_backtrace+0x0/0x178 [ 2.596888] show_stack+0x14/0x20 [ 2.600710] dump_stack+0x84/0xa4 [ 2.604007] panic+0x13c/0x2ec [ 2.607306] mount_block_root+0x1a0/0x284 [ 2.610254] mount_root+0x140/0x174 [ 2.614335] prepare_namespace+0x138/0x180 [ 2.617634] kernel_init_freeable+0x220/0x240 [ 2.621804] kernel_init+0x10/0x108 [ 2.626229] ret_from_fork+0x10/0x18 [ 2.629539] SMP: stopping secondary CPUs [ 2.633465] Kernel Offset: disabled [ 2.637253] CPU features: 0x002,20802008 [ 2.640463] Memory Limit: none [ 2.644645] ---[ end Kernel panic - not syncing: VFS: Unable to mount root fs on unknown-block(259,2) ]--- This problem was introduced by this change on Linux: commit ff1522bb7d98450c72aea729f0b4147bc9986aed Author: David Engraf <david.engraf@sysgo.com> Date: Thu Jan 3 15:28:31 2019 -0800 initramfs: cleanup incomplete rootfs Unpacking an external initrd may fail e.g. not enough memory. This leads to an incomplete rootfs because some files might be extracted already. Fixed by cleaning the rootfs so the kernel is not using an incomplete rootfs. Link: http://lkml.kernel.org/r/20181030151805.5519-1-david.engraf@sysgo.com Signed-off-by: David Engraf <david.engraf@sysgo.com> Cc: Dominik Brodowski <linux@dominikbrodowski.net> Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org> Cc: Philippe Ombredanne <pombredanne@nexb.com> Cc: Arnd Bergmann <arnd@arndb.de> Cc: Luc Van Oostenryck <luc.vanoostenryck@gmail.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org> interesting. how did you reproduce that, with your own kernel? did you notice that this commit was reverted in 5.0: a841c673f135 (revert "initramfs: cleanup incomplete rootfs") Yes, I had this on my kernel. Thank you i did not notice it was reverted. Then well, it is useless then.
gharchive/pull-request
2019-03-08T13:15:23
2025-04-01T04:35:10.610546
{ "authors": [ "lugu", "ndechesne" ], "repo": "ndechesne/meta-qcom", "url": "https://github.com/ndechesne/meta-qcom/pull/96", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
728798411
Document running hlint on changed files Closes #1151 Thanks to my coworker @drostron and @ldubinets for actually coming up with these commands Thanks!
gharchive/pull-request
2020-10-24T14:16:12
2025-04-01T04:35:10.612631
{ "authors": [ "MaxGabriel", "ndmitchell" ], "repo": "ndmitchell/hlint", "url": "https://github.com/ndmitchell/hlint/pull/1154", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
297829724
Refactored device location prediction Refactoring for #35 Codecov Report Merging #38 into 0629-device-location-prediction will increase coverage by 0.65%. The diff coverage is 39.79%. @@ Coverage Diff @@ ## 0629-device-location-prediction #38 +/- ## ===================================================================== + Coverage 15.47% 16.13% +0.65% - Complexity 64 66 +2 ===================================================================== Files 35 35 Lines 982 1004 +22 Branches 80 80 ===================================================================== + Hits 152 162 +10 - Misses 822 833 +11 - Partials 8 9 +1 Impacted Files Coverage Δ Complexity Δ ...exenio/bleindoorpositioning/IndoorPositioning.java 0% <0%> (ø) 0 <0> (ø) :arrow_down: .../location/distance/LocationDistanceCalculator.java 92.85% <100%> (+22.85%) 2 <1> (-1) :arrow_down: ...eindoorpositioning/location/LocationPredictor.java 18.86% <18.86%> (ø) 3 <3> (?) ...bleindoorpositioning/location/angle/AngleUtil.java 90.9% <87.5%> (+5.19%) 4 <2> (+2) :arrow_up: ...exenio/bleindoorpositioning/location/Location.java 66.66% <90.47%> (+14.49%) 16 <5> (+3) :arrow_up: Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update b2fff69...f8daf89. Read the comment docs.
gharchive/pull-request
2018-02-16T15:50:28
2025-04-01T04:35:10.627911
{ "authors": [ "Steppschuh", "codecov-io" ], "repo": "neXenio/BLE-Indoor-Positioning", "url": "https://github.com/neXenio/BLE-Indoor-Positioning/pull/38", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
76984224
Broken URL on game move Hello! I've just encountered a strange occurence. Steps: go to this game (I wasn't logged in, if it makes any difference): https://gokibitz.com/kifu/VkXK5jd7 go to the next move (either using display arrows or keyboard arrows) around move 18 or 19 the URL instead of being ../VkXK5jd7?path=19 becomes https://gokibitz.com/?path=19 this results in a redirect to the homepage I'm not familiar with the codebase. At first I thought you were doing some kind of media discovery to go to the next move, then I realized that you only fetch comments move by move, so you simply iterate through the SGF that you load when you access the game page the first time. I guess there's some weird corner case when you build the URI for the next move? Thank you and good luck! EDIT: Ok so... diving into the code a bit. I see that $scope.player is from a transcluded directive (sgf.js) which wraps wGo. Something tells me the issue is not upstream, why would wGo change the URL?! EDIT2: I've noticed something interesting! Go to this game: https://gokibitz.com/kifu/4JHZE2om I'm redirected at the very first move, however! If I go back and remove the "?path=xx" from the URL I can go on watching the game and the "?path" part doesn't reappear! Maybe there's a problem in these lines? (controllers/kifu.js ln 116) `` // Put the move in the query string for super sweet permalinks $location.search('path', pathFilter($scope.kifu.path, 'string')); `` Hi Sphaso! We are definitely the ones changing the URL. It's to keep it up to date and make it easier for people to share specific moves. Thanks for documenting it with a reproducible test case! If you feel like fixing it up yourself you are in the correct area. I think it must be the path filter messing up. If not someone will be sure to take a look at it soon since it should be simple. Christopher Christopher I'd love to tackle it myself, thanks! I tried setting up a dev env on my Arch box but I'm having a bit of an issue. It seems that the local website is not bringing up any *.js nor *.css besides app.js. I didn't touch the gulp file. I'm a bit inexperienced with the whole nodeJS toolchain, do you have any idea on how to troubleshoot this? Great! I set up on an Ubuntu box and Nate uses OS X. Can you show some pictures or output? I am new to nodejs as well but I can hopefully figure it out as I have hit a lot of the issues. And in case you didn't notice the instructions for development set up were just changed slightly with pull request #88. It might help! Perfect reproduction steps. This is very strange! I was able to reproduce the issue and will check it out tonight. Thanks for taking the effort to report this. -- Nate Eagle On May 16, 2015, at 4:38 AM, sphaso notifications@github.com wrote: Hello! I've just encountered a strange occurence. Steps: go to this game (I wasn't logged in, if it makes any difference): https://gokibitz.com/kifu/VkXK5jd7 move to the next move (either using display arrows or keyboard arrows) around move 18 or 19 the URL instead of being ../VkXK5jd7?path=19 becomes https://gokibitz.com/?path=19 this results in a redirect to the homepage I'm not familiar with the codebase. At first I thought you were doing some kind of media discovery to go to the next move, then I realized that you only fetch comments move by move, so you simply iterate through the SGF that you load when you access the game page the first time. I guess there's some weird corner case when you build the URI for the next move? Thank you and good luck! — Reply to this email directly or view it on GitHub. @levelonedev actually, it seems only the CSS is missing. With some trouble (due to the sketchy layout) I was able to register and upload a game, so I'm guessing all the functionalities are there on both client and server side. Not sure what screenshot \ ouput might give clues... gulp watch output? I've also noticed there's no map file, zo... no live debugging? good old console.log? @neagle is it cool if I try to fix it? consider it a sort of initiation ritual :P but if you think it's urgent please don't mind me. Make sure you've run gulp default (or just gulp) if it seems like some of the files are missing. Yeah, I'd love to have you give it a shot! Thank you. :D -- Nate Eagle On May 16, 2015, at 2:11 PM, sphaso notifications@github.com wrote: @levelonedev actually, it seems only the CSS is missing. With some trouble (due to the sketchy layout) I was able to register and upload a game, so I'm guessing all the functionalities are there on both client and server side. Not sure what screenshot \ ouput might give clues... gulp watch output? I've also noticed there's no map file, zo... no live debugging? good old console.log? @neagle is it cool if I try to fix it? consider it a sort of initiation ritual :P but if you think it's urgent please don't mind me. — Reply to this email directly or view it on GitHub. News from the trenches: the problem with the CSS was that "npm install" had skipped some modules (namely "require" and "stream-browserify") giving a warning I didn't notice the first time the bug seems to be strictly related to not being logged in. I cannot reproduce when logged in further proof: when the redirect occures I get a 401 (unauthorized) on /api/markdown, which makes me think pathFilter might be a red herring my current hypothesis is that, quite simply, /api/markdown needs to be accessed anonimously I'll keep digging. Well, it seems pretty straightforward now. The "commentBox" directive at line 43 calls /api/markdown $http.post('/api/markdown/', { markdown: element.val() }, { timeout: canceler.promise }) .success(function (data) { $scope.preview = data.markup; }); However, this route is protected (server/routes/markdown line 11): router.post('/', auth.ensureAuthenticated, function (req, res) { var html = marked(parseLabels(req.body.markdown)) || ''; res.json({ markup: html }); }); Is it safe to remove the authentication on this route? Was there a reason why it was put? It seems that this api simply translates markdown into HTML. I'm not much of an api designer, but shouldn't it be a GET? Nope, I don't think authentication is needed. Thanks for digging down to find this! Remove it and test localhost just to be sure and then submit a pull request! Hm, I think we need to peek a bit deeper. The comment box shouldn't be calling the markdown API when you're not logged in: you shouldn't even see the comment box if you're not logged in. I put it behind auth because I didn't want to make it available as an anonymous markdown service, even if such abuse is pretty unlikely. I think we need to figure out the root cause, which is what's causing a call to get sent in this case where the user's not logged in. Thanks for investigating this and big thanks to @levelonedev for helping get you ramped up this afternoon. You have no idea how happy I was to see this comment thread happen. :dancers: Ah ha: so the call that's getting triggered has the text from the game comment, in this case: "This was probably too greedy. I think the honte play is simply K8, planning to use the thickness to reduce the right later. White should remember that between the lower left shimari and the top side, it's still an even game. There's no need to rush." I only recently added parsing of the text in the game comment, so it's a good candidate to have a newly introduced bug. @neagle @levelonedev Thank you guys! I had fun and hopefully I'll have more :) So I'm guessing we don't want to show the commentBox when the user is not logged in? (i.e. add a condition here) div.panel.panel-default( ng-if='sgfComment || kifu.owner.username === currentUser.username' ) I'm not exactly sure if this will be enough to stop from calling the API. Where is the 'gkCommentBox' directive injected? We already don't show the editGameComment box to non-logged-in users, since it doesn't show for anybody but the owner. That textarea uses the gkCommentBox directive, as you note, but it doesn't include a preview element, since we're not actually parsing markdown for game comments (yet?). Some condition, however, is making preview get fired even though it doesn't seem obvious than it should. As I see it, the fix is: Figuring out why it's getting called, because we at least should know, right? :smile: Amending the conditional that fires preview to make sure it only fires when appropriate, and/or fixing some more mysterious cause, depending on the results of #1. Oh ok, sorry, I forgot to grep for the snake-case of the directive :) gets me everytime! Things are much clearer now, thanks. div(ng-show='editGameComment') This div will get rendered even if you're not the owner (in fact, even if you're not logged in). This means that it's bound to the directive, just hidden via CSS. At this point, my guess is that this will trigger the api call (commentBox.js) $scope.$watch('model', function (newValue, oldValue) { // @see http://stackoverflow.com/a/18915585/399077 if (newValue !== oldValue) { preview(); } }); (where preview() actually calls the api.) This means that replacing that ng-show for an ng-if will probably do the trick, like Christmas morning! I'll confirm \ disprove my hypothesis as soon as I get home from work. It seems like you already have some Angular experience: is that true? Or are you just picking up details about how it works astonishingly quickly? Good research on this. I think you're almost certainly right that changing div(ng-show='editGameComment') to div(ng-if='editGameComment') would fix the problem, but to me that feels like ignoring the problem's root cause, which is that the directive's not being smart about when to fetch a preview. I hope you'll forgive me, I didn't do this deliberately to usurp your task, but since I needed to refactor gkCommentBox for something @levelonedev is doing anyway, I slipped in a fix for this. Basically, I'm checking whether a preview value has actually been specified before doing any of the logic for fetching a preview, which is a performance improvement, anyway. Btw, you also inspired me to get lint / code style checkers in our gulp setup so that it's easier for other devs to jump in, as well as to get us set up with Travis CI so that new builds and pull requests are automatically tested. (Well, our tests right now consist solely of lint/code-style checks, but we'll improve that soon.) Thanks again for your help! Hey @neagle no problem! I have about a year of Angular experience, I'm not an ng-guru but I get by, thanks :) I agree that simply putting ng-if is a bit of an hack, I debated this in my head for a while. I think that ultimately it's not that bad, the problem is that the directive shouldn't be injected at all. You could put the whole HTML in a partial view and put this logic (don't show for users who are not logged in, don't own the game etc.) in the template controller so the behavior becomes more understandable (which, from what I see, is why the whole ng-if idea smells.) We should also consider trade-offs: ng-if won't render the HTML, so when a user actually can and want to edit a comment, the template will have to be rendered, linked etc. At the same time, leaving an ng-show will keep the directive bound and all the bindings will fire at any $scope change. But it's probably not worth thinking about all this for now :) Great news about Travis and gulp, can't wait for proper unit tests :) I'm very green on angular unit testing (the TDD church will banish me) so it will be fun to write some! I'm not an ng-guru, either, though I'm starting to get more comfortable. GoKibitz is my first and only significant project in Angular, though I have a couple smaller projects at work that I've used Angular for. In general, I'm pretty happy with it: even though there are lots of things that could use refactoring since I know more now than I did when I originally implemented them, having a framework like Angular makes changes and new features a lot easier than it might be otherwise. Proper unit tests will be great: I'm new to the TDD church, too, and most of my experience is with relatively simple mocha tests for NPM modules. But one of the main benefits of GoKibitz for me is the chance to grow as a developer, and if it helps some of the devs who contribute, too, then that's better still. :smile: Thanks again for your help!
gharchive/issue
2015-05-16T08:38:43
2025-04-01T04:35:10.655878
{ "authors": [ "levelonedev", "neagle", "sphaso" ], "repo": "neagle/gokibitz", "url": "https://github.com/neagle/gokibitz/issues/87", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1310523451
"Sorry, assemblyscript is not supported in your system, use --contract=rust " error in NEAR. Suggest me to use rust instead of assemblyscript when i give assemblyscript in command I just want to use create-near-app package from the npm registry as assemblyscript and react. But it gives https://i.stack.imgur.com/Lx36I.png error, how can i fix that ? That's due to this changes: https://github.com/near/create-near-app/commit/dd8ac16f74dc231c94c12a423ebafd60e1dda3bd#diff-e727e4bdf3657fd1d798edcd6b099d6e092f8573cba266154583a746bba0f346R62 which means if your system is not 64-bit MacOS or Linux you should use Rust instead AS. @gagdiez but why? If you want to avoid Apple's M1 better use this detection algorithm: const os = require('os'); const isM1 = return os.cpus()[0].model.includes("Apple"); const support_sandbox = !isM1 Also it will be great to make a reference to this issue https://github.com/near/workspaces-rs/issues/110 @gagdiez this error described in https://github.com/near/workspaces-rs/issues/110. It also affects to Rust template. I proposed a solution. Just properly detect M1 arch as descibed here. And plz remove unnecessary message "Sorry, assemblyscript is not supported in your system, use --contract=rust" which compleately block all users which use 32/64-bit windows for ANY platform sure, we are currently working on the new release, will see to change the message to "Sorry, NEAR-SDK-AS is not supported in your system" by then. Thanks for the heads up! @sametmollaoglu @MaxGraey the latest version of create-near-app has just been released, which enables assemblyscript in Mac and Linux. Our tool does not have support for windows, for which you cannot use it to start a NEAR project. You can try it now using npx create-near-app and that will download version 4.0.0. I will clarify info about Windows support: you can use all our tools if you are in the Windows subsystem for Linux.
gharchive/issue
2022-07-20T06:41:49
2025-04-01T04:35:10.666127
{ "authors": [ "MaxGraey", "gagdiez", "sametmollaoglu", "volovyk-s" ], "repo": "near/create-near-app", "url": "https://github.com/near/create-near-app/issues/1850", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
768426348
add subaccounts description / standardize usage Fixes #556 Thanks for the quick turnaround!
gharchive/pull-request
2020-12-16T03:55:33
2025-04-01T04:35:10.667031
{ "authors": [ "mikedotexe", "thisisjoshford" ], "repo": "near/docs", "url": "https://github.com/near/docs/pull/557", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1088802984
Fungible token transfer sometimes requires signing two storage deposit transactions Problem Fungible token transfer sometimes requires signing two storage deposit transactions. Probably this is happening due to the decrease of the storage costs, while the old contracts were using the hardcoded 0.0125 NEAR for the storage (and the new ones -- 0.00125). Example: https://explorer.near.org/transactions/E2tC36DMkTWf6Uo4yxCJNMy7taFBhXiZb539CPCfwHZM -- failed https://explorer.near.org/transactions/8HXBq2QwVPup5wypfynWEgZRXudLxaeX3EhVT36u8vU6 -- succeeded Expected Behavior Single storage deposit transaction. Proposed solution Instead of guessing the amount of $NEAR to be deposited for the storage, one need to execute a view call to the token contract and check out the storage balance bounds. CLI version of the call: $ near view 6b175474e89094c44da98b954eedeac495271d0f.factory.bridge.near storage_balance_bounds View call: 6b175474e89094c44da98b954eedeac495271d0f.factory.bridge.near.storage_balance_bounds() { min: '12500000000000000000000', max: '12500000000000000000000' } From this data one can extract the needed balance @MaximusHaximus I don't have clear what to do if the contract doesn't have this view method available, and how this could scale (I suppose this "view" can be called once and easily cached, instead of using the indexer every time). I'll have you take it from there for both analysis and handover to the team. From a product standpoint, I see risks in the latency of the operation - but they are lower than having this annoying "double signature" behavior, that would be amplified in presence of 2FA, Ledger and (soon) external services
gharchive/issue
2021-12-26T16:23:47
2025-04-01T04:35:10.670955
{ "authors": [ "alexauroradev", "stefanopepe" ], "repo": "near/near-wallet", "url": "https://github.com/near/near-wallet/issues/2338", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1806609897
142 solution reply post Solves issue #142 Preview version -User sees the same fields & preview in the solution reply as shown in the create solution post -User is able to select USDC and proceed successfully without a warning message Hmm when I try reply from your preview, it's still not showing USDC: Did you deploy something else that overrides this branch? @ailispI noticed that the previous preview was accidentally overwritten, I apologize for the late reaction. Here's the updated preview for your review preview
gharchive/pull-request
2023-07-16T14:50:41
2025-04-01T04:35:10.676129
{ "authors": [ "Tguntenaar", "ailisp" ], "repo": "near/neardevhub-widgets", "url": "https://github.com/near/neardevhub-widgets/pull/205", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1059437126
Commenting out profile items doesn't function as expected Basically as the title says. If I comment out a profile item by wrapping it in <!-- itemHere --> it does hide the link, but the font awesome icon is still present, causing this fun mess. Ugh, that's not cool. I can fix it soon though! Will update when done. This item is being tracked in the new omg.lol Feedback Forum. To follow its status, visit: https://feedback.omg.lol/16605846533313 This has been resolved.
gharchive/issue
2021-11-21T17:25:55
2025-04-01T04:35:10.678356
{ "authors": [ "chxseh", "newbold" ], "repo": "neatnik/omg.lol", "url": "https://github.com/neatnik/omg.lol/issues/167", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
418625390
Make all modules optional and separate them out in the codebase TL;DR: It should be possible to specify exactly what features you want using cargo, and it should "just work." Currently there is some intermingling between features2d and xfeatures2d. OpenCV also has several other modules and some have dependencies on each other. This depends on the completion of the sys crate refactor #115. The sys crate will only build the parts of OpenCV requested. This should result in being able to include just the components needed to avoid having to pay royalties to patent encumbered things in opencv_contrib. Moving all contrib features behind contrib feature would be sufficient. It just make no sense to me. Having 100+ features doesn't make crate easier to use but the opposite. Having minimal amount of switches that covers all cases is the best. Think of it as of Karnaugh map. Whats the status of this? I just compiled opencv3 from sources and ocr.hpp doesnt seem to exist? I also cant find it in the source repo. This might help these kind of issues @geoeo It seems @Pzixel might not be interested in accepting changes to enable only specific modules. However if you are interested in messing around with cv-rs where it builds OpenCV for you, you can use my #117 branch. Just point your git line to my repo and set `rev = "refactor". You can also use the system feature to cause it to use the system packages like the current master does. It requires clang to be installed as well though. Are you interested in making each OpenCV module optional? If you could provide some feedback, it might help determine what happens with this issue. The core issue is different. Its that it seems that the ocr.hpp file has been removed from the core opencv repo. This would have helped to mitigate the failed compilation. But im not 100% sure. @geoeo Make sure you compile the opencv submodule in this repository. It points to the right build of OpenCV. I thought I only have to do $cargo build What else do I have to do? I cant get this folder to compile. I have to install 100 dependencies which are not relevant. I would +1 this if that means anything. @geoeo After I merge the #117 branch, you wont really need any dependencies installed on Linux (except GTK3) or Windows. You can try it out now if you want, but I am still dealing with issues on Ubuntu 16.04 on Travis (so it passes CI). There's a typo in build.rs line 102 let cargo_rustc_link_search = env::var("OPENCV_LIB").unwrap_or("/use/local/lib".into()); I think it should be let cargo_rustc_link_search = env::var("OPENCV_LIB").unwrap_or("/usr/local/lib".into()); Didnt want to make a PR for this. Fixed
gharchive/issue
2019-03-08T04:20:16
2025-04-01T04:35:10.684420
{ "authors": [ "Pzixel", "geoeo", "vadixidav" ], "repo": "nebgnahz/cv-rs", "url": "https://github.com/nebgnahz/cv-rs/issues/119", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
913657550
футер Powered by Netlify, Hugo, and Doks from Params Добавил новый футер. Смотрится вроде неплохо. Футер классный! Еще и меню повторяется, очень здорово. Я подумаю может текст как-то продлить про факультет.
gharchive/issue
2021-06-07T15:08:37
2025-04-01T04:35:10.696436
{ "authors": [ "epogrebnyak", "necelentano" ], "repo": "necelentano/finec-mgimo-v2", "url": "https://github.com/necelentano/finec-mgimo-v2/issues/28", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1526392911
Branch condition covered but labeled as uncovered Bug Description The following code says a branch condition is never false when it is. To Reproduce Python code to reproduce the issue (parsers.py): import unittest def parse_sms(text): data = {} for i, split in enumerate(text.split(",")): if i == 0: data["name"] = split.title() print(len(data['name'].split()) == 1) if len(data['name'].split()) == 1: return None continue return data class TestParsers(unittest.TestCase): def test_parse_sms(self): messages = [ "John Doe, Jane", "Jane, John Doe", ] for message in messages: parse_sms(message) Command to reproduce the issue: python -m coverage erase python -m coverage run --branch -m unittest parsers.py python -m coverage html I expected line 9 to be both True and False, but the coverage report says it is never False. The print statement on line 8 prints both False and True. Relevant details: $ python --version Python 3.9.13 $ python -m coverage --version Coverage.py, version 7.0.3 with C extension $ python -m coverage debug sys -- sys ------------------------------------------------------- coverage_version: 7.0.3 coverage_module: C:\Users\...\AppData\Local\Packages\PythonSoftwareFoundation.Python.3.9_qbz5n2kfra8p0\LocalCache\local-packages\Python39\site-packages\coverage\__init__.py tracer: -none- CTracer: available plugins.file_tracers: -none- plugins.configurers: -none- plugins.context_switchers: -none- configs_attempted: .coveragerc setup.cfg tox.ini pyproject.toml configs_read: -none- config_file: None config_contents: -none- data_file: -none- python: 3.9.13 (tags/v3.9.13:6de2ca5, May 17 2022, 16:36:42) [MSC v.1929 64 bit (AMD64)] platform: Windows-10-10.0.19042-SP0 implementation: CPython executable: C:\Users\...\AppData\Local\Microsoft\WindowsApps\PythonSoftwareFoundation.Python.3.9_qbz5n2kfra8p0\python.exe def_encoding: utf-8 fs_encoding: utf-8 pid: 6636 cwd: ... path: ... environment: HOME = C:\Users\... PYTHONPATH = ... PYTHONUSERBASE = C:\Users\...\AppData\Local\Packages\PythonSoftwareFoundation.Python.3.9_qbz5n2kfra8p0\LocalCache\local-packages TEMP = C:\Users\...\AppData\Local\Temp TMP = C:\Users\...\AppData\Local\Temp command_line: C:\Users\...\AppData\Local\Packages\PythonSoftwareFoundation.Python.3.9_qbz5n2kfra8p0\LocalCache\local-packages\Python39\site-packages\coverage\__main__.py debug sys sqlite3_sqlite_version: 3.37.2 sqlite3_temp_store: 0 sqlite3_compile_options: ATOMIC_INTRINSICS=0, COMPILER=msvc-1929, DEFAULT_AUTOVACUUM, DEFAULT_CACHE_SIZE=-2000, DEFAULT_FILE_FORMAT=4, DEFAULT_JOURNAL_SIZE_LIMIT=-1, DEFAULT_MMAP_SIZE=0, DEFAULT_PAGE_SIZE=4096, DEFAULT_PCACHE_INITSZ=20, DEFAULT_RECURSIVE_TRIGGERS, DEFAULT_SECTOR_SIZE=4096, DEFAULT_SYNCHRONOUS=2, DEFAULT_WAL_AUTOCHECKPOINT=1000, DEFAULT_WAL_SYNCHRONOUS=2, DEFAULT_WORKER_THREADS=0, ENABLE_FTS3, ENABLE_FTS4, ENABLE_FTS5, ENABLE_JSON1, MALLOC_SOFT_LIMIT=1024, MAX_ATTACHED=10, MAX_COLUMN=2000, MAX_COMPOUND_SELECT=500, MAX_DEFAULT_PAGE_SIZE=8192, MAX_EXPR_DEPTH=1000, MAX_FUNCTION_ARG=127, MAX_LENGTH=1000000000, MAX_LIKE_PATTERN_LENGTH=50000, MAX_MMAP_SIZE=0x7fff0000, MAX_PAGE_COUNT=1073741823, MAX_PAGE_SIZE=65536, MAX_SQL_LENGTH=1000000000, MAX_TRIGGER_DEPTH=1000, MAX_VARIABLE_NUMBER=32766, MAX_VDBE_OP=250000000, MAX_WORKER_THREADS=8, MUTEX_W32, SYSTEM_MALLOC, TEMP_STORE=1, THREADSAFE=1 This issue does not occur on Python 3.11, but using your example I was able to reproduce it on Python 3.8. I believe this is a duplicate of #198.
gharchive/issue
2023-01-09T22:00:45
2025-04-01T04:35:10.701261
{ "authors": [ "aa-stevo", "kevin-brown", "nedbat" ], "repo": "nedbat/coveragepy", "url": "https://github.com/nedbat/coveragepy/issues/1530", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1670407455
relative_files & include broken since 7.0 Describe the bug Create simple test case with relative_files = true mkdir foo echo 'print()" > foo/bar.py echo "[run]\nrelative_files = true" > .coveragerc coverage run foo/bar.py coverage report --include=foo/bar.py Expected (as behavior was before 7.0): Name Stmts Miss Cover -------------------------------- foo/bar.py 1 0 100% -------------------------------- TOTAL 1 0 100% Actual: No data to report Additional analysis: It appears that the regex created from the foo/bar.py glob pattern includes the current working directory, which might explain why it wouldn't match against relative paths anymore. Hmm, I see what you mean. It will still work if you use --include='*/foo/bar.py'. I have a possible fix, but it's hard to work through all the permutations. I will investigate. Preferably, since I have a script invoke that report across repositories where different versions might be used. Also, I think the current syntax is intuitive for the relative files case. @marcgibbons This is fixed in ee6506f0. This is now released as part of coverage 7.2.4.
gharchive/issue
2023-04-17T04:58:16
2025-04-01T04:35:10.706020
{ "authors": [ "ikonst", "marcgibbons", "nedbat" ], "repo": "nedbat/coveragepy", "url": "https://github.com/nedbat/coveragepy/issues/1604", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
278427688
Notification moving on the screen Is it possible to move the notification around the screnn, like old windows save-screen ? Hi @mtugnoli, I'm not familiar with windows save-screen. Can you explain a little bit? Do you mean dragging notifications? If so, it's not possible.
gharchive/issue
2017-12-01T10:49:33
2025-04-01T04:35:10.773696
{ "authors": [ "mtugnoli", "needim" ], "repo": "needim/noty", "url": "https://github.com/needim/noty/issues/416", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1993483189
[Bug Report] Fix n_params counts Describe the bug The n_params counts calculated here are wrong. For example, LLAMA uses SwiGLU so the 2x factor in the linked code is wrong. Further this just ignores bias parameters I think? Code example # Load in Llama-7B llama.cfg.n_params # 5033164800 ... System Info N/A Additional context N/A Checklist [ x ] I have checked that there is no similar issue in the repo (required) Unclear what the solution should be. There are plausibly three different parameter counts that are helpful Parameters in training Parameters ignoring embeddings Parameters used now (e.g folding layer norm deletes some parameters) IMO this should be just total parameters for simplicity and alignment with the Pythia suite. Who cares about LayerNorm On Tue, 14 Nov 2023, 8:09 pm ArthurConmy, @.***> wrote: Unclear what the solution should be. There are plausibly three different parameter counts that are helpful: Parameters in training Parameters ignoring embeddings Parameters used now (e.g folding layer norm deletes some parameters) — Reply to this email directly, view it on GitHub https://github.com/neelnanda-io/TransformerLens/issues/448#issuecomment-1811155350, or unsubscribe https://github.com/notifications/unsubscribe-auth/ASRPNKNC7SNW2N2YHGVXGDDYEPFXFAVCNFSM6AAAAAA7LMEBR2VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTQMJRGE2TKMZVGA . You are receiving this because you are subscribed to this thread.Message ID: @.***>
gharchive/issue
2023-11-14T20:06:56
2025-04-01T04:35:10.787108
{ "authors": [ "ArthurConmy", "neelnanda-io" ], "repo": "neelnanda-io/TransformerLens", "url": "https://github.com/neelnanda-io/TransformerLens/issues/448", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2000066305
no longer working in most recent Ventura 13.6.1 update extension is in System Settings > Extensions > Added Extensions but no longer works. not sure when it stopped working but definitely no longer working in the most recent Ventura update 13.6.1 on an intel imac. was working a few months ago in earlier Ventura 13.x Not sure if related but for me the color option in the context menu is unavailable for iCloud folders only. Can still use the color option context menu on local and dropbox folders though. FWIW I forked this and added German localization and a dark blue colour. https://github.com/RedTartan04/Manila You'd have to build it yourself though, sorry. Tested on 12.7.3 and 13.6.4 on M1s. extension is in System Settings > Extensions > Added Extensions but no longer works. not sure when it stopped working but definitely no longer working in the most recent Ventura update 13.6.1 on an intel imac. was working a few months ago in earlier Ventura 13.x Seems like it's working again on Sonoma
gharchive/issue
2023-11-17T23:15:35
2025-04-01T04:35:10.803028
{ "authors": [ "1gcd", "RedTartan04", "artlambi", "maxoakland" ], "repo": "neilsardesai/Manila", "url": "https://github.com/neilsardesai/Manila/issues/14", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1328096213
Authentification 認証機能を作成 どのように認証済みユーザを保管するのか? sequenceDiagram autonumber actor User actor Admin participant Server as Server User->>Server: Some Request alt User is authenticated Note right of User: Do nothing! else User is not authenticated Server->>User: Message Note left of Server: "You are not authorized user,<br>Will you send auth request to admin?"<br>"あなたは未認証のユーザです。<br>adminに認証リクエストを送信しますか?" alt User's answer is "No" Server->>User: Message Note left of Server: "OK! good bye!"<br>"ほなさいなら" else User's answer is "Yes" Server->>Admin: Message Note right of Admin: "Auth Request From "User", Will you Authenticate "User"?"<br>"○○から認証リクエストきたけど認証しちゃう?" Admin-->>Server:the Answer alt Admin's Answer is "No" Server-->>User: Message Note right of User: "Sorry, You are denied from Admin"<br>"ごめん。あかんってさ" else Admin's Answer is "Yes" Server->Server:Save userId to Authenticated User Server-->>User: Message Note right of User: "Congratulations, You are authenticated!"<br>"よかったね。承認されたみたいよ。" end end end DBとしてSqliteを採用 ORMとしてSQLAlchemyを採用 シーケンス図をREADME.mdに転記して修正 #13
gharchive/issue
2022-08-04T05:42:03
2025-04-01T04:35:10.810764
{ "authors": [ "nekonisi" ], "repo": "nekonisi/baby_channel", "url": "https://github.com/nekonisi/baby_channel/issues/3", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2514713538
README.md specifies setting permissions to logs file to 777 This is a HUGE NO-NO on Linux and introduces a massive security vulnerability. OK, thanks for the report. Your recommendation? Most of the RPis are expected to not be multiuser, so this is rather theoretical. Hey, You're not talking about users but processes - which could be a worm/malware/virus etc. so never underestimate the risk. Good security also helps in other ways. Fortunately, it's easy to fix. Just give the logs directory the ownership of the PHP process - usually the same as your web server, and give that user write permissions; you can also give a group access to the directory if other processes need right to it for read and/or write. Depends on which processes access the directory. Regards, -Andy (cybervegan) On Mon, 9 Sept 2024 at 20:22, Galimatias Nekromoff @.***> wrote: OK, thanks for the report. Your recommendation? Most of the RPis are expected to not be multiuser, so this is rather theoretical. — Reply to this email directly, view it on GitHub https://github.com/nekromoff/rpi-monitor-dashboard/issues/7#issuecomment-2338902932, or unsubscribe https://github.com/notifications/unsubscribe-auth/AB2NTQ2NOVPQTRTNY5BRIG3ZVXYPZAVCNFSM6AAAAABN5FXCRKVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDGMZYHEYDEOJTGI . You are receiving this because you authored the thread.Message ID: @.***> Thank you, I'll change README accordingly. On Mon, Sep 9, 2024 at 9:29 PM cybervegan @.***> wrote: Hey, You're not talking about users but processes - which could be a worm/malware/virus etc. so never underestimate the risk. Good security also helps in other ways. Fortunately, it's easy to fix. Just give the logs directory the ownership of the PHP process - usually the same as your web server, and give that user write permissions; you can also give a group access to the directory if other processes need right to it for read and/or write. Depends on which processes access the directory. Regards, -Andy (cybervegan) On Mon, 9 Sept 2024 at 20:22, Galimatias Nekromoff @.***> wrote: OK, thanks for the report. Your recommendation? Most of the RPis are expected to not be multiuser, so this is rather theoretical. — Reply to this email directly, view it on GitHub < https://github.com/nekromoff/rpi-monitor-dashboard/issues/7#issuecomment-2338902932>, or unsubscribe < https://github.com/notifications/unsubscribe-auth/AB2NTQ2NOVPQTRTNY5BRIG3ZVXYPZAVCNFSM6AAAAABN5FXCRKVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDGMZYHEYDEOJTGI> . You are receiving this because you authored the thread.Message ID: @.***> — Reply to this email directly, view it on GitHub https://github.com/nekromoff/rpi-monitor-dashboard/issues/7#issuecomment-2338914719, or unsubscribe https://github.com/notifications/unsubscribe-auth/ACBHPTL4LML7VAVKMT3RLE3ZVXZKRAVCNFSM6AAAAABN5FXCRKVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDGMZYHEYTINZRHE . You are receiving this because you commented.Message ID: @.***>
gharchive/issue
2024-09-09T19:10:02
2025-04-01T04:35:10.824386
{ "authors": [ "cybervegan", "nekromoff" ], "repo": "nekromoff/rpi-monitor-dashboard", "url": "https://github.com/nekromoff/rpi-monitor-dashboard/issues/7", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
518482076
Update code base to PHP 7.1 features Seeing as the last few PRs got us on track for a major version bump, and I have a few minutes spare, I thought I'd — minus asking first :roll_eyes: — submit this. It's only a few minutes worth of work & verification, so please just close this if you're uninterested @Seldaek :smiley_cat: Eh why not do a big cleanup once thanks
gharchive/pull-request
2019-11-06T13:46:20
2025-04-01T04:35:10.826982
{ "authors": [ "GawainLynch", "Seldaek" ], "repo": "nelmio/NelmioCorsBundle", "url": "https://github.com/nelmio/NelmioCorsBundle/pull/137", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
860096940
fix #1432 fix harvesting status on node operator accounts Tested: Not fixed. When you switch to another view and go back you have to put node again. @cryptoBeliever this is a separate issue and fixed in a different pr
gharchive/pull-request
2021-04-16T19:12:29
2025-04-01T04:35:10.836613
{ "authors": [ "AnthonyLaw", "bassemmagdy", "cryptoBeliever" ], "repo": "nemgrouplimited/symbol-desktop-wallet", "url": "https://github.com/nemgrouplimited/symbol-desktop-wallet/pull/1439", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
161311679
REPEATED TRAINING I'm fairly a beginner in NN, what I would love to know is: will adding documents and training time after time have any negative effect on the results of the first batch of documents that were added? E.g. Day 1: add and train 20 docs... SAVE neural network to file Day 2: load neural network and add a new/different set of documents and train... SAVE ..... Day #n (~100): load neural network and classify some text... will it give the correct class for items added on day 1 all the way through let's say day 12? @ohenepee this isn't technically setup for continuous learning as the training process starts everything from scratch (as opposed to just doing a backward pass when a new document is added). The way we've structured it is to essentially just use the pre-trained models in production and do training in batches as new data becomes available once in a while. Okay thanks
gharchive/issue
2016-06-20T22:34:52
2025-04-01T04:35:10.838934
{ "authors": [ "nemo", "ohenepee" ], "repo": "nemo/natural-synaptic", "url": "https://github.com/nemo/natural-synaptic/issues/2", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
385047255
Add retreat counterpart of advance prior is to next what retreat is to advance There is no "retreat" in the standard library, so this could be confusing. I should probably ensure that utf8::advance() retreats for a negative distance. Thanks, I wasn't aware having only standard library-like methods was a design decision. There is no prior in the standard library either, but there is std::prev (and std::next can go backwards).
gharchive/pull-request
2018-11-28T00:27:15
2025-04-01T04:35:10.845615
{ "authors": [ "glebm", "nemtrif" ], "repo": "nemtrif/utfcpp", "url": "https://github.com/nemtrif/utfcpp/pull/20", "license": "BSL-1.0", "license_type": "permissive", "license_source": "github-api" }
374641596
Persitence Plugin for exporting storage changes First draft for Persistence plugin. [X] First draft [x] Fix compiling errors [x] Create a bool variable that says the task to be done on Persist, in our cases, storage dump on every block. [x] Sync mainnet [x] Tested @erikzhang, @igormcoelho, we are not being able to get the object class Trackable. Persistence.cs(24,22): error CS0246: The type or namespace name 'Trackable' could not be found (are you missing a using directive or an assembly reference?) [/opt/neo-plugins/Persistence/Persistence.csproj] Can you give a hand? By the way, is it really necessary to put this plugin into neo-project repos? Almost....aehauheauheaea Erik, if you think this is useful as example we can keep it, otherwise we just use it in our storage verification project. [{"state":"Added","key":"System.Byte[]","value":"System.Byte[]"},{"state":"Added","key":"System.Byte[]","value":"System.Byte[]"},{"state":"Added","key":"System.Byte[]","value":"System.Byte[]"},{"state":"Added","key":"System.Byte[]","value":"System.Byte[]"},{"state":"Added","key":"System.Byte[]","value":"System.Byte[]"}]}, Again... =/ New export {"block":1444843,"size":5,"storage": [{"state":"Added","key":"85104d0b1bc285289b17717a6facaa2cbd1712b3","value":"0060b7c6c5fe11"},{"state":"Added","key":"130b891dc5341bcef93c077fc7ec5624ee8776f8","value":"00f87f24795120"},{"state":"Added","key":"45bcb590e3e0fb0010d7bfde6f7cd39382fd86e9","value":"00c0b6403b6f0c"},{"state":"Added","key":"746f74616c537570706c79","value":"0018ee2b7abf3e"},{"state":"Added","key":"73616c65734e656f","value":"0041ad29410b"}]}, The previous one: {"block":1444843,"size":5,"storage": [{"key":"f91d6b7085db7c5aaf09f19eeec1ca3c0db2c6ec130b891dc5341bcef93c077fc7ec562400ee8776f80000000000000000000000000c","value":"000700f87f2479512000eb0b1600"},{"key":"f91d6b7085db7c5aaf09f19eeec1ca3c0db2c6ec45bcb590e3e0fb0010d7bfde6f7cd3930082fd86e90000000000000000000000000c","value":"000700c0b6403b6f0c00eb0b1600"},{"key":"f91d6b7085db7c5aaf09f19eeec1ca3c0db2c6ec73616c65734e656f000000000000000008","value":"00060041ad29410b00eb0b1600"},{"key":"f91d6b7085db7c5aaf09f19eeec1ca3c0db2c6ec746f74616c537570706c79000000000005","value":"00070018ee2b7abf3e00eb0b1600"},{"key":"f91d6b7085db7c5aaf09f19eeec1ca3c0db2c6ec85104d0b1bc285289b17717a6facaa2c00bd1712b30000000000000000000000000c","value":"00070060b7c6c5fe1100eb0b1600"}]}, I think it is still wrong. Values not match Done: :+1: {"block":1444843,"size":5,"storage": [{"state":"Added","key":"f91d6b7085db7c5aaf09f19eeec1ca3c0db2c6ec85104d0b1bc285289b17717a6facaa2c00bd1712b30000000000000000000000000c","value":"00070060b7c6c5fe1100"},{"state":"Added","key":"f91d6b7085db7c5aaf09f19eeec1ca3c0db2c6ec130b891dc5341bcef93c077fc7ec562400ee8776f80000000000000000000000000c","value":"000700f87f2479512000"},{"state":"Added","key":"f91d6b7085db7c5aaf09f19eeec1ca3c0db2c6ec45bcb590e3e0fb0010d7bfde6f7cd3930082fd86e90000000000000000000000000c","value":"000700c0b6403b6f0c00"},{"state":"Added","key":"f91d6b7085db7c5aaf09f19eeec1ca3c0db2c6ec746f74616c537570706c79000000000005","value":"00070018ee2b7abf3e00"},{"state":"Added","key":"f91d6b7085db7c5aaf09f19eeec1ca3c0db2c6ec73616c65734e656f000000000000000008","value":"00060041ad29410b00"}]}, Thanks @erikzhang, without your assistance and attention it would be impossible to reach these magic functions! PS: @igormcoelho, there is some minnor different between the storage reported here and the previous one we were getting before. Even the Key is different. The values is different (as expected) because we added that new field of Height. But I do know, what do you think? Congratulations brother @vncoelho for your Persistance xD The values seem correct, it's just the order that is not the same. https://raw.githubusercontent.com/NeoResearch/neo-storage-audit/master/BlockStorage_1500000/dump-block-1445000.json But order is not important for modified PatriciaMerkle tree, so we can use this valuable information to finally build that tree ;) I think we can merge this plugin with StatesDumper. Storage dump on top, @erikzhang. We gonna migrate to StatesDumper and test again. On top and syncing the blockchain. All tests are completed. Soon we gonna move to StateDumper and separate parameters. https://github.com/NeoResearch/neo-storage-audit
gharchive/pull-request
2018-10-27T12:44:56
2025-04-01T04:35:10.878508
{ "authors": [ "erikzhang", "igormcoelho", "vncoelho" ], "repo": "neo-project/neo-plugins", "url": "https://github.com/neo-project/neo-plugins/pull/21", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
225075034
NeoServletContainer,-1,false Hi there, I followed the instructions in your readme to the letter, but on restarting my neo4j server I found this error Starting Neo4j failed: org.neo4j.server.web.NeoServletContainer-673720ea@f0284693==org.neo4j.server.web.NeoServletContainer,-1,false I'm not sure where the error is neo4j version 3.1.3 Apache Maven 3.5.0 (ff8f5e7444045639af65f6095c62210b5713f426; 2017-04-03T20:39:06+01:00) Maven home: /usr/local/Cellar/maven/3.5.0/libexec Java version: 1.8.0_121, vendor: Oracle Corporation Java home: /Library/Java/JavaVirtualMachines/jdk1.8.0_121.jdk/Contents/Home/jre Default locale: en_GB, platform encoding: UTF-8 OS name: "mac os x", version: "10.12.1", arch: "x86_64", family: "mac" In which log did you see that? Anything in debug.log ? Did your server start? I saw this message in a popup when attempting to start the server from a gui. The server did not start. There seems to be no additional details in a log in the locations I would expect to see a log. Just investigating now... After an hour of digging and some desperation I restarted my computer. Now it's working. Nothing changed. No idea what the problem was. Thank you @jexp for taking a look too!
gharchive/issue
2017-04-28T13:05:41
2025-04-01T04:35:10.895537
{ "authors": [ "jexp", "smccamley" ], "repo": "neo4j-contrib/neo4j-graphql", "url": "https://github.com/neo4j-contrib/neo4j-graphql/issues/20", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
127596744
QueryFactory w/RelatedNode Same idea as https://github.com/neo4jrb/neo4j/pull/1109 but implemented differently. This fixes/permits: MyRelClass.create(activerel_object.from_node, activerel_object.to_node) @djvs @cheerfulstoic I'm going to cherry-pick this into 6.1.x once it passes, too. Awesome ;) Run-ran the one build because it looked like a transient error Then it happened again. Ugh. Let's start opening issues when we see these, then someone can take care of them when they have time. Transient(-ish) error fixed!
gharchive/pull-request
2016-01-20T03:32:11
2025-04-01T04:35:10.972670
{ "authors": [ "cheerfulstoic", "subvertallchris" ], "repo": "neo4jrb/neo4j", "url": "https://github.com/neo4jrb/neo4j/pull/1112", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
624658539
Feature Request: Ignore specific tsserver codes Would it be possible to add a configuration option for disabling specific tsserver disagnostics/suggestions codes. My specific usecase, is that I want to ignore 7016 - Could not find a declaration file for module ... But instead of disabling all suggestions, I would like to just ignore that. But maybe a configuration for disabling specific codes, could be more generic and helpful in other cases as well? PR welcome. Cool, would you implement it here? https://github.com/neoclide/coc-tsserver/blob/master/src/server/features/diagnostics.ts#L145 No, I don't like the idea to ignore specific diagnostics, it can be useful, the feature should be provided by configuration of language server, not client. tsserver doesn't seem to support this, from what I can find. If it is an opt in feature, would it be that bad, you are already doing specific checks for code 6133 It's same as VSCode, we will support ignore 7016 when VSCode support it. Use "javascript.suggestionActions.enabled": false, "typescript.suggestionActions.enabled": false,
gharchive/issue
2020-05-26T07:25:11
2025-04-01T04:35:10.978315
{ "authors": [ "chemzqm", "kibs" ], "repo": "neoclide/coc-tsserver", "url": "https://github.com/neoclide/coc-tsserver/issues/156", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1000934936
Speed up tsserver I am here due to the following note: **Note**: tsserver could be quite slow to initialize on big project, exclude unneunnecessary files in your jsconfig.json/tsconfig.json. Could someone help me set up an angular project folder to exclude an enormous node_modules folder? My tsconfig.json: There is https://github.com/iamcco/coc-angular for angular project, and we have typescript.preferences.includePackageJsonAutoImports configuration which might helps. That's not too many files, maybe you need SSD.
gharchive/issue
2021-09-20T12:53:34
2025-04-01T04:35:10.980896
{ "authors": [ "chemzqm", "pavlo-hadzheha" ], "repo": "neoclide/coc-tsserver", "url": "https://github.com/neoclide/coc-tsserver/issues/314", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1270696561
Incorrect behavior on completion import Problem If lines are together and trying to import - such bug appears. Before: After: versions vim version: NVIM v0.8.0-dev+404-g2b539d6fd node version: v16.15.1 coc.nvim version: 0.0.81-2522eee 2022-06-14 19:03:29 +0800 coc.nvim directory: /home/user/.local/share/nvim/plugged/coc.nvim term: xterm-256color platform: linux Log of coc.nvim 2022-06-14T15:57:48.661 INFO (pid:51042) [services] - registered service "languageserver.tailwind-lsp" 2022-06-14T15:57:49.004 INFO (pid:51042) [plugin] - coc.nvim initialized with node: v16.15.1 after 475ms 2022-06-14T15:57:49.772 INFO (pid:51042) [attach] - receive notification: openList [ 'files' ] 2022-06-14T15:57:51.474 INFO (pid:51042) [services] - tailwind-lsp state change: stopped => starting 2022-06-14T15:57:51.547 INFO (pid:51042) [services] - registered service "tsserver" 2022-06-14T15:57:51.611 INFO (pid:51042) [services] - service tsserver started 2022-06-14T15:57:51.634 INFO (pid:51042) [language-client-index] - Language server "languageserver.tailwind-lsp" started with 51080 2022-06-14T15:57:52.012 INFO (pid:51042) [services] - tailwind-lsp state change: starting => running 2022-06-14T15:57:52.018 INFO (pid:51042) [services] - service languageserver.tailwind-lsp started 2022-06-14T15:57:53.042 INFO (pid:51042) [core-watchman] - watchman watching project: /home/user/TS/app 2022-06-14T15:58:06.518 INFO (pid:51042) [snippets-session] - Content change before snippet, reset snippet position 2022-06-14T15:58:06.524 INFO (pid:51042) [snippets-session] - Content change before snippet, reset snippet position 2022-06-14T15:58:06.538 INFO (pid:51042) [attach] - receive notification: showSignatureHelp [] 2022-06-14T15:58:09.675 INFO (pid:51042) [attach] - receive notification: showInfo [] Try minimal vimrc. This is also happening to me. Specifically though, its in a file that has no imports yet. so when if my entire file contents are function useHook () { const a = useState() } and then i get auto completion from coc for useState and hit enter, than my file gets changed to import {useState} from "react"; const a = useState(initialState)() const a = useState(initialState)() } however, if i have an import at the top, and this is my file content import {useEffect} from 'react' function useHook () { const a = useState() } than when i auto complete for useState, i get import {useEffect, useState} from 'react' function useHook () { const a = useState(initialState)() } the latter is more correct (minus the double call to useState), but it doesn't mess up the imports. so for me it happens consistently when i have no imports and then try to import via auto completion. Fixed on latest release branch of coc.nvim.
gharchive/issue
2022-06-14T12:01:02
2025-04-01T04:35:10.990061
{ "authors": [ "chemzqm", "jeffshaver", "username664" ], "repo": "neoclide/coc-tsserver", "url": "https://github.com/neoclide/coc-tsserver/issues/381", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1741688498
Honkai: The Star Rail Railway in China cannot be started, cannot be installed, and the startup error System.FormatException: String '周三/23/4/05 上午 12:01:31' was not recognized as a valid DateTime. at System.DateTimeParse.Parse(ReadOnlySpan1, DateTimeFormatInfo, DateTimeStyles) at System.DateTime.Parse(String, IFormatProvider ) at System.Convert.ToDateTime(String, IFormatProvider) at System.String.System.IConvertible.ToDateTime(IFormatProvider) at System.Convert.ChangeType(Object, Type , IFormatProvider) at System.Convert.ChangeType(Object, Type ) at SevenZipExtractor.ArchiveFile.GetProperty[T](UInt32 fileIndex, ItemPropId name) in E:\myGit\Collapse\Hi3Helper.Core\Classes\Data\Tools\SevenZipTool\SevenZipExtractor\SevenZipExtractor\ArchiveFile.cs:line 297 at SevenZipExtractor.ArchiveFile.GetPropertySafe[T](UInt32 fileIndex, ItemPropId name) in E:\myGit\Collapse\Hi3Helper.Core\Classes\Data\Tools\SevenZipTool\SevenZipExtractor\SevenZipExtractor\ArchiveFile.cs:line 266 at SevenZipExtractor.ArchiveFile.get_Entries() in E:\myGit\Collapse\Hi3Helper.Core\Classes\Data\Tools\SevenZipTool\SevenZipExtractor\SevenZipExtractor\ArchiveFile.cs:line 226 at SevenZipExtractor.ArchiveFile..ctor(Stream archiveStream, Nullable1 format, String libraryFilePath) in E:\myGit\Collapse\Hi3Helper.Core\Classes\Data\Tools\SevenZipTool\SevenZipExtractor\SevenZipExtractor\ArchiveFile.cs:line 103 at Hi3Helper.Data.SevenZipTool.LoadArchive(String InputFile) in E:\myGit\Collapse\Hi3Helper.Core\Classes\Data\Tools\SevenZipTool.cs:line 30 at Hi3Helper.Data.SevenZipTool.GetUncompressedSize(String inputFile) in E:\myGit\Collapse\Hi3Helper.Core\Classes\Data\Tools\SevenZipTool.cs:line 49 at CollapseLauncher.InstallManager.Base.InstallManagerBase1.<>c__DisplayClass41_0.<StartPackageInstallation>b__1(GameInstallPackage x) in E:\myGit\Collapse\CollapseLauncher\Classes\InstallManagement\BaseClass\InstallManagerBase.cs:line 230 at System.Linq.Enumerable.Sum[TSource,TResult,TAccumulator](IEnumerable1, Func2) at System.Linq.Enumerable.Sum[TSource](IEnumerable1, Func2) at CollapseLauncher.InstallManager.Base.InstallManagerBase1.b__41_0() in E:\myGit\Collapse\CollapseLauncher\Classes\InstallManagement\BaseClass\InstallManagerBase.cs:line 230 at System.Threading.Tasks.Task.InnerInvoke() at System.Threading.Tasks.Task.<>c.<.cctor>b__273_0(Object) at System.Threading.ExecutionContext.RunFromThreadPoolDispatchLoop(Thread, ExecutionContext, ContextCallback, Object) --- End of stack trace from previous location --- at System.Threading.ExecutionContext.RunFromThreadPoolDispatchLoop(Thread, ExecutionContext, ContextCallback, Object) at System.Threading.Tasks.Task.ExecuteWithThreadLocal(Task&, Thread ) --- End of stack trace from previous location --- at CollapseLauncher.InstallManager.Base.InstallManagerBase`1.StartPackageInstallation() in E:\myGit\Collapse\CollapseLauncher\Classes\InstallManagement\BaseClass\InstallManagerBase.cs:line 223 at CollapseLauncher.Pages.HomePage.InstallGameDialog(Object sender, RoutedEventArgs e) in E:\myGit\Collapse\CollapseLauncher\XAMLs\MainApp\Pages\HomePage.xaml.cs:line 496 Hi @mazhentao1992, We've pushed a new hotfix to fix this issue on the latest stable release, which is v1.69.21 version. Please make sure you've updated your launcher to the latest version. We are going to close this issue in 24 hours and if you have any issues, feel free to let us know.
gharchive/issue
2023-06-05T12:39:21
2025-04-01T04:35:11.087134
{ "authors": [ "mazhentao1992", "neon-nyan" ], "repo": "neon-nyan/Collapse", "url": "https://github.com/neon-nyan/Collapse/issues/159", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1942293815
Make hot standby feedback work Currently walproposer receives from safekeepers, but doesn't register. This was recently implemented, right? Closing
gharchive/issue
2023-10-13T17:00:49
2025-04-01T04:35:11.088207
{ "authors": [ "arssher", "hlinnaka" ], "repo": "neondatabase/neon", "url": "https://github.com/neondatabase/neon/issues/5558", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
315167191
TASK: Use correct validation ResultsViewHelper The Form\ValidationResultsViewHelper was deprecated since Flow 2.1 and shouldn't be used anymore. Style failure is unrelated...
gharchive/pull-request
2018-04-17T17:54:26
2025-04-01T04:35:11.102525
{ "authors": [ "kitsunet" ], "repo": "neos/form", "url": "https://github.com/neos/form/pull/73", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1082214633
fix: remove vim.lsp.util.close_preview_autocmd Inline vim.lsp.util.close_preview_autocmd, it was removed from neovim. Duplicate of #1575, I should probably pay more attention lol
gharchive/pull-request
2021-12-16T13:44:34
2025-04-01T04:35:11.240990
{ "authors": [ "ii14" ], "repo": "neovim/nvim-lspconfig", "url": "https://github.com/neovim/nvim-lspconfig/pull/1578", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
935162155
Create event subscription documentation Context: I'm creating headless neovim child instance like so: self._nvim = attach( "child", argv=["/usr/bin/env", "nvim", "--embed", "--headless"] ) I'm looking into subscribing to User events (so that I can self._nvim.command(":map zp doau User blabla<CR>") and then react to it within my Python code), but the only documentation I have found is this: subscribe(event) Subscribe to a Nvim event. unsubscribe(event) Unsubscribe to a Nvim event. I assumed I could self.nvim.subscribe('User', lambda ev: …) but no. All I can do is self.nvim.subscribe('User'). So I went to look at the tests to see how I can actually receive those events in the userland, and there are subscriptions tests indeed: vim.subscribe('event2') vim.command('call rpcnotify(0, "event1", 1, 2, 3)') vim.command('call rpcnotify(0, "event2", 4, 5, 6)') vim.command('call rpcnotify(0, "event2", 7, 8, 9)') event = vim.next_message() assert event[1] == 'event2' assert event[2] == [4, 5, 6] However that is still not very helpful. Does that mean I now have to create an event loop that will poll vim.next_message and filter relevant events? I thought pynvim does that for me as per documentation of the Nvim class? class Nvim(object): … When this library is used on python3.4+, asyncio event loop is guaranteed to be used. It is available as the "loop" attribute of this class. unfortunately subscribe/unsubscribe are mostly useless and unrelated to the vim "events" (autocommands). Instead see :help autocmd. work to improve this is https://github.com/neovim/neovim/pull/11613 for example Instead see :help autocmd. Excuse me, I am at a loss as to now :help autocmd helps with OPs question, could you elaborate?
gharchive/issue
2021-07-01T19:47:38
2025-04-01T04:35:11.245355
{ "authors": [ "filmil", "justinmk", "rr-" ], "repo": "neovim/pynvim", "url": "https://github.com/neovim/pynvim/issues/488", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
335736842
Pre-generate thumbnails to improve performance Currently, thumbnails are generated at run-time as needed using the thumbnail tag. Each time this tag is used, first the app must check if the thumbnail already exists on the media storage, and if not, generate it. When the media storage is remote such as on S3, this can take a lot of time. It would be much more efficient to instead use the thumbnail_url tag which assumes the thumbnail already exists and just writes the appropriate url in the HTML. This just means the thumbnail should be pre-generated whenever the thumbnail_options for an image are modified. There could be a switch to enable both the current dynamic and the pre-generating behavior, but since I don't see any advantage for the dynamic approach in the context of this package, I suggest just changing the behavior to always pre-genrate the thumbnails. It looks like I was wrong - for non-local storage, both the source and thumbnail generation time are queried in the database instead of querying the modification time of the file in the storage. (I'm still seeing this taking a significant time on page load with just a few thumbnails, but it's probably not due to remote storage access.) @danra djangocms-blog uses easy_thumbnails for thumbnail generation, which caches the thumbnails status in the database and support async thumnails generation. If it's your use case, you can change the template to reflex this and have a look at east_thumbnails async pregeneration https://easy-thumbnails.readthedocs.io/en/stable/usage/#asynchronous-pregeneration
gharchive/issue
2018-06-26T09:36:51
2025-04-01T04:35:11.248777
{ "authors": [ "danra", "yakky" ], "repo": "nephila/djangocms-blog", "url": "https://github.com/nephila/djangocms-blog/issues/459", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
1696306651
Create initial version of UERANSIM kpt package Initial implementation of UERANSIM kpt package based on Towards5gs helm charts, as discussed in nephio-project/nephio#105 UERANSIM packages will require to be specialized mannualy, when N2 and N3 endpoints will be known. All service-level configuration stays as default Towards5gs configuration. will be moved to the nephio-project/nephio-example-packages
gharchive/pull-request
2023-05-04T16:22:10
2025-04-01T04:35:11.250838
{ "authors": [ "matysiaq" ], "repo": "nephio-project/nephio-packages", "url": "https://github.com/nephio-project/nephio-packages/pull/25", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1732501892
Workload cluster deployment Make sure clusters are up and ConfigSync is deployed See readme included in for https://github.com/nephio-project/test-infra/pull/56 - this is now working for GCE. It should work for other environments too, once #235 build the sandboxes for those environments. @n2vo @liamfallon Make similar issues for each environment you plan to support - for now, this is not different cluster providers, but instead using the Kind provisioning in different environments. So summarizing the decision for this and #235, scope for R1 is: Demo/dev environment (all run in single VM with Kind clusters): GCE VM OpenStack VM Local VM Bring-your-own-cluster: requires manual setup of all clusters, repos, and joining workload clusters to Nephio ideally we document and test this in a couple environments Automated e2e test for GCE is in this PR: https://github.com/nephio-project/test-infra/pull/63/files This issue can be closed as it is satisfied by our e2e provisioning and test case 001.sh and 002.sh.
gharchive/issue
2023-05-30T15:26:24
2025-04-01T04:35:11.255358
{ "authors": [ "gvbalaji", "johnbelamaric" ], "repo": "nephio-project/nephio", "url": "https://github.com/nephio-project/nephio/issues/236", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
150100199
Having some difficulty authenticating (401 error) HI there, I am still trying to authenticate to GitLab API, I have tried the method suggested on the instructions: let config = TokenConfiguration("https://code.tiferrei.com/api/v3/", token: "MY-TOKEN") But I immediately get an error saying the function requires the parameter url, and not token. So I tried the following: let config = TokenConfiguration("https://gitlab.example.com/api/v3/", token: "MY-TOKEN") let config = TokenConfiguration(url: "https://gitlab.example.com/api/v3/", "MY-TOKEN") let config = TokenConfiguration("https://gitlab.example.com/api/v3/", "MY-TOKEN") With these the compiler error got solved, but with a simple viewDidLoad() of: TanukiKit(config).me() { response in switch response { case .Success(let user): print(user.login) case .Failure(let error): print(error) } }``` I still always get the 401 error on the console. What am I doing wrong? Thanks Did you try without /api/v3/? Only as https://code.tiferrei.com? No, I supposed it required the api part as the example. Let me try it... This is what I do in Git2Go when the user is logged in: public static func client() -> TanukiKit { if let authToken = privateToken, baseURL = baseURL { let apiURL = baseURL.stringByAppendingURLPath("/api/v3") let config = PrivateTokenConfiguration(authToken, url: apiURL) return TanukiKit(config) } return TanukiKit() } Hum, without the /api/v3/ part I now get: Error Domain=NSCocoaErrorDomain Code=3840 "JSON text did not start with array or object and option to allow fragments not set." UserInfo={NSDebugDescription=JSON text did not start with array or object and option to allow fragments not set.} Guess it's a progress from a 401 hahah You have to use a PrivateTokenConfiguration if you are using a self hosted GitLab (authentication is different). See my example. Oh, makes sense now! Sorry! Could you please just explain this segment: if let authToken = privateToken, baseURL = baseURL what does it do? It's checking if the things are in the keychain (if the user is logged in) or not. Hum, but this one will be permeant, I mean I want it to only ask for auth once and save it to the keychain, then of course, if the user wants to log out, the credential is deleted and the login screen is presented. Do I still need the if condition? Sure, because the keychain entry is optional. It's there for logged in users and its absent for logged out users. OK, thanks for the explanation! Closing the issue as solved.
gharchive/issue
2016-04-21T14:59:17
2025-04-01T04:35:11.262542
{ "authors": [ "pietbrauer", "tiferrei" ], "repo": "nerdishbynature/TanukiKit", "url": "https://github.com/nerdishbynature/TanukiKit/issues/7", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
111206567
when using rez-build --ba pass the string to bez as well like the title says. Currently this change would make all the previous bez build fail, as the arguments to bez's build function has one more argument. Also it would be up to the implementer of rezbuild.py to re-build arguments as they would get as a string, ie build_args="foo=1 bar=cat". Closing, merged with master.
gharchive/pull-request
2015-10-13T15:34:48
2025-04-01T04:35:11.264565
{ "authors": [ "holofermes", "nerdvegas" ], "repo": "nerdvegas/rez", "url": "https://github.com/nerdvegas/rez/pull/252", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2341097294
Is there a quick way of leveraging this codebase for CPU instead of GPU? I know this is a CUDA-accelerated library, but I am just wondering if I can easily migrate the code functionalities like fitting on images to the CPU implementation? It would be really helpful for my use-case. Thanks! Not really -- you would need to implement a few CPU kernels to make it work on CPU I believe you can use code from _torch_impl.py (https://github.com/nerfstudio-project/gsplat/blob/main/gsplat/cuda/_torch_impl.py) at the cost of compute time I believe you can use code from _torch_impl.py (https://github.com/nerfstudio-project/gsplat/blob/main/gsplat/cuda/_torch_impl.py) at the cost of compute time Yes most of the functions have its pure torch version with only two warnings that I would like to point out: _rasterize_to_pixels() still relies on CUDA: It relies on rasterize_to_indices_in_range() which is compiled from a customized cuda kernel, and it also relies on nerfacc which only works with CUDA. the pure torch implementation of _isect_tiles() is via a forloop so could be extremely slow when using it in an actual rendering process. Other than that the functions in _torch_impl.py should be fairly good (implemented in a proper batch way). They would certainly run slower tho but should be acceptable. Thanks a lot!
gharchive/issue
2024-06-07T20:13:22
2025-04-01T04:35:11.269099
{ "authors": [ "ChaosAdmStudent", "ivanpuhachov", "liruilong940607" ], "repo": "nerfstudio-project/gsplat", "url": "https://github.com/nerfstudio-project/gsplat/issues/205", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
649748437
Don't use [*]Names fields in data Currently, getting data related to module uploads is done like so: { package(name: "eggs") { packageUploadNames } } As you can see, if you want to get the uploads, one must query again using the names. "Names" should be removed, including for users with the packageNames fields and instead allow these fields to directly query the packages and package uploads. [x] Direct queries to packages have been added. [ ] Direct query to users added too. To summarise: There should not be a seperate query for packageUploads like the following: { packageUploads(name: "eggs@v0.1.0") {} } Instead: { packages(name: "eggs") { uploads(name: "0.1.0") {} } } The packageNames field on the packages object would be removed in favour of this. We should also support querying a certain amount of uploads and potentially pagination of them. This same process should be applied to the packageNames field on the user object as well. I might change the issue name and description to reflect this. @zorbyte https://github.com/nestdotland/api/commit/f606dc1109d90fe8a38ba4bfbda2f58f6c0fb4f5 has been merged. It adds: { packages(name: "eggs") { uploads(name: "0.1.0") { name } } } and { packages(name: "eggs") { uploads { name } } } And removes the global query for uploads @zorbyte Can we close this for now we have https://github.com/nestdotland/api/pull/32 merged.
gharchive/issue
2020-07-02T08:40:01
2025-04-01T04:35:11.286347
{ "authors": [ "divy-work", "nestdotland", "zorbyte" ], "repo": "nestdotland/api", "url": "https://github.com/nestdotland/api/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2243167085
Dashboard GUI - problem with saving changes Describe the problem Using google Chrome 123.0.6312.123 (latest) and netbird dashboard 2.2.0 (latest) There is problem saving any changes in gui, had to save it couple times. Same problem with adding security keys, have to repeat it few times. To Reproduce Steps to reproduce the behavior: Go to 'Setup Keys' Click on 'Create Setup Key' Fill in name 'TestKey' Click 'Create Setup Key' Click 'Copy to clipboard' Click 'Close' There is no key added, no error, nothing. Have to repeat steps 1 - 6 couple times to finally add it. Expected behavior Adding keys or saving changes successful every time. Netbird self hosted NetBird version netbird dashboard 2.2.0 @GandiKB can you check if there are any management error logs? you can see that with docker compose logs management It's related with browser itself, it works on Firefox. I have the same issue, the problem appears on Firefox and Chrome.
gharchive/issue
2024-04-15T09:25:44
2025-04-01T04:35:11.344955
{ "authors": [ "Enailis", "GandiKB", "mlsmaycon" ], "repo": "netbirdio/dashboard", "url": "https://github.com/netbirdio/dashboard/issues/371", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2333651107
Feature Request: Support for Always-On VPN Tunneling with Netbird on Android Dear GitHub Team, I hope this message finds you well. I would like to request a new feature for the Android platform that would significantly enhance the security and flexibility of VPN usage. The feature I'm proposing is the ability to have an always-on VPN connection while using Netbird Android client to tunnel through the my primary VPN. Feature Details: Always-On VPN: Ensure that the primary VPN connection is always active, providing continuous security and privacy protection. Secondary VPN Tunneling (Netbird): Allow users to establish a secondary VPN connection that tunnels through the always-on primary VPN. This feature would enable users to add an extra layer of encryption and route their traffic through multiple VPN servers. User Interface: Integrate a simple and intuitive user interface within the VPN settings on Android, allowing users to configure and manage both the primary and secondary VPN connections easily. Compatibility: Ensure compatibility with a wide range of VPN protocols and services to provide users with flexibility in choosing their preferred VPN providers. Performance Optimization: Implement performance optimizations to minimize any potential latency or bandwidth issues that might arise from using multiple VPN connections simultaneously. Use Case: This feature would be particularly beneficial for users who require an additional layer of security and privacy, such as journalists, activists, and users in regions with strict internet censorship. By allowing a secondary VPN to tunnel through the primary always-on VPN, users can enhance their online anonymity and protect their data from potential threats. Conclusion: Implementing this feature would provide a significant security enhancement for Android users and cater to those who need advanced VPN configurations. I believe this addition would make the Android platform more robust and secure for all users. Thank you for considering this feature request. I look forward to the possibility of seeing this implemented in future updates. Best regards, ~ Johnny Double VPNs mode will not be more security than single VPN mode. I'm more interested in routing traffic through Netbird without interrupting my primary VPN connection than in security. I understand that double VPN can offer additional layers of privacy, but it doesn't always translate to significantly greater security. For most users, a single, well-configured VPN connection from a reliable provider is sufficient for both security and performance.
gharchive/issue
2024-06-04T14:12:04
2025-04-01T04:35:11.349381
{ "authors": [ "johnny-IoT", "nxtreaming" ], "repo": "netbirdio/netbird", "url": "https://github.com/netbirdio/netbird/issues/2091", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1380008804
Script create_vm.py error when Cluster have site. Looks like when a selected cluster have a site the script returns an error. I'm running Netbox 3.3.2 with Python 3.10. Maybe it is something related to the newer Netbox version. I have fixed this by adding the following two lines to the end of class and def run: class NewVM(Script): class Meta: name = "New VM" description = "Create a new VM" vm_name = StringVar(label="VM name") .... site = ObjectVar(model=Site, required=True)` def run(self, data, commit): vm = VirtualMachine( name=data["vm_name"], .... site=data.get("site"), )
gharchive/issue
2022-09-20T21:20:05
2025-04-01T04:35:11.391474
{ "authors": [ "celeroncool", "joaolucasmacedo" ], "repo": "netbox-community/customizations", "url": "https://github.com/netbox-community/customizations/issues/70", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1384946183
Frontend: account page typo fix This PR fix an account page typo. /deploy
gharchive/pull-request
2022-09-25T08:36:06
2025-04-01T04:35:11.497809
{ "authors": [ "cheina97" ], "repo": "netgroup-polito/CrownLabs", "url": "https://github.com/netgroup-polito/CrownLabs/pull/813", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
394851687
Where do we edit the CSS? The admin page is really helpful but have not been successful finding the CSS responsible for the layout. The left panel shows the paragraphs nicely divided but when it appears on the actual site, it becomes a big lump of words. Thanks. @erquhart Hi, Thank for sharing this awesome theme. Currently, I'm using this theme for my portfolio. But somehow I need to change the size of the single poss / single page. Cause by the default it looks bit small, can you provide me about how to change the "width" of the single post is? There are a few custom styles here: https://github.com/netlify-templates/gatsby-starter-netlify-cms/blob/master/src/components/all.sass The blog post template itself is here: https://github.com/netlify-templates/gatsby-starter-netlify-cms/blob/master/src/templates/blog-post.js The template component uses classes from the Bulma framework: https://bulma.io/ Hope it helps! Sent with GitHawk
gharchive/issue
2018-12-30T07:04:09
2025-04-01T04:35:11.512518
{ "authors": [ "erquhart", "iColdPlayer", "swallowingspecialist" ], "repo": "netlify-templates/one-click-hugo-cms", "url": "https://github.com/netlify-templates/one-click-hugo-cms/issues/103", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1124704925
feat: Show edit graph link on dev startup, send operations library edits to UI, and run prettier (if available) on code output 🎉 Thanks for submitting a pull request! 🎉 Summary Fixes #4180 Makes editing the Graph operations library file locally work while also sending updates to the ui to make the DX much smoother via file monitoring. It also runs prettier (if available) on the output so the code conforms to the user's local expectations. Shows a link when starting netlify dev --graph that takes the user to the graph editing UI (requested by @biilmann) For us to review and ship your PR efficiently, please perform the following steps: [ ] Open a bug/issue before writing your code 🧑‍💻. This ensures we can discuss the changes and get feedback from everyone that should be involved. If you`re fixing a typo or something that`s on fire 🔥 (e.g. incident related), you can skip this step. [ ] Make sure the status checks below are successful ✅ Oh, great note, I'll fix that up tomorrow. Thank you! On Sat, Feb 5, 2022 at 8:05 AM ehmicky @.***> wrote: @.**** commented on this pull request. In src/lib/one-graph/cli-netlify-graph.js https://github.com/netlify/cli/pull/4183#discussion_r800071878: @@ -241,6 +241,31 @@ const ensureFunctionsPath = (netlifyGraphConfig) => { fs.mkdirSync(fullPath, { recursive: true }) } +let disablePrettierDueToPreviousError = false + +const runPrettier = async (filePath) => { if (disablePrettierDueToPreviousError) { return } const command = prettier --write ${filePath} try { const commandProcess = execa.command(command, { Small note: if filePath has a space in its path (which is common on Windows, e.g. C:\Program Files), it will be interpreted as two different arguments by execa.command(). Instead, using execa('prettier', ['--write', filePath]) will prevent this problem. — Reply to this email directly, view it on GitHub https://github.com/netlify/cli/pull/4183#pullrequestreview-873945184, or unsubscribe https://github.com/notifications/unsubscribe-auth/AAAITYCGE5MSYO22M5CVO4DUZVDELANCNFSM5NTBVPNQ . You are receiving this because you were assigned.Message ID: @.***>
gharchive/pull-request
2022-02-04T23:28:50
2025-04-01T04:35:11.521273
{ "authors": [ "sgrove" ], "repo": "netlify/cli", "url": "https://github.com/netlify/cli/pull/4183", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
424460306
Working on drafts locally? Is there a way to create or edit drafts locally? Basically, I would like to start working on a draft locally on my own machine, then make it available to NetlifyCMS so others can contribute. I saw that NetlifyCMS saves drafts in their own branch, but trying to create and push a draft branch manually myself doesn't seem to work. Is this part of the workflow documented anywhere? Editorial workflow entries are created as pull requests, but metadata in an orphan ref is us Sent with GitHawk Oh cool. Btw I'm sure you've considered this but what about having a status: draft field in your markdown file's frontmatter? This could coexist with the more complex GitHub branch workflow as a simpler way to work with drafts locally? @SachaG there's also netlify-cms-backend-fs which provides an easy way to work on files in the local filesystem when in development mode (although that still doesn't solve integration with the editorial workflow of course). Yeah I've considered the need for draft functionality as a separate entity from editorial workflow for a while now, less about a workflow and more about parity with built in draft functionality for SSG's like Jekyll and Hugo, which use a flag as you mentioned. Feature request here: https://github.com/netlify/netlify-cms/issues/942 This would be resolved by #2977, closing in favor of that.
gharchive/issue
2019-03-23T04:01:29
2025-04-01T04:35:11.525176
{ "authors": [ "SachaG", "bjrn", "erquhart" ], "repo": "netlify/netlify-cms", "url": "https://github.com/netlify/netlify-cms/issues/2236", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
315282391
Display public_folder path in Media Library - Summary It can be useful for content editors to see the path to uploaded media / files. I have added on the public_folder to the displayed filename. This may not be desired for all users, maybe we can add an option in the config? - Test plan Visual confirmation: - Description for the changelog Display public_folder path in Media Library - A picture of a cute animal (not mandatory but encouraged) Deploy preview for cms-demo ready! Built with commit 7961027681583c225d6923a5ce6dc9b4c71280cc https://deploy-preview-1281--cms-demo.netlify.com Deploy preview for netlify-cms-www ready! Built with commit 7961027681583c225d6923a5ce6dc9b4c71280cc https://deploy-preview-1281--netlify-cms-www.netlify.com Can you say the specific benefit(s) for you of seeing the path? Do you ever need to copy/paste it for some reason? Yeh, we've had clients who wish to link to the asset directly in the markdown editor. Sorry, kinda slow over here I guess - help me understand exactly what they're wanting to do. Link to the asset directly how? And to what end? No worries, I didn't explain it well. For example: User uploads a PDF file to the Media Library User wishes to link to this file in Markdown Editor: More info [/assets/uploads/file.pdf](here) Displaying the path in the Media Library makes it more obvious for non-devs to find the file's path. Is there a reason they can't just upload through an image block in MarkDown? @tech4him1 this was my first response as well 😅 We created a file block component for them (the image block only allows image files). However, editor blocks cannot be placed inline like this: This is a paragraph, the user wants to link to [filename.pdf](/assets/uploads/filename.pdf) inside this paragraph. If we could link to media using the Link button in the Markdown Editor, this would also solve the problem. That's a reasonable feature request. Sent with GitHawk @Jinksi I don't think we should go with this PR as the solution, but the problem is definitely worth solving. Can you convert this to an issue? @erquhart Will do. Agreed, the ability to add a dropdown with useful media meta-info would be great for certain client projects.
gharchive/pull-request
2018-04-17T23:59:57
2025-04-01T04:35:11.534209
{ "authors": [ "Jinksi", "erquhart", "tech4him1", "verythorough" ], "repo": "netlify/netlify-cms", "url": "https://github.com/netlify/netlify-cms/pull/1281", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
358115546
Update Docs “uploadcare” Automatically generated by Netlify CMS Deploy preview for cms-demo ready! Built with commit 3039334acbe438ad8b110da88370c8270657f0af https://deploy-preview-1728--cms-demo.netlify.com
gharchive/pull-request
2018-09-07T15:35:48
2025-04-01T04:35:11.536055
{ "authors": [ "erquhart", "verythorough" ], "repo": "netlify/netlify-cms", "url": "https://github.com/netlify/netlify-cms/pull/1728", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
280228696
Fix position of editor view controls These were so far off to the right tht they got cut off by the screen. This aligns them with the user icon - Summary - Test plan - Description for the changelog - A picture of a cute animal (not mandatory but encouraged) Deploy preview ready! Built with commit b7d3717fda8ee3ee3f89c70dc20455118cdbdef8 https://deploy-preview-889--netlify-cms-www.netlify.com
gharchive/pull-request
2017-12-07T18:16:53
2025-04-01T04:35:11.538916
{ "authors": [ "erquhart", "verythorough" ], "repo": "netlify/netlify-cms", "url": "https://github.com/netlify/netlify-cms/pull/889", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1760244712
Fix makefile image-build with podman @msherif1234 this is the same fix as in the operator. Also adding a 2>/dev/null on which to avoid unnecessary errors displayed repeatedly with podman /lgtm /approve
gharchive/pull-request
2023-06-16T09:14:59
2025-04-01T04:35:11.556416
{ "authors": [ "jotak", "msherif1234" ], "repo": "netobserv/network-observability-console-plugin", "url": "https://github.com/netobserv/network-observability-console-plugin/pull/345", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2032411378
Feature: Interface Pretty-Names Is your feature request related to a problem? Please describe. I'm using softflowd to export to goflow2 and use grafana-agent/grafana-promtail to process their data. To get a basic throughput metric we can add-up the flow-bytes per host, if-in and if-out. That looks like that: flow_throughput{application="netflow_data",low_if_in="0",flow_if_out="0",flow_vlan_in="0",hostname="xxx"} 149368 Example grafana-agent config Describe the solution you'd like What would be very nice to have is a interface pretty name instead of the generic interface-id we see. I would guess that would be easy to implement using the golang net.Interfaces function. The functionality could be triggered using the -format.selector flag. Describe alternatives you've considered Using rsyslog magic to script it.. but that would not be a very clean approach. duplicate.. sorry
gharchive/issue
2023-12-08T10:40:41
2025-04-01T04:35:11.577964
{ "authors": [ "superstes" ], "repo": "netsampler/goflow2", "url": "https://github.com/netsampler/goflow2/issues/255", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2027671199
transport: mute errors when too verbose This adds two CLI arguments that will by default reduce the amount of logs if there is a transport issue. Arguments: transport.err.int: interval to log errors transport.err.cnt: threshold of errors per interval If the threshold of errors during an interval is crossed, subsequent errors will be muted until the interval clears. One of the common case is when Kafka is down: ERRO[0028] transport error error="kafka transport kafka: Failed to produce message to topic flow-messages: circuit breaker is open" ERRO[0028] transport error error="kafka transport kafka: Failed to produce message to topic flow-messages: circuit breaker is open" ERRO[0028] transport error error="kafka transport kafka: Failed to produce message to topic flow-messages: circuit breaker is open" WARN[0029] too many transport errors ERRO[0037] transport error error="kafka transport kafka: Failed to produce message to topic flow-messages: circuit breaker is open" ERRO[0037] transport error error="kafka transport kafka: Failed to produce message to topic flow-messages: circuit breaker is open" ERRO[0038] transport error Will require updating https://github.com/netsampler/goflow2/pull/199 Idea: add "skipped x error messages" when counter is reset
gharchive/pull-request
2023-12-06T05:09:22
2025-04-01T04:35:11.580719
{ "authors": [ "lspgn" ], "repo": "netsampler/goflow2", "url": "https://github.com/netsampler/goflow2/pull/249", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2271981518
A suggestion about specifying a source of secure randomness for better security Hello everyone! I'm new to netty and when I was browsing the code I found the following code in JdkSslServerContext.java here: ctx.init(keyManagerFactory.getKeyManagers(), wrapTrustManagerIfNeeded(trustManagerFactory.getTrustManagers()), null); The code here use a "null", which means using the default securerandom. Although securerandom has already provided enough security, specifying one will ensure better security. For example, using SunJSSE.cryptoProvider for fip mode will be better. Looks valid. @normanmaurer WDYT? @hyperxpro @Gax-c I think we could add another constructor that take one ? Yup. Shall I go ahead with a PR? @hyperxpro sure Hi @normanmaurer, @hyperxpro, @Gax-c I'm interested in this so I created a PR. Could you please review this?
gharchive/issue
2024-04-30T16:29:12
2025-04-01T04:35:11.622471
{ "authors": [ "Gax-c", "hyperxpro", "normanmaurer", "thxwelchs" ], "repo": "netty/netty", "url": "https://github.com/netty/netty/issues/14026", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
46468192
If there are some Chinese characters in url, HttpRequest.getUri(), the return is messy code code example: Initializer: cp.addLast("httpReqDecoder", new HttpRequestDecoder()); cp.addLast("httpObjectAggregator", new HttpObjectAggregator(65536)); cp.addLast("httpResEncoder", new HttpResponseEncoder()); cp.addLast("chunkedWriteHandler", new ChunkedWriteHandler()); the request url in IE is: http://127.0.0.1:11299/someurl.jsp?keyword=中国汉字 the return result of HttpRequest.getUri() is /someurl.jsp?keyword=ᅱ￐ᄍᄎ츄ᅱ What can I do to resovle this problem? tips: IE does not encode the url, but chrome encode the url with UTF-8. for example in chrome: http://127.0.0.1:11299/someurl.jsp?keyword=中国汉字 the result is: /someurl.jsp?keyword=%E4%B8%AD%E5%9B%BD%E6%B1%89%E5%AD%97 @zhanggang807 这个主要是因为IE传输URL时,会采用ISO8859-1编码格式传输,需要在服务端进行一次转码才可以获取到正确的中文。Chrome会把URL先encoding一下,所以没关系
gharchive/issue
2014-10-22T03:03:38
2025-04-01T04:35:11.625915
{ "authors": [ "wlwinston", "zhanggang807" ], "repo": "netty/netty", "url": "https://github.com/netty/netty/issues/3035", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2232943602
Fix voidPromise in Http2FrameCodec.writeHeadersFrame Motivation: Writing a Http2HeadersFrame on a new stream with a void promise would throw an IllegalStateException. Modification: Unvoid the promise in the code that requires adding a listener. Result: No IllegalStateException. Thanks!
gharchive/pull-request
2024-04-09T08:55:37
2025-04-01T04:35:11.627575
{ "authors": [ "chrisvest", "yawkat" ], "repo": "netty/netty", "url": "https://github.com/netty/netty/pull/13958", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
310008638
Using dependency injection with annotations While looking at examples, I saw some references of DI, but I couldn't get constructor injection working by default. Is there a way to do DI either spring-boot way or guice or dagger framework. Some reference would be nice. @rohit-zabbed We have evaluated all existing DI frameworks before building something new. Most of our customers are using Docker container to deliver their services to the cloud environment and they need to change the injections in certain environments with just configuration changes in service.yml file. With annotations, you have to change the source code and repackage the application. The following is a link to several tutorials on how to use the service module in light-4j. https://www.networknt.com/tutorial/common/service/
gharchive/issue
2018-03-30T08:29:50
2025-04-01T04:35:11.632200
{ "authors": [ "rohit-zabbed", "stevehu" ], "repo": "networknt/light-4j", "url": "https://github.com/networknt/light-4j/issues/182", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1128462449
Fix examples/features/jaeger example Description Examples/features/jaeger example works with the old version of NSM, which supports tracing via OpenTracing. The current version of NSM supports tracing only via Opentelemetry Collector. @szvincze Be free to ping us if the problem is still actual for you
gharchive/issue
2022-02-09T12:12:54
2025-04-01T04:35:11.633356
{ "authors": [ "NikitaSkrynnik", "denis-tingaikin" ], "repo": "networkservicemesh/deployments-k8s", "url": "https://github.com/networkservicemesh/deployments-k8s/issues/4622", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2287607379
Is the author considering when to release the evaluation code? Is the author considering when to release the evaluation code? Thanks for your attention, we're going to release to the evaluation code this month. Thanks for your attention, we're going to release to the evaluation code this month. Is the author considering when to release the evaluation code?
gharchive/issue
2024-05-09T12:36:37
2025-04-01T04:35:11.648749
{ "authors": [ "wuxiaofei01", "xiaogangpeng" ], "repo": "neu-vi/HOI-Diff", "url": "https://github.com/neu-vi/HOI-Diff/issues/18", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1637657351
Add support for java.util.logging I tend to use java.util.logging for logging. Could you add its support ? I never used java.util.logging. I guess, that makes me to a bad person to implement it. :) But I like the idea and would be happy to check and changes. The PR above will use the existing interface of ExpectedLoggingAssert, i.e. it uses hasErrorMessage when checking for log messages with level "SEVERE", because j.u.l. has no error level. To stay withing the j.u.l. domain we could add hasSevereMessage with the same functionality as hasErrorMessage. Using hasErrorMessage to check for a severe message could be confusing to the test reader. Since I don't use java.util.logging either, @Riduidel , could you help here and say what you would prefer? Bonus question: Would you remove hasErrorMessage in the assertj-logging-jul artefact? Of course this goes for all the other methods that handle error/severe levels, too.
gharchive/issue
2023-03-23T14:27:18
2025-04-01T04:35:11.659317
{ "authors": [ "Riduidel", "achimgrimm", "akutschera" ], "repo": "neuland/assertj-logging", "url": "https://github.com/neuland/assertj-logging/issues/6", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
266095092
Getting Exception while Parsing jade template While Parse the jade template. I got the following errors. expected class de.neuland.jade4j.lexer.token.Indent but got class de.neuland.jade4j.lexer.token.Newline How could i resolve it? Hard to say without the template. Have you tried narrowing it down?
gharchive/issue
2017-10-17T11:51:34
2025-04-01T04:35:11.660601
{ "authors": [ "LamaBimal", "naltatis" ], "repo": "neuland/jade4j", "url": "https://github.com/neuland/jade4j/issues/163", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
145530120
GPU memory leak Probably not clearing old textures Appears to have been fixed in https://github.com/neurodata/NeuroDataViz/commit/d4837db99e8f4ff3f5ff3bd81674988a8b0e5c1d
gharchive/issue
2016-04-03T19:02:53
2025-04-01T04:35:11.679545
{ "authors": [ "alexbaden" ], "repo": "neurodata/NeuroDataViz", "url": "https://github.com/neurodata/NeuroDataViz/issues/91", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
316974015
Phase Enconding direction issue Hello, I am working to implement distortion correction in my MRI analysis for epi data using fmriprep. Unfortunately, for all my combinations of i, i-, j and j-, I get an error as follows: raise Exception("None of the discovered fieldmaps has the right " Exception: None of the discovered fieldmaps has the right phase encoding direction. Possibly a problem with metadata. If not, rerun with '--ignore fieldmaps' to skip distortion correction step. Any help will be greatly appreciated. Thank you. Hi Jahanikia, with ndmg, we don't use distortion correction, as only a select number of datasets provide the required metadata to include this in analysis (you need some baseline files for adding distortion correction). I would recommend asking Chris over at fmriprep for help regarding distortion correction, or just avoiding it entirely (--ignore fieldmaps option you noted in their pipeline). Let us know if you need any help with ndmg. As this question does not have to do with ndmg, I'm going to mark the issue as "closed", but feel free to post follow ups you might have.
gharchive/issue
2018-04-23T20:38:05
2025-04-01T04:35:11.682229
{ "authors": [ "JahanikiaSahar", "ebridge2" ], "repo": "neurodata/ndmg", "url": "https://github.com/neurodata/ndmg/issues/162", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2618111852
Add wheel as dependency to check-manifest Problem With a Python 3.12 system (base?) installation, pre-commit run check-manifest returns an error of missing dependencies. According to this issue this is due to a change in behaviour in Python 3.12, in which apparently wheel is no longer preinstalled with virtualenv / venv. See release notes too. Proposed solution Option 1: adding wheel as an additional dependency in .pre-commit-config.yaml. This is implemented in this PR. - repo: https://github.com/mgedmin/check-manifest rev: "0.49" hooks: - id: check-manifest args: [--no-build-isolation] additional_dependencies: [setuptools-scm, wheel] Option 2: specifying a language version for Python in check-manifest. Feels less flexible. - repo: https://github.com/mgedmin/check-manifest rev: "0.49" hooks: - id: check-manifest args: [--no-build-isolation] language_version: "3.12" additional_dependencies: [setuptools-scm] Question In CI, we run check-manifest with Python 3.10 and without the --no-build-isolation flag. Do we need to make CI and the pre-commit hook more similar? Why do we run the --no-build-isolation flag in the pre-commits? Option 1 sounds sensible to me
gharchive/pull-request
2024-10-28T11:15:15
2025-04-01T04:35:11.698971
{ "authors": [ "adamltyson", "sfmig" ], "repo": "neuroinformatics-unit/python-cookiecutter", "url": "https://github.com/neuroinformatics-unit/python-cookiecutter/pull/126", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
242769749
Display in radiological convention. The current software always shows images using neurological convention (left anatomy on left of display). Request feature to optionally display in radiological convention, left anatomy on right of display. Mirror Coronal and axial slices, but not rendering. Ensure drawings are loaded/saved correctly. Included in latest release.
gharchive/issue
2017-07-13T17:08:41
2025-04-01T04:35:11.700551
{ "authors": [ "neurolabusc" ], "repo": "neurolabusc/MRIcroGL", "url": "https://github.com/neurolabusc/MRIcroGL/issues/12", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
221316418
add optic [ ] create wrapper for this binary: sct_detect_spinalcord [x] replace centerline detection with optic in propseg (Ben) [ ] add tests for binaries, test for different OS distribs [ ] package binary (done for osx) Model generated for each contrast (t1, t2, t2s and dmri). Available under /Public_JCA/temp/charley/linux/models/ I've create the branch issue_1249. In this branch, I integrated OptiC into PropSeg, without the use of a wrapper. It is a first step for creating a wrapper for OptiC that would be called by PropSeg. Link to PropSeg in new branch: https://github.com/neuropoly/spinalcordtoolbox/blob/issue1249/scripts/sct_propseg.py MacOs Binaries available under /Public_JCA/temp/charley/mac/10_7/ and /Public_JCA/temp/charley/mac/10_11/ does version 10_7 work for all >10.7 osx? if so, any advantage in copying 10_11? i've updated uploaded binaries on OSF and updated the branch for the install part. Please update your local installation to reflect changes: ./install_sct -m I have tested 10_7 on my machine (El Capitan, 10.11.5): it works well! I do not see advantage to use 10_11 Unix binary of isct_spine_detect seems to use FSL: django:t1 bedelb$ isct_spine_detect -ctype=dpdt -lambda=1.0 "/Users/bedelb/code/spinalcordtoolbox/data/models/t1_model.yml" "t1_int16_RPI_imghdr" "t1_int16_RPI_imghdr_optic" ERROR:: Environment variable FSLOUTPUTTYPE is not set! Please make sure that the appropriate configuration file is sourced by your shell (e.g. by putting it in .profile). e.g. bash or sh users add the line ". ${FSLDIR}/etc/fslconf/fsl.sh" e.g. tcsh or csh users add the line "source ${FSLDIR}/etc/fslconf/fsl.csh" Here are the results for sct_testing/large: T1-weighted Command: "sct_propseg -i t1/t1.nii.gz -c t1 -init-centerline optic Dataset: /Volumes/Public_JCA/sct_testing/large/ GLOBAL RESULTS: Duration: 1483s Passed: 75/116 Crashed: 0/116 Mean: {'dice_segmentation': 0.8661204544467307, 'duration [s]': 81.44691399048115} STD: {'dice_segmentation': 0.18441609014819368, 'duration [s]': 30.866210507926827} Command: "sct_propseg -i t1/t1.nii.gz -c t1 Dataset: /Volumes/Public_JCA/sct_testing/large/ GLOBAL RESULTS: Duration: 1066s Passed: 24/116 Crashed: 1/116 Mean: {'dice_segmentation': 0.6692521998875399, 'duration [s]': 52.99569157074238} STD: {'dice_segmentation': 0.30516104566549374, 'duration [s]': 32.57508440274568} T2-weighted Command: "sct_propseg -i t2/t2.nii.gz -c t2 -init-centerline optic Dataset: /Volumes/Public_JCA/sct_testing/large/ GLOBAL RESULTS: Duration: 1558s Passed: 183/256 Crashed: 0/256 Mean: {'dice_segmentation': 0.8965265611318537, 'duration [s]': 41.221134138293564} STD: {'dice_segmentation': 0.15291260300075263, 'duration [s]': 13.274626862061824} Command: "sct_propseg -i t2/t2.nii.gz -c t2 Dataset: /Volumes/Public_JCA/sct_testing/large/ GLOBAL RESULTS: Duration: 1549s Passed: 118/256 Crashed: 36/256 Mean: {'dice_segmentation': 0.7161338428029882, 'duration [s]': 39.81973373796791} STD: {'dice_segmentation': 0.3738910200186539, 'duration [s]': 18.235122280997768} T2*-weighted Command: "sct_propseg -i t2s/t2s.nii.gz -c t2s -init-centerline optic Dataset: /Volumes/Public_JCA/sct_testing/large/ GLOBAL RESULTS: Duration: 1374s Passed: 132/267 Crashed: 0/267 Mean: {'dice_segmentation': 0.8312970396151269, 'duration [s]': 32.0688976955771} STD: {'dice_segmentation': 0.1911104445470129, 'duration [s]': 11.959219802345736} Command: "sct_propseg -i t2s/t2s.nii.gz -c t2s Dataset: /Volumes/Public_JCA/sct_testing/large/ GLOBAL RESULTS: COMING SOON I have enabled OptiC by default for PropSeg. To use old detection method, use option -init-centerline hough (should we hide it from the documentation or display it, knowing it is less robust?) We should also replace the documentation concerning the contrast: type of image contrast, t2: cord dark / CSF bright ; t1: cord bright / CSF dark Proposition: type of image contrast, if your contrast is not in the available options (t1, t2, t2s, dmri), use t1 (cord bright / CSF dark) or t2 (cord dark / CSF bright) Je remplacerais dmri par dwi, car des gens pourraient penser qu'il sagit de la b=0 (qui est en fait T2w) --Julien -------- Original message --------From: Benjamin De Leener notifications@github.com Date: 2017-04-15 12:04 PM (GMT-05:00) To: neuropoly/spinalcordtoolbox spinalcordtoolbox@noreply.github.com Cc: Julien Cohen-Adad jcohen@polymtl.ca, Assign assign@noreply.github.com Subject: Re: [neuropoly/spinalcordtoolbox] add optic (#1249) I have enabled OptiC by default for PropSeg. To use old detection method, use option -init-centerline hough (should we hide it from the documentation or display it, knowing it is less robust?) We should also replace the documentation concerning the contrast: type of image contrast, t2: cord dark / CSF bright ; t1: cord bright / CSF dark Proposition: type of image contrast, if your contrast is not in the available options (t1, t2, t2s, dmri), use t1 (cord bright / CSF dark) or t2 (cord dark / CSF bright) — You are receiving this because you were assigned. Reply to this email directly, view it on GitHub, or mute the thread. {"api_version":"1.0","publisher":{"api_key":"05dde50f1d1a384dd78767c55493e4bb","name":"GitHub"},"entity":{"external_key":"github/neuropoly/spinalcordtoolbox","title":"neuropoly/spinalcordtoolbox","subtitle":"GitHub repository","main_image_url":"https://cloud.githubusercontent.com/assets/143418/17495839/a5054eac-5d88-11e6-95fc-7290892c7bb5.png","avatar_image_url":"https://cloud.githubusercontent.com/assets/143418/15842166/7c72db34-2c0b-11e6-9aed-b52498112777.png","action":{"name":"Open in GitHub","url":"https://github.com/neuropoly/spinalcordtoolbox"}},"updates":{"snippets":[{"icon":"PERSON","message":"@benjamindeleener in #1249: I have enabled OptiC by default for PropSeg. To use old detection method, use option -init-centerline hough (should we hide it from the documentation or display it, knowing it is less robust?)\r\n\r\nWe should also replace the documentation concerning the contrast:\r\n\r\ntype of image contrast, t2: cord dark / CSF bright ; t1: cord bright / CSF dark\r\n\r\n\r\nProposition:\r\n\r\ntype of image contrast, if your contrast is not in the available options (t1, t2, t2s, dmri), use t1 (cord bright / CSF dark) or t2 (cord dark / CSF bright)\r\n"}],"action":{"name":"View Issue","url":"https://github.com/neuropoly/spinalcordtoolbox/issues/1249#issuecomment-294302181"}}} Bonne idée. Il faudrait changer le nom du modèle dans data/ aussi. effectivement, je m'en occupe Ca a l'air de crasher sur Travis, et je ne sais pas trop pourquoi: https://travis-ci.org/neuropoly/spinalcordtoolbox/jobs/222376744#L514 Une idée? sct_testing fonctionne bien chez moi. les binaires linux ne sont pas encore sur osf ;-) je fais ca asap Cool. Par contre ca fait la même erreur sur Travis/osx ah mince... il faudrait faire un printout de l'erreur @jcohenadad Oops (on Travis: https://travis-ci.org/neuropoly/spinalcordtoolbox/jobs/222478501#L655) Checking test_sct_propseg........................... OUTPUT: -- sct_propseg -v 1 -i /home/travis/build/neuropoly/spinalcordtoolbox/sct_testing_data/t2/t2.nii.gz -c t2 -r 1 -ofolder sct_propseg_sct_testing_data_170416025213_413774/ Running /home/travis/build/neuropoly/spinalcordtoolbox/scripts/sct_propseg.py -v 1 -i /home/travis/build/neuropoly/spinalcordtoolbox/sct_testing_data/t2/t2.nii.gz -c t2 -r 1 -ofolder sct_propseg_sct_testing_data_170416025213_413774/ Check folder existence... Detecting the spinal cord using OptiC /bin/sh: 1: isct_spine_detect: not found Works on OSX now. Works on linux (debian) as well. I removed some todo: add optic test --> already tested via propseg add wrapper --> not urgent. Running propseg already output the centerline
gharchive/issue
2017-04-12T16:18:50
2025-04-01T04:35:11.739933
{ "authors": [ "benjamindeleener", "charleygros", "jcohenadad" ], "repo": "neuropoly/spinalcordtoolbox", "url": "https://github.com/neuropoly/spinalcordtoolbox/issues/1249", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1206548812
如何对mysql进行某字段的检索呢? 是否要将数据库导出写入文件才行? 目前只支持全文检索,就是text字段,如果你想检索mysql的字段,可以把多个字段的文本内容拼接到一起,然后放到text字段中,这样就可以被搜索到了。
gharchive/issue
2022-04-18T01:17:23
2025-04-01T04:35:11.771497
{ "authors": [ "cute-angelia", "newpanjing" ], "repo": "newpanjing/gofound", "url": "https://github.com/newpanjing/gofound/issues/1", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
747842833
newrelic-infrastructure: needs affinity values add .Values.affinity to the Daemonset's spec.template.spec.affinity Is your feature request related to a problem? Please describe. We need to prevent this from getting deployed to fargate on eks and the most sane way is an affinity rule to avoid fargate's lavel Describe the solution you'd like Add {{- if .Values.affinity }} affinity: {{ toYaml .Values.affinity | indent 8 }} {{- end }} to the Daemonset's template like https://github.com/helm/charts/blob/90590531b829dbc1142833a5d124cef7cc17b893/stable/prometheus-node-exporter/templates/daemonset.yaml#L80 Describe alternatives you've considered We considered adding a standard label to all our nodes which wouldn't be applied to the fargate instance which we could use with the newrelic-infrastructure's nodeSelector but that would require a redeploy of all of our clusters and doesn't make much sense to have a node label that applies to all nodes (except fargate) We need this functionality too. We use newrelic to monitor only specific pods, and we do know their location. It's kind of overhead to deploy the daemonset on all the nodes. We at TableCheck Japan use the newrelic-infrastructure chart extensively, and want to expand our usage to our mixed AWS EKS Managed Node Groups & Fargate cluster. Please let me know if you want me to open a new MR and sign the CLA (as I see this is unsigned @freehck-codefresh). H there! Thanks for raising your concerns. Jut wanted to let you know that we're working on EKS fargate support now. As a part of this effort, we're also adding the ability to add custom nodeAffinity items to the daemonSet :) This feature was implemented in #348. Definiing global.fargate will now automatically add affinity rules to exclude the DaemonSet from fargate nodes. in the future, this label will also enable other features for instrumenting EKS fargate, in which we're still working. Additionally, this PR also allwos to manually define node affinity rules. However please keep in mind that if these rules are manually overriden, global.fargate will no longer add the fargate exclusion automatically. @roobre Thank you so much, we’re going to use this immediately. Keep up the great work and we 💜 New Relic.
gharchive/issue
2020-11-20T22:47:36
2025-04-01T04:35:11.792427
{ "authors": [ "DragonStuff", "cweidinger", "freehck-codefresh", "roobre" ], "repo": "newrelic/helm-charts", "url": "https://github.com/newrelic/helm-charts/issues/210", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
702290916
Events → BatchMode → Flush doesn't block and wait Description The documentation for the Events.Flush method says: Flush gives the user a way to manually flush the queue in the foreground. This is also used by watchdog when the timer expires. I expect foreground to be synonymous with block and wait, and the client doesn't appear to be doing that. This is the culmination of research that was originally discussed in #464. Go Version go version go1.15.2 darwin/amd64 Current behavior Reading thousands of records from one system, and pushing them into New Relic via pkg/events in batch mode. Code is very similar to what was posted in #464, but essentially boils down to something like this: (I've stripped out some event handling and logging code, so there may be some phantom variables here.) func main() { ctx := context.Background() cfg := config.New() cfg.InsightsInsertKey = os.Getenv("NEWRELIC_INSIGHTS_INSERT") cfg.Compression = config.Compression.Gzip cfg.LogLevel = "trace" client := events.New(cfg) logger := cfg.GetLogger() // Start batch events mode if err := client.BatchMode( ctx, nrAccountID, events.BatchConfigTimeout(batchTimeout), ); err != nil { log.Fatalln("error starting batch mode:", err) } // Spawn goroutines... wgRawData.Add(1) go profileRaw(rawChannel, &client, &cfg) // Wait for goroutines to return... go func() { wgRawData.Wait() close(rawChannel) }() for rawData := range rawChannel { spew.Dump(rawData) } // Force-flush the events when we're done with this batch of data so that // the function ends and returns instead of listening indefinitely. if err := client.Flush(); err != nil { logger.Error(fmt.Sprintf("error flushing event queue: %s", err.Error())) } logger.Info("==> Events instructed to flush.") } // Leverages Go channels to execute the processing concurrently. func profileRaw(c chan Response, client *events.Events, cfg *config.Config) { // Prepare to close the channel defer wgRawData.Done() ctx := context.Background() logger := cfg.GetLogger() logger.Info("==> Kicking off the profiler...") // Run the profiler output, ok := profiler(profilerDuration) if !ok { logger.Error("Profiler function failed.") c <- resp } logger.Info("==> Parsing profiler data...") payloads := strings.Split(output, "\n\n") logger.Info("==> Sending data chunks to New Relic as events...") for i := range payloads { payload := payloads[i] // Read the JSON into the appropriate Go struct err = json.Unmarshal([]byte(payload), &event) if err != nil { logger.Error(fmt.Sprintf("[] error unmarshaling JSON into struct: %s %s", err.Error(), payload)) break } // ...do stuff to transform raw profiler event data into a flattened // structure that the New Relic Events API wants... // Queue a custom event if err := client.EnqueueEvent(ctx, flatEvent); err != nil { logger.Error(fmt.Sprintf("error posting custom event: %s", err.Error())) } } logger.Info("==> Sending events completed.") // Return data to the channel, closing it c <- resp } What happens is that client.Flush() isn't preventing the program from ending, so main() exits before the queue flushes (or, more accurately, even starts). Expected behavior My expectation is that client.Flush() should block and wait — preventing main() from exiting until the flushed events have responded successfully (or at least have been sent in the first place). Steps To Reproduce Create a new events client. Put it into batch mode. Set it to flush every second. Enqueue a few thousand event messages. Trigger a flush. Program ends. Additional Context I was seeing cases where sometimes events made it, and other times they did not. The difference appeared to be related to which log level I was setting on the client. Once I narrowed-down the variables in the puzzle, I discovered that it wasn't the log level so much as the time it took for those log levels to write data to stdout/stderr. debug and trace wrote a lot more data to stdout/stderr, which slowed-down the program enough to allow the goroutines to complete. info wrote less data, enabling the program to run faster, and the goroutines didn't have enough time to finish (or even start). A workaround has been to sleep at the end of the main() function. time.Sleep(5 * time.Second) But what this means is that it always ends up sleeping either too long or not long enough. Too long means that I'm missing out on other live events while I'm waiting for the program to finish sleeping before listening again. Too short means that some events I've read aren't making their way to the Events API. Ideally, the client should intrinsically understand what it needs to block for, then stop blocking once the queue flushing has completed (for whatever definition of "completed" is appropriate — request has been sent, or response has been received). References or Related Issues #464 Thank you for raising the issue. It looks to me what we need is a new blocking Shutdown() method that will wait until all the messages are sent before returning, and prevent new messages from entering the buffer. Looks like the Flush() method only write a control message to the channel, which triggers the listener routine to perform the actual sending. But this is not quite right either, since the grabAndConsumeEvents() method also spawns a go routine to perform the actual sending. The batchWorker is maintaining some state about the count, which I'm not sure how to handle if the flusher logic were moved out. Too, a Flush() doesn't stop the execution, so even while a flush is happening, it looks like more events could be written to the buffer, which is probably less than desirable. So perhaps if we took the route of writing a new Shutdown() method, this could terminate the context, but that might not be right since the context is passed to batch worker from the caller. Perhaps a combination of sync.Wait usage within grabAndConsumeEvents and copying the flush logic into a shutdown method could help. I wonder too, if we might be able to leverage the context object to close down the rest of the system while the post completes. I don't expect to be able to look into this in the near term though. @skyzyx This looks like a documentation issue as Flush was meant to allow a lever to manually initiate sending data, but not block/wait. (The initial use-cases for this were all services, and not expected to shutdown). Adding logic as @zlesnr suggests to a new func seems reasonable. Bump. Hi @skyzyx Were you able to work around this or is it still an active issue?
gharchive/issue
2020-09-15T21:36:28
2025-04-01T04:35:11.804677
{ "authors": [ "jthurman42", "kidk", "skyzyx", "zlesnr" ], "repo": "newrelic/newrelic-client-go", "url": "https://github.com/newrelic/newrelic-client-go/issues/476", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1741469864
[Product Partnerships] Added AWS Transcribe Entity Summary A concise description of the changes being introduced. Please review the pre-merge checklist section to validate this pull request is ready for review and merge. If it is not ready, please mark the pull request as a draft. The owners of this repo are not experts in the subject matter of the quickstarts. We review for the quickstart to be functional and for security risks. If you are seeking feedback on the content of the quickstart, please seek out a subject matter expert. If you are not an internal NR contributor, we can do our best to connect you with a content reviewer. Pre merge checklist [X] Did you check you NRQL syntax? - Does it work? [X] Did you include a Data source and Documentation reference? [X] Are all documentation links publicly accessible? [X] Did you check your descriptive content for voice and tone? [X] Did you check your descriptive content for spelling and grammar errors? [X] Did you review your content with a subject matter expert? (e.g. a Browser agent quickstart is reviewed with a member of the Browser Agent team) Dashboards [X] Does the PR contain a screenshot for each of your dashboards? [X] Do your screenshots show data? [X] Has the sanitization script been run on each dashboard? Alerts [X] Did you check that your alerts actually work? @brammerl , can you please merge the PR
gharchive/pull-request
2023-06-05T10:26:39
2025-04-01T04:35:11.812210
{ "authors": [ "rahul188" ], "repo": "newrelic/newrelic-quickstarts", "url": "https://github.com/newrelic/newrelic-quickstarts/pull/1826", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
567874305
chore: update README content Updates README content to match App Catalog and developer website descriptions. Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it. :tada: This PR is included in version 1.1.0 :tada: The release is available on GitHub release Your semantic-release bot :package::rocket:
gharchive/pull-request
2020-02-19T21:59:58
2025-04-01T04:35:11.815550
{ "authors": [ "claassistantio", "devfreddy", "jbeveland27" ], "repo": "newrelic/nr1-browser-analyzer", "url": "https://github.com/newrelic/nr1-browser-analyzer/pull/22", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1299701219
Add support for images in enclosure tags in Atom Newsboat version (copy the output of newsboat -v or the first line of git show): Newsboat 2.28.0 - https://newsboat.org/ Copyright (C) 2006-2015 Andreas Krennmair Copyright (C) 2015-2022 Alexander Batischev Copyright (C) 2006-2017 Newsbeuter contributors Copyright (C) 2017-2022 Newsboat contributors Newsboat is free software licensed under the MIT License. (Type `newsboat -vv' to see the full text.) It bundles: - JSON for Modern C++ library, licensed under the MIT License: https://github.com/nlohmann/json - optional-lite library, licensed under the Boost Software License: https://github.com/martinmoene/optional-lite - expected-lite library, licensed under the Boost Software License: https://github.com/martinmoene/expected-lite Newsboat 2.28.0 System: Linux 5.18.9-arch1-1 (x86_64) Compiler: g++ 12.1.0 ncurses: ncurses 6.3.20211021 (compiled with 6.3) libcurl: libcurl/7.84.0 OpenSSL/1.1.1q zlib/1.2.12 brotli/1.0.9 zstd/1.5.2 libidn2/2.3.2 libpsl/0.21.1 (+libidn2/2.3.0) libssh2/1.10.0 nghttp2/1.48.0 (compiled with 7.83.1) SQLite: 3.39.0 (compiled with 3.38.5) libxml2: compiled with 2.9.14 Config file (copy from ~/.newsboat/config or ~/.config/newsboat/config): # Key bindings unbind-key J unbind-key K unbind-key UP unbind-key DOWN unbind-key g unbind-key G unbind-key HOME unbind-key END bind-key j next bind-key k prev bind-key j down bind-key k up bind-key g home bind-key G end # Formats articlelist-format "%4i %n %D %t" feedlist-format "%4i %n %11u %t" # Dialog title formats articlelist-title-format "Articles in feed '%T' - unread: %u/%t" feedlist-title-format "Feeds - unread: %u/%t" help-title-format "Help" itemview-title-format "Article '%T'" # Miscellaneous feed-sort-order title browser "mpv %u > /dev/null 2>&1 &" datetime-format "%a %d-%m-%Y %H:%M" download-retries 3 download-timeout 60 error-log "~/.local/share/newsboat/error.log" goto-next-feed no refresh-on-startup yes show-keymap-hint no reload-threads 100 # Notifications notify-program "/bin/notify-send" notify-format "Newsboat: %d new articles loaded" # Colors color listfocus default default standout color listnormal_unread default default bold color listfocus_unread default default standout bold color title default default standout bold color end-of-text-marker color240 default highlight article "https?://[^ ]+" default default bold # Filters ignore-mode "display" ignore-article "https://www.youtube.com/feeds/videos.xml?channel_id=UC2N0bFREZDPYg5H-t7LKgnA" "title =~ \".*Creepypasta.*\"" ignore-article "https://www.youtube.com/feeds/videos.xml?channel_id=UCmHoLDhqgvDaXWm8w1ZJzZw" "title =~ \".*#shorts.*\"" ignore-article "https://www.youtube.com/feeds/videos.xml?channel_id=UCKoFUWEIFF33-92tBA0Cvzg" "title =~ \".*#Shorts.*\"" ignore-article "https://www.youtube.com/feeds/videos.xml?channel_id=UC4V3oCikXeSqYQr0hBMARwg" "title =~ \".*LIVE\ INTEGRALE.*\"" Steps to reproduce the issue: Add this feed: http://0x0.st/oQ7F.xml Open any article As you can see at the bottom there is the list of URLs embedded in the HTML, but there aren't those included in the link tag, which all point to images. Example: <link rel="enclosure" type="image/jpeg" href="https://cdn.hytale.com/variants/blog_cover_03f4f83af5ca1255ca51d8fa767d5be1_winter_2021_development_update_header.jpg"/> I've taken a look at #1615 but it's not clear to me if support has been added, and if so if it's only for RSS or Atom too. My request is to add support for images, perhaps listing the URLs at the end of the article as suggested in that issue. Other info you think is relevant: Hey, thanks for the report! No support was added yet, which is why #1615 is still open. I'll close this as a duplicate then. Please subscribe to that other issue for updates ;)
gharchive/issue
2022-07-09T14:59:45
2025-04-01T04:35:11.824256
{ "authors": [ "Minoru", "llamasblade" ], "repo": "newsboat/newsboat", "url": "https://github.com/newsboat/newsboat/issues/2120", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1718454166
Exception: videoQuality cannot be empty (again) #82 have been fixed for Vimeo, but for YouTube live streams the bug is still there. For example, with pod_player 0.1.2 for this stream: https://youtu.be/86YLFOog4GM I got: E/flutter (26777): [ERROR:flutter/runtime/dart_vm_initializer.cc(41)] Unhandled Exception: Exception: videoQuality cannot be empty E/flutter (26777): #0 PodPlayerController._checkAndWaitTillInitialized (package:pod_player/src/controllers/pod_player_controller.dart:73:9) E/flutter (26777): #1 PodPlayerController._checkAndWaitTillInitialized (package:pod_player/src/controllers/pod_player_controller.dart:82:11) E/flutter (26777): <asynchronous suspension> E/flutter (26777): #2 PodPlayerController._checkAndWaitTillInitialized (package:pod_player/src/controllers/pod_player_controller.dart:82:5) E/flutter (26777): <asynchronous suspension> E/flutter (26777): #3 PodPlayerController.initialise (package:pod_player/src/controllers/pod_player_controller.dart:61:5) E/flutter (26777): <asynchronous suspension> E/flutter (26777): To play live videos you should pass live = true To play live videos you should pass live = true My fault, thanks, it helped.
gharchive/issue
2023-05-21T10:37:40
2025-04-01T04:35:11.830991
{ "authors": [ "newtaDev", "xurc" ], "repo": "newtaDev/pod_player", "url": "https://github.com/newtaDev/pod_player/issues/113", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1963760468
Mystery Text Discussion: ww.txt Post your screenshots and discuss your findings about ww.txt here! I chose this science-fiction novel for this text analysis assignment, which is about martians. When I analyzed this text using Voyant, I was surprised that "martians" was the second-most mentioned term in the entire text; "said" was mentioned the most! Then, I looked at the relationships with the terms "said" and "martians" in the Termsberry, where I observed that more related terms show up when I hover to "martians" than when I hovered to "said". Using AntConc, on the other hand, I have found that the NGram size of 3 gives me a lot of words with a frequency count in double digits, whereas an NGram size of 4 gives me many words with a frequency count of more than 5. With an NGram size of 5, "the edge of the pit" is mentioned the most, in which it was written 8 times. Using the KWIC feature in AntConc, I experimented with the terms "the martians", in which I have found 137 phrases that include these terms, most of these sentences that were written about the martians have just done as the story progressed. After looking through the various texts we were given, I decided to analyze ww.txt. And sorry if this is cheating but after reading through this I am fairly certain this is H.G. Wells War of the Worlds. However, even though I knew what this story was, I tried my best to make sure that my knowledge did not interfere with my analysis. I'll start by showing the simplest thing, being the "word cloud" generated by Voyant. When creating the word cloud I made sure that it excluded the obvious words like "the" and "and", because I didn't think that they would show me anything interesting. When looking at the word cloud the most noticeable words are "martian", "came", and "people". What's interesting about this is that it puts man and martian on equal footing. For a story about aliens visiting Earth, the aliens get just as much attention as the people panicking on the planet. In fact when setting the N-Gram size to 1, the word "martians" was only used 5 more times total compared to the word "people". When setting the N-Gram size to something larger, like 3, the most used phrase becomes "out of the" with a whopping 53 uses! When looking into their uses, the most common use of the phrase "out of the" is not in reference to aliens leaving their ships, but instead people panicking to get away from said aliens or foreign objects. Overall I think what I found most interesting from this reading was the focus on words like "people" and "man". This is very clearly a story about Martians visiting Earth, but to make the story feel more immersive or realistic, the author chose to focus on the people that we can more easily relate to, rather than the aliens. I wanted to do the ghost stories radio text, but I couldn't get it to open up with AntConc so ww.txt is my next choice! Here's the answers for the questions asked on the assignment: 1, Ngram sizes 1 through 5 give frequency counts of 5 or above. 2. Ngram sizes 1 through 4 give frequency counts of 10 or more. 3. These are the top five phrases seen throughout the text out of the(53), there was a(30), of the martians(28), the heat ray(27), of the pit(26) 4. Out of the - There are a lot of needing to get out of something situations. This makes me curious about what kind of situation the protagonist is put in. Of the pit - Of the pit pops up 26 times, the pit seems as a very important location in the story. When viewing the text in Voyant Tools I noticed the top five most frequent words are said (166), martians (163), people (159), came (151), and saw (129). These words being the top five make a lot of sense since the story is about Martins. Here's a screenshot of "of the pit" being used multiple times throughout the story. I chose this science fiction text since the genre has always been a fascination to me. I was interested in analyzing the frequency of various phrases throughout the text and "ulla ulla ulla ulla" as well as "of the heat ray" had a frequency of 7 and 8 respectively. When investigating further, I discovered that the phrase "of the heat ray" in context throughout the text had 8 different uses at various points throughout the story! This is interesting because it leads me to believe that heat rays may be a weapon utilized by the martians in the novel. Finally, when inserting the text into Voyant, I created a wordcloud of the text in its' entirety. The words "brother," "smoke" and "pit" caught my eye as being frequently used. This text must have themes of family ties and war. A question that I have based on this Voyant analysis is what "pit" could mean? I analyzed the ww.txt option. This text seems to be about Martians and their time on Earth. When I put the text into Voyant tools, "Martian" is the second most used word throughout the text. "Said" is the first most used word. The phrases "of the Martians", "of the pit", and "of the cylinder" are the 3 most used phrases. The KWIC option is really interesting because you can learn much more about your text, just by a 3-word Ngram. I like using AntConc to learn more about the story because it gives you more insight on what is happening throughout the text without having to read it. After looking through the various texts we were given, I decided to analyze ww.txt. And sorry if this is cheating but after reading through this I am fairly certain this is H.G. Wells War of the Worlds. However, even though I knew what this story was, I tried my best to make sure that my knowledge did not interfere with my analysis. I'll start by showing the simplest thing, being the "word cloud" generated by Voyant. When creating the word cloud I made sure that it excluded the obvious words like "the" and "and", because I didn't think that they would show me anything interesting. When looking at the word cloud the most noticeable words are "martian", "came", and "people". What's interesting about this is that it puts man and martian on equal footing. For a story about aliens visiting Earth, the aliens get just as much attention as the people panicking on the planet. In fact when setting the N-Gram size to 1, the word "martians" was only used 5 more times total compared to the word "people". When setting the N-Gram size to something larger, like 3, the most used phrase becomes "out of the" with a whopping 53 uses! When looking into their uses, the most common use of the phrase "out of the" is not in reference to aliens leaving their ships, but instead people panicking to get away from said aliens or foreign objects. Overall I think what I found most interesting from this reading was the focus on words like "people" and "man". This is very clearly a story about Martians visiting Earth, but to make the story feel more immersive or realistic, the author chose to focus on the people that we can more easily relate to, rather than the aliens. Wow, I am surprised that "martians" were written a little more often than "people", especially for a sci-fi story! @OH-ThatGuy Wow, I am surprised that "martians" were written a little more often than "people", especially for a sci-fi story! Looking at ww.txt, I found that using ngrams: 2,3,4, mostly returned frequency's above 5. Ngrams of 2-3 typically give frequency counts above the double digits. the phrase "out of the" was repeated the most in the text. If I limit phrases to only those that appear 10+ times, I get 65 results. when looking at "there was a" I start to think this text is some sort of thriller that keeps the reader in the dark. I chose to continue using this text, even though we went over it in class. I started by continuing to look at the "out of the" n-gram, since I figured that would be a good place to start. I noticed that in the KWIC view they were ordered by what word immediately followed the n-gram. I thought it was interesting that in a book about Martians attacking (hmmm, I wonder what book that could be?), the highest matches for "out of the" were actually people emerging out of water. Granted, I think Martians (and Martian-related things) still emerge out of things more frequently than people, but still. There's a lot of things coming out of pits, which I personally find interesting as a lot of these pits seem to be in reference to fighting or taking cover from the Martians. World War 1 wouldn't occur for quite some time after this text was first written, but I think its interesting to think about that the tactic of digging in and hiding in holes in the ground seemed to be such a valid tactic all the way back then that it was employed against alien invaders. Switching to Voyant, the longest string was "I was walking down the road to clear my brain." In a book describing alien invasions, the longest single sentence is a moment were seemingly nothing is happening, just a stroll down the road. Not much of a surprise here... I first chose the ww text because I thought the name would be easy to find once I had some clues. The very first thing I did was slap the text into Voyant. I can't help but notice the word "martians" in the very front. Reminding me of a certain classic dystopian novel with "war" and "worlds" in the title. Next I put the text into Antconc. I chose to use a N-Gram size of 4 because I wanted a different clue than "martians". "the heat ray" reminds me a lot of a science fiction novel. In War of the Worlds, the weapon of choice for the martians are heat rays. In Antconc all the phrases with a 3 ngram give you a frequency count of above 5 for everything. The more you go up in ngrams the fewer phrases will have a frequency count of above 5. 2 ngrams will give all the phrases a frequency count of double digits and going up in ngrams will give you fewer phrases with a frequency count in the double digits. Antconc shows that “out of the”, “there was a”,” of the Martian”,” the edge of the”, and ”for the most part” are all common phrases. In Voyant, the words said, Martians, people, came, and saw were the most common words with said being mentioned 3 more times than Martians. Voyant takes out a lot more words than what Antconc does s a lot of phases have the same starting word but the rest of the phrase is different.
gharchive/issue
2023-10-26T14:53:54
2025-04-01T04:35:11.854596
{ "authors": [ "Fkhan2027", "KaitlynScutella", "MystKitteh", "OH-ThatGuy", "Rainbow7779", "VocaloidOtaku39", "creaturepsu", "ebeshero", "gylertaydos", "vnichols16" ], "repo": "newtfire/introDH-Hub", "url": "https://github.com/newtfire/introDH-Hub/issues/90", "license": "CC0-1.0", "license_type": "permissive", "license_source": "github-api" }
188383834
Create installable builds for Linux, macOSX and Windows (e.g. Setup Travis and Appveyor CI/builds) We need to setup for now a build (tests will come later) such that we can have as an end result an installable archive for each OS. check with Travis that we have macOs enabled create a build script that only builds each of Linux, macOS and Windows respectively setup the travis and appevyor yaml to build and make available the results for download A key part of the build is to be able to upload somewhere the final built apps. For Appveyor this is something I worked out here: https://github.com/pombreda/thirdparty/blob/e9f593bdaa1e51ea0a514c8df643ef0371e69e30/appveyor.yml#L37 this uploads to bintray basically anything that is in this directory: https://github.com/pombreda/thirdparty/blob/e9f593bdaa1e51ea0a514c8df643ef0371e69e30/appveyor.yml#L34 For Travis, the part for uploading is there: https://github.com/pombreda/thirdparty/blob/e9f593bdaa1e51ea0a514c8df643ef0371e69e30/.travis.yml#L56 and this use this JSON file to upload anything in the wheels directory https://github.com/pombreda/thirdparty/blob/e9f593bdaa1e51ea0a514c8df643ef0371e69e30/bintray_travis.json There may be other ways, but I am not sure that it is worth it: For instance appveyor has a link to the latest build "artifact": https://www.appveyor.com/docs/packaging-artifacts/ .... but only the latest. Hence the need to publish it somewhere instead. The builds could be published to Github too btw, but this would need a tad work https://www.appveyor.com/docs/deployment/github/ and this would mean basically having a tag on each build The ref docs for travis are there: https://docs.travis-ci.com/user/deployment/ We need though to define a proper scheme for versions as each built version should be unique (e.g. reference the commit) Yesterday's commit https://github.com/nexB/aboutcode-manager/commit/58f9a82ac469b25682a5ebc455b7c56368af7830 contains my initial draft appveyor.yml. The push triggered AppVeyor build 216 (https://ci.appveyor.com/project/nexB/aboutcode-manager/build/216). Seemed to install OK including sqlite3. The build installed npm 3.10.10 and node 6.11.0. Exited with code 1 on line 1735. Failed upon running python build.py. The failure block: python build.py => BUILDING AboutCode App release: 2.0.0 Traceback (most recent call last): File "build.py", line 221, in <module> build() File "build.py", line 167, in build shutil.rmtree(build_dir) File "C:\Python27\lib\shutil.py", line 239, in rmtree onerror(os.listdir, path, sys.exc_info()) File "C:\Python27\lib\shutil.py", line 237, in rmtree names = os.listdir(path) WindowsError: [Error 3] The system cannot find the path specified: u'dist\\*.*' Command exited with code 1 This is similar, but not identical, to the error I encountered running locally on Windows 10. The AppVeyor error seems to say the path to the dist directory can’t be found, reminiscent of my initial Ubuntu 16.04 error, where the dist directory in fact did not exist. However, examining our code tree in Git Bash as well as Visual Studio Code, I see that a dist directory already exists. The initial appveyor.yml file: version: '{build}' image: Visual Studio 2017 # Test against the latest version of this Node.js version # environment: # nodejs_version: "6" platform: - x64 cache: - node_modules - '%APPDATA%\npm-cache' init: - git config --global core.autocrlf input # Install scripts (runs after repo cloning) install: - ps: Install-Product node 6 x64 # - git reset --hard HEAD - npm install - npm prune build_script: - node --version - npm --version - python build.py test: off At this stage we have a proper build that is done on Windows and that creates some archive @jdaguil contacted Travis to enable mac. The build is failing on Linux with the npm bin command (with shell=True) and is failing on mac because of apt-get calls. We now have the build running on all three OSses and producing some archive. Next we need to upload them somewhere so that we can test them. I will also add a SHA1 and SHA256 checksum such that these uploads can be verified. The application that is created from the new script is now working on Mac. Linux and Windows still need to be tested and verified. Testing on Windows 10, the application created by running python build.py (in the Windows command prompt) seems to work well. With my current setup, this requires: checking out branch 16 in Git Bash, running git pull to update, opening the Windows command prompt, navigating to the aboutcode-manager directory note that the command prompt does not reflect which branch is currently checked out -- running git status, for example, produces 'git' is not recognized as an internal or external command, operable program or batch file-- Git Bash does that work running python build.py. When I ran python build.py in Git Bash, the result was this: Tue Jul 11, 2017 10:01 AM /c/code/nexb/dev/aboutcode-manager JMH (16-python-build-script) $ python build.py => BUILDING AboutCode App release: 2.0.0 platform: Windows-10-10.0.15063 sys.platform: win32 Traceback (most recent call last): File "build.py", line 243, in <module> build() File "build.py", line 179, in build os.makedirs(build_dir) File "C:\Python27\lib\os.py", line 157, in makedirs mkdir(name, mode) WindowsError: [Error 5] Access is denied: u'dist' The application created by running python build.py seems to work well on Ubuntu 16.04. I started to have at least some downloads pushed for Mac and Linux from Travis. Here are some quick notes: Project at sf: https://sourceforge.net/projects/aboutcode-manager/ User with "SCP" right (aka "release technician") https://sourceforge.net/u/aboutcode-rel/ I generated a dedicated SSH key pair with ssh-keygen -t rsa -C "aboutcode-rel@shell.sf.net and asked it to be sored in the root dir of the ABCM checkout This par of keys was created (I named them this): deploy_rsa deploy_rsa.pub I then installed travis globally This is a pain. $ travis login $ travis encrypt-file deploy_rsa --add This created deploy_rsa.enc and updated the travis.yml (and damaged it a bit along the way) with: - openssl aes-256-cbc -K $encrypted_906d16ae5679_key -iv $encrypted_906d16ae5679_iv -in deploy_rsa.enc -out deploy_rsa -d I then fixed the travis script I also tested the SCP with this: scp -i deploy_rsa aboutcode.ABOUT aboutcode-rel@web.sourceforge.net:/home/frs/project/aboutcode-manager/builds/ We updated the Travis script and bingo: https://sourceforge.net/projects/aboutcode-manager/files/builds/ Issues: - version/name of the build is not correct: git fails - we need to compute a sha1 ATM on Linux, I get this error: Uncaught Error: Please install sqlite3 package manually ... so we are getting there! This is for a build I fetched from sf.net We still have one issues in that the build does not run yet properly on Mac... We are getting make: g++-4.8: No such file or directory At this stage all OS do build: Linux and Mac are uploaded to https://sourceforge.net/projects/aboutcode-manager/files/builds/ Windows appveyor handles this through Artifact for each build: For instance this contains the latest build download https://ci.appveyor.com/project/nexB/aboutcode-manager/build/artifacts @jdaguil the mac build needs to be verified for sanity @johnmhoran same for the linux build The upload takes place on EVERY commit and we may want to limit this to tags only for sanity? @pombredanne will do. And I agree I think we should limit to tags. I am having issues with the mac build from sourceforge: When first extracted, I am prompted with the following message: After I go to system preferences and change the settings from Mac App Store and identified developers to Anywhere I have to re-extract the archive to get the error message in (1) to go away. The above error message persists with the initial extracted archive even after preferences are changed. I then get another error message: OS Details: gcc --version Configured with: --prefix=/Applications/Xcode.app/Contents/Developer/usr --with-gxx-include-dir=/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX10.10.sdk/usr/include/c++/4.2.1 Apple LLVM version 6.0 (clang-600.0.56) (based on LLVM 3.5svn) Target: x86_64-apple-darwin15.6.0 Thread model: posix usr/bin/xcodebuild -version Xcode 6.1.1 Build version 6A2008a I am also having issues running the build.py script locally. At https://github.com/nexB/aboutcode-manager/issues/16#ref-commit-9858320 I was able to run the script and produce a working app. Now the the script will not run locally, and I get the following error: $ build.py DEBUG: get_version: tag: 2.0.0-rc3 distance: 112 commit: f858339 dirty: False DEBUG: get_version: dirty, using timestamp: 2.0.0-rc3.post112.f858339 ############################################################# => BUILDING AboutCode App release: 2.0.0-rc3.post112.f858339 platform: Darwin-15.6.0-x86_64-i386-64bit sys.platform: darwin using NPM bin at: /aboutcode-manager/node_modules/.bin Running electron-rebuild... Running command: u'/aboutcode-manager/node_modules/.bin/electron-rebuild'... ✔ Rebuild Complete Running electron-packager... Running command: u'/aboutcode-manager/node_modules/.bin/electron-packager . AboutCode-Manager --prune --ignore=thirdparty/* --ignore=dist/* --ignore=/\\.idea --ignore=/\\.gitignore --ignore=/test --ignore=/tmp --ignore=/bower.json --platform=darwin --arch=x64 --icon=assets/app-icon/mac/aboutcode.icns --version=1.4.0 --out=dist --asar=true --overwrite=true'... Packaging app for platform darwin x64 using electron v1.4.0 Command failed: npm prune --production npm ERR! May not delete: /private/var/folders/6d/7nz7bkx92n38xwc2frhqrsjm0000gn/T/electron-packager/darwin-x64/AboutCode-Manager-darwin-x64/Electron.app/Contents/Resources/app/node_modules/.bin npm ERR! A complete log of this run can be found in: npm ERR! /Users/jilliandaguil/.npm/_logs/2017-08-09T17_49_17_717Z-debug.log Failed to execute command: u'/aboutcode-manager/node_modules/.bin/electron-packager . AboutCode-Manager --prune --ignore=thirdparty/* --ignore=dist/* --ignore=/\\.idea --ignore=/\\.gitignore --ignore=/test --ignore=/tmp --ignore=/bower.json --platform=darwin --arch=x64 --icon=assets/app-icon/mac/aboutcode.icns --version=1.4.0 --out=dist --asar=true --overwrite=true' Aborting... @pombredanne Changing my local version of npm to 5.2.0 fixed the issue in https://github.com/nexB/aboutcode-manager/issues/16#issuecomment-321332090. For reference see, https://github.com/npm/npm/issues/17781 @pombredanne Regarding https://github.com/nexB/aboutcode-manager/issues/16#issuecomment-321329244, latest work in a229cfa resolves the issue when downloading the built archive from sourceforge. However, now the following message is displayed when trying to open the app: The app is functional if one of the following steps is taken: Security preferences are changed or Right click on the app and click Open @pombredanne @jdaguil Just a note re a temporary build failure when working a few minutes ago in issue #154 -- see my comment. @jdaguil FWIW, here are somoe pointers for the macOS issue that are likely related: https://github.com/travis-ci/travis-ci/issues/6935 and https://github.com/travis-ci/travis-ci/issues/7065 https://github.com/Daemon-Devarshi/MedicationSchedulerSwift3.0/issues/10 https://developer.apple.com/support/certificates/expiration/ There are likely other sources of details that would matter to us. I support the idea that we get the iCloud app certificate assuming that we can use it at the project level.
gharchive/issue
2016-11-09T23:55:06
2025-04-01T04:35:11.888564
{ "authors": [ "jdaguil", "johnmhoran", "mjherzog", "pombredanne" ], "repo": "nexB/aboutcode-manager", "url": "https://github.com/nexB/aboutcode-manager/issues/16", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1328973840
Task/nxc 57 Descrição Implementação dos novos designs tokens e criação dos components com as cores respectivas do design token. Task # nxc-57 Tipo de alteração [ ] Correção de bug [x] Melhoria [x] Novo componente Checklist: [x] Meu código segue as diretrizes de estilo deste projeto [x] Realizei uma auto-revisão do meu próprio código [ ] Eu comentei meu código, particularmente em áreas difíceis de entender [ ] Fiz as alterações correspondentes na documentação [ ] Minhas alterações não geram novos erros [ ] Adicionei testes que comprovam que minha correção é eficaz ou que minha feature funciona [ ] Testes unitários novos e existentes continuam passando localmente com minhas alterações Mandou bem! Só vou pedir para deletar o package-lock.json e adicionar no .gitignore, ok? Estamos usando o yarn para gerenciar as dependências. @soarexs Parabéns! estamos realmente emocionados por sua primeira task! Quase chorei!
gharchive/pull-request
2022-08-04T17:38:13
2025-04-01T04:35:11.914679
{ "authors": [ "daniloras", "pedro-cartco", "soarexs" ], "repo": "next-cart-dev/N-C-components-library", "url": "https://github.com/next-cart-dev/N-C-components-library/pull/7", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1509499955
2단계 - 인덱스 설계 안녕하세요. 2단계 - 인덱스 설계 미션 리뷰 부탁드립니다. 감사합니다. https://github.com/next-step/infra-subway-db/pull/23 1단계 미션에서 질문 남긴게 있는데요. 여유 있으실 때 확인 좀 부탁드려도 될까요? order by 정렬도 인덱스를 사용하는 지 확인 할 수 있나요? 3번째 요구사항이 0.3 s 밑으로 내려가질 않습니다.. 질문 남기신 부분이 어떤 것일까요? 이 Comment 인가요? https://github.com/next-step/infra-subway-db/pull/23#issuecomment-1357309690 order by 정렬도 인덱스를 사용하는 지 확인 할 수 있나요? 실행계획에서 filesort가 발생하지 않는다면, 해당 인덱스를 사용하여(key를 보면 되겠죠?) 굳이 다시 정렬할 필요가 없기 때문입니다. 질문 남기신 부분이 어떤 것일까요? 이 Comment 인가요? #23 (comment) 아래 위치에 코맨트 남겨놨습니다! 혹시 본인한테만 보이나요? https://github.com/next-step/infra-subway-db/pull/23/commits/925cd23be7064cc31fa4e5a573a8803a566a0e8d https://github.com/next-step/infra-subway-db/pull/23/commits/925cd23be7064cc31fa4e5a573a8803a566a0e8d 리뷰 확인했고 일부 내용 수정했습니다. 마지막 리뷰 요청이 될 것 같습니다! 감사합니다.
gharchive/pull-request
2022-12-23T16:01:12
2025-04-01T04:35:11.922285
{ "authors": [ "OhSeungCheol", "brainbackdoor" ], "repo": "next-step/infra-subway-db", "url": "https://github.com/next-step/infra-subway-db/pull/28", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1983839033
Don't need to base64-encode the username and passoword for basic auth. The docs (README) suggests that the user should base64-encode their USERNAME:PASSWORD string passed in via weblog configuration. This encoding is unnecessary, and will break, becauase the encoding is already handled by the nextflow.util.SimpleHttpClient class: https://github.com/nextflow-io/nextflow/blob/f5362a7b067173a29d684663df22bb48fbbf5659/modules/nextflow/src/main/groovy/nextflow/util/SimpleHttpClient.groovy#L133 con.setRequestProperty("Authorization","Basic ${basicToken.bytes.encodeBase64()}") Ah, this is weblog. I'm retiring my previous comment 😅
gharchive/pull-request
2023-11-08T15:16:31
2025-04-01T04:35:12.555231
{ "authors": [ "pditommaso", "robsyme" ], "repo": "nextflow-io/nf-weblog", "url": "https://github.com/nextflow-io/nf-weblog/pull/8", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1651991923
Readme > usage: mention build! I expected building of static html files to be included in the usage section (at least, since it is not mentioned in the Book). @holyjak thank you! Will follow-up with mentioning the build functionality in the book as well.
gharchive/pull-request
2023-04-03T12:39:45
2025-04-01T04:35:12.586566
{ "authors": [ "holyjak", "mk" ], "repo": "nextjournal/clerk", "url": "https://github.com/nextjournal/clerk/pull/448", "license": "ISC", "license_type": "permissive", "license_source": "github-api" }
1867921012
🛑 Nextpost.tech is down In 0256a7e, Nextpost.tech (https://nextpost.tech) was down: HTTP code: 0 Response time: 0 ms Resolved: Nextpost.tech is back up in 427fdca after 488 days, 5 hours, 37 minutes.
gharchive/issue
2023-08-26T05:47:26
2025-04-01T04:35:12.588921
{ "authors": [ "sergeykomlev" ], "repo": "nextpost-tech/status", "url": "https://github.com/nextpost-tech/status/issues/299", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1939072993
🛑 Nextpost.tech is down In 4d8da1a, Nextpost.tech (https://nextpost.tech) was down: HTTP code: 0 Response time: 0 ms Resolved: Nextpost.tech is back up in 25254ef after 30 minutes.
gharchive/issue
2023-10-12T02:39:12
2025-04-01T04:35:12.591209
{ "authors": [ "sergeykomlev" ], "repo": "nextpost-tech/status", "url": "https://github.com/nextpost-tech/status/issues/364", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1773239780
🛑 /searchUser - Search for user by keyword is down In 38c5845, /searchUser - Search for user by keyword (https://nextpost.tech/tiktok-api/searchUser?keyword=apple&date_posted=all-time&sort_by=relevance&cache_timeout=0&license_key=$API_KEY) was down: HTTP code: 400 Response time: 5696 ms Resolved: /searchUser - Search for user by keyword is back up in 2efe8f7.
gharchive/issue
2023-06-25T11:55:28
2025-04-01T04:35:12.593969
{ "authors": [ "sergeykomlev" ], "repo": "nextpost-tech/tiktok-status", "url": "https://github.com/nextpost-tech/tiktok-status/issues/793", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }