added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T04:54:50.523966
2023-02-18T17:08:33
1590405397
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13572", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/4326" }
gharchive/issue
⚠️ Bofrost.de has degraded performance In 4db2ee8, Bofrost.de (https://www.bofrost.de) experienced degraded performance: HTTP code: 200 Response time: 1235 ms Resolved: Bofrost.de performance has improved in 0b609b8.
2025-04-01T04:54:50.527209
2023-02-19T11:49:53
1590658437
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13573", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/4364" }
gharchive/issue
⚠️ Bofrost.at has degraded performance In 703cae7, Bofrost.at (https://www.bofrost.at) experienced degraded performance: HTTP code: 200 Response time: 1766 ms Resolved: Bofrost.at performance has improved in 39b1a1d.
2025-04-01T04:54:50.531506
2022-11-18T22:43:57
1455927746
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13574", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/451" }
gharchive/issue
⚠️ Bofrost.at has degraded performance In aecf138, Bofrost.at (https://www.bofrost.at) experienced degraded performance: HTTP code: 200 Response time: 1158 ms Resolved: Bofrost.at performance has improved in 874afbe.
2025-04-01T04:54:50.535078
2024-05-20T22:25:06
2306899331
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13575", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/47228" }
gharchive/issue
⚠️ Bofrost.at has degraded performance In a8d2f3a, Bofrost.at (https://www.bofrost.at) experienced degraded performance: HTTP code: 200 Response time: 2107 ms Resolved: Bofrost.at performance has improved in 7c462da after 14 minutes.
2025-04-01T04:54:50.539095
2024-05-25T04:53:26
2316697288
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13576", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/47949" }
gharchive/issue
⚠️ Bofrost.de has degraded performance In 5bcc909, Bofrost.de (https://www.bofrost.de) experienced degraded performance: HTTP code: 200 Response time: 1840 ms Resolved: Bofrost.de performance has improved in 2e77bde after 1 hour, 17 minutes.
2025-04-01T04:54:50.542666
2024-06-24T09:41:50
2369748819
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13577", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/52078" }
gharchive/issue
⚠️ Bofrost.at has degraded performance In c2237db, Bofrost.at (https://www.bofrost.at) experienced degraded performance: HTTP code: 200 Response time: 875 ms Resolved: Bofrost.at performance has improved in 190d8b7 after 13 minutes.
2025-04-01T04:54:50.545853
2024-06-26T18:25:00
2376007398
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13578", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/52355" }
gharchive/issue
⚠️ Bofrost.lu has degraded performance In de077d9, Bofrost.lu (https://www.bofrost.lu) experienced degraded performance: HTTP code: 200 Response time: 1468 ms Resolved: Bofrost.lu performance has improved in 81b3c40 after 23 minutes.
2025-04-01T04:54:50.548813
2024-06-29T11:26:56
2381685951
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13579", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/52756" }
gharchive/issue
⚠️ Bofrost.es has degraded performance In bcccc17, Bofrost.es (https://www.bofrost.es) experienced degraded performance: HTTP code: 200 Response time: 1444 ms Resolved: Bofrost.es performance has improved in bac3398 after 14 minutes.
2025-04-01T04:54:50.551729
2022-11-23T14:58:36
1461915551
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13580", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/630" }
gharchive/issue
⚠️ Bofrost.at has degraded performance In 6612228, Bofrost.at (https://www.bofrost.at) experienced degraded performance: HTTP code: 200 Response time: 1027 ms Resolved: Bofrost.at performance has improved in ea3d4bf.
2025-04-01T04:54:50.561710
2024-02-19T11:34:59
2142162515
{ "authors": [ "Spasitjel" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13581", "repo": "ClassicUO/ClassicUO", "url": "https://github.com/ClassicUO/ClassicUO/issues/1669" }
gharchive/issue
Potential race condition in PacketHandlers._clilocRequests Describe the bug (WIP sorry, posted this prematurely) Sorry nvm, should've waited anyway to open this. I thought I saw it being called possyble from the render thread and the thread receiving packets from plugins, but it's not the case.
2025-04-01T04:54:50.570609
2015-07-04T08:56:17
92991852
{ "authors": [ "hyguyz", "mattrmiller", "sparkeleven" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13582", "repo": "Classy-Geeks/laravel-potion", "url": "https://github.com/Classy-Geeks/laravel-potion/issues/13" }
gharchive/issue
path not specified whenever i am making command php artisan potion:make-assets its saying the system canot find the path specified........ may be more docmentation is needed If you pull my fork of this project, you can use --debug to show you all the paths and files it attempts to make so you can work out what is failing. Laravel 5.1 only https://github.com/sparkeleven/laravel-potion i am using latest version of laravel my css and js files are present in resources/assets folder when ever i am running this command php artisan potion:make-assets its saying system cant find the path specified .... please tell am i doing something wrong or i have to add something more in command I have plans to add better output, so that you can see what might be going wrong. I will try to find some time today or tonight to get this into the main branch. I just pushed v1.0.3 which adds more output. It will output what directory it's trying to create, but more importantly it will output what it is pulling from your config for assets and resource paths. Just a reminder: assets: is the folder that will be used to store final assets that are created by potion. resources: are files/folders that are used to create the final assets. ok let me try it
2025-04-01T04:54:50.598247
2021-02-06T07:17:22
802633019
{ "authors": [ "Exper1mental", "Jay-Madden", "zepthro" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13583", "repo": "ClemsonCPSC-Discord/ClemBot", "url": "https://github.com/ClemsonCPSC-Discord/ClemBot/issues/281" }
gharchive/issue
Calculator Command Can't Handle Commas? When using the calculator, it appears it doesn't handle commas; or rather, it ignores numbers included before commas. e.g. !calc 1,100 * 2 outputs 200.0 cc @mp7786 Can we close this issue. Issue seems resolved since #282. Closed in #282
2025-04-01T04:54:50.951831
2020-11-16T18:45:18
744073606
{ "authors": [ "AmiciA", "curranjw", "drej1", "foolioo", "ohassidi", "y0y0z" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13585", "repo": "Clipy/Clipy", "url": "https://github.com/Clipy/Clipy/issues/428" }
gharchive/issue
big sur keyboard firstly, i wanna thank you for the clipy. i am using clipy version 1.2.1 and i've upgraded to big sur. now (it used to work prior to the big sur upgrade) when i access the clipy history using a keyboard shortcut, i am unable to select from history using the keyboard (arrows and number keys). In addition to this issue, input "COMMAND" + "SHIFT" stopped working when you click the item in the history or snippet. i was able to bypass the issue by restarting clipy (quitting clipy and starting it again). does it work for you? i also noticed that even when i bypass the issue by restarting clipy, the issue reoccur when the mouse is hovering the history list Same occured to me on Big Sur with M1 SoC. Restart helped but I think it will occur again... Same occured to me on Big Sur with M1 SoC. Restart helped but I think it will occur again... i was able to bypass the issue by restarting clipy (quitting clipy and starting it again). does it work for you? Still annoying to restart the app, each time I am still having this issue with Monterey 12.0.1 and Clipy v1.2.1. Keyboard navigation with arrow keys works after a restart of Clipy. to all those who follow this thread, look at https://github.com/Clipy/Clipy/issues/319#issuecomment-874115468
2025-04-01T04:54:50.964222
2023-09-30T15:42:16
1920303829
{ "authors": [ "ASimpleTechie", "Clooos" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13586", "repo": "Clooos/Bubble-Card", "url": "https://github.com/Clooos/Bubble-Card/issues/86" }
gharchive/issue
Pop-ups/Buttons not being contained within designated card area I'm sorry to be a pain but I just cannot figure this out. I have my pop up cards and horizontal button stack sitting in a vertical stack with other cards on the left of my tablet dashboard. I cannot figure out how to get the buttons and the pop ups to stay within the area of the vertical stack. This image below shows the buttons sprawled across the bottom of the dashboard when I want them contained within the area that the other cards are. I would like the buttons to be scrollable from within the width of that panel (same width as the weather card). I would also like the pop ups to stay within the same area when they are open, but instead they pop up in the centre (please see image below) I am unsure what I need to do to achieve this. I am really liking the look and features of this card so I really want to figure this out. Thank you for your constant hard work. Hi! Just add this to your pop-up to remove the centering of the pop-up, but you will need to edit its width_desktop too to fit your column (something like 500px). styles: | #root { left: inherit !important; } Thanks! Do I add it to each individual pop-up within the stack? You can probably just add your cards in horizontal-stacks, tell me if it's what you are looking for.
2025-04-01T04:54:50.980889
2018-06-16T13:32:16
332987507
{ "authors": [ "fristonio" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13587", "repo": "Cloud-CV/origami-lib", "url": "https://github.com/Cloud-CV/origami-lib/pull/37" }
gharchive/pull-request
Refactor origami-lib Update the whole project to make it more modular Added robust exception handling. Added better caching and connection mechanism. Documented the whole code. No python linting errors. Better managed server and handlers @uttu357 here ^
2025-04-01T04:54:50.986028
2024-02-07T20:39:40
2123844982
{ "authors": [ "jackdbd" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13588", "repo": "CloudCannon/pagefind", "url": "https://github.com/CloudCannon/pagefind/issues/558" }
gharchive/issue
Inline SVG in search form blocked by Content-Security-Policy Description I have a website with a very strict Content-Security-Policy. I read the documentation and managed to allow the Pagefind WebAssembly module. I noticed there is another small issue though: the Svelte component for the search form (see the .pagefind-ui__form here) uses a data URL to inline an SVG (the magnifying glass icon). This SVG is blocked by my CSP. A possible solution To allow only the SVG used by the Pagefind UI search form, we can include the following data: scheme source in the img-src directive of our Content-Security-Policy header: img-src 'self' data:image/svg+xml,%3Csvg width='18' height='18' viewBox='0 0 18 18' fill='none' xmlns='http://www.w3.org/2000/svg'%3E%3Cpath d='M12.7549 11.255H11.9649L11.6849 10.985C12.6649 9.845 13.2549 8.365 13.2549 6.755C13.2549 3.165 10.3449 0.255005 6.75488 0.255005C3.16488 0.255005 0.254883 3.165 0.254883 6.755C0.254883 10.345 3.16488 13.255 6.75488 13.255C8.36488 13.255 9.84488 12.665 10.9849 11.685L11.2549 11.965V12.755L16.2549 17.745L17.7449 16.255L12.7549 11.255ZM6.75488 11.255C4.26488 11.255 2.25488 9.245 2.25488 6.755C2.25488 4.26501 4.26488 2.255 6.75488 2.255C9.24488 2.255 11.2549 4.26501 11.2549 6.755C11.2549 9.245 9.24488 11.255 6.75488 11.255Z' fill='%23000000'/%3E%3C/svg%3E%0A To allow any data URL, we can write this img-src directive instead: img-src 'self' data:; I guess it should also be possible to allow only inline SVGs and no other MIME types (e.g. image/jpeg) using the directive below (I haven't tested it though). img-src 'self' data:image/svg+xml; I forgot to mention that allowing a data: scheme is something that the user has to do. Instead, something that Pagefind could do is not embedding the SVG, and hosting it somewhere (e.g. AWS S3, Google Cloud Storage, Cloudflare R2, etc). Or even letting the user self-host the SVG.
2025-04-01T04:54:51.026422
2023-08-01T18:36:29
1831820895
{ "authors": [ "DivyanshiNarang", "ighoshsubho" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13591", "repo": "Clueless-Community/fintech-api", "url": "https://github.com/Clueless-Community/fintech-api/pull/490" }
gharchive/pull-request
README.md enhanced closes #475 Added links of official documentation of the respective Tech stack used, in the README.md file. Good @DivyanshiNarang! Keep contributing to fintech API repo like this.
2025-04-01T04:54:51.030999
2022-10-02T19:03:41
1393866602
{ "authors": [ "AKACHI-4", "Rajdip019", "codeswithroh" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13592", "repo": "Clueless-Community/first-contribution", "url": "https://github.com/Clueless-Community/first-contribution/issues/164" }
gharchive/issue
Add Contributor's Profile Images !! Update Contributor's profile image will appear at the bottom of README.md CONTRIBUTING.MD will be modified to aware newcomers about contribution. Yes, let's work on this now! @AKACHI-4 @AKACHI-4 Should the images be added on a new folder or should it be given from a link ? It should be in the README.md #167 PR generated and #164 issue solved as I am working on it under the last issue but it gets closed so I generated it here merge PR so changes will get visible for everyone Thanks, @Rajdip019 Yup that's better. And more clear. Yep 😄👍 @AKACHI-4 I couldn't find the place to add the contributors image in the Readme.md file Hey @codeswithroh actually I have already generated a PR regarding this .....checkout their you will get some idea about it.
2025-04-01T04:54:51.032023
2023-05-29T08:03:16
1730288447
{ "authors": [ "karthikbhandary2", "nikhil25803" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13593", "repo": "Clueless-Community/scrape-up", "url": "https://github.com/Clueless-Community/scrape-up/issues/229" }
gharchive/issue
Getting Exact followers from Instagram profiles. I would like to improve the followers() function by making it return the exact number of followers a profile has. This method has already been added.
2025-04-01T04:54:51.037736
2024-05-12T11:14:36
2291286704
{ "authors": [ "Arnavop", "Cheshta17", "Disha-Baghel", "Nithincoderboy", "nikhil25803" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13594", "repo": "Clueless-Community/scrape-up", "url": "https://github.com/Clueless-Community/scrape-up/issues/913" }
gharchive/issue
Feat: scrape-up for Unstop.com Describe the feature Scrape the details of upcoming contests / jobs from unstop.com Add ScreenShots Record [X] I agree to follow this project's Code of Conduct [X] I'm a GSSoC'24 contributor [X] I want to work on this issue @nikhil25803 Assign this issue to me I would able to work on it Go ahead @Disha-Baghel Note Please create a separate module for this, as in the folder and project structure (if it is already created, just add your features as functions in the same module). Do not use the `selenium web driver as it is incompatible with all devices and cloud platforms. Before making any changes, please check whether the module you want to add exists. If yes, then you can add your functionality as a method only make a separate module and class for it. All the best 👨‍💻 @nikhil25803 Could you assign this issue to me. I want to work on this project. @nikhil25803 Could you assign this issue to me. I want to work on this project. Hello, I wanted to kindly inform you that I am already actively working on the issue at hand. Thank you for your understanding is this done?
2025-04-01T04:54:51.054096
2015-05-07T14:33:30
74007765
{ "authors": [ "adamtheturtle", "itamarst", "jongiddy", "tomprince" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13595", "repo": "ClusterHQ/flocker", "url": "https://github.com/ClusterHQ/flocker/pull/1380" }
gharchive/pull-request
[FLOC-1742] Ubuntu CLI installation instructions Primarily adds a task to install the CLI in Ubuntu. Modifies acceptance test framework to allow starting of a single node, with a --type flag to select client (testing installation of client on a single node) or cluster (testing node installation on a 2-node cluster). Acceptance test can be run using: admin/run-acceptance-tests --distribution ubuntu-14.04 --provider aws --config-file config.yml --branch ubuntu-cli-FLOC-1742 --type client or cd docs; make html and follow the installation instructions in indepth/installation.html#installing-flocker-cli This can be a follow-up but I think that the Getting Started Ubuntu CLI installations should be updated and the release process changed to say that they do not need to be manually tested. Thanks, please resubmit when the comments have been addressed. Added FLOC-1893 to update Getting Started guide to use DEB :question: It is getting impossible to keep the state in my head (maybe not @tomprince 's) of what is being tested and what isn't. This might be a good opportunity to create a document which gives a brief outline of what is being tested, starting with just these tests. For example, for some things the S3 archive is being tested, for some things it isn't. For some things the CLI instructions are being tested (Ubuntu CLI instructions on the tutorial) for some things they aren't. Buildbot builder names only say so much. Or make a follow-up issue if you prefer. I didn't do the most thorough review ever, but given passing builder and addressing comments above (adam and mine), you can merge. Update https://clusterhq.atlassian.net/browse/FLOC-1309 as appropriate. Can you confirm whether the following acceptance criteria for the has been tested?: Removing the package should perform the expected cleanup: Remove all files and config changes which have been added to the system. This should also happen if the package was not installed cleanly, e.g. if there was an error. If not, either do that here and resubmit (uninstall instructions?) or create a follow-up issue. This is worthy of a change to What's New and NEWS. Sorry I know Itamar said that this can be merged but I've spotted a few things since and would like a look before this is merged. I'll be as quick as possible. Please update https://clusterhq.atlassian.net/browse/FLOC-1325 as appropriate (I'm not sure if it can be closed or is still relevant to some other instructions). Either do or create a follow-up story blocker: update the release process sections which test the Getting Started Guide client installation and the tutorial client installation - we shouldn't manually test things which have automated testing. Thanks @jongiddy - there are a lot of comments but quite a few of them are existing problems highlighted by this branch. Please address (probably mostly by making many follow-up issues) and resubmit for a hopefully quick final review. Though what @itamarst said is right - perhaps get a staging builder running tests against this before merging. Regarding uninstall instructions, I think it is reasonable to assume that people know how to remove packages from their operating system. Regarding uninstall instructions, I think it is reasonable to assume that people know how to remove packages from their operating system. I'd say not with Homebrew at least - that is, Homebrew is a third party package manager, and people have had problems uninstalling our Homebrew packages (you have to remember which version you installed). However that is not part of this issue and not a product requirement. Either do or create a follow-up story blocker: update the release process sections which test the Getting Started Guide client installation and the tutorial client installation - we shouldn't manually test things which have automated testing. The Release Process instructions for CLI require going through the whole process described in the Getting Started Guide. As only one step is now automatically tested, it is still appropriate to keep this step in the release process. And the occasional manual run of the CLI installation steps on Ubuntu probably doesn't hurt, to ensure unexpected environmental conditions aren't causing problems. The Release Process instructions for CLI require going through the whole process described in the Getting Started Guide. @jongiddy There's that and there's also "Test the client install instructions work on all supported platforms by following the instructions and checking the version:". It only hurts in that this would allow us to automate one more small bit of the release process - helping us towards the goal of an entirely automated process. I think adding a note "The Ubuntu install instructions are already tested automatically so they don't need to be tested manually" would be suitable. @adamtheturtle Do you know, are the instructions on this page ever tested automatically? The instructions that are tested on Buildbot specify the branch, and hence use the Buildbot repo. During the release process, do the acceptance tests get run without a branch specifier, hence using the ClusterHQ repo? Regarding automation of Getting Started Guide testing in release process, I have added a comment to FLOC-1309 to make the assignee of that issue aware that automated testing must avoid using the Buildbot additions. This needs master to be merged in (sorry I saw you did this 5 hours ago!). Please update https://clusterhq.atlassian.net/browse/FLOC-1325 as appropriate (I'm not sure if it can be closed or is still relevant to some other instructions). Nothing was done on this. Maybe nothing needed to be done. This is worthy of a change to What's New and NEWS. Create a follow-up issue for expanding the installed package tests (either --version for all CLIs or Tom said "(We'll probably end up running the cli test suite, instead).") The non-effective, but not an error, format is in preparation for the change commented above. As it is clearly marked with an issue, I think this is OK for now. :question: Code which does nothing is confusing. Whoever puts ARCHIVE_BUCKET there knows how to use .format. Please address the above, fix merge conflicts and merge once this is passing on the new builder. (And resubmit if this fails on the new builder and non-trivial changes need to be made). A release has been done so the ARCHIVE_BUCKET can be used I think. Idea (follow-up?): Perhaps running this should be part of the release process, with some kind of flag and version option which means "does exactly what the instructions say and doesn't touch Buildbot" and then the rest of the time it will run with everything possible from Buildbot.
2025-04-01T04:54:51.056934
2015-08-25T10:48:53
102998374
{ "authors": [ "adamtheturtle", "dwgebler" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13596", "repo": "ClusterHQ/flocker", "url": "https://github.com/ClusterHQ/flocker/pull/1882" }
gharchive/pull-request
[FLOC-2921] Release Flocker 1.2.0 FLOC-2921 Awaiting build results... Assuming that you are happy with the build results when GitHub is back, continue. Please continue Finally, a successful build!
2025-04-01T04:54:51.085732
2023-01-16T16:46:03
1535224261
{ "authors": [ "ffried", "gruffins" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13597", "repo": "CocoaLumberjack/CocoaLumberjack", "url": "https://github.com/CocoaLumberjack/CocoaLumberjack/pull/1328" }
gharchive/pull-request
Add Birch to extensions New Pull Request Checklist [x] I have read and understood the CONTRIBUTING guide [x] I have read the Documentation [x] I have searched for a similar pull request in the project and found none [x] I have updated this branch with the latest master to avoid conflicts (via merge from master or rebase) [x] I have added the required tests to prove the fix/feature I am adding [x] I have updated the documentation (if necessary) [x] I have run the tests and they pass [x] I have run the lint and it passes (pod lib lint) This merge request fixes / refers to the following issues: ... Pull Request Description Hey, I run a free remote logging service and I've created bindings for CocoaLumberjack. Wondering if it would be ok to link to it in the extensions section? Thanks! Absolutely! I've activated auto-merge, so it should get merged once the GitHub Actions pipeline is through. @ffried thanks for the approval! It looks like one of the specs are failing which I would think is unrelated to documentation change. Let me know if you need anything. @gruffins Thanks for the reminder! It's an unrelated (spurious) failure. I'll merge it manually.
2025-04-01T04:54:51.119798
2021-06-14T08:13:25
920150693
{ "authors": [ "Shaishav0507", "hkaur008" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13598", "repo": "CodeFlow201/GitMarkonics", "url": "https://github.com/CodeFlow201/GitMarkonics/pull/39" }
gharchive/pull-request
Added Project Admin, Code of Conduct and License Hey @Yashnatani021-coder I have added Code of conduct, Project Admin and License to the readme file. Please Check and update. I didn't know whose pic to put on project admin, so used a gif. If you could provide with the picture I'll update it. Thanks Cool , I liked gif stuff
2025-04-01T04:54:51.125971
2017-07-07T01:49:33
241134454
{ "authors": [ "abettermap", "dgendill" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13600", "repo": "CodeForFoco/solar-scorecard", "url": "https://github.com/CodeForFoco/solar-scorecard/pull/18" }
gharchive/pull-request
chartJS Stairstep Chart I thought I would try out chartJS and see how it compares to d3. This PR adds the "stairstep chart" that Matt and I talked about last week and a starting point for chartJS (stairstep.html). I also made some changes to where typescript gets compiled (js-compiled folder), and tested the npm install and npm run on a fresh checkout. Looks like everything is working. Nice work on automating all of that, @abettermap that makes development a lot nicer. chartJS made it pretty easy to this up. Yearly kW amounts and the goals plotted as a separate line. On vacation for a couple weeks but code looks good at a skim so I'll pull it in and look more closely later.
2025-04-01T04:54:51.128937
2016-09-19T02:43:53
177695423
{ "authors": [ "gladiatorAsh", "nilamdeka23" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13601", "repo": "CodeGladiators/CatchMeIfUCan", "url": "https://github.com/CodeGladiators/CatchMeIfUCan/issues/9" }
gharchive/issue
Game images Creation of game images like thief and houses Added thief image to the images folder, Resized thief image to fit requirement; to maintain aspect ratio with the game board Added game board(Game World) image
2025-04-01T04:54:51.138338
2023-08-11T08:16:40
1846401843
{ "authors": [ "jshergal", "martinchege324" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13603", "repo": "CodeMazeBlog/CodeMazeGuides", "url": "https://github.com/CodeMazeBlog/CodeMazeGuides/pull/1337" }
gharchive/pull-request
CM-714: Difference Between String, FormattableString, IFormattable in C# First draft @jshergal, everything should be fine now. @jshergal it should be fine now. @jshergal, I have applied your feedback; let me know if it's okay now. @jshergal, I have applied your feedback; let me know if it's okay now. Yes, the change in that unit test looks good. There is other feedback I have added that still needs to be addressed. One thing, in looking at the unit test where you made the recent changes, it feels to me like the check for the type should be the first assertion, since the other two assertions are based on the fact that the return value is actually a string. As a whole I am not sure the value that is added by asserting that the return value is a string, especially given that we are calling Contains on the return value, which will basically fail to compile if the object was not a string (with the exception of course that some other return type could have a Contains method). As a whole, if we want to keep that assertion, then from a logical flow perspective it should be the first assertion. We can think of the ordering in this way: Setup Run some operation which gives a result Verify the result is of an expected type Now verify that the result value matches what is expected @jshergal, it's sorted; thanks for your feedback. @jshergal, let me know if I have missed anything. @jshergal, it should be fine now
2025-04-01T04:54:51.156450
2024-07-10T16:02:36
2401148925
{ "authors": [ "shehabattia96" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13604", "repo": "CodeToCAD/CodeToCAD", "url": "https://github.com/CodeToCAD/CodeToCAD/issues/344" }
gharchive/issue
AST Parse @supported decorators and add supported information for each provider to the docs. Extension of #336 . Now that every provider has the @supported decorator, we need to pull the information into the docs and intellisense syntax highlighting We'll hold off on the intellisense highlighting - this could be done in an IDE extension instead.
2025-04-01T04:54:51.188536
2021-01-05T17:57:23
779361033
{ "authors": [ "codecademydev", "dreamwasp" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13605", "repo": "Codecademy/client-modules", "url": "https://github.com/Codecademy/client-modules/pull/1267" }
gharchive/pull-request
feat(Interstitial): add focus mgmt to component Overview This PR adds focus management to our Interstitial components. Basically, when an Interstitial pops up, the focus will be moved to the heading of that Interstitial. This mimics the way Links works when using screen readers. I will also make it explicit in the description for this component that it provides focus management and that only one interstitial should be rendered at a time. PR Checklist [ ] Related to designs: [x] Related to JIRA ticket: LX-4518 [ ] I have run this code to verify it works [ ] This PR includes unit tests for the code change 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.3.8-alpha.45086a.0@codecademy/gamut-labs@7.6.4-alpha.45086a.0@codecademy/gamut@19.8.1-alpha.45086a.0<EMAIL_ADDRESS> 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.3.8-alpha.45086a.0@codecademy/gamut-labs@7.6.4-alpha.45086a.0@codecademy/gamut@19.8.1-alpha.45086a.0<EMAIL_ADDRESS> 🚀 Styleguide deploy preview ready!https://5ff4df869b215c1ac7dd7233--gamut-preview.netlify.appDeploy Logs 🚀 Styleguide deploy preview ready!https://5ff4df869b215c1ac7dd7233--gamut-preview.netlify.appDeploy Logs 🚀 Styleguide deploy preview ready!https://5ff4e550ea6fa22627c9bc8e--gamut-preview.netlify.appDeploy Logs 🚀 Styleguide deploy preview ready!https://5ff4e550ea6fa22627c9bc8e--gamut-preview.netlify.appDeploy Logs 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.3.8-alpha.6fb3f4.0@codecademy/gamut-labs@7.6.4-alpha.6fb3f4.0@codecademy/gamut@19.8.1-alpha.6fb3f4.0<EMAIL_ADDRESS> 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.3.8-alpha.6fb3f4.0@codecademy/gamut-labs@7.6.4-alpha.6fb3f4.0@codecademy/gamut@19.8.1-alpha.6fb3f4.0<EMAIL_ADDRESS> 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.3.8-alpha.3369df.0@codecademy/gamut-labs@7.6.4-alpha.3369df.0@codecademy/gamut@19.8.1-alpha.3369df.0<EMAIL_ADDRESS> 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.3.8-alpha.3369df.0@codecademy/gamut-labs@7.6.4-alpha.3369df.0@codecademy/gamut@19.8.1-alpha.3369df.0<EMAIL_ADDRESS> 🚀 Styleguide deploy preview ready!https://5ff4e7c816e5d6265a928f46--gamut-preview.netlify.appDeploy Logs 🚀 Styleguide deploy preview ready!https://5ff4e7c816e5d6265a928f46--gamut-preview.netlify.appDeploy Logs 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.3.9-alpha.0e6a50.0@codecademy/gamut-labs@7.6.5-alpha.0e6a50.0@codecademy/gamut@19.8.2-alpha.0e6a50.0<EMAIL_ADDRESS> 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.3.9-alpha.0e6a50.0@codecademy/gamut-labs@7.6.5-alpha.0e6a50.0@codecademy/gamut@19.8.2-alpha.0e6a50.0<EMAIL_ADDRESS> 🚀 Styleguide deploy preview ready!https://5ff5e1e0efedb11242ec3a79--gamut-preview.netlify.appDeploy Logs 🚀 Styleguide deploy preview ready!https://5ff5e1e0efedb11242ec3a79--gamut-preview.netlify.appDeploy Logs 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.3.9-alpha.775576.0@codecademy/gamut-labs@7.6.5-alpha.775576.0@codecademy/gamut@19.8.2-alpha.775576.0<EMAIL_ADDRESS> 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.3.9-alpha.775576.0@codecademy/gamut-labs@7.6.5-alpha.775576.0@codecademy/gamut@19.8.2-alpha.775576.0<EMAIL_ADDRESS> 🚀 Styleguide deploy preview ready!https://5ff5f62b8e93ac09f459ef60--gamut-preview.netlify.appDeploy Logs 🚀 Styleguide deploy preview ready!https://5ff5f62b8e93ac09f459ef60--gamut-preview.netlify.appDeploy Logs 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.3.9-alpha.9e0c67.0@codecademy/gamut-labs@7.6.5-alpha.9e0c67.0@codecademy/gamut@19.8.2-alpha.9e0c67.0<EMAIL_ADDRESS> 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.3.9-alpha.9e0c67.0@codecademy/gamut-labs@7.6.5-alpha.9e0c67.0@codecademy/gamut@19.8.2-alpha.9e0c67.0<EMAIL_ADDRESS> 🚀 Styleguide deploy preview ready!https://5ff695789dc17143696415fb--gamut-preview.netlify.appDeploy Logs 🚀 Styleguide deploy preview ready!https://5ff695789dc17143696415fb--gamut-preview.netlify.appDeploy Logs 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.4.3-alpha.93a3e2.0@codecademy/gamut-labs@7.10.2-alpha.93a3e2.0@codecademy/gamut@19.10.2-alpha.93a3e2.0<EMAIL_ADDRESS> 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.4.3-alpha.93a3e2.0@codecademy/gamut-labs@7.10.2-alpha.93a3e2.0@codecademy/gamut@19.10.2-alpha.93a3e2.0<EMAIL_ADDRESS> 🚀 Styleguide deploy preview ready!https://6002301301aff12dccdc7086--gamut-preview.netlify.appDeploy Logs 🚀 Styleguide deploy preview ready!https://6002301301aff12dccdc7086--gamut-preview.netlify.appDeploy Logs 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.4.4-alpha.74606e.0@codecademy/gamut-labs@7.10.4-alpha.74606e.0@codecademy/gamut@19.10.3-alpha.74606e.0<EMAIL_ADDRESS> 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.4.4-alpha.74606e.0@codecademy/gamut-labs@7.10.4-alpha.74606e.0@codecademy/gamut@19.10.3-alpha.74606e.0<EMAIL_ADDRESS> 🚀 Styleguide deploy preview ready!https://6007153131a9423d2cf58977--gamut-preview.netlify.appDeploy Logs 🚀 Styleguide deploy preview ready!https://6007153131a9423d2cf58977--gamut-preview.netlify.appDeploy Logs 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.4.4-alpha.5355e1.0@codecademy/gamut-labs@7.10.4-alpha.5355e1.0@codecademy/gamut@19.10.3-alpha.5355e1.0<EMAIL_ADDRESS> 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.4.4-alpha.5355e1.0@codecademy/gamut-labs@7.10.4-alpha.5355e1.0@codecademy/gamut@19.10.3-alpha.5355e1.0<EMAIL_ADDRESS> 🚀 Styleguide deploy preview ready!https://6007178d2d238c3a4a965bb9--gamut-preview.netlify.appDeploy Logs 🚀 Styleguide deploy preview ready!https://6007178d2d238c3a4a965bb9--gamut-preview.netlify.appDeploy Logs 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.4.4-alpha.ff4424.0@codecademy/gamut-labs@7.10.4-alpha.ff4424.0@codecademy/gamut@19.10.3-alpha.ff4424.0<EMAIL_ADDRESS> 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.4.4-alpha.ff4424.0@codecademy/gamut-labs@7.10.4-alpha.ff4424.0@codecademy/gamut@19.10.3-alpha.ff4424.0<EMAIL_ADDRESS> 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.4.4-alpha.e3a98a.0@codecademy/gamut-labs@7.10.4-alpha.e3a98a.0@codecademy/gamut@19.10.3-alpha.e3a98a.0<EMAIL_ADDRESS> 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.4.4-alpha.e3a98a.0@codecademy/gamut-labs@7.10.4-alpha.e3a98a.0@codecademy/gamut@19.10.3-alpha.e3a98a.0<EMAIL_ADDRESS> 🚀 Styleguide deploy preview ready!https://60071c0318447343bb5953ab--gamut-preview.netlify.appDeploy Logs 🚀 Styleguide deploy preview ready!https://60071c0318447343bb5953ab--gamut-preview.netlify.appDeploy Logs 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.4.4-alpha.d3a4e7.0@codecademy/gamut-labs@7.10.4-alpha.d3a4e7.0@codecademy/gamut@19.10.3-alpha.d3a4e7.0<EMAIL_ADDRESS> 📬Published Alpha Packages:@codecademy/gamut-illustrations@0.4.4-alpha.d3a4e7.0@codecademy/gamut-labs@7.10.4-alpha.d3a4e7.0@codecademy/gamut@19.10.3-alpha.d3a4e7.0<EMAIL_ADDRESS> 🚀 Styleguide deploy preview ready!https://600722d7ced2c060f42a8ee4--gamut-preview.netlify.appDeploy Logs 🚀 Styleguide deploy preview ready!https://600722d7ced2c060f42a8ee4--gamut-preview.netlify.appDeploy Logs 📬Published Alpha Packages:@codecademy/eslint-config@3.1.1-alpha.b52bc1.0@codecademy/gamut-illustrations@0.4.4-alpha.b52bc1.0@codecademy/gamut-labs@7.10.4-alpha.b52bc1.0@codecademy/gamut@19.10.3-alpha.b52bc1.0@codecademy/macros@1.1.2-alpha.b52bc1.0<EMAIL_ADDRESS> 📬Published Alpha Packages:@codecademy/eslint-config@3.1.1-alpha.b52bc1.0@codecademy/gamut-illustrations@0.4.4-alpha.b52bc1.0@codecademy/gamut-labs@7.10.4-alpha.b52bc1.0@codecademy/gamut@19.10.3-alpha.b52bc1.0@codecademy/macros@1.1.2-alpha.b52bc1.0<EMAIL_ADDRESS> 🚀 Styleguide deploy preview ready!https://600726d434af4847fb52d066--gamut-preview.netlify.appDeploy Logs 🚀 Styleguide deploy preview ready!https://600726d434af4847fb52d066--gamut-preview.netlify.appDeploy Logs
2025-04-01T04:54:51.291426
2017-06-14T15:06:10
235911831
{ "authors": [ "Demonstrandum", "meiamsome" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13607", "repo": "CodingTrain/Rainbow-Code", "url": "https://github.com/CodingTrain/Rainbow-Code/issues/178" }
gharchive/issue
Coding Challenge: Plotting electron occurrence probability of a hydrogen atom using the hydrogen quantum wave function. Create a hydrogen's electron density plot at different energies, using the hydrogen wave function. Resources: https://en.wikipedia.org/wiki/Hydrogen_atom#Schr.C3.B6dinger_equation https://en.wikipedia.org/wiki/Wave_function http://hyperphysics.phy-astr.gsu.edu/hbase/quantum/hydwf.html Like so, where a brighter colour means a greater likelihood of finding an electron there: Hello, this repository is for the code from tutorials only, please but all topic suggestions/requests in https://github.com/CodingTrain/Rainbow-Topics thanks. Thanks
2025-04-01T04:54:51.292320
2019-02-27T02:09:26
414910561
{ "authors": [ "gruselhaus", "jcponce" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13608", "repo": "CodingTrain/website", "url": "https://github.com/CodingTrain/website/pull/1167" }
gharchive/pull-request
Adding my contribution I just re wrote it in p5.js and added the noise to the each particle. :) ❤️
2025-04-01T04:54:51.300400
2019-04-17T00:05:23
434035471
{ "authors": [ "gruselhaus", "shiffman", "vladig98" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13609", "repo": "CodingTrain/website", "url": "https://github.com/CodingTrain/website/pull/1282" }
gharchive/pull-request
Code refactored and Moves implemented I refactored the code so that all rotations are in one function. I've implemented the real moves of the Rubik's cube. Colors of the cube updated. I love this idea! Thank you for taking the time to do this work! I like to keep the code in this repo matching the video so that viewers can follow along. I just posted the same thoughts in #1282 but I'll repeat here anyway! You could submit your own version as a community contribution. I could create a community "Rubik's Cube" project separate from the website that viewers could develop. I like doing this, but it often becomes difficult to manage. What do you think? (I'll also note that I am releasing part 3 tomorrow). Hi Dan, The reason behind this pull request is mostly the implementation of the moves and updating the colors. Also you said that you wanted to refactoring the rotation functions so I did that (not saying that this is the best way to do it but..... anyways it's there). So please think about implementing the moves for part 4 so that it will be easier once you try to do the AI (if you intend to use any algorithms of course). And yes I'll try to submit my version on the web. Thank you You're amazing. @shiffman I updated part 3. Please take a look. This is amazing! Could it be posted as a community contribution and linked from the challenge? Then I could refer to it if / when I continue the challenge. This is amazing! Could it be posted as a community contribution and linked from the challenge? Then I could refer to it if / when I continue the challenge. Yes, of course. But please do continue this series. The coding challenge turns out great. I know that the AI part might be difficult but you'll figure it out. You always do. I'd like to leave this open for a little while pending a new pull request with a link to the code this code or even just a diff? This will be a helpful reminder if/when I return to the project! Implemented along https://github.com/CodingTrain/website/pull/1373
2025-04-01T04:54:51.331731
2023-03-09T03:16:57
1616328365
{ "authors": [ "jyooi" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13610", "repo": "CoinAlpha/hbui", "url": "https://github.com/CoinAlpha/hbui/issues/3" }
gharchive/issue
Hbui Modal component not pop up even the showOpen props is true I recently tried out the Modal component with the passive trigger props showOpen as a true value and the model is not popping up. it resolved by updating the hbui version
2025-04-01T04:54:51.337760
2017-03-21T11:09:00
215701670
{ "authors": [ "KrzysiekJ", "blackstat" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13611", "repo": "CoinBlack/blackcoin", "url": "https://github.com/CoinBlack/blackcoin/issues/2" }
gharchive/issue
OP_RETURN outputs with zero value are rejected OP_RETURN outputs with zero value are nonstandard and are not accepted to enter the mempool. ‘ERROR: AcceptToMemoryPool : nonstandard transaction: dust’. Fees are paid by other outputs. Flood prevention should be done using fee policy (vide the idea of dynamic minimum transaction fees). Enforcing minimum transaction value would reasonably work only for OP_RETURN transactions.
2025-04-01T04:54:51.346799
2022-10-04T15:39:02
1396487273
{ "authors": [ "Chweyaa", "DaryKeruita", "Dimfred" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13612", "repo": "Cojodi/openseapy", "url": "https://github.com/Cojodi/openseapy/issues/1" }
gharchive/issue
Disconnects Randomly Can't run for a day straight After a few hours randomly disconnects. @DaryKeruita so a big part of the issues was probably related to the streaming models, apparently they are very different on the mainnet, I have only tested them on testnet. I added now some basic reconnection mechanisms found two events which I have found only on mainnet rightnow. And I run the stream and monitor it for errors the next days and see how it plays out. I guess it should work better now. Ping me if you find issues preferably with logs next time. Okay so I found some other issues, as of rightnow I think while I can reconnect to the stream, probably every subscriptions has to be reapplied, I think it will take some time when I can fix this. Hi. Is this fixed? Hi I think it is running quite stable for some time now, so I'd say yes? I mean it still diconnects, but then tries to reconnect, and that can take some time. Sometimes a couple minutes, but then resumes as before. But thats more an issue with opensea not with the lib (I think). Ping me once you have tested it.
2025-04-01T04:54:51.355322
2023-11-30T16:38:02
2019054132
{ "authors": [ "ColinFay", "jospueyo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13613", "repo": "ColinFay/webrtools", "url": "https://github.com/ColinFay/webrtools/issues/1" }
gharchive/issue
Error installing packages After reading your posts, I'm experimenting with webR in node.js. However, when trying to use your package to install R packages, it raise an error: $ Rscript ./node_modules/webrtools/r/install.R dplyr Skipping install of 'webrtools' from a github remote, the SHA1 (063164d4) has not changed since last install. Use `force = TRUE` to force installation Error in contrib.url(repos, type) : trying to use CRAN without setting a mirror Calls: <Anonymous> ... <Anonymous> -> <Anonymous> -> unique -> contrib.url Ejecución interrumpida I'm working on Windows 11 and R 4.3.2. Let me know if I can help anyhow. Closed via #2
2025-04-01T04:54:51.406705
2022-05-21T07:01:11
1243893706
{ "authors": [ "deepankar-sajwan" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13614", "repo": "ColoredCow/portal", "url": "https://github.com/ColoredCow/portal/issues/1441" }
gharchive/issue
When adding team members in the project details error occurs if designation of team member not added Describe the bug In updating project details , when adding details of the team members if we do not select designation of the team member, and save the form error is displayed on the screen. Refer screenshot The error message still exist The issue has been fixed
2025-04-01T04:54:51.420794
2018-03-29T14:34:23
309779595
{ "authors": [ "CLAassistant", "maxmellen" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13615", "repo": "Comcast/dialyzex", "url": "https://github.com/Comcast/dialyzex/pull/8" }
gharchive/pull-request
Do not crash when no warnings are emitted Hi there 👋 This PR aims to address the bug mentioned in #7. Best 🤗 Thank you for your submission, we really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
2025-04-01T04:54:51.422770
2023-06-13T08:16:46
1754308251
{ "authors": [ "CLAassistant", "barthap" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13616", "repo": "CommE2E/comm", "url": "https://github.com/CommE2E/comm/pull/239" }
gharchive/pull-request
[CI] Add blob unit tests GH action Test CI Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
2025-04-01T04:54:51.438536
2024-05-20T14:17:02
2306102438
{ "authors": [ "cameronmore", "gregfowlerphd", "swartik" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13617", "repo": "CommonCoreOntology/CommonCoreOntologies", "url": "https://github.com/CommonCoreOntology/CommonCoreOntologies/issues/253" }
gharchive/issue
Artifact Function subclass justifications Issue #251 made me wonder about the many descendants of class Artifact Function and the rationale for including all of them in CCO. I'm not saying they're unnecessary classes, just questioning whether they should be in a mid-level ontology. Consider: Only two descendants of Artifact Function, Payload Capacity and Sensor Modality Function, have inheres-in subclass restrictions. I haven't examined most of the class definition annotations closely, but the ones I've seen seldom explicitly mention the class of artifact used in the function. CCO offers lots of functions to express how an artifact is used, but it does not guide modelers in how to tie functions to artifacts. This is an important design decision with significant consequences for the content of a mid-level ontology. Was this design decision deliberate or organic? It seems like there should be a roughly one-to-one correspondence of artifacts to their functions in the ontologies. If there are more of one than the other, or significant underlap, then this is worth investigating. Why include a function but not include (at least one) corresponding artifact? Same with artifacts--at least one vehicle-related function should exist, even if there are a number of vehicle subclasses with no more specific functions. Reflecting the decision to under-axiomatize the ontology to allow users to craft their own data models with less restriction, the subclass restrictions seems to be an oversight and should be deleted to align with the other functions. CCO offers lots of functions to express how an artifact is used, but it does not guide modelers in how to tie functions to artifacts. This is an important design decision with significant consequences for the content of a mid-level ontology. Was this design decision deliberate or organic? I'm not sure what you mean here. CCO follows the BFO pattern, that material entities bear functions, and those are realized in processes, and they may be 'used-by' agents in those processes, and further the usage of these artifacts may produce certain cco:effects. @cameronmore: I don't want to speak for @swartik here, but I think what he means is: Unlike his proposed definition of 'Life Support Artifact Function', the definitions for subclasses of Artifact Function don't mention subclasses of Material Artifact (and vice versa). The subclasses of Artifact Function don't have associated axioms linking them to subclasses of Material Artifact (and vice versa). @swartik this is an interesting issue. My own personal take: As an artifact with a history, the ontologies developed more in some areas than others. There are terms which we all think are too specific for a mid level ontology (looking at you cco:Flywheel), but they're there, and I personally hope that domain ontologies take over some specific terms for which they may be the better and more authoritative curator. Regarding the perhaps functional relationship between functions and artifacts, I personally think that the way the terms are defined now allows the ontologies to follow the construction and experimentation of real world artifacts, not the other way around. Every infomercial seems to invent a new artifact which solves an old problem in some unique way. Now, the latest vegetable mincer can bear the same 'Food Slicing Artifact Function' as the other five mincers on the market, while being its own unique artifact (defined, perhaps, in terms of how it performs that function). Functions are realized through the manifestations, effects. Freezers, fridges, laboratory storage fridges, mini-coolers, ice-machinies, ice houses, air conditioners, etc., all share the same or highly similar functions, and I think the current taxonomy handles such variation well by not more directly linking them. If functions had normative statements in them, references to specific artifacts, then I can imagine users getting discourages when asserting that a cooling function inheres in something that's not a cooling system. On the other hand, you may say that the functions could be linked up to generic classes, rather than a specific cooling artifact, but just to the generic set of artifacts that cool. This is a good strategy, but I am concerned that it would still bake in too much rigidity into the models and subtract from some flexibility. Of course, the saying goes 'bend, don't break,' and I can't say whether this flexibility is a universal benefit or if there are steps we can take to shore up the artifact hierarchy or the function hierarchy. I can't speculate much without seeing more domains and their choices re function and artifact. @cameronmore I wish I could offer wisdom; I tend to think about artifact/function classes one at a time, on an as-needed basis. I think we've taken this discussion as far as is useful in the issues. I'll close it.
2025-04-01T04:54:51.440826
2022-01-05T06:27:44
1094024704
{ "authors": [ "Bruzay", "Evavic44" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13618", "repo": "CommunityPro/support", "url": "https://github.com/CommunityPro/support/issues/73" }
gharchive/issue
Please invite me to the Community Organization Name Samuel Ayoade Discord Username (if applicable) E.g John#412 No response Questions & suggestions? No response Welcome to Communitypro @Bruzay we are so pleased to have you join us 🙌🏾🤩
2025-04-01T04:54:51.459785
2024-05-22T21:06:47
2311483085
{ "authors": [ "rhmdnd", "yuumasato" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13619", "repo": "ComplianceAsCode/compliance-operator", "url": "https://github.com/ComplianceAsCode/compliance-operator/pull/522" }
gharchive/pull-request
Reduce timeout from 30 minutes to 10 minutes in serial tests Before, we were waiting for 30 minutes for a scan to timeout. 30 minutes is a long time, and we have some retry logic in the operator to rekick scans after a timeout. Let's reduce the timeout for tests to see if we can improve resilience to flaky scans. /test e2e-aws-serial Is this ironic? 2024/05/22 23:04:33 waiting until suite test-remove-profile-scan-binding reaches target status 'DONE'. Current status: RUNNING 2024/05/22 23:04:33 waiting until suite test-remove-profile-scan-binding reaches target status 'DONE'. Current status: RUNNING main_test.go:1766: timed out waiting for the condition --- FAIL: TestRemoveProfileScan (1800.04s)
2025-04-01T04:54:51.571333
2021-11-04T14:26:42
1044821677
{ "authors": [ "concordium-cl" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13620", "repo": "Concordium/concordium-reference-wallet-android", "url": "https://github.com/Concordium/concordium-reference-wallet-android/issues/54" }
gharchive/issue
Changing passcode/biometrics not possible when failed identities exist Description When entering the "Update Passcode & Biometrics" flow, a msg pops up: "All identities and accounts must be finalized before passcode can be changed. Please wait, then try again." All my accounts/identities are finalized but some identities failed. First after deleting those identity cards, the user is allowed to proceed. Might be good to tell the user to delete the failed identity cards as well. Steps to Reproduce let an identity fail more > update passcode & biometrics > continue Expected Result Msg to user should include to actively delete his failed identities. The sentence about waiting should be reworked. Actual Result Msg only talks about finalized and pending accounts/identities. It also tells the user to wait, which doesn't help in case of failed identities. Versions Concordium wallet 1.0.14 Android 9 Huawei y6s Add to message that failed identities must be manually deleted.
2025-04-01T04:54:51.575034
2024-06-03T07:49:38
2330423506
{ "authors": [ "Radiokot", "mh-concordium" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13621", "repo": "Concordium/cryptox-android", "url": "https://github.com/Concordium/cryptox-android/issues/58" }
gharchive/issue
Freshly installed app starts in a strange way To reproduce: Install new build 1.1.0-qa.5. Start the app. Result: It seems that app stops and starts again. Expected: App should just start. startApp.mp4 I can't reproduce it. Can you provide more details? Although I don't think this is a significant issue. I added a video. I could try it on another device. There is not much more to add, I just install fresh app (no seed phrase) and start the app. Hm, okay. I can also see it only on Android 11 for now. I marked it as won't do since it doesn't reproduce consistently.
2025-04-01T04:54:51.615096
2021-02-12T11:56:35
807177069
{ "authors": [ "CLAassistant", "darioAnongba" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13622", "repo": "ConsenSys/doc.common", "url": "https://github.com/ConsenSys/doc.common/pull/22" }
gharchive/pull-request
Exclude github orchestrate private repo Exclude github orchestrate private repo Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
2025-04-01T04:54:51.616802
2022-12-23T22:07:29
1509812287
{ "authors": [ "alexandratran" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13623", "repo": "ConsenSys/doctools.action-builder", "url": "https://github.com/ConsenSys/doctools.action-builder/pull/33" }
gharchive/pull-request
Enable snippets Add extensions to enable Snippets and code highlighting, in order to accommodate https://github.com/ConsenSys/doc.zk-evm/pull/7. Merging this now since everyone is on break and the bridge content in https://github.com/ConsenSys/doc.zk-evm/pull/7 must be merged by Christmas. This can be updated later if needed.
2025-04-01T04:54:51.621410
2019-06-21T20:16:21
459356786
{ "authors": [ "vanessabridge" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13624", "repo": "ConsenSys/wittgenstein", "url": "https://github.com/ConsenSys/wittgenstein/issues/59" }
gharchive/issue
POW Sim Build a protocol that simulates pow algorithms, uses difficulty calculations and includes uncles: For Uncle creation Before adding as uncle: [x] Validate that block received has the same parent as existing block at same height [x] Check block by height in blockchain and add as uncle if same parent and hasn’t been included and there are less than 2 uncles at that height [ ] Prevent uncles past height+6 to be added Refactoring [x] Remove uncle reference from the node Mining [x] Adjust difficulty calculation to include uncles [x] Validate calculations using real data from main net in a test [ ] Create reward functions that allocates funds for successfully mining: 1) block + transaction fees 2) an uncle ** Data collection** [ ] Generate stats for uncles and record them in a separate file (csv) [ ] Select minimal parameters necessary to feed to RL agent [ ] Set method that calls agent at decision points Generate data and store it outside with minimum parameters necessary #58
2025-04-01T04:54:51.631445
2020-09-23T05:40:31
707063132
{ "authors": [ "caiyuan1222", "davidmjiang" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13625", "repo": "ContactAssistApp/dashboard", "url": "https://github.com/ContactAssistApp/dashboard/pull/62" }
gharchive/pull-request
addTimeZone Admin can select the time zone when they create an announcement, so we can use this time zone to change as UTC string. Only take a four-time zone across U.S. now. When loading the page, we convert all those time as local time. I cannot create the new announcement, are we still allowed to create now? Want to test the landing page. What issue are you running into when trying to create the announcement? A few observations: Need to pass in timezone to the Card components in CreatePsa.jsx so that it shows up in the mobile preview and confirmation page. Timezones are hardcoded as PDT, EDT, etc. What happens when daylight savings ends in November? I think you may want do some research into what external libraries or browser APIs are available to help with this A few observations: Need to pass in timezone to the Card components in CreatePsa.jsx so that it shows up in the mobile preview and confirmation page. Timezones are hardcoded as PDT, EDT, etc. What happens when daylight savings ends in November? I think you may want do some research into what external libraries or browser APIs are available to help with this Hello @davidmjiang For the question I believe I have added timeZone in createPSA file. To my understanding, as user provides timeZone, javascript could detect the time saving. Attach a picture to better understand: The senario is user select 9:00 am to 6:00pm in EST time, which is 3 hours ahead for the PDT time. Since I was in PDT time, so the card will show the time based on my location and also indicate the time zone. I am trying to understand the picture. Is the EDT time in gray shown for testing purposes and 6:00am-3:00pm is what actually shows up on the card? As for daylight savings, it will change from PDT for example to PST, so will we handle that? yes, David! the gray is when I create the announcement, I take 9:00 am to 6:00pm EDT. Since I was in PDT time, so the card will show 6:00am to 3:00pm PDT (the card will always show the local time). Oh, I see. after day saving, it should be PST? I did not realize it will change as PST. will add more conditions. https://stackoverflow.com/questions/11887934/how-to-check-if-dst-daylight-saving-time-is-in-effect-and-if-so-the-offset add link for better explanation
2025-04-01T04:54:51.633874
2022-02-24T12:10:32
1149212151
{ "authors": [ "MukeshS-hexaware", "amouat" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13626", "repo": "ContainerSolutions/trow", "url": "https://github.com/ContainerSolutions/trow/issues/317" }
gharchive/issue
Request to release v0.3.5 I am currently using the Helm chart version of Trow, and when I deploy Trow to AKS and try to pull images from Trow, I'm facing this error: Failed to pull image "trow/test:latest": rpc error: code = FailedPrecondition desc = failed to pull and unpack image "trow/test:latest": failed commit on ref "layer-sha256:60324950a964e825b7fa78043fbe2330be484170e14f6ae6b7baeb4950f4bc9b": "layer-sha256:60324950a964e825b7fa78043fbe2330be484170e14f6ae6b7baeb4950f4bc9b" failed size validation: 318653 != 318654: failed precondition I'm thinking this has been already fixed in e9e2a83de4da77962aef0728ec2e34a6e678fc98, which has not been included in any release yet. Please publish a new release with this commit included. Sure, I'll take a look later. Hi, any updates on this? Hey, this should be done now. Apologies for the delay, I was having some trouble with releases and new machines. Let me know if there are any issues!
2025-04-01T04:54:51.668196
2018-09-06T14:41:41
357687855
{ "authors": [ "jbednar", "martindurant", "mmccarty", "seibert" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13627", "repo": "ContinuumIO/intake", "url": "https://github.com/ContinuumIO/intake/issues/151" }
gharchive/issue
Configure Intake to cache files in the current directory From @jbednar In most cases I'd rather have the files associated with the notebook or project where I'm working, not dumped into an ever-increasing global cache heap. I would be nice to have a simple flag for configuring intake (something intake.local_cache=True) to store cached items in a subdirectory like .intake_cache/ in the current directory. I hadn't thought of this usage model, but I guess it would work in some cases. I would definitely not want to make it the default behavior, as these cache directories are both invisible and hard for a CLI to discover without scanning your whole hard drive. This sounds pretty reasonable to me. The CLI would get file locations from the metadata, which would be small and remain in the (global) config directory. I would suggest config['cache_dir'] = "datadir" could do this (better names accepted). Yeah, the CLI could look at the metadata which should avoid the need to scan the hard drive. I would definitely not want to make it the default behavior, as these cache directories are both invisible and hard for a CLI to discover without scanning your whole hard drive. It might make sense to make the directory not be invisible, e.g. ./intake_cache/; there are pros and cons to doing that. But there are problems with both approaches, i.e. putting cache files in cwd and in the home directory; in one case it's clear what would be affected by deleting that file, but hard to find all of them (though one can do find ~ -name intake_cache), and in the other it's easy to find all of them but hard to tell what any particular file is and whether it can be deleted without a major download cost in your future. Yeah, the CLI could look at the metadata which should avoid the need to scan the hard drive. That makes sense -- the metadata should be tiny in any reasonable case, and presumably no one should object to having that in their homedir.
2025-04-01T04:54:51.676928
2020-08-06T19:44:00
674557982
{ "authors": [ "CreeperPlanet26" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13628", "repo": "Cool-Bot-List/Cool-Bot-List", "url": "https://github.com/Cool-Bot-List/Cool-Bot-List/pull/55" }
gharchive/pull-request
POST api/users/token/:id I still need Zyleaf to make the logic that creates the token. I only setup the boilerplate for it. Should we change it to no 7 day thing This is ready for merge
2025-04-01T04:54:51.724542
2022-07-31T19:49:14
1323559002
{ "authors": [ "CoopPlayzz-Bot" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13629", "repo": "CoopPlayzz/Webstatus-electronicbrands", "url": "https://github.com/CoopPlayzz/Webstatus-electronicbrands/issues/2301" }
gharchive/issue
⚠️ Samsung has degraded performance In 6452b5b, Samsung (https://samsung.com) experienced degraded performance: HTTP code: 200 Response time: 3027 ms Resolved: Samsung performance has improved in 32825fa.
2025-04-01T04:54:51.727348
2022-05-07T21:23:14
1228726060
{ "authors": [ "CoopPlayzz-Bot" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13630", "repo": "CoopPlayzz/Webstatus-electronicbrands", "url": "https://github.com/CoopPlayzz/Webstatus-electronicbrands/issues/827" }
gharchive/issue
⚠️ Samsung has degraded performance In f34ad1d, Samsung (https://samsung.com) experienced degraded performance: HTTP code: 200 Response time: 2181 ms Resolved: Samsung performance has improved in 7502ec0.
2025-04-01T04:54:51.729828
2022-07-26T07:50:35
1317841472
{ "authors": [ "CoopPlayzz-Bot" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13631", "repo": "CoopPlayzz/Webstatus-games", "url": "https://github.com/CoopPlayzz/Webstatus-games/issues/1380" }
gharchive/issue
⚠️ Fortnite has degraded performance In cdfa796, Fortnite (https://fortnite.com) experienced degraded performance: HTTP code: 200 Response time: 1599 ms Resolved: Fortnite performance has improved in 0f955ba.
2025-04-01T04:54:51.731749
2022-12-30T11:47:19
1514466272
{ "authors": [ "CoopPlayzz-Bot" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13632", "repo": "CoopPlayzz/Webstatus-games", "url": "https://github.com/CoopPlayzz/Webstatus-games/issues/2718" }
gharchive/issue
⚠️ Fortnite has degraded performance In 7b65845, Fortnite (https://fortnite.com) experienced degraded performance: HTTP code: 200 Response time: 1556 ms Resolved: Fortnite performance has improved in af47af9.
2025-04-01T04:54:51.735864
2016-07-07T11:11:58
164285212
{ "authors": [ "CooperRS", "dafdafydd" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13634", "repo": "CooperRS/RMDateSelectionViewController", "url": "https://github.com/CooperRS/RMDateSelectionViewController/issues/70" }
gharchive/issue
Dismiss parent view controller after picking the time Hello, I wonder if it is possible to dismiss the picker and the parent view controller at the same time. I experimented with putting dismissViewControllerAnimated in the Done action but it does not dismiss the parent - only RMDateSelectionViewController disappears. Is there any way it could be done? Hi there, I think you can do this by calling dismissViewControllerAnimated twice. However, this might be bad style :D. Best regards, Roland Thank you! This is how I overcome it for now but I wonder if there a "cleaner" way to do it.
2025-04-01T04:54:51.743164
2022-02-20T23:40:55
1145169463
{ "authors": [ "Raychanan", "jpwchang" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13635", "repo": "CornellNLP/Cornell-Conversational-Analysis-Toolkit", "url": "https://github.com/CornellNLP/Cornell-Conversational-Analysis-Toolkit/issues/145" }
gharchive/issue
Sample notebook reports errors Hi, many thanks for the development of the great package! I'm trying to run this sample notebook Predicting Conversations Gone Awry With Convokit on Google Colab here. I did no modifications except for the first chunk I added ! pip -q install convokit ! pip uninstall spacy -y ! pip install -U spacy==3.1.4 !python -m spacy download en_core_web_sm However, an error occurred in the second cell from the bottom: TypeError: __init__() takes from 1 to 2 positional arguments but 3 were given. Would it be possible for you to point out how to correct the error? Many thanks! Running prediction task for feature set politeness_strategies Generating labels... Computing paired features... Using 38 features Running leave-one-page-out prediction... --------------------------------------------------------------------------- RemoteTraceback Traceback (most recent call last) RemoteTraceback: """ Traceback (most recent call last): File "/usr/lib/python3.7/multiprocessing/pool.py", line 121, in worker result = (True, func(*args, **kwds)) File "/usr/lib/python3.7/multiprocessing/pool.py", line 44, in mapstar return list(map(*args)) File "<ipython-input-37-de914fca85cc>", line 11, in run_pred_single base_clf = Pipeline([("scaler", StandardScaler()), ("featselect", SelectPercentile(f_classif, 10)), ("logreg", LogisticRegression(solver='liblinear'))]) TypeError: __init__() takes from 1 to 2 positional arguments but 3 were given """ The above exception was the direct cause of the following exception: TypeError Traceback (most recent call last) [<ipython-input-38-9704095ec82e>](https://localhost:8080/#) in <module>() 4 for combo in feature_combos: 5 combo_names.append("+".join(combo).replace("_", " ")) ----> 6 accuracy = run_pipeline(combo) 7 accs.append(accuracy) 8 results_df = pd.DataFrame({"Accuracy": accs}, index=combo_names) 6 frames [<ipython-input-37-de914fca85cc>](https://localhost:8080/#) in run_pipeline(feature_set) 97 y = labeled_pairs_df.first_convo_toxic.values 98 print("Running leave-one-page-out prediction...") ---> 99 accuracy, coefs, scores, hyperparams, pvalue = run_pred(X, y, feature_names, labeled_pairs_df.page_id) 100 print("Accuracy:", accuracy) 101 print("p-value: %.4e" % pvalue) [<ipython-input-37-de914fca85cc>](https://localhost:8080/#) in run_pred(X, y, fnames, groups) 33 34 with Pool(os.cpu_count()) as p: ---> 35 prediction_results = p.map(partial(run_pred_single, X=X, y=y), splits) 36 37 fselect_pvals_all = [] [/usr/lib/python3.7/multiprocessing/pool.py](https://localhost:8080/#) in map(self, func, iterable, chunksize) 266 in a list that is returned. 267 ''' --> 268 return self._map_async(func, iterable, mapstar, chunksize).get() 269 270 def starmap(self, func, iterable, chunksize=None): [/usr/lib/python3.7/multiprocessing/pool.py](https://localhost:8080/#) in get(self, timeout) 655 return self._value 656 else: --> 657 raise self._value 658 659 def _set(self, i, obj): [/usr/lib/python3.7/multiprocessing/pool.py](https://localhost:8080/#) in worker() 119 job, i, func, args, kwds = task 120 try: --> 121 result = (True, func(*args, **kwds)) 122 except Exception as e: 123 if wrap_exception and func is not _helper_reraises_exception: [/usr/lib/python3.7/multiprocessing/pool.py](https://localhost:8080/#) in mapstar() 42 43 def mapstar(args): ---> 44 return list(map(*args)) 45 46 def starmapstar(args): [<ipython-input-37-de914fca85cc>](https://localhost:8080/#) in run_pred_single() 9 y_train, y_test = y[train_idx], y[test_idx] 10 ---> 11 base_clf = Pipeline([("scaler", StandardScaler()), ("featselect", SelectPercentile(f_classif, 10)), ("logreg", LogisticRegression(solver='liblinear'))]) 12 clf = GridSearchCV(base_clf, {"logreg__C": [10**i for i in range(-4,4)], "featselect__percentile": list(range(10, 110, 10))}, cv=3) 13 TypeError: __init__() takes from 1 to 2 positional arguments but 3 were given Hi @Raychanan, It appears that this is caused by a change to scikit-learn's SelectPercentile class in the 1.x scikit-learn release. I've committed an updated version of the notebook to deal with this change. The change is small, so if you don't want to re-upload the notebook to colab from scratch, you can simply change one line in your existing colab notebook. Find the following line: base_clf = Pipeline([("scaler", StandardScaler()), ("featselect", SelectPercentile(f_classif, 10)), ("logreg", LogisticRegression(solver='liblinear'))]) And change it to: base_clf = Pipeline([("scaler", StandardScaler()), ("featselect", SelectPercentile(score_func=f_classif, percentile=10)), ("logreg", LogisticRegression(solver='liblinear'))]) That should resolve the error! This helps a lot! Thanks so much!
2025-04-01T04:54:51.775420
2024-08-02T09:10:21
2444478726
{ "authors": [ "JoelLourenco" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13636", "repo": "Cosmo-Tech/terraform-kubernetes-cosmotech-tenant", "url": "https://github.com/Cosmo-Tech/terraform-kubernetes-cosmotech-tenant/pull/24" }
gharchive/pull-request
Jlou/feat analyse variables Description of the PR: Jira ticket ref : Jira ticket In order to easily change variables values, a refactoring of values has been made. The majority of hardcoded values in child modules now use variables and all default variables values have been moved to root variables. Variables for the remote module Cosmo-Tech/platform-core/cosmotech have been created, as part of the refactoring of the corresponding repository. All root variables have been sorted into module-related variables.tf files Format fixed, but I can't rename branch to azure
2025-04-01T04:54:51.779367
2021-08-02T12:07:31
958075660
{ "authors": [ "Highlander-maker", "JakeHartnell", "blockcreators", "nullmames" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13637", "repo": "CosmosContracts/hack-juno", "url": "https://github.com/CosmosContracts/hack-juno/issues/15" }
gharchive/issue
[Nomination] giansalex for telegram bot Who are you nominating? @giansalex (Giancarlos Salas) What are you nominating them for? His incredible work on Juno Telegram Bot and creation of the code that allows a user to upload a wasm file on chain Please provide links to their work. https://github.com/Highlander-maker/juno-tg-bot Nice. Would be cool if it supported token gated groups. I.E. own a certain amount of cw20 tokens to gain access to a telegram group. interesting... So smart contract that checks the balance of a wallet that knows how much of the token that can gain access ? interesting... So smart contract that checks the balance of a wallet that knows how much of the token that can gain access ? You shouldn't need to write any custom smart contact. Can be handled entirely by the bot. Basically, have the user connect their wallet, get the balance from the cw20 token contract. If the user's amount is larger than the minimum set in bot config, then they get access to the token gated channel. you should have also nominated yourself @Highlander-maker. you have spent a great amount of time on the project and started it also. @JakeHartnell ok Jake. Let me look in to it. :) @JakeHartnell pretty sure omniflix have this working already for telegram: juno.omniflix.com @JakeHartnell pretty sure omniflix have this working already for telegram: juno.omniflix.com If they open source it, will nominate them. They deserve some $JUNO for that. Does theirs work with cw20s? This sounds like https://collab.land/ @giansalex has been one of the most outstanding contributors during Hack Juno. :clap: :clap: :clap: This ticket is a duplicate of his main nomination #13. So closing, but will link to this issue from there.
2025-04-01T04:54:51.854748
2014-12-31T12:24:36
53181656
{ "authors": [ "rippler", "robby-dermody" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13638", "repo": "CounterpartyXCP/counterpartyd_build", "url": "https://github.com/CounterpartyXCP/counterpartyd_build/issues/172" }
gharchive/issue
Problem with Fed Node installer from develop branch 2014-12-31 12:18:20,353|DEBUG: RUNNING COMMAND: cd /home/xcp/counterpartyd_build && git pull origin develop From https://github.com/CounterpartyXCP/counterpartyd_build * branch develop -> FETCH_HEAD *** Please tell me who you are. Run git config --global user.email<EMAIL_ADDRESS> git config --global user.name "Your Name" to set your account's default identity. Omit --global to set the identity only in this repository. fatal: unable to auto-detect email address (got 'root@ubuntu.(none)') 2014-12-31 12:18:22,020|ERROR: Command failed: 'cd /home/xcp/counterpartyd_build && git pull origin develop' I thnk so...it's a public repo so it shouldn't ask you for that, unless it thinks you are somehow committing...I'll close this and if it's an issue again, reopen it Ah, yes, now I know why - prolly because of https://github.com/CounterpartyXCP/counterwallet/issues/627
2025-04-01T04:54:51.885325
2016-01-12T13:44:40
126182427
{ "authors": [ "CrackerakiUA", "Mobius77x", "camposcristianeze", "cleever", "mpmlj", "rickerd" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13639", "repo": "CrackerakiUA/ngImgCropFullExtended", "url": "https://github.com/CrackerakiUA/ngImgCropFullExtended/issues/48" }
gharchive/issue
Angular 2.0 I believe this plugin must be made in angular 2.0 too. I am planing to recreate it or add code to this project branch, when angular 2 get the release. Is there someone who are interesting in this? +1 +1 I will start this soon, when i do that i will show the link here. +1. Have you started work on this? Yes but process it stopped. I will continue with it when angular 2 is released. We have tried to make small app and we saw that ng2 is not ready. Any update on this ? I think no. @CrackerakiUA did you do any research on Angular 2.0? @rickerd same as before, my team tried couple projects on ng2. They are changing things, we didn't like it. We decide to have a look again after 6 months. This repo is dead. Please move to the other project: https://github.com/CrackerakiUA/ui-cropper
2025-04-01T04:54:51.892145
2023-04-18T15:09:02
1673335968
{ "authors": [ "CravateRouge", "jsdhasfedssad" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13640", "repo": "CravateRouge/bloodyAD", "url": "https://github.com/CravateRouge/bloodyAD/issues/37" }
gharchive/issue
"get writable" errors with "Logon failure" in some cases Using the command get writable sometimes fails with "Logon failure". I do not know exactly what triggers this but I get this error when authenticating using regular low priviledge accounts but not when authenticating using DAs so this may be related to privileges or lack there of. In the screenshot you can see that the credentials for domainuser1 are correct but that the command get writable still fails when authenticating using the same account. FYI. The account domainuser1 does not have write rights of any sort to any object: Just to be certain, using the command get writable will list objects the authenticating account has GenericAll or GenericWrite rights on, correct? Perhaps WriteDACL too? @jsdhasfedssad The username is different between your two commands @embargod Please open your own issue, it's a not related to the current issue Ooops. You are correct. My bad! What about the privileges? Will get writable list objects the authenticating account has GenericAll or GenericWrite rights on?Perhaps WriteDACL too? It will output you what you can write on the object or child if you can create a child to the object. If you want to know exactly what you can write on the object, use --detail. E.g. if it outputs a WRITE permission on nTSecurityDescriptor it means you have WRITEDACL . If you want a Security Descriptor way of permissions you can alway use get object --resolve-sd This is an interesting feature. Would you say it can partially replace Bloodhound/Sharphound? At least when it comes to identify and abuse edges that grants you write access? Is there a way of using get writable in combination with --detail on a single object? Not only filtered by type. It seems some attributes cannot be written to regardless of access rights. For example the attribute badPwdCount. BloodyAD is not showing that I have write access to that but I wanted to try it anyway using a DA account. Are some attributes "protected"? I get the below error when trying the --resolve-sd command. The account I use is a DA: It can help to identify interesting rights yes. Especially the write child which is not part of bloodhound edges if I'm not mistaken. But of course it doesn't replace the powerful graph db used by bloodhound. You can't do it for a single object or use get object with attributes allowedAttributesEffective or allowedChildClassesEffective. You could also use resolve-sd. For your error can you add a print(self.control_enum) before l.254 in utils.py and show me the output? @jsdhasfedssad can you confirm your issue is fixed? otype filtering on user should be fixed too
2025-04-01T04:54:51.896968
2022-05-05T20:16:37
1227134393
{ "authors": [ "erl-hpe", "rbak-hpe" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13641", "repo": "Cray-HPE/craycli", "url": "https://github.com/Cray-HPE/craycli/pull/50" }
gharchive/pull-request
CASMCMS-7876: Update BOS cli for v2 Summary and Scope This adds the BOS v2 api to the cli. It also adds a new way to inject code into the generator so that it's possible to access data at a point that is not otherwise possible. This is necessary in order to setup a new bulk update endpoint for the cli. Issues and Related PRs Resolves CASMCMS-7876 Testing Tested on: Hela and others Test description: We have been using this to call all the commands while testing BOS v2 Risks and Mitigations None Pull Request Checklist [X] Version number(s) incremented, if applicable [X] Copyrights updated [X] License file intact [X] Target branch correct [X] CHANGELOG.md updated [X] Testing is appropriate and complete, if applicable [X] HPC Product Announcement prepared, if applicable This needs to have the lint failures in the build fixed.
2025-04-01T04:54:51.899828
2022-10-21T17:44:44
1418665161
{ "authors": [ "ndavidson-hpe" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13642", "repo": "Cray-HPE/docs-csm", "url": "https://github.com/Cray-HPE/docs-csm/pull/2702" }
gharchive/pull-request
CASMTRIAGE-4425: Incorrect node command - Update the desired configuration for all NCNs. Description Mistyped option in command needed a capitlization. Checklist Before Merging [x] If I added any command snippets, the steps they belong to follow the prompt conventions (see example). [x] If I added a new directory, I also updated .github/CODEOWNERS with the corresponding team in Cray-HPE. [x] My commits or Pull-Request Title contain my JIRA information, or I do not have a JIRA. /backport release/1.3
2025-04-01T04:54:51.902874
2023-02-23T23:09:05
1597649753
{ "authors": [ "kimjensen-hpe" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13643", "repo": "Cray-HPE/docs-csm", "url": "https://github.com/Cray-HPE/docs-csm/pull/3341" }
gharchive/pull-request
CASMTRIAGE-4986 1.4 : Modify the VCS per flight service check Description Resolves CASMTRIAGE-4986 for CSM 1.4 The previous VCS pre flight service check does not work if other services such as COS are not also installed. The new VCS pre flight service check will instead check that the service is able to return the api version Checklist [ ] If I added any command snippets, the steps they belong to follow the prompt conventions (see example). [ ] If I added a new directory, I also updated .github/CODEOWNERS with the corresponding team in Cray-HPE. [x] My commits or Pull-Request Title contain my JIRA information, or I do not have a JIRA. /backport main
2025-04-01T04:54:51.998695
2021-07-09T16:51:11
940923572
{ "authors": [ "BhavikSojitra", "jainaayush01" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13650", "repo": "Crio-WFH/Chrome-extensions", "url": "https://github.com/Crio-WFH/Chrome-extensions/pull/32" }
gharchive/pull-request
To-Do Extension Pull Request Checklist Go through the check boxes given below and make sure you mark the relevant ones. To check put x inside [ ] eg. [x] Extension Is your project an [x] Extension [ ] Application Given below are the mandatory requirements we expect from your project. Your Pull Request will be reviewed only if all the minimum requirements relevant to your project are satisfied. [x] Added readme.md to my project folder with relevant information. [x] I have provided the demo video link (showing the functionality in action) or the public deployment link of my extension of an application within the readme.md (if Applicable). [x] I have explained the process in a readable manner within readme.md. Terms and Conditions Plagiarism check from Team Crio will be solid and if your code is found to be plagiarized, the team has every right to reject your Pull Request and take necessary actions. [x] I can assure you that this is my own contribution and I did not involve in any kind of plagiarism activities. Buckle up and wait till you receive communication from the review team. Keep a tab on this pull request page for comments on your project. Any necessary suggestions will also be provided here. All the best :) Once your PR gets merged, take some time ⏲️, celebrate 🥳 and share it with the community on Slack. To-Do Extension Hey @BhavikSojitra , I can see that there are many to-do list chrome extensions. We are sorry to say that we are not moving forward with this PR as the deadline is over I have already submitted my project two days from the deadline so there is nothing no issue with the deadline.
2025-04-01T04:54:52.015899
2023-12-05T14:15:48
2026307753
{ "authors": [ "harmen-xb", "willem-xbreeze" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13652", "repo": "CrossBreezeNL/PowerDesigner-MDDE-Extension", "url": "https://github.com/CrossBreezeNL/PowerDesigner-MDDE-Extension/pull/14" }
gharchive/pull-request
Feature/stereotyped mappings Add stereotype on datasource to enable using Mappings for different purposes besides MDDE modeling. For example when there is horizontal and vertical lineage defined in the model @willem-xbreeze I had a look at the changes, which seem ok this way. I only thought you probably wanted to force the stereotype. Because things like model checks, form, etc. in the BaseClassifierMapping and BaseStructuralFeatureMapping metaclass (https://github.com/CrossBreezeNL/PowerDesigner-MDDE-Extension/tree/main/decomposed/extensions/MDDE_LDM/Profile/BaseClassifierMapping & https://github.com/CrossBreezeNL/PowerDesigner-MDDE-Extension/tree/feature/stereotyped_mappings/decomposed/extensions/MDDE_LDM/Profile/BaseStructuralFeatureMapping) should all also be moved into a Criterea on the DataSource stereotype. Right? Otherwise all behavior modelling on the BaseClassifierMapping and BaseStructuralFeatureMapping will be applied to all mappings. Based on your story on being able to use mappings for different purposed won't work I think. @willem-xbreeze I had a look at the changes, which seem ok this way. I only thought you probably wanted to force the stereotype. Because things like model checks, form, etc. in the BaseClassifierMapping and BaseStructuralFeatureMapping metaclass (https://github.com/CrossBreezeNL/PowerDesigner-MDDE-Extension/tree/main/decomposed/extensions/MDDE_LDM/Profile/BaseClassifierMapping & https://github.com/CrossBreezeNL/PowerDesigner-MDDE-Extension/tree/feature/stereotyped_mappings/decomposed/extensions/MDDE_LDM/Profile/BaseStructuralFeatureMapping) should all also be moved into a Criterea on the DataSource stereotype. Right? Otherwise all behavior modelling on the BaseClassifierMapping and BaseStructuralFeatureMapping will be applied to all mappings. Based on your story on being able to use mappings for different purposed won't work I think. Latest checkin made the following changes: Moved all MDDE content in the extension under BaseClassifierMapping and BaseStructuralFeatureMapping to a new criterion. Note: This gives a lot of changes in Git which are only relocations of existing items in the extension Changed the initialize event handler under BaseStructuralFeatureMapping (since event handlers cannot be moved under criterion) so it only runs on appropriate attribute mappings Made the export of mappings conditional, the MDDE XML export only includes MDDE mappings Added an example model with business concepts and a local extension in the ExampleDWH model to illustrate how this can work with an additional form for "business concept mappings" The changes look fine, what I am missing is a change in the add new mapping method, where you would set the stereotype of the new data source to mdde_DataSource. Or how is this taken care of for new users of the extension that the stereotype is set on creation of the first mapping? The criterion allows datasource stereotype to be empty. But i have added a check now in the new mapping method to make sure that the datasource has a valid stereotype, either empty or mdde_datasource. See commit https://github.com/CrossBreezeNL/PowerDesigner-MDDE-Extension/pull/14/commits/5a269365f8d102aa46f54b66b6be73c27b70a535
2025-04-01T04:54:52.026515
2020-05-30T16:45:48
627771670
{ "authors": [ "alexisVidal29", "husseinshaib1", "rdelrosario" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13653", "repo": "CrossGeeks/GoogleClientPlugin", "url": "https://github.com/CrossGeeks/GoogleClientPlugin/issues/57" }
gharchive/issue
exception on release im using xamarin forms, its works fine on debug mode but when i generate the apk file and install on device, it fails with the exception: The google sign in could not completed it's process correctly have you added the sha-1 of the release key ? have you added the sha-1 of the release key ? i mean my release apk works fine but when i try google login shows that exception, the app works fine only the google login fails What exception are you getting? What exception are you getting? after it shows account for pick up, the exception is: The google sign in could not completed it's process correctly in debug mode have no problem It seems you have an issue with release configuration did you add the sha1 of the keystore you are using for release in firebase console portal? On Sat, May 30, 2020, 8:22 PM Alexis Vidal<EMAIL_ADDRESS>wrote: What exception are you getting? after it shows account for pick up, the exception is: The google sign in could not completed it's process correctly in debug mode have no problem — You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/CrossGeeks/GoogleClientPlugin/issues/57#issuecomment-636402174, or unsubscribe https://github.com/notifications/unsubscribe-auth/AATOAJ6NU67XTKXVY6I376LRUGPNRANCNFSM4NOYN4ZA . excuse can you help with these configuration Are you using a local keystore to sign the apk on release or using PlayStore auto-signing? On Sat, May 30, 2020, 8:32 PM Alexis Vidal<EMAIL_ADDRESS>wrote: excuse can you help with these configuration — You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/CrossGeeks/GoogleClientPlugin/issues/57#issuecomment-636403056, or unsubscribe https://github.com/notifications/unsubscribe-auth/AATOAJZNS7ZMOOLO4D2CF4LRUGQRBANCNFSM4NOYN4ZA . im using a local keystore(ah-doc) i already create on firebase console an android application with the sha1 key To get the SHA1 for the release keystore: keytool -exportcert -list -v -alias -keystore On Sat, May 30, 2020, 8:42 PM Alexis Vidal<EMAIL_ADDRESS>wrote: im using a local keystore(ah-doc) i already create on firebase console an android application with the sha1 key — You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/CrossGeeks/GoogleClientPlugin/issues/57#issuecomment-636403845, or unsubscribe https://github.com/notifications/unsubscribe-auth/AATOAJ4UGBIMCIZT462YRJDRUGRWXANCNFSM4NOYN4ZA .
2025-04-01T04:54:52.032196
2018-04-23T05:51:46
316669852
{ "authors": [ "CrowdHailer", "varnerac" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13654", "repo": "CrowdHailer/raxx", "url": "https://github.com/CrowdHailer/raxx/pull/109" }
gharchive/pull-request
Fix Raxx.body spec Raxx.body was String.t(). Changes it to binary. String.t() implies it's a string. It may be a string, but could just as easily be a raw binary that's not a string. Well you learn something new everyday. This looks good, but as I know people who run dialyzer over raxx projects I still think this needs a mention in the changelog. Perhaps I should also add something to an ISSUE_TEMPLATE saying about a changlog. Also if chunked responses/requests can have binary data then should the spec of Raxx.Data should be changed? yeah I think this also needs changing. https://github.com/CrowdHailer/raxx/blob/master/lib/raxx.ex#L233 Although I am curious to see if dialyzer is smart enough to pick it up
2025-04-01T04:54:52.035746
2024-10-04T00:44:58
2565211424
{ "authors": [ "fiona8953", "redhatrises" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13655", "repo": "CrowdStrike/falcon-boshrelease", "url": "https://github.com/CrowdStrike/falcon-boshrelease/issues/16" }
gharchive/issue
falcon-linux-sensor pre-start job failed because it can't find systemctl to restart the service When using Crowdstrike v1.3 and other service broker tiles on Cloud Foundry, we tried to remove the old aws broker v1.4.20 tile but delete-errand failed due to pre-start job falcon-linux-sensor. failed. We re-ran the errand manually with keep-alive then ssh on the VM and found below error message in the /var/vcap/sys/log/falcon-linux-sensor/falcon-installer.log - 2024/10/03 05:41:09 Error restarting Falcon sensor service: Could not find systemctl: /usr/bin/systemctl It looks like the falcon sensor installation worked but just failed because it can't find systemctl to restart the service, however it's still installed, so when we re-ran the errand manually with the same vm this time it passed and succeeded (because falcon sensor was already installed...). Could you please fix this issue with Crowdstrike in case it might impact other errands? Thanks. Hello, Can you provide the OS and OS version that the errand was run on? Also, can you provide the PATH to systemctl on the VM? Hi @redhatrises , The stemcell is light-bosh-aws-xen-hvm-ubuntu-xenial-go_agent-621.969.tgz, OS: ubuntu-xenial, Version: 621.969. PATH to systemctl on the VM is /usr/bin/systemctl. Thanks for providing that info. This is fixed in the installer and will be fixed in the next release. Closing.
2025-04-01T04:54:52.046367
2021-02-02T23:16:44
799807285
{ "authors": [ "CrustyJew", "derekantrican" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13656", "repo": "CrustyJew/RedditSharp", "url": "https://github.com/CrustyJew/RedditSharp/issues/237" }
gharchive/issue
Error reading integer. Unexpected token: StartObject. Path 'data.accounts_active', line 1, position 1601. Just started failing within the last few hours. Looks like something may have changed with the Reddit API? Or maybe it's just an intermittent error? I'm using the latest pre-release (v2.0.0-CI00075) Code: var subreddit = await new Reddit().GetSubredditAsync("/r/videos"); Exception: JsonReaderException: Error reading integer. Unexpected token: StartObject. Path 'data.accounts_active', line 1, position 1601. at Newtonsoft.Json.JsonReader.ReadAsInt32() at Newtonsoft.Json.JsonReader.ReadForType(JsonContract contract, Boolean hasConverter) at Newtonsoft.Json.Serialization.JsonSerializerInternalReader.PopulateObject(Object newObject, JsonReader reader, JsonObjectContract contract, JsonProperty member, String id) at Newtonsoft.Json.Serialization.JsonSerializerInternalReader.Populate(JsonReader reader, Object target) at Newtonsoft.Json.JsonSerializer.PopulateInternal(JsonReader reader, Object target) at RedditSharp.Helpers.PopulateObject(JToken json, Object obj) at RedditSharp.Things.Thing.Populate(JToken json) at RedditSharp.Things.Thing.Parse(IWebAgent agent, JToken json) at RedditSharp.Things.Thing.Parse[T](IWebAgent agent, JToken json) at RedditSharp.Helpers.d__01.MoveNext() at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at System.Runtime.CompilerServices.TaskAwaiter1.GetResult() at WatchNow.SourceManagement.d__13.MoveNext() in C:\Users\derek\source\repos\WatchNow\WatchNow\SourceManagement.cs:line 161 Is this still ongoing? This previously occurred when Reddit broke some shit and started returning a null or missing the json for it entirely. Seems to be working now, I guess? I guess it was some sort of intermittent error. Odd that it was accounts_active in the error, but probably is the first key parsed. accounts_active is the actual JSON key they screw up every once in a while. I've complained about it before but apparently it's regressed yet again.
2025-04-01T04:54:52.058724
2024-03-13T02:05:19
2182975308
{ "authors": [ "WingZer0o" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13657", "repo": "Cryptographic-API-Services/cas-api", "url": "https://github.com/Cryptographic-API-Services/cas-api/issues/40" }
gharchive/issue
Devise unlimited user's development API key and one user for production key. When attaching a development key in the SDK there must be some sort of determination when the key get sent with the OS information if it is a development key or a production key. User should be able to connect multiple SDK's with the development key and only one instance of the production key unless they are paying for a higher subscription (this should be coded in another ticket). User's should be able to see all operating system information of the developers currently using the SDK in development. (To be coded in another ticket). This was written and pushed already. When storing the information in cache upon sending it to the API. if (existingCacheInformation != null) { JsonSerializerOptions options = new JsonSerializerOptions() { PropertyNameCaseInsensitive = false, }; OSInfoRedisEntry cacheInformation = JsonSerializer.Deserialize<OSInfoRedisEntry>(existingCacheInformation, options); // TODO: perform other checks besides IP address and Operating System // also perform check based upon the API key. if (!cacheInformation.IsApiKeyProd) { result = new OkObjectResult(new { message = "Enjoy using your development key" }); }
2025-04-01T04:54:52.072033
2022-07-17T00:20:03
1306946252
{ "authors": [ "Greelan" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13658", "repo": "CubeCoders/AMP", "url": "https://github.com/CubeCoders/AMP/issues/684" }
gharchive/issue
Add "Pause" update stage to Generic This is not for requesting support for new games/applications To do this you should go to https://github.com/CubeCoders/AMPTemplates and first attempt to build a configuration yourself - otherwise you can request a template from this repo. Feature Request Feature Information: It would be useful to have a configurable "Pause" update stage. This would pause the update stages for a specified number of seconds. One use case is the new Stationeers server build. The updated draft template includes start/wait/stop stages to generate the server files. However the console output that appears then is different to subsequent starts. If a console line that is present on all starts is selected as the app ready regex, then on first start the stop stage kicks in too quickly, because the server does some additional work after that point on first start. This causes the stop stage to not work correctly. This would be avoided if a pause could be introduced after the wait update stage, that operates once the regex is matched and the server is otherwise considered to be started. This might also help in some cases where a start stage seems to hang if run immediately after a set executable stage. I confirm: [x] that I have searched for an existing feature request matching the description. This could also help the case (eg Vintage Story) where an executable fails to be started if that update stage occurs immediately after the server files are unzipped/extracted.
2025-04-01T04:54:52.076965
2023-04-15T16:55:32
1669482751
{ "authors": [ "IceOfWraith" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13659", "repo": "CubeCoders/AMP", "url": "https://github.com/CubeCoders/AMP/issues/870" }
gharchive/issue
SteamCMD Steam Guard Issues Bug Report System Information Operating System - Ubuntu 22.04 LTS AMP version and build date - v<IP_ADDRESS>, built 31/03/2023 20:24 Which AMP release stream you're using - Bleeding I confirm: [x] that I have searched for an existing bug report for this issue. [x] that I am using the latest available version of AMP. [x] that my operating system is up-to-date. Symptoms What are you trying to do? Update a Steam game for the first time. What are you expecting to happen? AMP should appropriately handle the Steam Guard code. What is actually happening? ('Nothing' is not an acceptable answer!) The Steam Guard prompt happens twice and hangs if "Cancel" is pressed. Reproduction If you setup a game like Assetto Corsa Compatizione which requires login for the first time, AMP prompts for the Steam Guard code as expected. If you press "Cancel" the process just hangs. If you type the code correctly AMP accepts it but then prompts again with the code pre-filled. It seems AMP is using the regex of Steam Guard which is in the request from SteamCMD and the response as seen below: This computer has not been authenticated for your account using Steam Guard. Please check your email for the message from Steam, and enter the Steam Guard code from that message. Steam Guard code:OK Heh... I fixed this one a while ago.
2025-04-01T04:54:52.085833
2018-12-05T22:38:22
387976939
{ "authors": [ "JDemler", "barbeque" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13661", "repo": "CurrySoftware/elm-datepicker", "url": "https://github.com/CurrySoftware/elm-datepicker/pull/18" }
gharchive/pull-request
Some grammar, spelling fixes It looks like I set up my last PR too early! The name of the nightwatch example had become out-of-date in the README. Thanks again :+1:
2025-04-01T04:54:52.100193
2021-05-27T02:54:48
903149495
{ "authors": [ "CutFlame", "Jaypamieukomnom", "Maksman2000", "PhoenixMC-0", "Viinyl", "lordgecko10", "megaenderdragon" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13662", "repo": "CutFlame/MCDSaveEdit", "url": "https://github.com/CutFlame/MCDSaveEdit/issues/58" }
gharchive/issue
MCD SE not unable to launch since the Hidden Depths update Yeah, they changed the encryption keys again. Next release (version 1.3.5) will work. I will make sure it gets published next week sometime. how close is it to being done There is a way around this, if you only need to edit item power levels and/or character level, or item rarity, then you can just switch install location. Simply uninstall, and make sure you (1) back up your character folder and (2) make sure the keep save file is turned on in the uninstall screen, then reinstall in a non-default location (I used my primary games folder) and then, when the editor fails to find MCD, simply press no game files (forget what the actual button is called) and find your character folder by going to File > Open > your save game. You cannot create items, change enchant slots, or anything like that, but it will work for most things. There is a way around this, if you only need to edit item power levels and/or character level, or item rarity, then you can just switch install location. Simply uninstall, and make sure you (1) back up your character folder and (2) make sure the keep save file is turned on in the uninstall screen, then reinstall in a non-default location (I used my primary games folder) and then, when the editor fails to find MCD, simply press no game files (forget what the actual button is called) and find your character folder by going to File > Open > your save game. You cannot create items, change enchant slots, or anything like that, but it will work for most things. ^ This is true ^ Yeah, they changed the encryption keys again. Next release (version 1.3.5) will work. I will make sure it gets published next week sometime. can't wait!!! Going to place some new armor properties onto my all-in-one gilded opluent armor Yeah, they changed the encryption keys again. Next release (version 1.3.5) will work. I will make sure it gets published next week sometime. But when... Yeah, they changed the encryption keys again. Next release (version 1.3.5) will work. I will make sure it gets published next week sometime. But when... You actually can make your Opulent Armor gilded right now by doing the method I mentioned above, as long as you have the right enchantments. If not, then you can look for an armor piece that has Luck of the Sea on it, max out that enchant, and restock merchants until the restock cost is more than 10k emeralds. This works super well, and I've gotten about ~50 uniques just from this. Seeing as you can set your player level (to get more enchant points) and set your player emerald and gold count (for restocking and buying items), with this limited version, I'd say its worth it. (side note, before this, I had never gotten a verdant robe, but with luck of the sea and high emerald count, I now have 3. So yea.) But when will it come out? From what I understand, finding the key can be time consuming, but I don't really know cause I'm not coding this. On a side note, this program still works for the most part with this bug using the method that I mentioned, but as Cutflame said, it should be coming out this week. this is the AES key, @CutFlame : 0x0A01138AC7553ECEA7E9E257B5A45FA6665679D360B270278AE5D5367B172E40 There is a way around this, if you only need to edit item power levels and/or character level, or item rarity, then you can just switch install location. Simply uninstall, and make sure you (1) back up your character folder and (2) make sure the keep save file is turned on in the uninstall screen, then reinstall in a non-default location (I used my primary games folder) and then, when the editor fails to find MCD, simply press no game files (forget what the actual button is called) and find your character folder by going to File > Open > your save game. You cannot create items, change enchant slots, or anything like that, but it will work for most things. I've tried this and It just closes when I click "OK" on the error box, tried running as admin and all the ins and outs, can't get to this "no game files" section as what I said earlier, it just closes itself after closing the error. It did the same thing for me, however my install of MCD has been messed up for a long time because I transferred my save data (all of it) from another computer to my new one back when MCD had its own launcher. Did the error change from looking like the picture at the top of this thread to something like this? Yeah it looks like that, when I click "OK" it just closes the entire thing, i'm just wanting to set my gear up to the current max power level but none of my old editors are working either. When the program will be updated When the program will be updated I have no Idea. He said this week New version is released (1.3.6)
2025-04-01T04:54:52.110975
2023-10-30T17:48:33
1968915068
{ "authors": [ "CyanideData2", "Wojtas1PL" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13663", "repo": "CyanideData2/Sun-O-Matic", "url": "https://github.com/CyanideData2/Sun-O-Matic/issues/3" }
gharchive/issue
Communication protocol I've been thinking about how to make sure that C# objects can be easily understood by the Arduino, and I figured the easiest way would be to not use objects at all. Thus, I propose the following communication standard, which would ensure the datastream sent over BLE can be easily received, stored, and processed: comms protocol v1.txt I looked a bit at the way C# handles serial communication, and I think it's gonna be easier if we instead do it like this: Each event is a 32 bit number, first bit representing the target setting (or override signal for the sync time signal), the other bits representing the 31 bit integer - just add a leading zero to get a standard 32 bit one, we won't ever need more than 17 bits anyway. So instead of a proper protocol we just create our own standard?
2025-04-01T04:54:52.122054
2019-03-21T10:52:32
423670559
{ "authors": [ "CyberZHG", "coveralls", "nsmetanin" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13664", "repo": "CyberZHG/keras-bert", "url": "https://github.com/CyberZHG/keras-bert/pull/37" }
gharchive/pull-request
Split building and loading the model This can be useful if you want to build the BERT model from given config, but don't want to load pretrained weights. Pull Request Test Coverage Report for Build 97 5 of 5 (100.0%) changed or added relevant lines in 1 file are covered. No unchanged relevant lines lost coverage. Overall coverage remained the same at 100.0% Totals Change from base Build 96: 0.0% Covered Lines: 340 Relevant Lines: 340 💛 - Coveralls Thanks for the PR~ I've published a new version. Great, that was fast, thanks!
2025-04-01T04:54:52.126652
2023-01-31T15:54:03
1564581293
{ "authors": [ "Cyclenerd" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13665", "repo": "Cyclenerd/google-cloud-pricing-cost-calculator", "url": "https://github.com/Cyclenerd/google-cloud-pricing-cost-calculator/issues/53" }
gharchive/issue
Feature request: Committed use discounts for RHEL images Compute Engine committed use discounts are now also available for Red Hat Enterprise Linux (RHEL) image licenses. Add: [x] Committed use discounts for RHEL images [x] Committed use discounts for RHEL for SAP images Source: https://cloud.google.com/compute/disks-image-pricing#premiumimages Commit: https://github.com/Cyclenerd/google-cloud-pricing-cost-calculator/commit/a245e9612ab64a6b3913c6cb625731d1ede9cfba New release: https://github.com/Cyclenerd/google-cloud-pricing-cost-calculator/releases/tag/v1.1.0
2025-04-01T04:54:52.135136
2024-06-10T06:54:15
2343018014
{ "authors": [ "met4000", "rubensworks" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13666", "repo": "CyclopsMC/IntegratedScripting", "url": "https://github.com/CyclopsMC/IntegratedScripting/issues/12" }
gharchive/issue
Error handling when returning undefined Issue type: :heavy_plus_sign: Feature request Short description: Functions that return undefined (e.g. function f() { return undefined; }, or function f() {} ) give a rather verbose error as follows when apply0ed; It would be desireable to either have: a dedicated error for a function returning undefined, rather than a translation error occuring (e.g. 'error: function returned undefined instead of a value'); or for an explicit 'undefined' singleton to be the output instead of erroring, allowing for logic such as equals(f(), undefined) on the InDy side. Thanks for the suggestion!
2025-04-01T04:54:52.139203
2017-03-26T16:03:41
217062008
{ "authors": [ "Cylix", "vivekvrao" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13667", "repo": "Cylix/tacopie", "url": "https://github.com/Cylix/tacopie/issues/6" }
gharchive/issue
Trying to serve html to browser on tacopie server I am trying to serve up some html content to a browser over a tacopie server. This was my message handler - where I tried both sync and async writes and trying to close the connection when done. Neither works, maybe because I dont understand how the write/disconnect works. Any ideas? thank you! void on_new_message(const std::shared_ptr<tacopie::tcp_client>& client, const tacopie::tcp_client::read_result& res) { if (res.success) { std::string resp = "this is the response"; std::string fullResp = "HTTP/ 1.1 200 OK\r\nContent-Type: type/html\r\nConnection: Closed\r\n Content-Length: " + std::to_string(resp.length()) + std::string("\r\n\r\n") + resp; std::vector<char> buf(fullResp.begin(), fullResp.end()); client->get_socket().send(buf, fullResp.length()); //client->async_write({ buf, nullptr }); client->disconnect(); } else { info("WebClient disconnected"); client->disconnect(); } } Hi, Sorry for the time to reply, I finally could have a look. I tried your code and it first worked well over curl and netcat, but apparently not with a browser (I tried chrome and safari). So my guess was that there is an issue with the content of your request. I double checked with the expected HTTP format, and there is indeed an issue in your response. HTTP/ 1.1 There is an extra space after the / \r\n Content-Length: : There is an extra space Before the Content When I changed your output into: HTTP/1.1 200 OK\r\nContent-Type: type/html\r\nConnection: Closed\r\nContent-Length: it worked perfectly fine :) I close this issue, but feel free to re-open it if necessary :) Best BTW, my final code is the following one: void on_new_message(const std::shared_ptr<tacopie::tcp_client>& client, const tacopie::tcp_client::read_result& res) { if (res.success) { std::cout << "write" << std::endl; std::string resp = "this is the response"; std::string fullResp = "HTTP/1.1 200 OK\r\nContent-Type: type/html\r\nConnection: Closed\r\nContent-Length: " + std::to_string(resp.length()) + std::string("\r\n\r\n") + resp; std::vector<char> buf(fullResp.begin(), fullResp.end()); client->async_write({buf, [=](tacopie::tcp_client::write_result&) { std::cout << "write cb" << std::endl; client->disconnect(); }}); } else { std::cout << "disconnect" << std::endl; client->disconnect(); } }
2025-04-01T04:54:52.164459
2015-09-06T22:42:50
105121191
{ "authors": [ "9rnsr", "MartinNowak" ], "license": "BSL-1.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13668", "repo": "D-Programming-Language/druntime", "url": "https://github.com/D-Programming-Language/druntime/pull/1382" }
gharchive/pull-request
Merge remote-tracking branch 'upstream/stable' into merge_stable Conflicts: src/rt/profilegc.d Auto-merge toggled on
2025-04-01T04:54:52.168085
2015-06-29T02:50:17
91677728
{ "authors": [ "ajvincent", "rro4785", "s-ludwig" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13669", "repo": "D-Programming-Language/dub", "url": "https://github.com/D-Programming-Language/dub/issues/610" }
gharchive/issue
Build/Run with DMD on Windows with a username with white spaces Hi! I'm trying to build a DlangIDE with dub and dmd on Windows 7 Pro, but show me an error like that: Error: cannot read file C:\Users\My.d When my realy path is like that: *C:\Users\My Username* Searching in internet I found that my problem is my username because it has white spaces, but never found how can I change my dub's local repository. Anybody can help me with that? Can you try out the latest beta (0.9.24-beta.1, installer, zip) and see if it works there? This is most likely the same root cause as #540. Ok, I'll try and I'll let you know. Great!!! That's all!!! Thank you!!! I'm experiencing essentially the same bug, and I can confirm the beta fixes the bug.
2025-04-01T04:54:52.180512
2023-10-08T21:24:07
1932072451
{ "authors": [ "skrawcz", "zilto" ], "license": "BSD-3-Clause-Clear", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13670", "repo": "DAGWorks-Inc/hamilton", "url": "https://github.com/DAGWorks-Inc/hamilton/issues/448" }
gharchive/issue
Autogenerate argparser to execute Driver from command line Is your feature request related to a problem? Please describe. When deploying scripts in SageMaker or VertexAI, configuration needs to be passed via CLI / argparse. This requires writing tedious argparse code, which often ends up being poorly typed, documented, and maintained. Describe the solution you'd like Automatically generate argparser (or OmegaConf/Hydra) for Hamilton nodes of the instantiated driver. Each parsing argument could include the type and docstring (if not a top node). It is possible to resolve what should be in inputs and in overrides. final_vars could also be specified. import argparse class Driver: ... def with_cli(self): parser = argparse.ArgumentParser(prog="HamiltonCLI", description="Generated CLI") for n in self.graph.get_nodes(): parser.add_argument(f"--{n.name}") self.args = parser.parse_args() def resolve_args(self): inputs, overrides = resolve_node_value(self.args) action, kwargs = resolve_action(self.args) # visualize, execute if action == "execute": self.execute(inputs=inputs, overrides=overrides, **kwargs) if __name__ == "__main__": dr = ( driver.Builder() .with_module(transforms) .with_cli() # include arguments to limit supported operations (e.g., execute only) .build() ) Alternative A simpler and more explicit approach could be passing two list of strings to expose input and override nodes. This prevents having a CLI flooded with irrelevant args if __name__ == "__main__": dr = ( driver.Builder() .with_module(transforms) .with_cli(inputs=list(), overrides=list()) .build() ) The nodes supported would be limited to primitives that can be expressed on the command line. One challenge is properly coercing args, which are all strings, into the correct Hamilton type. This could be done efficiently with Pydantics yep -- https://typer.tiangolo.com/ could perhaps make this simpler? Another clean pattern could be to decorate functions with @cli and then .with_cli() collects that when building the driver. However, it wouldn't be possible to annotate top level nodes Another clean pattern could be to decorate functions with @cli and then .with_cli() collects that when building the driver. However, it wouldn't be possible to annotate top level nodes Yeah without instantiating a driver and knowing the requested outputs we wouldn't know what they are. But that doesn't mean we couldn't have something dynamic... or alternatively we just have a command line that creates a CLI file for a given driver set up...
2025-04-01T04:54:52.182824
2017-04-07T16:46:10
220268338
{ "authors": [ "andrewufrank", "thvitt" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13671", "repo": "DARIAH-DE/DARIAH-DKPro-Wrapper", "url": "https://github.com/DARIAH-DE/DARIAH-DKPro-Wrapper/issues/29" }
gharchive/issue
use of DKPro-Wrapper as a server (or at least with stdin and stdout is it possible to use the DKPro-Wrapper as a server (similar to coreNLP) or at least as continously running, expecting text on stdin and producing stdout, without closing. The startup time is a heavy overhead if the pipelline is started for each small snippet of text as it occurs and i would prefer not to collect all input and then treat it in a single file. thank you! it is possible to run the pipeline on a folder of text or xml files, the wrapper will then create a filename.csv for each file in the given output folder. Just pass a directory to the -input option. There are currently no plans on our side to implement a server mode. (Running a full analysis on texts of notable length takes quite a while, so there's a certain risk of timeouts) A real streaming mode implementation isn't possible since the document must be loaded completely before further analysis steps begin. It might be possible to implement something that could accept multiple documents on stdin separated by sth like \0 …
2025-04-01T04:54:52.186145
2024-06-16T07:40:33
2355509163
{ "authors": [ "JantsoP", "Pdzly" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13672", "repo": "DC-Resources-Wiki/Discord-Resources-Wiki", "url": "https://github.com/DC-Resources-Wiki/Discord-Resources-Wiki/issues/20" }
gharchive/issue
Deprecated "Programming Related" link The autocode library got shut down https://discordresources.com/resources/tools/programming#autocode https://discordresources.com/resources/moderation#a-guide-to-civil-conversation this one is broken too This is an open project. Anyone can do PR to fix these issues :)
2025-04-01T04:54:52.197442
2020-09-11T10:15:03
699128758
{ "authors": [ "allemangD", "bpaniagua", "jcfr", "luciacev" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13673", "repo": "DCBIA-OrthoLab/ShapeVariationAnalyzer", "url": "https://github.com/DCBIA-OrthoLab/ShapeVariationAnalyzer/issues/54" }
gharchive/issue
cannot see ShapeVariationAnalyzer module @jcfr @allemangD @bpaniagua Hi ShapeVariationAnalyzer is not working/cannot find module after installing SlicerCMF. Thanks! Per the log in https://github.com/DCBIA-OrthoLab/SlicerCMF/issues/18, the issue seems to be that sklearn is not configured properly for Mac, and this causes errors when it is imported by shapepca, which is imported by ShapeVariationAnalysis. The same happens if only ShapeVariationAnalyzer is installed via the extension manager (without SlicerCMF). Here is a full log from my machine (using yesterday's Slicer preview release) Slicer_29370_20200917_132211_450.log When I build ShapeVariationAnalyzer myself, no such error occurs. $ git clone https://github.com/DCBIA-OrthoLab/ShapeVariationAnalyzer.git $ cd ShapeVariationAnalyzer $ mkdir build $ cd build $ cmake -DShapeVariationAnalyzer_BUILD_SLICER_EXTENSION=ON -DSlicer_DIR=~/src/Slicer/build/Slicer-build .. $ ninja $ inner-build/SlicerWithShapeVariationAnalyzer This tells me there may be some issue in the factory machine causing sklearn not to be configured correctly. @jcfr, do you have any thoughts on this? If I understand the superbuild correctly, scikit-learn is installed via pip in the default configuration. https://github.com/DCBIA-OrthoLab/ShapeVariationAnalyzer/blob/master/SuperBuild/External_python-ShapeVariationAnalyzer-requirements.cmake#L43-L54 If I manually install via pip with the same hashes, I get an error that the hash does not match: ERROR: THESE PACKAGES DO NOT MATCH THE HASHES FROM THE REQUIREMENTS FILE. If you have updated the package versions, please update the hashes. Otherwise, examine the package contents carefully; someone may have tampered with them. scikit-learn==0.23.1 from https://files.pythonhosted.org/packages/ea/47/078b5ef83ccff1def48c29158be998375321397be53af6dc10f22c6dc08d/scikit_learn-0.23.1-cp38-cp38-macosx_10_9_x86_64.whl#sha256=0c3464e46ef8bd4f1bfa5c009648c6449412c8f7e9b3fc0c9e3d800139c48827 (from -r req.txt (line 1)): Expected sha256 e585682e37f2faa81ad6cd4472fff646bf2fd0542147bec93697a905db8e6bd2 Expected or 058d213092de4384710137af1300ed0ff030b8c40459a6c6f73c31ccd274cc39 Expected or e9879ba9e64ec3add41bf201e06034162f853652ef4849b361d73b0deb3153ad Got 0c3464e46ef8bd4f1bfa5c009648c6449412c8f7e9b3fc0c9e3d800139c48827 This is confusing to me, since the superbuild succeeds locally without issue and cdash does not report any such issue. The issue doesn't seem to affect the Linux build of Slicer or the extension manager. Here is the log from my Ubuntu 20 machine after installing the SlicerCMF extension bundle. ShapeVariationAnalyzer and the Population Analysis module load without issue. Slicer_29363_20200917_135032_863.log Do you mind testing in your macbook? The above logs are from my macbook. Everything works correctly if I build the extension myself, but loading from the extension manager causes the problem. This makes me think there is some issue with the factory machine. In the second comment I was just noting that the issue is never present for me on linux, to emphasize I think this issue is specific to the mac factory machine and not the others, although I can't confirm for Windows. Thanks for checking, David! @sjh26 can you please check in any issues in the macOS factory? If I manually install via pip with the same hashes, I get an error that the hash does not match: @allemangD and I looked at this, and python 3.6 (version also used in Slicer) need to be used to ensure the hashes match. It looks like the .dylibs directory containing libomp.dylib is properly packaged in the .tar.gz file but there is any extracting the content of the archive into its final destination. This has been fixed in https://github.com/Slicer/Slicer/pull/5307 @bpaniagua @allemangD @jcfr ShapeVariationAnalyzer does not show and instead in the SlicerCMF index when we click on ShapeVariationAnalyzer a module Population Analysis starts that does not seem to allow computing group averages or group differences , and is quite different from SVA. Not sure how to report this issue and will further clarify with Bea in June. @luciacev The fix for this bug required a change to Slicer, so the stable version from last year we had told you to use () will not contain that fix. I've confirmed that SVA loads correctly on Mac in the latest stable version (4.11.20210226). As best I can tell, all the other modules work correctly there, too. I would recommend we try to use that 2021 version moving forward. Regarding https://github.com/DCBIA-OrthoLab/SlicerCMF/issues/13, SPHARM-PDM is available on Mac in that version, but not RigidAlignment. I'm working on fixing that next. I've made notes about these in the weekly updates document, and will send that out to the team this Friday. Downloads for that revision here. Closing as the issue is fixed in the current stable release 4.11.20210226.
2025-04-01T04:54:52.211894
2023-06-07T15:19:32
1746150800
{ "authors": [ "JoyfulGen", "annamorphism", "jacobdgm" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13674", "repo": "DDMAL/CantusDB", "url": "https://github.com/DDMAL/CantusDB/issues/712" }
gharchive/issue
Images link for first two manuscripts in "Browse Chants" page lead to a page that never loads To reproduce: Go to "Browse Chants" page; Click on the "Images" link for either A-Gu 29 or A-Gu 30; A new tab opens, but the url is weird and the site never loads. Demo (New vs. Old comparison): https://github.com/DDMAL/CantusDB/assets/83373378/fa8da2fc-0d77-4e7b-bb9a-73a384936b68 I've clicked on random other manuscript images and they all load fine. From my casual random selection, it seems to be just these two that are badly behaved. That weird URL shows up elsewhere on the internet, so I'm assuming the library changed it at some point; I think it was still floating around on OldCantus last October, so it should go away once the data is updated (but will be worth checking.) OldCantus links to the images at literature.at : http://www.literature.at/viewer.alo?objid=1138&page=9&viewmode=fullscreen&rotate=&scale=1 It would actually be much better to link to the university's images: https://unipub.uni-graz.at/obvugrscript/content/pageview/6789708 , both because the URL is less ugly and also because the literature.at ones are in a completely bonkers order (which will be relevant to serving them on CU: @dchiller take note!) But maybe that is a problem for Cantus to solve and not DDMAL! We can't do too much about other sites not loading well, but if there are different urls, this is a data issue that we can expect to be resolved once we sync the data from OldCantus to New. @annamorphism, I'm assigning you to this, in case you want to change the urls for these sources as you suggested (either do it on OldCantus in the next few days so we can sync from there, or wait for the full switchover and make the change both on Staging and Production). Please close this once you're done (or whenever you feel it's appropriate) As there is nothing for NewCantus developers to do about this issue, I'm closing it.
2025-04-01T04:54:52.244233
2018-01-11T13:23:01
287784375
{ "authors": [ "mbeckerle" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13676", "repo": "DFDLSchemas/GIF", "url": "https://github.com/DFDLSchemas/GIF/issues/1" }
gharchive/issue
Add "unit" tests. src/test/scala and src/test/resources currently empty Needs basic tests and example image data. Fixed in commit e48f8d2ca0096e52491dc5c5d93268604878829e of GIF git repository.
2025-04-01T04:54:52.248951
2020-09-15T09:59:03
701790337
{ "authors": [ "tijmenb" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13677", "repo": "DFE-Digital/apply-for-teacher-training", "url": "https://github.com/DFE-Digital/apply-for-teacher-training/pull/2902" }
gharchive/pull-request
Check if find is down only once on providers page Context The list of available courses on Apply (https://www.apply-for-teacher-training.service.gov.uk/candidate/providers) is a bit slow according to Skylight: https://www.skylight.io/app/applications/t8bEzG0cuIkd/recent/6h/endpoints/CandidateInterface::ContentController%23providers?responseType=html Changes proposed in this pull request Query is find is down only once. This saves a lot of SiteSetting queries and calculation, and might solve the performance issues with this page. Guidance to review Link to Trello card https://trello.com/c/x0OxYLae/2143-speed-up-the-providers-page-in-the-candidate-ui Things to check [x] This code does not rely on migrations in the same Pull Request [x] If this code includes a migration adding or changing columns, it also backfills existing records for consistency [x] API release notes have been updated if necessary [x] New environment variables have been added to the Azure config @davidgisbey it was my bad - I introduced the SiteSetting lookup in the timetable, which made this slow. @tvararu should have mentioned that, but it's a class method, so memoizing it would persist across requests and not change if today's date or the cycle schedule changes.
2025-04-01T04:54:52.254988
2024-04-04T16:28:30
2225998783
{ "authors": [ "elceebee" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13678", "repo": "DFE-Digital/apply-for-teacher-training", "url": "https://github.com/DFE-Digital/apply-for-teacher-training/pull/9262" }
gharchive/pull-request
[1514] Part 1 add unique index to Vender name Context Currently, we store and validate the Vendor name as an enum on the Vendor model. There are problems with this: Adding a new Vender requires a code change. This is also a potential security risk as all our vendors are located in our public repo. We want to replace the enum with basic presence and uniqueness validation. Changes proposed in this pull request Migration to add a unique index to vendor name. I've checked in production that all of our vendors have unique names as it is (there are only 6). Once this has been merged in, I will open the PR here which actually gets rid of the enum on the model Guidance to review Link to Trello card https://trello.com/c/jMnBxRIb Things to check [x] If the code removes any existing feature flags, a data migration has also been added to delete the entry from the database [x] This code does not rely on migrations in the same Pull Request [x] If this code includes a migration adding or changing columns, it also backfills existing records for consistency [ ] If this code adds a column to the DB, decide whether it needs to be in analytics yml file or analytics blocklist [ ] API release notes have been updated if necessary [ ] If it adds a significant user-facing change, is it documented in the CHANGELOG? [ ] Required environment variables have been updated added to the Azure KeyVault Tech debt: Remove Enum of Vendor names
2025-04-01T04:54:52.268982
2018-10-02T14:02:09
365912714
{ "authors": [ "defong", "timabell" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13679", "repo": "DFE-Digital/search-and-compare-ui", "url": "https://github.com/DFE-Digital/search-and-compare-ui/pull/187" }
gharchive/pull-request
Fixed socket exception Context SocketException Changes proposed in this pull request Exhausted socket Guidance to review Same vein as https://github.com/DFE-Digital/search-and-compare-ui/pull/186 Craig David bug causing exceptions - 21 seconds... timeout
2025-04-01T04:54:52.272667
2021-12-09T15:20:13
1075693897
{ "authors": [ "bencmitchell", "cesidio" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13680", "repo": "DFE-Digital/teaching-vacancies", "url": "https://github.com/DFE-Digital/teaching-vacancies/pull/4407" }
gharchive/pull-request
[TEVA-3546] Allow education support roles to be apply via TV Jira ticket URL https://dfedigital.atlassian.net/browse/TEVA-3546 Screenshots of UI changes: @cesidio I've got jobseeker open on left and hiring staff on right. I still see the Professional status question for a Teaching Assistant role I don't see this for SENDCo or education support though Look good, section no longer there for Teaching Assistant
2025-04-01T04:54:52.312443
2023-02-21T22:13:28
1594164235
{ "authors": [ "smythp" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13681", "repo": "DHRI-Curriculum/Dhrift-Homepage", "url": "https://github.com/DHRI-Curriculum/Dhrift-Homepage/issues/10" }
gharchive/issue
Set up jekyll build teston PRs We should probably test that the build doesn't fail before merging. Set this up. In theory, when we create PRs where, a check should run to see if the PR breaks the jekyll build.
2025-04-01T04:54:52.362892
2016-10-24T20:18:53
184943820
{ "authors": [ "jonasbn" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13682", "repo": "DK-Hostmaster/epp-service-specification", "url": "https://github.com/DK-Hostmaster/epp-service-specification/issues/8" }
gharchive/issue
Exchange XML examples for renew domain The examples was originally lifted from the RFCs to support the developers of the extension of the service, these examples should be exchanged for real working examples lifted from the actual service so all details are captured. The examples for this particular command have been updated: ref: https://github.com/DK-Hostmaster/epp-service-specification#renew-domain
2025-04-01T04:54:52.379753
2021-06-16T15:22:12
922757145
{ "authors": [ "codecov-commenter", "rainman110" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13683", "repo": "DLR-SC/tigl", "url": "https://github.com/DLR-SC/tigl/pull/811" }
gharchive/pull-request
Respect PointBehaviour for tiglWingGetSegmentEtaXsi Description I had to change the class structure / method signature slightly to allow to pass, how GetEtaXsi is evaluated. Unfortunately, GetEtaXsi on the paramter surfaces is almost a factor of thousand slower, which could be a problem for the VTK export with metadata. Therefore I decided to use the old eta / xsi interpretation for the exports for now. closes #808 How Has This Been Tested? To validate, I changed the consistency test, added a performance test and added a function to tiglviewer to play around with the function. Screenshots, that help to understand the changes(if applicable): Checklist: [x] A test for the new functionality was added. [ ] All tests run without failure. [x] The new code complies with the TiGL style guide. [ ] New classes have been added to the Python interface. [ ] API changes were documented properly in tigl.h. Codecov Report Merging #811 (1833e84) into master (e05dfb1) will increase coverage by 0.07%. The diff coverage is 85.93%. @@ Coverage Diff @@ ## master #811 +/- ## ========================================== + Coverage 62.17% 62.25% +0.07% ========================================== Files 291 291 Lines 23438 23466 +28 ========================================== + Hits 14572 14608 +36 + Misses 8866 8858 -8 Impacted Files Coverage Δ src/wing/CCPACSWing.h 100.00% <ø> (ø) src/wing/CCPACSWingSegment.h 50.00% <ø> (ø) src/wing/CCPACSWing.cpp 74.55% <46.15%> (+0.99%) :arrow_up: src/wing/CCPACSWingSegment.cpp 82.64% <95.55%> (+1.65%) :arrow_up: src/api/tigl.cpp 54.39% <100.00%> (-0.03%) :arrow_down: src/geometry/CTiglTriangularizer.cpp 86.90% <100.00%> (+0.07%) :arrow_up: src/wing/CCPACSWingComponentSegment.cpp 50.09% <100.00%> (-0.19%) :arrow_down:
2025-04-01T04:54:52.381475
2023-11-28T16:50:58
2014876095
{ "authors": [ "johntruckenbrodt", "maawoo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13684", "repo": "DLR-terrabyte/demo-bids23", "url": "https://github.com/DLR-terrabyte/demo-bids23/issues/1" }
gharchive/issue
Suggestion to include nbviewer-link in readme The experience of looking at Jupyter Notebooks on GitHub is unfortunately not the best. It might be a good idea to include the following link in the Readme: https://nbviewer.org/github/DLR-terrabyte/demo-bids23/blob/main/demo.ipynb Good call @maawoo. Thanks!
2025-04-01T04:54:52.385263
2021-11-04T15:26:19
1044887627
{ "authors": [ "akberg" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13685", "repo": "DMPRO2021-Audio/mcu", "url": "https://github.com/DMPRO2021-Audio/mcu/issues/9" }
gharchive/issue
Update synth_t structure Update synth_t struct according to current FPGA structure, which sums to a size of 1244 bytes: package shape_pkg; typedef enum logic [7:0]{ SAWTOOTH = 0, SQUARE, SIN, PIANO } wave_shape; endpackage package protocol_pkg; //---------------------------------------------------------------------------------------------/ // Definition of protocol for communication between MCU and FPGA //---------------------------------------------------------------------------------------------/ import shape_pkg::*; typedef struct packed { logic [31:0] gain /* FIXED POINT */; logic [31:0] duration; } envelope_t; typedef struct packed { logic [31:0] freq /* FIXED POINT */; logic [31:0] velocity; envelope_t [0:`ENVELOPE_LEN-1] envelopes; wave_shape shape; logic [7:0] cmds; } wavegen_t; // typedef struct packed { // logic [31:0] delay; // logic [31:0] feedback; // } echo_t; typedef struct packed { logic signed [31:0] balance /* FIXED POINT */; } pan_t; // typedef struct packed { // logic [31:0] playback_volume; // logic [7:0] playback_speed; // logic [7:0] cmds; // } looper_t; typedef struct packed { // Tau values are delays. 0, 1, 2 and 3 are for comb filters, selected in a 1:1.5 range where // the smallest determines the initial delay. Tau 4 and 5 are for all-pass filters, usually // set to 5ms and 1.7 ms. Tau is computed as time [ms] * sample rate // Gain values are always < 1.0. 0, 1, 2 and 3 are for comb filters = 10 ^ (-3 t / T), where // T is the reverb time. logic signed [31:0][0:5] tau; logic signed [31:0][0:6] gain /* FIXED POINT */; } reverb_t; typedef struct packed { wavegen_t [0:`N_OSCILLATORS-1] wave_gens; logic [31:0] master_volume; reverb_t reverb; pan_t pan; // looper_t looper; // echo_t echo; } synth_t; endpackage Possible changes to save bandwidth: typedef struct packed { logic [7:0] gain /* FIXED POINT */; logic [7:0] duration; } envelope_t; Takes size from 1244 bytes to 524 bytes.
2025-04-01T04:54:52.390846
2018-07-05T09:25:56
338494380
{ "authors": [ "stephaniesimms", "xsrust" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13686", "repo": "DMPRoadmap/roadmap", "url": "https://github.com/DMPRoadmap/roadmap/issues/1661" }
gharchive/issue
Configure Plan readability through branding.yml Currently, plans are readable by a super_admin, an admin of any of the organizations to which co-owners belong, or to any user with a role in the plan. For some services, it is not wanted for org-admins to be able to view the full text of any plan within their organization, and this should be restricted to super-admins, or org-admins with requested reviews. Note that a user requesting a review adds a role for the associated org-admin users on the plan. We should add a configuration option for whether or not org-admins can view plans from users of their org. This will affect whether the plans at /org_admin/plans have active links to the full DMP content or not. Where all plans can't be viewed by org admins, they will only have access to plans where a specific review has been requested (see notification panel above plans table) i tested as a super admin and org admin for UCSF and noted some weird behaviors: as a super admin I see the notifications panel w/many plans awaiting feedback (screenshot 1). i can also click the links to open each plan below and view the content. as an org admin I don't see the notifications panel, but I should (screenshot 2). when i click the links to open the plans in the main table i receive an error msg (screenshot 3) and am taken to My Dashboard. desired behavior is to make the default configuration to give Org Admins authorization to open and view plans. I think this was a requirement for UK unis in fact, and is the current behavior. Org Admins should always see plans awaiting feedback in the Notifications panel regardless of the configuration setting. if the tool is configured to prevent Org Admins from viewing the contents of all plans, the error message should not redirect them to My Dashboard. instead they should stay on the Plans page. Super Admin UCSF view of plans page Org Admin UCSF view of plans page Error message when I click the link to view a plan as an Org Admin default configuration looks good now. will just need to keep an eye on whether newly privileged admins can see a complete list of plans awaiting feedback. closing issue.
2025-04-01T04:54:52.395305
2022-04-25T17:15:40
1214808100
{ "authors": [ "johnpinto1", "mariapraetzellis", "pherterich" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:13687", "repo": "DMPRoadmap/roadmap", "url": "https://github.com/DMPRoadmap/roadmap/issues/3163" }
gharchive/issue
Conditional questions causing plans to disapear Plans that include a conditional question are disappearing. For example, a user can successfully create a plan from a template including a conditional question, but when they go back to view the plan, the plan will disappear from the page. We temporarily fixed this issue by removing the conditional question from the template, but I believe this bug will be affecting all plans that include conditional questions. @raycarrick-ed This seems to be related to issue I noticed when dealing with a bug for Maastricht. The plan sections container continues to have style="display:none". Will investigate. @mariapraetzellis, @briri Not yet identified cause, but working on it. As I have an example which fails. Just proved in broken case cause this rendering cuases everything in content_for to be surrouned by a display:none block. <%= render partial: 'phases/edit_plan_answers', locals: { plan: plan, phase: phase, answers: answers, readonly: readonly, base_template_org: base_template_org, guidance_presenter: guidance_presenter, } %> Thanks @briri. A cursory test shows commenting out hide questions fixes issue. Will tomorrow dig deeper as suggested in your comment. I think @raycarrick-ed added this to the latest release, so I will close out here. @pherterich I can't see it in https://github.com/DigitalCurationCentre/roadmap/commits/deploy/dmponline?before=99d780716fca202f19a6dcb7da40ce62edb52e76+35&branch=deploy%2Fdmponline&qualified_name=refs%2Fheads%2Fdeploy%2Fdmponline It is possible has been patched by @raycarrick It is patched and part of release 3.1.1 proper