added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T04:10:16.841500
2024-02-16T08:47:22
2138138683
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13802", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/35469" }
gharchive/issue
⚠️ Bofrost.gr has degraded performance In f14b00c, Bofrost.gr (https://www.bofrost.gr) experienced degraded performance: HTTP code: 200 Response time: 4760 ms Resolved: Bofrost.gr performance has improved in b907cc0 after 6 minutes.
2025-04-01T04:10:16.844402
2023-02-08T10:44:11
1575889188
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13803", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/3610" }
gharchive/issue
⚠️ Bofrost.at has degraded performance In 4b988c9, Bofrost.at (https://www.bofrost.at) experienced degraded performance: HTTP code: 200 Response time: 945 ms Resolved: Bofrost.at performance has improved in 2625fda.
2025-04-01T04:10:16.847325
2024-03-01T06:48:04
2162726031
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13804", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/36423" }
gharchive/issue
⚠️ Bofrost.de has degraded performance In 2485319, Bofrost.de (https://www.bofrost.de) experienced degraded performance: HTTP code: 200 Response time: 866 ms Resolved: Bofrost.de performance has improved in 65528b9 after 6 minutes.
2025-04-01T04:10:16.850261
2024-03-06T17:23:28
2172054268
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13805", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/36665" }
gharchive/issue
⚠️ Bofrost.com has degraded performance In 13c32b8, Bofrost.com (https://www.bofrost.com) experienced degraded performance: HTTP code: 200 Response time: 962 ms Resolved: Bofrost.com performance has improved in 1d57c79 after 6 minutes.
2025-04-01T04:10:16.853374
2024-03-13T14:41:08
2184185096
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13806", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/37256" }
gharchive/issue
⚠️ Bofrost.at has degraded performance In 2b92d87, Bofrost.at (https://www.bofrost.at) experienced degraded performance: HTTP code: 200 Response time: 865 ms Resolved: Bofrost.at performance has improved in 83c70cb after 6 minutes.
2025-04-01T04:10:16.856293
2024-03-18T18:17:56
2192942728
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13807", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/37908" }
gharchive/issue
⚠️ Bofrost.lu has degraded performance In d3f0e5f, Bofrost.lu (https://www.bofrost.lu) experienced degraded performance: HTTP code: 200 Response time: 1468 ms Resolved: Bofrost.lu performance has improved in 1a38b9c after 8 minutes.
2025-04-01T04:10:16.859182
2024-04-01T20:51:16
2219040731
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13808", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/39853" }
gharchive/issue
⚠️ Bofrost.be has degraded performance In e8fb670, Bofrost.be (https://www.bofrost.be) experienced degraded performance: HTTP code: 200 Response time: 1112 ms Resolved: Bofrost.be performance has improved in a9ac0b3 after 24 minutes.
2025-04-01T04:10:16.865276
2024-04-27T08:48:45
2266919267
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13810", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/43442" }
gharchive/issue
⚠️ Bofrost.de has degraded performance In 0bf161c, Bofrost.de (https://www.bofrost.de) experienced degraded performance: HTTP code: 200 Response time: 1389 ms Resolved: Bofrost.de performance has improved in afefd49 after 52 minutes.
2025-04-01T04:10:16.868151
2024-05-13T10:41:41
2292453024
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13811", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/45979" }
gharchive/issue
⚠️ Bofrost.at has degraded performance In 47ad293, Bofrost.at (https://www.bofrost.at) experienced degraded performance: HTTP code: 200 Response time: 1256 ms Resolved: Bofrost.at performance has improved in 2095d6e after 6 minutes.
2025-04-01T04:10:16.871062
2024-05-16T02:30:12
2299203578
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13812", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/46404" }
gharchive/issue
⚠️ Bofrost.be has degraded performance In 814e1c5, Bofrost.be (https://www.bofrost.be) experienced degraded performance: HTTP code: 200 Response time: 1220 ms Resolved: Bofrost.be performance has improved in d69836c after 16 minutes.
2025-04-01T04:10:16.873967
2024-06-23T09:32:55
2368470146
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13813", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/51996" }
gharchive/issue
⚠️ Bofrost.lu has degraded performance In 757cbd2, Bofrost.lu (https://www.bofrost.lu) experienced degraded performance: HTTP code: 200 Response time: 2141 ms Resolved: Bofrost.lu performance has improved in b300b3b after 6 minutes.
2025-04-01T04:10:16.877044
2024-07-11T13:43:53
2403230939
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13814", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/54662" }
gharchive/issue
⚠️ Bofrost.de has degraded performance In 19860cd, Bofrost.de (https://www.bofrost.de) experienced degraded performance: HTTP code: 200 Response time: 836 ms Resolved: Bofrost.de performance has improved in 20b338e after 21 minutes.
2025-04-01T04:10:16.879910
2023-03-29T07:37:49
1645206724
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13815", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/6793" }
gharchive/issue
⚠️ Bofrost.be has degraded performance In 695a07b, Bofrost.be (https://www.bofrost.be) experienced degraded performance: HTTP code: 200 Response time: 3744 ms Resolved: Bofrost.be performance has improved in ccb32c5.
2025-04-01T04:10:16.882801
2023-04-08T22:34:32
1659665041
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13816", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/7522" }
gharchive/issue
⚠️ Bofrost.de has degraded performance In d09da3a, Bofrost.de (https://www.bofrost.de) experienced degraded performance: HTTP code: 200 Response time: 1310 ms Resolved: Bofrost.de performance has improved in df699a1.
2025-04-01T04:10:16.885679
2023-04-12T18:46:11
1665069109
{ "authors": [ "Ckal" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13817", "repo": "Ckal/Uptime", "url": "https://github.com/Ckal/Uptime/issues/7796" }
gharchive/issue
⚠️ Bofrost.be has degraded performance In f684792, Bofrost.be (https://www.bofrost.be) experienced degraded performance: HTTP code: 200 Response time: 1550 ms Resolved: Bofrost.be performance has improved in e3f701b.
2025-04-01T04:10:16.920105
2023-06-06T22:43:07
1744724826
{ "authors": [ "Clauvin" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13818", "repo": "Clauvin/ambient_ui_clock", "url": "https://github.com/Clauvin/ambient_ui_clock/issues/1" }
gharchive/issue
Build basic clock [x] Research minimalistic clocks; [x] Choose a clock design to implement; [x] Created a circle [x] Make code get time from the OS where the program is running (initially, that's Windows); [x] Add hour, minute and second hands, with no movement; [x] Add rotational movement of hands, with an added-at-start initial time; [x] Refactor code creating functions for rotational movement; [ ] Add test code of rotational movement; And chaosprint#5340 from Discord basically did the whole code. Since the objective here is to learn, let's gradually built our code understanding parts of his. And almost there, only need the test code.
2025-04-01T04:10:16.923006
2024-05-27T11:25:56
2318925088
{ "authors": [ "lauracchen", "srmsoumya", "yellowcap" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13819", "repo": "Clay-foundation/model", "url": "https://github.com/Clay-foundation/model/pull/256" }
gharchive/pull-request
Remove clay-v0-tutorials & add visualize embeddings for v1 Remove deprecated tutorial notebooks for clay-v0 Add a tutorial notebook to show how embedding space looks like for Clay Encoder (v1) Show examples for 2 sensors, i.e NAIP & Sentinel-2 hey @srmsoumya is there any reason to keep but "archive" the older tutorial notebooks? I doubt there's a reason someone wants to use v0 over v1 at this point, but maybe in the future we should think about keeping deprecated notebooks so people can still make use of the older model versions if they want to? hey @srmsoumya is there any reason to keep but "archive" the older tutorial notebooks? @lauracchen the notebooks stay in the repo, so that is definitively a way to archive them. We were also thinking of re-organising the content once we have merged the main pieces and we see what structure would make sense. Let's consider linking to the old notebooks at that point. Merging to have the updates notebooks out asap. Let's figure out the linters afterwards.
2025-04-01T04:10:17.177587
2022-10-27T14:42:19
1425755886
{ "authors": [ "charleskawczynski" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13821", "repo": "CliMA/ClimaAtmos.jl", "url": "https://github.com/CliMA/ClimaAtmos.jl/pull/1037" }
gharchive/pull-request
Try to fix edmf inconsistency job This PR breaks up compute_ref_state into compute_ref_pressure! and compute_ref_density so that we can call compute_ref_pressure in sgs_flux_tendency! when anelastic and test_consistency. This doesn't fix the problem, yet, but it's a clear correction. bors r+ bors r+
2025-04-01T04:10:17.179100
2022-11-01T17:38:38
1431818859
{ "authors": [ "charleskawczynski", "sriharshakandala" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13822", "repo": "CliMA/ClimaAtmos.jl", "url": "https://github.com/CliMA/ClimaAtmos.jl/pull/1067" }
gharchive/pull-request
Use log2 y-axis for SYPD scaling plot Purpose Use log10 y axis for SYPD scaling plot [x] I have read and checked the items on the review checklist. bors r+ bors r+
2025-04-01T04:10:17.187907
2022-08-16T23:31:56
1340981057
{ "authors": [ "charleskawczynski" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13823", "repo": "CliMA/Thermodynamics.jl", "url": "https://github.com/CliMA/Thermodynamics.jl/pull/126" }
gharchive/pull-request
Add microbenchmarks for Float32 and Float64 This PR adds microbenchmarks for Float32. I'm hoping that this provides more information on the impact of #125. However, we won't know the true impact until we test in ClimaAtmos since performance seems to depend on the actual values used. bors r+
2025-04-01T04:10:17.373855
2022-10-12T19:10:47
1406682723
{ "authors": [ "mcarbonneaux", "zhicwu" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13824", "repo": "ClickHouse/clickhouse-jdbc", "url": "https://github.com/ClickHouse/clickhouse-jdbc/issues/1107" }
gharchive/issue
JDBC Url need support urls with underscores (RFC3986)... Generaly when you use docker-compose the hosts are formed with underscore between component... like <docker-compose directory name>_<app name>_<instance number>... This url are rejected by clickhouse-jdbc, with error "host is missed or wrong" or "port is missed or wrong" depend on clickhouse-jdbc version... or not get the good port (respond the default port by default = 8123)... The problem seem to come from a bug/strictness in java.net.URI from jdk used by clickhouse-jdbc... java.net.URI are used in https://github.com/ClickHouse/clickhouse-jdbc/blob/master/clickhouse-jdbc/src/main/java/ru/yandex/clickhouse/ClickhouseJdbcUrlParser.java#L42 to decode jdbc url... Is related to bugs in jdk : https://bugs.openjdk.org/browse/JDK-8221675 and https://bugs.openjdk.org/browse/JDK-8019345 and https://stackoverflow.com/questions/22203111/is-javas-uri-resolve-incompatible-with-rfc-3986-when-the-relative-uri-contains. In fact java.net.URI are based on old dns name RFC2396 that not authorize underscore... them this RFC had been updated in RFC3986 that authorize underscore... Also spring had the same pb... https://github.com/spring-projects/spring-framework/issues/24439, and had create alternate class that resolve the pb... The prefered solution while be to fix in java.net.URI... but the bug are old, maybe may corrected in jdk18+... but i think that can continue to wait for long time... https://bugs.openjdk.org/browse/JDK-8291591 and https://cr.openjdk.java.net/~dfuchs/writeups/updating-uri/ The other solution are to use alternative URI parse in clickhouse jdbc driver like in spring... with org.springframework.web.util.UriComponentsBuilder... https://docs.spring.io/spring-framework/docs/current/javadoc-api/org/springframework/web/util/UriComponentsBuilder.html Thanks for the report @mcarbonneaux. URI should not be mandatory except for the optional HttpClient implementation. URL on the other hand is less strict and it works for the case. I'll add a custom class(with toJavaUri() method to create a patched URI for HttpClient or such) and a loose parser to replace URI. Meanwhile, if you're using docker-compose, can you use service name(app name in your example) instead? Since service name may contain underscore as well, depending on your scenario, perhaps a custom DNS like consul or gateway like traefik can be of use? Is not the only the underscore the problem, if you see the rfc matrix https://cr.openjdk.java.net/~dfuchs/writeups/updating-uri/... For my docker compose the default search domain are not the domain of the network... And the network also had underscore too....
2025-04-01T04:10:17.387109
2024-03-27T18:10:41
2211475580
{ "authors": [ "JMGilbert", "brews", "kemccusker" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13825", "repo": "ClimateImpactLab/dscim", "url": "https://github.com/ClimateImpactLab/dscim/issues/232" }
gharchive/issue
CI tests fail when upgrading pandas from 2.1.4 to >=2.2.0 PR #216 suggests the package fails tests when the CI environment bumps pandas from v2.1.4 to the latest version (currently v2.2.1). It started failing since the pandas v2.2.0 release. I can't judge why it's failing from a quick look at the testing logs. Archived logs from the failed CI run are here: logs_22096399592.zip Looks like all the test_*_points tests failed. The first failure is test_adding_up_points with an AssertionError on assert_frame_equal. Looks like something is out of order somehow. I updated the title and description with this info. I think it's probably related to this: https://pandas.pydata.org/docs/whatsnew/v2.2.0.html#merge-and-dataframe-join-now-consistently-follow-documented-sort-behavior #216 now includes a fix. I appreciate the fix @JMGilbert! Stop me if I'm reading this wrong but on quick glance it looks like the issue isn't with the dscim package code itself. The issue is more that the tests were written to read the expected results from an older file of stashed results. This expected results file was written with an older version of pandas and so it needs to get sorted before it can be compared with the results generated from the newer pandas version? Looking at this ... hmm... It's not something that needs to get fixed here, but It would be ideal if the tests didn't need to handle this type of logic, instead focusing on what's needed to test some target behavior. The hacky workaround stuff makes it harder to reason what the test is actually trying to test. It's going to keep getting muddled as time marches forward. Again, I think this is absolutely something that shouldn't be addressed in this issue and PR. The code that needs updating had this hacky stuff before this issue. Depending on priorities, this is maybe something to consider for future cleanup before legacy slows things down too much, @kemccusker. It might come back to bite later. @brews would love your input on how better to write the tests to avoid these types of issues, for future cleanup. For this PR, sorting seems like a reasonable fix, yeah? Closed by #216
2025-04-01T04:10:17.419213
2024-06-14T13:36:48
2353406337
{ "authors": [ "Clivern", "mateusz-spychala" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13826", "repo": "Clivern/Lynx", "url": "https://github.com/Clivern/Lynx/issues/83" }
gharchive/issue
Unhealthy container because of lack of curl package in docker image When I launch docker-compose from master, container is unhealthy. This is caused by lack of curl package in docker image. docker image clivern/lynx:0.11.13 is release which has curl Issue is resolved by https://github.com/Clivern/Lynx/pull/84
2025-04-01T04:10:17.421019
2016-10-13T09:12:58
182735768
{ "authors": [ "CloCkWeRX", "fand" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13827", "repo": "CloCkWeRX/tipsy", "url": "https://github.com/CloCkWeRX/tipsy/pull/10" }
gharchive/pull-request
Add package.json Added package.json so that we can install this package via npm. By the way, I couldn't run rake build because I don't have project-kit. I googled a minute but couldn't find out project-kit. Should I replace Rakefile with npm scripts? Should I replace Rakefile with npm scripts? Sure, why not
2025-04-01T04:10:17.448145
2015-03-10T00:16:12
60431755
{ "authors": [ "VortixDev", "unusualcrow" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13828", "repo": "CloudSixteen/Clockwork", "url": "https://github.com/CloudSixteen/Clockwork/issues/193" }
gharchive/issue
Broken scoreboard, error upon opening menu with TAB http://i.imgur.com/H8xe0EK.png Multiple people were on the server. Not moving the scoreboard down as it should. http://i.imgur.com/0EDU8mi.png When opening the tab menu. http://i.imgur.com/8ewAcyJ.png Selection color stays there after running the mouse over an option. http://i.imgur.com/dmF7LQu.jpg I equipped and held a suitcase. Please ask for help regarding your own server's issues on the forums (http://forums.cloudsixteen.com)
2025-04-01T04:10:17.460498
2022-10-13T13:32:56
1407816026
{ "authors": [ "Pritam991", "nikhil25803" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13829", "repo": "Clueless-Community/Datasets", "url": "https://github.com/Clueless-Community/Datasets/pull/192" }
gharchive/pull-request
Add files via upload Description Issue Ticket Number Fixes #(issue_number) Checklist: [x] I have followed the contributing guidelines of this project as mentioned in CONTRIBUTING.md [x] I have cross-checked that the dataset I'm uploading is not present in the project [x] I have cross-checked that the dataset is present in the correct folder Hi folks, My first contribute in this community. Glad to part of this community. Thank You all. Hey @Pritam991 Congrats 🎉, your PR got merged. To be eligible for the swags from ClueLess. Follow these steps. Star this repository. ⭐ SignIn / SignUp in [[ClueLess Website](https://www.clueless.tech/)](https://www.clueless.tech/) . Go to the profile section from the top right-hand corner. Comment your profile URL under this comment tagging the person who merged your PR. That’s it. Follow the leaderboard! And don’t forget to share your achievement with your friends. https://www.clueless.tech/profile/admin/116003764215798886346 @nikhil25803
2025-04-01T04:10:17.467455
2024-05-15T04:57:12
2296864927
{ "authors": [ "Siddhesh-Agarwal", "nikhil25803" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13830", "repo": "Clueless-Community/scrape-up", "url": "https://github.com/Clueless-Community/scrape-up/issues/976" }
gharchive/issue
Add support for proxies Describe the feature As a web scrapping library, supporting Proxies while sending requests would be useful. I propose a RequestConfig class where we can set the request timeout, proxies, redirect, etc. Add ScreenShots The class would look something like this: class RequestConfig: timeout: int = 10 proxy: dict[str, str] = {} redirect: bool = 10 It should also be passable to Different Scrapers. Record [X] I agree to follow this project's Code of Conduct [X] I'm a GSSoC'24 contributor [X] I want to work on this issue Go ahead @Siddhesh-Agarwal Note Please create a separate module for this, as in the folder and project structure (if it is already created, just add your features as functions in the same module). Do not use the `selenium web driver as it is incompatible with all devices and cloud platforms. Before making any changes, please check whether the module you want to add exists. If yes, then you can add your functionality as a method only make a separate module and class for it. All the best 👨‍💻 @nikhil25803 Here is how I am thinking of solving this problem: create a config folder in src/scrape_up Create the class in that folder Here is the step I have some confusion about: should I add a config parameter to all scraping classes? if not, then what other way can we use the config? Yes go ahead. But for now add request timeout feature only. Hey, Rewriting every request in the project, every time more features are added, doesn't seem very maintainable. I am adding a new get function that utilizes request.get() and RequestConfig. @nikhil25803 I have added support for timeout, allow_redirect and headers parameters for now, I will add support for proxies if you think everything till this point is fine. Hey @nikhil25803 Can I create a PR for Proxy support as well? It is a small 7-8 line addition in one file Yeah go ahead @Siddhesh-Agarwal
2025-04-01T04:10:17.470382
2022-10-17T14:51:07
1411723419
{ "authors": [ "Rajdip019", "rohan9024" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13831", "repo": "Clueless-Community/web-ui-kit", "url": "https://github.com/Clueless-Community/web-ui-kit/issues/272" }
gharchive/issue
Create a Mobile Filter One need to make this component using HTML and Tailwind CSS. I would suggest to use Tailwind Playgrounds to make things faster and quicker. Here is a reference to the component. After building the component please raise a PR with a screenshot of the component and add the component in path of the file. If you need to use any icon please use it from Hero Icons Good luck. @Rajdip019 please assign this Go ahead @rohan9024 @Rajdip019 done👍
2025-04-01T04:10:17.531614
2020-04-29T00:19:44
608695381
{ "authors": [ "bkase", "ghost-not-in-the-shell" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13832", "repo": "CodaProtocol/coda", "url": "https://github.com/CodaProtocol/coda/issues/4807" }
gharchive/issue
BuildKite CI: New integration tests pipeline(s) Depends on us finishing the first integration test. This entails adding a new pipeline for the test and possibly dhall infrastructure for helping to define future integration tests (for example a function that takes the name of the test and sets up all the boilerplate for you). This will become more clear after we have a few integration tests implemented. Epic: #4762 generate/obtain a docker image for the current build pipeline build test_executive.exe load secret key json file for a service account with permissions to deploy authenticate service account using gcloud cli (e.g. gcloud auth activate-service-account set export GCLOUD_API_KEY="..." to a valid gcloud api key with permissions to pub/sub and logging execute test like so ./_build/default/src/app/test_executive/test_executive.exe --coda-image <docker-image-id> <test-name> for each test --nholland94
2025-04-01T04:10:17.534673
2021-12-17T18:27:33
1083519002
{ "authors": [ "Code-Hex", "aoxn" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13833", "repo": "Code-Hex/vz", "url": "https://github.com/Code-Hex/vz/issues/19" }
gharchive/issue
support for BridgedNetwork Hi, do you have plan to support for BridgedNetwork? @aoxn Thanks for asking me and creating a new PR (https://github.com/Code-Hex/vz/pull/20) Yes, I want to support BridgedNetwork. However, I was wondering how to implement this. Could you tell me how do you think how to use the BridgedNetwork? A simple example code would be helpful. Thanks/
2025-04-01T04:10:17.536418
2020-10-17T12:35:59
723748440
{ "authors": [ "TravelTimN", "stefdworschak" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13834", "repo": "Code-Institute-Community/ci-hackathon-app", "url": "https://github.com/Code-Institute-Community/ci-hackathon-app/pull/73" }
gharchive/pull-request
Pull request template Description: Adding a template for Pull Requests with suggested headings to simplify the process of creating a new Pull Request. @stefdworschak are you waiting for me on this one? Yep. Please review and if you're okay with it approve it. @TravelTimN thanks for reviewing. Could you please have another look. I added the rest of the forking instructions. Also, the images are linked already, but will only be working once the PR is merged as I have added the link to where the files will be on master.
2025-04-01T04:10:17.557570
2018-03-12T02:59:58
304226801
{ "authors": [ "CorneeSean", "codecov-io" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13835", "repo": "CodeForPoznan/volontulo", "url": "https://github.com/CodeForPoznan/volontulo/pull/872" }
gharchive/pull-request
Fix improper usage of BehaviorSubject throughout Angular app Description: Fixes improper usage of BehaviorSubject class and replaces it with Subject where apropriate, making overall code more readable and less error-prone. New imports / dependencies: N/A Unit Tests: N/A What tests do I need to run to validate this change: Login page: try to login with wrong mail/password - login error message should appear Reset password page: after form submission success message should appear Confirm reset pasword page: After confirmation form is sent successfully, you should be redirected to 'login' page Organizations list: ogranizations list should appear (warning: it will not work until #870 is fixed) Organization contact form: message is shown after form submission Codecov Report Merging #872 into master will increase coverage by 0.03%. The diff coverage is 59.09%. @@ Coverage Diff @@ ## master #872 +/- ## ========================================== + Coverage 91.71% 91.74% +0.03% ========================================== Files 106 106 Lines 3004 3003 -1 Branches 10 10 ========================================== Hits 2755 2755 + Misses 248 247 -1 Partials 1 1 Impacted Files Coverage Δ frontend/src/app/login/login.component.ts 85.71% <0%> (ø) :arrow_up: ...src/app/password-reset/password-reset.component.ts 72.22% <0%> (+3.8%) :arrow_up: ...ntend/src/app/organization/organization.service.ts 64.51% <100%> (+1.18%) :arrow_up: frontend/src/app/auth.service.ts 57.14% <50%> (-0.86%) :arrow_down: Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 393d311...abecb2b. Read the comment docs.
2025-04-01T04:10:17.569780
2022-02-08T05:49:29
1126823132
{ "authors": [ "mathewgrabau" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13837", "repo": "CodeMazeBlog/CodeMazeGuides", "url": "https://github.com/CodeMazeBlog/CodeMazeGuides/pull/313" }
gharchive/pull-request
Added sample code for BasicDelegates, Action and Func<EMAIL_ADDRESS>Mathew Grabau Sample code that goes with my article. Closing it per the guidelines in the email.
2025-04-01T04:10:17.579144
2024-01-15T04:52:58
2081247801
{ "authors": [ "patchworquill", "shehabattia96" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13838", "repo": "CodeToCAD/CodeToCAD", "url": "https://github.com/CodeToCAD/CodeToCAD/issues/298" }
gharchive/issue
Blender: one-direction extrude Blender's built-in extrude thickens the shape in two directions. We need a way to extrude in only one direction normal to a face. Blender's docs on Extrude explain the behaviour and implementation in more detail for reference: https://docs.blender.org/manual/en/latest/modeling/meshes/editing/duplicating/extrude.html This stackoverflow post has code for the API with a simple cube face extrusion example: https://stackoverflow.com/questions/37808840/selecting-a-face-and-extruding-a-cube-in-blender-via-python-api I was thinking a decent solution is to internally duplicate the face, translate it the extrude length along the normal vector, and loft the two surfaces using our existing loft method. The trickiest part here is to translate along the normal vector imo. This thread has the formulas for calculating a normal based on the face data using vector crossproduct and vector subtraction: https://blenderartists.org/t/getting-face-normals-from-python/309648 It also seems you are correct that the API doesn't provide access to them... @patchworquill Thanks for the reference, it helped! We can now project a face a certain distance away normal to another face. There is still a bug with rotated faces - I need to find the translation component due to rotation and add that to the translation vector before projecting it normal to the source face:
2025-04-01T04:10:17.634433
2019-12-22T17:11:54
541478152
{ "authors": [ "DavertMik", "koushikmohan1996" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13839", "repo": "Codeception/CodeceptJS", "url": "https://github.com/Codeception/CodeceptJS/pull/2098" }
gharchive/pull-request
Add grab*FromAll method Motivation/Description of the PR Applicable helpers: [x] Webdriver [x] Puppeteer [x] Nightmare [ ] REST [x] Appium [x] Protractor [x] TestCafe Description of this PR, which problem it solves Resolves #2093 Type of change [ ] Breaking changes [ ] New functionality [ ] Bug fix [ ] Documentation changes/updates [ ] Hot fix [ ] Markdown files fix - not related to source code Checklist: [ ] Tests have been added [ ] Documentation has been added (Run npm run docs) [ ] Lint checking (Run npm run lint) [ ] Local tests are passed (Run npm test) I have created a new document under webapi folder. But when I try to generate document It also makes changes which I did not change. Maybe it would have been missed in previous merges. (So I did not add any md files in the PR) Here is an overview of what got changed by this pull request: Clones added ============ - lib/helper/WebDriver.js 1 - lib/helper/Puppeteer.js 1 See the complete overview on Codacy Here is an overview of what got changed by this pull request: Clones added ============ - lib/helper/WebDriver.js 1 - lib/helper/Puppeteer.js 1 See the complete overview on Codacy Here is an overview of what got changed by this pull request: Clones added ============ - lib/helper/Nightmare.js 2 - lib/helper/Puppeteer.js 6 - lib/helper/WebDriver.js 7 - lib/helper/Protractor.js 6 - lib/helper/TestCafe.js 2 See the complete overview on Codacy Here is an overview of what got changed by this pull request: Clones added ============ - lib/helper/Nightmare.js 2 - lib/helper/Puppeteer.js 6 - lib/helper/TestCafe.js 2 - lib/helper/Protractor.js 6 - lib/helper/WebDriver.js 7 See the complete overview on Codacy Here is an overview of what got changed by this pull request: Clones added ============ - lib/helper/Puppeteer.js 6 - lib/helper/WebDriver.js 7 - lib/helper/TestCafe.js 2 - lib/helper/Nightmare.js 2 - lib/helper/Protractor.js 6 See the complete overview on Codacy Here is an overview of what got changed by this pull request: Clones added ============ - lib/helper/Puppeteer.js 6 - lib/helper/WebDriver.js 7 - lib/helper/TestCafe.js 2 - lib/helper/Protractor.js 6 - lib/helper/Nightmare.js 2 See the complete overview on Codacy Here is an overview of what got changed by this pull request: Clones added ============ - lib/helper/Nightmare.js 2 - lib/helper/Protractor.js 6 - lib/helper/TestCafe.js 2 - lib/helper/WebDriver.js 7 - lib/helper/Puppeteer.js 6 See the complete overview on Codacy Should add GrabCss for nightmare helper too after ##2111 merges
2025-04-01T04:10:17.689361
2016-10-04T00:46:36
180782283
{ "authors": [ "jhoffner", "kgraves" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13840", "repo": "Codewars/codewars.com", "url": "https://github.com/Codewars/codewars.com/issues/600" }
gharchive/issue
javascript 'WeIrD StRiNg CaSe' kata test case broken Many people are talking about it here The test passes this is a test into the written function, and here's the output: Expected: 'ThIs Is A TeSt', instead got: 'ThIs Is A tEsT' You can see the error between the 'A' and the 'T' of test. The algorithm is supposed to ignore spaces, as evidence by the beginning of the expected answer. Please log this issue on the kata discussion section, as an issue. Github issues is for site functionality only. @jhoffner It already is. It has been tagged as an issue at least 7 times in the discussion, and has a multitude of comments about the problems with the tests.
2025-04-01T04:10:17.697764
2023-08-16T15:30:55
1853475354
{ "authors": [ "CGNonofr", "kaisalmen", "killpack", "mehulmpt" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13841", "repo": "CodinGame/monaco-vscode-api", "url": "https://github.com/CodinGame/monaco-vscode-api/issues/164" }
gharchive/issue
Unable to open file \tmp\index.js Thank you for the fantastic work on this library. I'm using Next.js to bundle this (which is using webpack internally) and this full setup works fine - except on Windows. On windows, here is how it crashes inside edge browser: This is the minimal reproduction: Clone this repo: https://github.com/mehulmpt/next-monaco-demo Do yarn install Open on edge on windows Or visit this URL where it is already deployed, on edge on windows: https://nextdemo-5yofq3tfg-mehulmpt.vercel.app/ What is wrong here exactly - I cannot figure out. Could you help? Indeed, it was not tested on windows Is there the same error on the demo here? I'll guess I'll need to find a window somewhere The official demo works fine, but the above build fails. I couldn't understand if the bundler could cause this because my demo above works fine when used on macOS, Linux (chrome, safari, Firefox) @mehulmpt I can confirm your observation with Chrome / Edge on Windows, but it only occurs with your stack. We do similar things even with react in monaco-languageclient and no issues there. Yes, it's working fine with the official demo setup. I wonder why because it crashes on the in-memory filesystem and only on specific conditions (doesn't happen on chrome on mac for example) which is weird. I gave it another shot today morning and figured it out. Next.js/webpack injects process variable inside browser itself which causes confusion in this script: node_modules/monaco-editor/esm/vs/base/common/process.js This script is responsible for initializing the vscode process as win32 or posix. However, because typeof process is not really undefined (when it should have been - like it is in your demo example), it thinks it is running inside native Node.js environment. After this, we're probably firing off code branches depending on if you're reading from process or from variables like isWindows which use user agent to determine which causes those crashes. TL;DR - this can also be reproduced by setting your user agent to the following windows useragent on your Mac using Next.js build because Next internally also fills process variable: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/<IP_ADDRESS> Safari/537.36 I am not sure what the official solution should be like, but for now I have patched the file itself to force it into using browser safeProcess always. Here's the diff file if someone else gets stuck here: diff --git a/node_modules/monaco-editor/esm/vs/base/common/process.js b/node_modules/monaco-editor/esm/vs/base/common/process.js index 47b3de1..865041d 100644 --- a/node_modules/monaco-editor/esm/vs/base/common/process.js +++ b/node_modules/monaco-editor/esm/vs/base/common/process.js @@ -3,37 +3,14 @@ * Licensed under the MIT License. See License.txt in the project root for license information. *--------------------------------------------------------------------------------------------*/ import { globals, isMacintosh, isWindows } from './platform.js'; -let safeProcess; -// Native sandbox environment -if (typeof globals.vscode !== 'undefined' && typeof globals.vscode.process !== 'undefined') { - const sandboxProcess = globals.vscode.process; - safeProcess = { - get platform() { return sandboxProcess.platform; }, - get arch() { return sandboxProcess.arch; }, - get env() { return sandboxProcess.env; }, - cwd() { return sandboxProcess.cwd(); } - }; -} -// Native node.js environment -else if (typeof process !== 'undefined') { - safeProcess = { - get platform() { return process.platform; }, - get arch() { return process.arch; }, - get env() { return process.env; }, - cwd() { return process.env['VSCODE_CWD'] || process.cwd(); } - }; -} -// Web environment -else { - safeProcess = { - // Supported - get platform() { return isWindows ? 'win32' : isMacintosh ? 'darwin' : 'linux'; }, - get arch() { return undefined; /* arch is undefined in web */ }, - // Unsupported - get env() { return {}; }, - cwd() { return '/'; } - }; -} +const safeProcess = { + // Supported + get platform() { return isWindows ? 'win32' : isMacintosh ? 'darwin' : 'linux'; }, + get arch() { return undefined; /* arch is undefined in web */ }, + // Unsupported + get env() { return {}; }, + cwd() { return '/'; } +}; /** * Provides safe access to the `cwd` property in node.js, sandboxed or web * environments. * Can't you prevent process from being injected? starting from the v5, webpack stopped injecting "builtins" like process, I guess next is configured to continue doing it Could try, but Next.js uses it for a variety of things like injecting public envs into browser. @mehulmpt Did you end up finding a fix or workaround for this? I am running into a similar issue. Your reproducing example has been deleted and I'm at a loss for how to proceed.
2025-04-01T04:10:17.723259
2023-02-27T14:24:27
1601292472
{ "authors": [ "CGNonofr", "kaisalmen" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13842", "repo": "CodinGame/monaco-vscode-api", "url": "https://github.com/CodinGame/monaco-vscode-api/pull/69" }
gharchive/pull-request
Expand to npm workspace Transforms to npm workspace. It does not move the main code into a sub-folder. The question if this is desired or doesn't have any weird side effects: "workspaces": [ ".", "demo" ] What was done in addition: Demo monaco-vscode-api contraint is a wildcard npm install will no longer pull vscode if version info file with same version content is found. It is checked before the download starts and the file is written at the end rollup plugin-typescript uses its own config (tsconfig.build.json) vscode+eslint works now for main and demo code (tsconfig.json contains base config and is used only for that) I have updated the main description! @CGNonofr I have moved the demo code to src/demo. It excluded from the rollup build via exclusion in tsconfig.build.json Code is still considered by vscode / eslint because it is contained in tsconfig.json vite requires the project to be build because in package.json use the project itself as devDependency "vscode": "." Build is green now. vite resolves the packages, so there is no need to self-reference in the package.json now. Please check if the demo works as expected. I have the ProductQualityContext.notEqualsTo error, did you rebase on main? also the README needs to be updated I'm not sure to understand how it works, why is every import of vscode/** red? also how does vite resolves the import? It resolve the stuff from the dist folder. That's why the local code has to be built first: export default defineConfig({ resolve: { alias: { 'vscode/services': resolve(__dirname, './dist/services.js'), 'vscode/service-override/messages': resolve(__dirname, './dist/messages.js'), 'vscode/service-override/notifications': resolve(__dirname, './dist/notifications.js'), 'vscode/service-override/dialogs': resolve(__dirname, './dist/dialogs.js'), 'vscode/service-override/modelEditor': resolve(__dirname, './dist/modelEditor.js'), 'vscode/service-override/configuration': resolve(__dirname, './dist/configuration.js'), 'vscode/service-override/keybindings': resolve(__dirname, './dist/keybindings.js'), 'vscode/service-override/textmate': resolve(__dirname, './dist/textmate.js'), 'vscode/service-override/languageConfiguration': resolve(__dirname, './dist/languageConfiguration.js'), 'vscode/service-override/theme': resolve(__dirname, './dist/theme.js'), 'vscode/service-override/tokenClassification': resolve(__dirname, './dist/tokenClassification.js'), 'vscode/service-override/snippets': resolve(__dirname, './dist/snippets.js'), 'vscode/service-override/languages': resolve(__dirname, './dist/languages.js'), 'vscode/monaco': resolve(__dirname, './dist/monaco.js'), vscode: resolve(__dirname, './dist/api.js'), 'webpack-loader': resolve(__dirname, './dist/webpack-loader.js') } } }) services': resolve(__dirname, './dist/ser I see! FYI it seems you can use regexes in there or even implement a custom resolver Can't we do something about it in the tsconfig? I see! FYI it seems you can use regexes in there or even implement a custom resolver Yes, it could be written nicer. Wanted to make this work first. Question is: Is this approach better / nicer than having a separated npm package in the demo folder (as it was before)? Can't we do something about it in the tsconfig? vite does not make use of a tsconfig here I see! FYI it seems you can use regexes in there or even implement a custom resolver Yes, it could be written nicer. Wanted to make this work first. Question is: Is this approach better / nicer than having a separated npm package in the demo folder (as it was before)? Can't we do something about it in the tsconfig? vite does not make use of a tsconfig here vite doesn't but vscode does for the intellisense btw, what happens if you alias vscode/* to .? won't it use the exports in the package.json? I updated the vite config and the tsconfig used by vscode Does the demo code needs to be included into the main lib then? Isn't the vite aliases enough? Does the demo code needs to be included into the main lib then? Isn't the vite aliases enough? It is not contained in the lib. I updated the rollup config to rely on tsconfig.build.json which defines an exclude src/demo. The tsconfig.json is only the base config and is used by eslint/vscode. It is not contained in the lib. I updated the rollup config to rely on tsconfig.build.json which defines an exclude src/demo. The tsconfig.json is only the base config and is used by eslint/vscode. But the demo dependencies still need to be included in the root dev dependencies, right? But the demo dependencies still need to be included in the root dev dependencies, right? True, but they are only devDependencies. But the demo dependencies still need to be included in the root dev dependencies, right? True, but they are only devDependencies. I'm not sure to see the advantage compared to having a separate folder with its own node_modules though. It's only devDependencies but it still complexify the package.json and it's not easy to know what dependency is used for what I'm not sure to see the advantage compared to having a separate folder with its own node_modules though. Yes, I agree. The solution did not turn out to be simple. A npm workspace project is the better idea, but moving the lib to a sub-package only to separate it from the demo is overkill. There are some enhancements that can be kept/incorporated like the version check and the eslint improvements @CGNonofr I have reverted the vite or npm workspace changes and only kept enhancements (see updated list in PR description: https://github.com/CodinGame/monaco-vscode-api/pull/69#issue-1601292472) Yesterdays status I kept on branch kaisalmen/demo-integration-backup Pipeline fails because of this. I have no idea why suddenly 🤷‍♂️ Look at this. They broke the GH action: Last working build: Broken build: ⬆️ vite has to be run from root of project when demo uses monaco-vscode-api locally otherwise codicons cannot be loaded. If vite is launched from demo folder top level node_modules cannot be read. I update the start script and README, so npm run start does npm i in demo and runs vite! Now this is good to go finally. I don't have an idea what to do about the broken GH action!? Pipeline fails because of this. I have no idea why suddenly man_shrugging Yeah I tried to install the runner on another machine but it's missing yarn... I tried to install the self-hosted runner on a server but I got that weird error I don't have the time to investigate right now FYI, I saw this this morning, too, but it worked then after a runner picked it: I have tested the build on a fork I created this morning and used ubuntu-latest as runner and it works nicely: https://github.com/kaisalmen/monaco-vscode-api/actions/runs/4311997614/jobs/7522012650 I have tested the build on a fork I created this morning and used ubuntu-latest as runner and it works nicely: https://github.com/kaisalmen/monaco-vscode-api/actions/runs/4311997614/jobs/7522012650 You mean you use the official runners? I switched to hosted one because there wasn't enough memory and it was crashing. You mean you use the official runners? Yes, we don't have self-hosted runners and I forked it under my GH account Did you (I mean someone at your company) update the runners recently, like yesterday 😉 ? Did you (I mean someone at your company) update the runners recently, like yesterday wink ? What do you mean? I have to run the runner on my own machine. yesterday I tried to create a docker of it to make it run on one of our dev server but I failed for a mysterious reason. Can you try going back to official runner here? What do you mean? I have to run the runner on my own machine. yesterday I tried to create a docker of it to make it run on one of our dev server but I failed for a mysterious reason. Oh, really on your own machine!? Any idea, when you have time to fix it. Jobs aren't currently picked up. The job from yesterday evening was not picked up over night and I tried to re-run with no luck, so far. Can you try going back to official runner here? I did it on a fork for testing only, nothing changed here and I do not intend to 🙂 Oh, really on your own machine!? Yeah, that's the only project that requires it and I didn't take the time to run it elsewhere, is required a private token so I can't just run it on any machine, it needs a ansible conf, I tried yesterday without success. I did it on a fork for testing only, nothing changed here and I do not intend to feel free to do it if it works now! feel free to do it if it works now! Ok, I will try it now It already worked without doing anything Yes I restarted the runner on my machine This PR has became way to lengthy. 😆 Are you ok with the changes now? Runner config adjustments (if you want any) should be done independently, I think.
2025-04-01T04:10:17.727757
2017-05-14T10:54:10
228533624
{ "authors": [ "CodingBash" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13843", "repo": "CodingBash/FlyCAGE", "url": "https://github.com/CodingBash/FlyCAGE/issues/29" }
gharchive/issue
Research into deploying on ilstu AWS servers Research into strategy for deploying application on ILSTU's AWS servers. TASKS: [x] Get in contact with ILSTU's web development team Met with ISU research department. Currently building dev environment
2025-04-01T04:10:17.744364
2018-01-24T13:32:10
291209710
{ "authors": [ "kwichmann", "nilsweb", "shiffman" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13844", "repo": "CodingTrain/website", "url": "https://github.com/CodingTrain/website/pull/404" }
gharchive/pull-request
R implementation of A* algorithm Since I had to code an implementation of this algorithm anyway, I thought I might as well add it here, in case anyone's interested. @shiffman Should I try to resolve those conflicts and merge it or should it just be added as a community contribution (without merging the code)? For now, let's have any ports other than Processing or p5.js be "community contributions"?
2025-04-01T04:10:17.755662
2023-07-17T07:43:46
1807180397
{ "authors": [ "CodiumAI-Agent", "mrT23", "okotek", "zmeir" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13845", "repo": "Codium-ai/pr-agent", "url": "https://github.com/Codium-ai/pr-agent/pull/76" }
gharchive/pull-request
Optimization of Inline Comments Publishing Type of PR: Refactoring PR Description: This PR optimizes the process of publishing inline comments by making a single API call instead of multiple calls. It also includes a deprecation warning for the publish_inline_comment method, encouraging the use of the new publish_inline_comments method. Additionally, the PR updates the PyGithub version in the requirements. PR Main Files Walkthrough: -pr_agent/git_providers/github_provider.py: Refactored the publish_inline_comment method to log a deprecation warning and use the new publish_inline_comments method. Introduced the create_inline_comment method to create an inline comment and the publish_inline_comments method to publish multiple inline comments in a single API call. Also, updated the Github authentication method. -pr_agent/tools/pr_reviewer.py: Modified the _publish_inline_code_comments method to create and publish multiple inline comments in a single API call. Also, updated the _prepare_pr_review method to keep only code suggestions that can't be submitted as inline comments. -requirements.txt: Updated the PyGithub version. Resolves https://github.com/Codium-ai/pr-agent/issues/49 PR Analysis 🎯 Main theme: Optimizing the publishing of GitHub review comments 📌 Type of PR: Enhancement 🧪 Relevant tests added: No ✨ Focused PR: Yes, the PR is focused as it aims to optimize the publishing of GitHub review comments by making a single API call instead of multiple ones. PR Feedback 💡 General PR suggestions: The PR is well-structured and the changes are clear. However, it would be beneficial to add some tests to ensure the new functionality works as expected. Also, the PR description is missing, it would be helpful to provide a brief description of the changes made for better understanding. @hussam789 FYI - it works :) /describe i thought about upgrading PyGithub to 1.59 i think there is some breaking change there, where they don't support 'position' anymore (the use 'line' field instead) @zmeir questions: are you sure all the existing functionalities still work (including one you haven't changed) ? probably there isn't backward compatibility, right ? (people with 1.58.2 won't be able to use) So.. let's merge? @mrT23 @hussam789 Preparing review... PR Analysis 🎯 Main theme: Optimization of Inline Comments Publishing 📌 Type of PR: Refactoring 🧪 Relevant tests added: No ‣ Insights from user's answer: No user answers were provided ✨ Focused PR: Yes, the PR has a clear and coherent title and description, and all PR code diff changes are properly derived from the title and description. 🔒 Security concerns: No, the PR does not introduce any obvious security issues. The changes are mostly related to refactoring and optimization, and do not involve handling user input or data storage. PR Feedback 💡 General PR suggestions: The PR is well-structured and the changes are clearly explained in the description. However, it would be beneficial to include tests for the new functionality. This would ensure that the changes work as expected and prevent potential regressions in the future. Thanks for picking up my slack everyone! Was planning on dealing with this later this week but y'all are too fast :)
2025-04-01T04:10:17.762428
2019-06-26T22:13:54
461205178
{ "authors": [ "claudia-romano", "rgaufman" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13846", "repo": "CodyHouse/codyhouse-framework", "url": "https://github.com/CodyHouse/codyhouse-framework/issues/26" }
gharchive/issue
How to use with webpack? Hi there, I'm trying to get https://codyhouse.co/gem/vertical-timeline/ working, but it depends on codyhouse-framework. I'm using webpack and added this to my application.js: require("codyhouse-framework/main/assets/js/util") require("vertical-timeline/assets/js/main") require("vertical-timeline/assets/css/style.scss") When I run webpack, I get: ERROR in ./node_modules/vertical-timeline/assets/css/style.scss (./node_modules/css-loader/dist/cjs.js??ref--7-1!./node_modules/postcss-loader/src??ref--7-2!./node_modules/sass-loader/lib/loader.js??ref--7-3!./node_modules/vertical-timeline/assets/css/style.scss) Module build failed (from ./node_modules/sass-loader/lib/loader.js): @import '../../../../../codyhouse-framework/main/assets/css/style.scss'; // ⚠️ make sure to import the CodyHouse framework ^ File to import not found or unreadable: ../../../../../codyhouse-framework/main/assets/css/style.scss. in /Users/hackeron/Dropbox/Development/timeline2/node_modules/vertical-timeline/assets/css/style.scss (line 1, column 1) @ ./node_modules/vertical-timeline/assets/css/style.scss 2:14-173 @ ./app/javascript/packs/application.js How do I import the framework in WebPack? Hi there, in your application.js file, you'll need to import the style file of the framework as well: require("codyhouse-framework/main/assets/css/style.scss"); Also, please make sure to remove the following line from vertical-timeline/assets/css/style.scss file: @import '../../../../../codyhouse-framework/main/assets/css/style.scss'; This was used to import the scss file of the framework but you are already taking care of this in the application.js.
2025-04-01T04:10:17.801218
2020-04-09T19:58:49
597518614
{ "authors": [ "Col-E", "Ilolka1337" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13847", "repo": "Col-E/Recaf", "url": "https://github.com/Col-E/Recaf/issues/193" }
gharchive/issue
Font Colors Customize Font Colors Short description of the feature users can customize the colors of the strings - methods - values etc. ​​to their taste Extra details: More verbose details of features This is actually supported, but is done externally for now. I wrote a small guide on it here: https://coley.software/recaf-creating-custom-themes/ @Col-E i hope you add so that it can be changed through the program interface I plan to eventually have a live editor with a preview mode. Its not that hard to set that up. But that'll be done later. I'm currently focusing on other issues. @Col-E thx 👍 Information on how to create a custom theme can be found here: https://col-e.github.io/Recaf/doc-advanced-theme.html The plugin that activates the theme editor (which isn't done yet, but is usable at this point) is in the #plugins channel on discord until I create a proper list of plugins somewhere.
2025-04-01T04:10:17.832236
2016-07-28T15:38:11
168127503
{ "authors": [ "ankon", "kovacicmilan" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13848", "repo": "Collaborne/paper-stepper", "url": "https://github.com/Collaborne/paper-stepper/pull/8" }
gharchive/pull-request
Styling change of disabled nav buttons Make disabled buttons visible, but change style to signal that they are disabled. Original discussion: https://m.collaborne.com/#stream/06b033a7-94e9-4502-82f6-ffc5df3c3489 . Can you please rebase to master and update this PR? The build issue should be fixed now. Merged, and released in version 1.1.0, as this is a change breaking the existing API/behavior.
2025-04-01T04:10:17.886520
2024-04-17T03:35:33
2247291973
{ "authors": [ "66ling66", "Mr-Harry", "dongzhuoyao" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13849", "repo": "CompVis/zigma", "url": "https://github.com/CompVis/zigma/issues/7" }
gharchive/issue
code is error TypeError: Mamba.init() got an unexpected keyword argument 'scan_type'. Did you solve this problem? Hi, could you tell me how to run into this error? If you follow our installing tutorial, you should be able to run it successfully: https://github.com/CompVis/zigma/blob/19ef944dd87673d258455d8f226a08ac8fa868ca/model_zigma.py#L1215
2025-04-01T04:10:17.894805
2023-09-22T07:33:30
1908350653
{ "authors": [ "EktaBharti1998" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13850", "repo": "ComplianceAsCode/compliance-operator", "url": "https://github.com/ComplianceAsCode/compliance-operator/issues/420" }
gharchive/issue
compliance operator scans result doesn't contain results for CIS control 4.2 Kubelet I am using compliance operator to scan OpenShift cluster. There is a list of properties which are present in some of the cluster scan results while missing in other cluster scan results. Following is the list of missing controls: 1.2.8 1.2.10 1.2.14 1.2.15 1.2.16 1.2.17 1.2.19 1.3.6 1.4.2 4.2 Could anyone please help in understanding why some of the clusters would not be scanned for the above listed controls and how can I fix this? Here is the evidence that the scan result of our cluster(refer the status corresponding to cluster mycluster-us-south-1-bx2.4x16 as this is the healthy cluster, do not refer to overall status as there are some other cluster which are not healthy or don't have compliance operator installed and affecting the overall status) contains following control results: <img width="1164" alt="Screenshot 2023-09-26 at 10 26 55 AM" src="https://github.com/ComplianceAsCode/compliance-operator/assets/73580551/ 0ed6a041-bfb0-43b9-a138-cf07f7e62eac"> <img width="1168" alt="Screenshot 2023-09-26 at 10 27 59 AM" src="https://github.com/ComplianceAsCode/compliance-operator/assets/73580551/1c5b196 8-ae7c-415f-be71-b68340b33b4d">
2025-04-01T04:10:17.916073
2021-01-14T08:56:31
785806510
{ "authors": [ "jklmnn" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13851", "repo": "Componolit/AZ3", "url": "https://github.com/Componolit/AZ3/issues/12" }
gharchive/issue
Iterators Create iterators for Z3 expressions to allow looping through their terms. Fixed by #13. Fixed by #13.
2025-04-01T04:10:18.030424
2024-08-24T20:43:43
2484855932
{ "authors": [ "khanti42" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13852", "repo": "Consensys/starknet-snap", "url": "https://github.com/Consensys/starknet-snap/pull/335" }
gharchive/pull-request
fix: improve error message in rpc input validation PR Description: Improved Error Messages for Union Struct Validation Before: When testing a union struct with multiple levels, such as: const multiLevelStruct = union([ object({ field1: string(), }), object({ field3: boolean(), }), ]); If the input was incorrect, for example: { "field1": true } The error message provided was not clear and didn't offer much guidance: "Expected the value to satisfy a union of `object | object`, but received: [object Object]" This generic message made it difficult to diagnose the specific issue within the nested structure, particularly when working with complex schemas. Now: The error handling has been improved to provide much clearer and more informative messages. With the same incorrect input: { "field1": true } The new error message provides detailed information about where the validation failed: At path: field1 -- Expected a string, but received: true Expected a value of type `never`, but received: `true` At path: field3 -- Expected a value of type `boolean`, but received: `undefined` Benefits: Enhanced Clarity: The error messages now pinpoint the exact field and expected type, making it easier to debug and correct invalid input. Detailed Feedback: By providing detailed feedback at each level of the validation, developers can quickly identify and resolve issues with their input data. Better Support for Complex Schemas: This improvement is especially valuable when dealing with nested or complex data structures, where understanding the root cause of validation failures is crucial. This PR improves the developer experience and reduce time in debugging. This is closed, because complex superstruct are implemented as : https://github.com/MetaMask/keyring-api/blob/c7068d1d585c116cd8cd2aa2e838429215800de5/src/api.ts#L63-L76 instead of unions.
2025-04-01T04:10:18.041649
2020-12-23T10:35:21
773656709
{ "authors": [ "AntoineSIMTEK", "pastewka", "sannant", "sindhu2993" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13853", "repo": "ContactEngineering/ContactMechanics", "url": "https://github.com/ContactEngineering/ContactMechanics/issues/28" }
gharchive/issue
Define clearly what forces are It often takes me a while, scrolling thru the code and looking at some simulation results, to figure out what the forces are actually meant to be. I think our convention is that substrate forces are $- \partial E_el / \partial u$, with u positive into the halfspace. These forces are negative in repulsive contacts. We usually like to represent our forces as being positive inside the contact area, and these would be the forces applied on the substrate in the positive u direction. That is why I get confused so easily. Also sometimes the force is returned, but it is called pressure. Note that inside the "minimization code", using the gradient of the energy, rather then the force, is the most natural thing. I think the current convention makes sense after all but it needs to be clarified here: https://computationalmechanics.github.io/ContactMechanics/usage.html and in the individual docstrings. @sindhu2993 , @pastewka what are your opinions ? I thought that we essentially follow the convention that's in Johnson, i.e. positive displacements are into the half space and positive pressures lead to positive displacements. This is what was discussed with @tjunge at birth of this code. I agree that forces and pressures are not clearly distinguished and we should make clear from variable names whether it is a force or a pressure. Yes positive forces applied onto the halspace lead to positive displacements. But evaluate_force is exactly the oposite of that force. def evaluate_elastic_energy(self, forces, disp): """ computes and returns the elastic energy due to forces and displacements Arguments: forces -- array of forces disp -- array of displacements """ # pylint: disable=no-self-use return .5 * self.pnp.dot(np.ravel(disp), np.ravel(-forces)) I agree that this looks wrong and should be inverted. Is this function ever used anywhere? I guess that you need it now for the optimizers. It is used a lot. I mean evaluate_force, or evaluate_disp,that is used as input for forces. What we could think about is to switch progressively to evaluate_gradient, that is always unambigous. inside evaluate gradient we can make a note saying that positive values correspond to forces pressing into the substrate. I like the idea of switching to evaluate_gradient. This sounds like a major refactor. We should probably do this together some time in the new year. Yes. The forces that are produced by the half space of course need to carry the minus sign. Calling it gradient will make this all unabiguous. I will be happy to implement these changes in the code. It was a real struggle for me to figure out the minus signs. Having this information in the docstring of evaluate_gradient would helpful. Okay, whoever feels responsible is of course welcome to implement that refactor... Just let us know if someone is working on this so we don't start twice. Yes. Just read this thread because I'm preparing v1.0. Did anybody ever start working on this? Should this go into v1.0? I would prefer to avoid major changes because this may break TopoBank. No. I won't find time for this during the next weeks
2025-04-01T04:10:18.044476
2022-12-16T15:02:40
1500379254
{ "authors": [ "mcrot", "pastewka" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13854", "repo": "ContactEngineering/topobank", "url": "https://github.com/ContactEngineering/topobank/issues/944" }
gharchive/issue
Statistics tables look empty Since a while we're using django-trackstats for collecting metrics without using a third-party service like Google. However, this is no longer working with version 0.92.0, the metric tables are empty .. don't know why yet. The tests concerning the metrics have passed. Maybe this is related to the high age of django-trackstats?! However, I have just tested the statistics export and it is working, also the old data is still there. When looking into the PG tables, they look empty, not sure why. The statistics table hasn't been mailed out in a long time and appears broken. I've disabled trackstats entirely as the module appears to be a major source of delays in delivering response. Essentially, every request to deliver data from the database also entails a write which appears to make things slow.
2025-04-01T04:10:18.189317
2015-04-30T20:29:07
72272878
{ "authors": [ "TomAugspurger", "cpcloud" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13855", "repo": "ContinuumIO/blaze", "url": "https://github.com/ContinuumIO/blaze/pull/1070" }
gharchive/pull-request
Improve the ReLabel repring and add some info to the docstring cc @TomAugspurger closes #1069 :+1:
2025-04-01T04:10:18.202465
2021-09-05T13:48:29
988514181
{ "authors": [ "DanielNoord", "Pidgeot" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13856", "repo": "CookieMonsterTeam/CookieMonster", "url": "https://github.com/CookieMonsterTeam/CookieMonster/issues/976" }
gharchive/issue
[Steam] Grimoire timer not being added correctly Describe the bug The grimoire timer gets appended multiple times onto the maigc meter text before being cleared and readded. This appears to be a change specific to the Steam version of the game: the grimoire bar only redraws itself every 5 frames compared to every frame on the web version. To Reproduce The steps needed to reproduce the behavior: Load Cookie Monster into the Steam version. Make sure the grimoire timer is turned on. Open the grimoire and cast a spell if necessary. Screenshots Save file Mi4wNHx8MTYzMDcwOTMwNzQyMzsxNjMwNzA5MzA3NDIzOzE2MzA4NDg5NDQ5ODY7TWlrbyBLaXNhaTtyeGdnZHwwMTExMTEwMTExMDEwMTEwMDEwMTExMTAwfDQzMjM0NDEyMzcwMDg1Ljc2Njs3ODcyNTg4NzAxOTcyNi43NzsyMTYyMDsyNTE7MjEzOTE4MDE3NjIwODAuNzU7MTU7MDswOzA7MDswOzA7MDswOzA7MjUxOzA7MDswOzA7MDswOzswOzA7MDswOzA7MDswOy0xOy0xOy0xOy0xOy0xOzA7MDswOzA7NzU7MDswOzA7MTsxNjMwNzk4NDczOTk3OzA7MDs7MTAwOzA7MDs4MDQxOTY2NDYuNDYzODUxMzsyNTt8MTI2LDEyNiw0NTE4MDI4MjM3MiwwLCwwLDEyNjsxNDYsMTQ2LDg2MTYxMDA3NjYxLDAsLDAsMTQ2Ozg3LDg3LDQwNjgxNDI1MzgsMCwsMCw4Nzs3NSw3NSwxMzgzMDA5MTA4MCwwLCwwLDc1OzY3LDY3LDU0MjE1MjA2MzM3LDAsLDAsNjc7NjEsNjEsMjE3OTgzNjUxOTM2LDAsLDAsNjE7NTQsNTQsNTYzNjQ5NDE4ODMxLDAsLDAsNTQ7NDEsNDEsMjAyNjA2MjY3NjI5NywxLDQuNTUyNjYxODgzMDc5OTUzIDcgNyAxLDAsNDE7MzMsMzMsNTQ3MTU1ODkyMjI1MywwLCwwLDMzOzIyLDIyLDg5OTE5NTg0NDg2MzQsMCwsMCwyMjs1LDUsMjE0OTQxNDY3NzI3NSwwLCwwLDU7MCwwLDAsMCwsMCwwOzAsMCwwLDAsLDAsMDswLDAsMCwwLCwwLDA7MCwwLDAsMCwsMCwwOzAsMCwwLDAsLDAsMDswLDAsMCwwLCwwLDA7MCwwLDAsMCwsMCwwO3wxMTExMTExMTExMTEwMDExMTExMTExMTExMTExMTExMTExMTExMTExMTExMTExMTEwMDEwMTAwMDAwMDAwMDExMTExMTExMTExMTExMTExMTExMTExMTAwMTExMTExMTEwMDAwMDAwMDExMTExMDExMTExMTExMTExMTExMDAwMDEwMDAwMDAwMDAwMDAwMDAwMDAwMDAxMTExMTExMTAwMTExMTAwMDAwMDAwMTAwMDEwMTAxMDAwMTAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMTAxMDEwMTAwMDAwMTEwMDAwMDAwMDAwMDAwMDAwMTAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDEwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDExMTExMTExMDAwMDExMTExMTEwMDAwMDExMTExMTAwMDAwMDExMTExMTAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDExMTExMTAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDExMTF8MTExMTExMTEwMDAwMDAwMDExMTExMTExMDAwMDAwMTExMTExMTEwMDExMTExMDExMDExMDEwMDEwMDEwMDAwMDAwMDExMTAxMTExMTExMDAwMTAwMDAwMDAwMDAwMTEwMDAwMDEwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMTEwMDAwMDAwMDAwMDEwMDAwMDAwMDAwMDAwMTEwMDAwMDAwMDAxMTEwMDAxMTAwMDEwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMTAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMTAwMDAwMDAwMDAwMDAwMDAwMDAxMTAwMDAwMTAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMTAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMHwwLDQ0NjAsMzIyNCw3O3xDb29raWVNb25zdGVyOnsic2V0dGluZ3MiOnsiQ1BTTW9kZSI6MSwiQXZnQ1BTSGlzdCI6MywiQXZnQ2xpY2tzSGlzdCI6MCwiQ2FsY1dyaW5rIjowLCJTY2FsZSI6MiwiU2NhbGVEZWNpbWFscyI6MiwiU2NhbGVTZXBhcmF0b3IiOjAsIlNjYWxlQ3V0b2ZmIjoiOTk5OTk5IiwiVGltZUZvcm1hdCI6MCwiRGV0YWlsZWRUaW1lIjoxLCJQUERpc3BsYXlUaW1lIjowLCJCdWlsZENvbG91ciI6MSwiUFBPbmx5Q29uc2lkZXJCdXlhYmxlIjowLCJQUEV4Y2x1ZGVUb3AiOjAsIlBQUmlnaWRlbE1vZGUiOjAsIlBQU2Vjb25kc0xvd2VyTGltaXQiOiIwIiwiQ29sb3VyQmx1ZSI6IiM0YmI4ZjAiLCJDb2xvdXJHcmVlbiI6IiMwMGZmMDAiLCJDb2xvdXJZZWxsb3ciOiIjZmZmZjAwIiwiQ29sb3VyT3JhbmdlIjoiI2ZmN2YwMCIsIkNvbG91clJlZCI6IiNmZjAwMDAiLCJDb2xvdXJQdXJwbGUiOiIjZmYwMGZmIiwiQ29sb3VyR3JheSI6IiNiM2IzYjMiLCJDb2xvdXJQaW5rIjoiI2ZmMTQ5MyIsIkNvbG91ckJyb3duIjoiIzhiNDUxMyIsIkJvdEJhciI6MSwiVGltZXJCYXIiOjEsIlRpbWVyQmFyUG9zIjowLCJUaW1lckJhck92ZXJsYXkiOjEsIkF1dG9zYXZlVGltZXJCYXIiOjEsIlVwQmFyQ29sb3VyIjoxLCJVcGdyYWRlQmFyRml4ZWRQb3MiOjEsIlNvcnRCdWlsZGluZ3MiOjEsIlNvcnRVcGdyYWRlcyI6MSwiVXBncmFkZXNOZXZlckNvbGxhcHNlIjoxLCJEcmFnb25BdXJhSW5mbyI6MSwiR3JpbW9pcmVCYXIiOjEsIkdDVGltZXIiOjEsIkZhdmljb24iOjEsIldyaW5rbGVyQnV0dG9ucyI6MSwiSGlkZVNlY3Rpb25zQnV0dG9ucyI6MCwiVG9vbHRpcEJ1aWxkVXBncmFkZSI6MSwiVG9vbHRpcEFtb3IiOjEsIlRvb2xXYXJuTHVja3kiOjEsIlRvb2xXYXJuTHVja3lGcmVuenkiOjEsIlRvb2xXYXJuQ29uanVyZSI6MSwiVG9vbFdhcm5Db25qdXJlRnJlbnp5IjoxLCJUb29sV2FybkVkaWZpY2UiOjEsIlRvb2xXYXJuVXNlciI6MCwiVG9vbFdhcm5Cb24iOjEsIlRvb2xXYXJuUG9zIjoxLCJUb29sdGlwR3JpbSI6MSwiVG9vbHRpcFdyaW5rIjoxLCJUb29sdGlwTHVtcCI6MSwiVG9vbHRpcFBsb3RzIjoxLCJUb29sdGlwUGFudGhlb24iOjEsIlRvb2x0aXBBc2NlbmRCdXR0b24iOjEsIlN0YXRzIjoxLCJNaXNzaW5nVXBncmFkZXMiOjEsIk1pc3NpbmdBY2hpZXZlbWVudHMiOjAsIlVwU3RhdHMiOjEsIkhlYXZlbmx5Q2hpcHNUYXJnZXQiOiI0NDAiLCJTaG93TWlzc2VkR0MiOjAsIlRpdGxlIjoxLCJHZW5lcmFsU291bmQiOjEsIkdDTm90aWZpY2F0aW9uIjowLCJHQ0ZsYXNoIjowLCJDb2xvdXJHQ0ZsYXNoIjoiI2ZmZmZmZiIsIkdDU291bmQiOjEsIkdDVm9sdW1lIjoxMDAsIkdDU291bmRVUkwiOiJodHRwczovL2ZyZWVzb3VuZC5vcmcvZGF0YS9wcmV2aWV3cy82Ni82NjcxN185MzE2NTUtbHEubXAzIiwiRm9ydHVuZU5vdGlmaWNhdGlvbiI6MCwiRm9ydHVuZUZsYXNoIjowLCJDb2xvdXJGb3J0dW5lRmxhc2giOiIjZmZmZmZmIiwiRm9ydHVuZVNvdW5kIjoxLCJGb3J0dW5lVm9sdW1lIjoxMDAsIkZvcnR1bmVTb3VuZFVSTCI6Imh0dHBzOi8vZnJlZXNvdW5kLm9yZy9kYXRhL3ByZXZpZXdzLzE3NC8xNzQwMjdfMzI0MjQ5NC1scS5tcDMiLCJTZWFOb3RpZmljYXRpb24iOjAsIlNlYUZsYXNoIjowLCJDb2xvdXJTZWFGbGFzaCI6IiNmZmZmZmYiLCJTZWFTb3VuZCI6MSwiU2VhVm9sdW1lIjoxMDAsIlNlYVNvdW5kVVJMIjoiaHR0cHM6Ly93d3cuZnJlZXNvdW5kLm9yZy9kYXRhL3ByZXZpZXdzLzEyMS8xMjEwOTlfMjE5MzI2Ni1scS5tcDMiLCJHYXJkRmxhc2giOjAsIkNvbG91ckdhcmRGbGFzaCI6IiNmZmZmZmYiLCJHYXJkU291bmQiOjEsIkdhcmRWb2x1bWUiOjEwMCwiR2FyZFNvdW5kVVJMIjoiaHR0cHM6Ly9mcmVlc291bmQub3JnL2RhdGEvcHJldmlld3MvMTAzLzEwMzA0Nl84NjE3MTQtbHEubXAzIiwiTWFnaWNOb3RpZmljYXRpb24iOjAsIk1hZ2ljRmxhc2giOjAsIkNvbG91ck1hZ2ljRmxhc2giOiIjZmZmZmZmIiwiTWFnaWNTb3VuZCI6MSwiTWFnaWNWb2x1bWUiOjEwMCwiTWFnaWNTb3VuZFVSTCI6Imh0dHBzOi8vZnJlZXNvdW5kLm9yZy9kYXRhL3ByZXZpZXdzLzIyMS8yMjE2ODNfMTAxNTI0MC1scS5tcDMiLCJXcmlua2xlck5vdGlmaWNhdGlvbiI6MCwiV3JpbmtsZXJGbGFzaCI6MCwiQ29sb3VyV3JpbmtsZXJGbGFzaCI6IiNmZmZmZmYiLCJXcmlua2xlclNvdW5kIjoxLCJXcmlua2xlclZvbHVtZSI6MTAwLCJXcmlua2xlclNvdW5kVVJMIjoiaHR0cHM6Ly9mcmVlc291bmQub3JnL2RhdGEvcHJldmlld3MvMTI0LzEyNDE4Nl84MDQzLWxxLm1wMyIsIldyaW5rbGVyTWF4Tm90aWZpY2F0aW9uIjowLCJXcmlua2xlck1heEZsYXNoIjowLCJDb2xvdXJXcmlua2xlck1heEZsYXNoIjoiI2ZmZmZmZiIsIldyaW5rbGVyTWF4U291bmQiOjEsIldyaW5rbGVyTWF4Vm9sdW1lIjoxMDAsIldyaW5rbGVyTWF4U291bmRVUkwiOiJodHRwczovL2ZyZWVzb3VuZC5vcmcvZGF0YS9wcmV2aWV3cy8xNTIvMTUyNzQzXzE1NjYzLWxxLm1wMyIsIkJ1bGtCdXlCbG9jayI6MSwiRmF2b3VyaXRlU2V0dGluZ3MiOjJ9LCJoZWFkZXJzIjp7IkZhdm91cml0ZSI6MSwiQ2FsY3VsYXRpb24iOjEsIk5vdGF0aW9uIjoxLCJDb2xvdXJzIjoxLCJCYXJzRGlzcGxheSI6MSwiVG9vbHRpcCI6MSwiU3RhdGlzdGljcyI6MSwiTm90aWZpY2F0aW9uIjoxLCJOb3RpZmljYXRpb25HZW5lcmFsIjoxLCJOb3RpZmljYXRpb25HQyI6MSwiTm90aWZpY2F0aW9uRkMiOjEsIk5vdGlmaWNhdGlvblNlYSI6MSwiTm90aWZpY2F0aW9uR2FyZCI6MSwiTm90aWZpY2F0aW9uTWFnaSI6MSwiTm90aWZpY2F0aW9uV3JpbmsiOjEsIk5vdGlmaWNhdGlvbldyaW5rTWF4IjoxLCJNaXNjZWxsYW5lb3VzIjoxLCJMdWNreSI6MSwiQ2hhaW4iOjEsIlNwZWxscyI6MSwiR2FyZGVuIjoxLCJQcmVzdGlnZSI6MSwiV3JpbmsiOjEsIlNlYSI6MSwiQWNoaWV2cyI6MSwiTWlzYyI6MSwiaW5mb01lbnUiOjEsIm9wdGlvbnNNZW51IjoxfSwiZmF2b3VyaXRlU2V0dGluZ3MiOltdLCJsb2NrZWRNaW5pZ2FtZXMiOltdLCJ2ZXJzaW9uIjoiMi4wMzEuMTAifTtjb29raWVNb25zdGVyRnJhbWV3b3JrOnsic2V0dGluZ3MiOnt9LCJoZWFkZXJzIjp7ImluZm9NZW51IjoxLCJvcHRpb25zTWVudSI6MX0sImZhdm91cml0ZVNldHRpbmdzIjpbXX07TUVUQTpjb29raWUgbW9uc3RlciwqY29vbGVyIHNhbXBsZSBtb2QsKmxhbmcgc2FtcGxlIG1vZCwqc2FtcGxlIG1vZDs%3D%21END%21 Browser Steam release, game version 2.04 Suggested fix Either store the old magic bar text before running the normal draw funciton, and only add the timer if it changed (more generic, but probably slower?): Game.Objects['Wizard tower'].minigame.draw = function () { oldText = minigame.magicBarTextL.innerHTML; BackupGrimoireDraw(); if ( Game.mods.cookieMonsterFramework.saveData.cookieMonsterMod.settings.GrimoireBar === 1 && minigame.magic < minigame.magicM && oldText != minigame.magicBarTextL.innerHTML ) { minigame.magicBarTextL.innerHTML += ` (${FormatTime( CalculateGrimoireRefillTime(minigame.magic, minigame.magicM, minigame.magicM), )})`; } }; or add a different check for the Steam version to match the update frequency there: Game.Objects['Wizard tower'].minigame.draw = function () { BackupGrimoireDraw(); if ( Game.mods.cookieMonsterFramework.saveData.cookieMonsterMod.settings.GrimoireBar === 1 && minigame.magic < minigame.magicM && (typeof Steam === 'undefined' || Game.drawT % 5 == 0) ) { minigame.magicBarTextL.innerHTML += ` (${FormatTime( CalculateGrimoireRefillTime(minigame.magic, minigame.magicM, minigame.magicM), )})`; } }; The Steam part of the above could be excluded if you assume the web version will add this change as well (which it probably will eventually, but I assume you wouldn't want to break this on the web version in the meantime). Since I can't run the Steam version on my Mac pc, can you test and confirm that your second suggestion works? I don't have the tools available to rebuild the actual JS file (or I would have just made this a pull request), but I was able to hack it in to the built file and this is the result (i.e. it does work): I guess there's still some changes needed to make the bar stretch properly, but I'm not sure I could tell you what exactly and that should probably a separate change anyway - this is at least more usable than without it. Thanks! Going to add this right away! Added!
2025-04-01T04:10:18.207270
2024-11-23T07:17:55
2685553991
{ "authors": [ "CookiedOutMonster", "ariceeee" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13857", "repo": "CookiedOutMonster/COSC441", "url": "https://github.com/CookiedOutMonster/COSC441/pull/30" }
gharchive/pull-request
Separate variable types from values by making new blocks to represent specific values Relevant issue This closes #28. We're changing how the blocks work/are structured so that types and values are now separate blocks. An example will help explain: say you want to make an int variable and assign it a value of 5. Currently it's like IntegerBlock.SetValue("5") but w this change it becomes IntegerBlock, EqualsBlock, FiveBlock. Much more straightforward/ intuitive. Summary of changes VariableBlock is now VarTypeBlock, to show that it represents a variable type, not a whole variable (type + value) Hence it no longer has a Value field New block types! These represent actual values/literals, for example, the number 1. There are 3 new parent types (abstract classes), each w a bunch of children classes for spawnable blocks: TextBlock: ABlock, BBlock, CBlock, XBlock, YBlock, ZBlock NumberBlock: OneBlock, TwoBlock, ThreeBlock, FourBlock, FiveBlock, SixBlock, SevenBlock, EightBlock, NineBlock, DecimalBlock TrueFalseBlock: TrueBlock, FalseBlock BlockSpawner has been updated to reflect these new block types, so that when it spawns a block representing a value (ex: letter X), its type is XBlock, as opposed to StringBlock with .Value = "X" LGTM, though, your characters have numbers as the button text lol but ill fix it
2025-04-01T04:10:18.234752
2022-01-18T16:19:31
1107103904
{ "authors": [ "antoinejeannot" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13858", "repo": "Cornerstone-OnDemand/modelkit", "url": "https://github.com/Cornerstone-OnDemand/modelkit/pull/132" }
gharchive/pull-request
Fix callbacks While playing with callbacks to monitor prediction timers, I encountered some issues whose fixes are adressed in this PR: the callback batch_predictions are now a list, instead of an empty iterator callbacks are now GeneratorExit-proof What's your opinion ? Thanks in advance for reviewing. I might use some help to deal with the missing async test 🤔
2025-04-01T04:10:18.237858
2023-07-11T14:31:12
1799075365
{ "authors": [ "KonradUdoHannes" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13859", "repo": "CorrelAid/engage-2", "url": "https://github.com/CorrelAid/engage-2/issues/22" }
gharchive/issue
Implement minimal project creation and retrieval endpoints The main focus here should be to implement a minimal project model together with the following endpoints POST /projects GET /projects GET /projects/{id} The focus should not be on completeness of the project data model, but on the endpoints. There are no constraints on the path parameters for the endpoint as long as uniqueness is easy to achieve. done in #36
2025-04-01T04:10:18.243443
2021-06-28T20:56:08
931985094
{ "authors": [ "CLAassistant", "orkunkl" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13860", "repo": "CosmWasm/cosmwasm-plus", "url": "https://github.com/CosmWasm/cosmwasm-plus/pull/310" }
gharchive/pull-request
cw20-escrow: Upgrade to 0.15 Depends on https://github.com/CosmWasm/cosmwasm-plus/pull/304 Resolves https://github.com/CosmWasm/cosmwasm-plus/issues/309 Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you all sign our Contributor License Agreement before we can accept your contribution.1 out of 2 committers have signed the CLA.:white_check_mark: orkunkl:x: uintYou have signed the CLA already but the status is still pending? Let us recheck it.
2025-04-01T04:10:18.248391
2024-01-07T20:23:24
2069274676
{ "authors": [ "maurolacy", "webmaster128" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13861", "repo": "CosmWasm/cosmwasm", "url": "https://github.com/CosmWasm/cosmwasm/pull/1983" }
gharchive/pull-request
Secp256r1 support Completes the work done in #1877, which is in turn a follow-up of #1083. Basically, rebased from main, added some test vectors, and reorganised the code a bit for generality. And made CI pass. Who needs this implementation for what? I was planning to add it to 2.1 to be prepared for passkey apps. I just saw the pending PR and decided to jump in. Don't we need pubkey recovery here? Will take a look. Seen some comments about lack of recovery for this, but that may be wrong. I was thinking not requiring low-S signatures in secp256k1 by default was a mistake as all modern blockchain applications require low-S. Right now in CosmWasm you have to check low-S yourself in the contract if your protocol needs it. Maybe it's a good opportunity to make a better decision here. Any thoughts? I think there's a normalisation step related to that. Do you want to fail in case of high S? I think there's a normalisation step related to that. Do you want to fail in case of high S? For non-Ethereum and non-Bitcoin ECDSA, high S is perfectly valid. It's really up to the application to limit it. This is why we accept both in secp256k1_verify. The comments in https://twitter.com/simon_warta/status/1744814900164059555 are also not really convincing me to change it. Maybe it is better to just create a secp256k1_verify_strict and secp256r1_verify_strict later on. Don't we need pubkey recovery here? Will take a look. Seen some comments about lack of recovery for this, but that may be wrong. Added to the crypto package. Will work in the rest later.
2025-04-01T04:10:18.260152
2023-07-04T18:24:35
1788322560
{ "authors": [ "AshitaBidkar", "harshita-2003" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13862", "repo": "Counselllor/Counsellor-Web", "url": "https://github.com/Counselllor/Counsellor-Web/issues/230" }
gharchive/issue
added hovering to sign up & login Is your feature request related to a problem? Please describe. A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] Describe the solution you'd like A clear and concise description of what you want to happen. Describe alternatives you've considered A clear and concise description of any alternative solutions or features you've considered. Additional context Add any other context or screenshots about the feature request here. I want to work on this issue, kindly assign it to me.
2025-04-01T04:10:18.263564
2014-12-16T22:43:22
52177692
{ "authors": [ "adamkrellenstein", "ouziel-slama" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13863", "repo": "CounterpartyXCP/counterparty-core", "url": "https://github.com/CounterpartyXCP/counterparty-core/issues/537" }
gharchive/issue
Remove Rock-Paper-Scissors [ ] Disable RPS [ ] Remove most functionality [ ] Hard-code existing games [ ] Update API [ ] Update docs See #487 for inspiration. Want to back this issue? Post a bounty on it! We accept bounties via Bountysource. done here https://github.com/CounterpartyXCP/counterparty-core/pull/1821
2025-04-01T04:10:18.315499
2022-10-14T15:12:33
1409500637
{ "authors": [ "codecov-commenter", "insi-eb" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13864", "repo": "CppMicroServices/CppMicroServices", "url": "https://github.com/CppMicroServices/CppMicroServices/pull/750" }
gharchive/pull-request
701 Update DS and CA versions this is basically catching up with #668 I thought a bit about whether the component versions should be the same as for 3.7.2 or different, and in the end decided to use the same component versions as the API is the same (ABI is a different topic, but I'm not sure whether that should play into the VERSION here). If you disagree, please let me know what you would prefer. The example change and the slightly different overall VERSION update from #668 I already merged when setting the overall VERSION to 3.6.1. Codecov Report Merging #750 (a17d4cc) into c++14-compliant (8476383) will decrease coverage by 0.64%. The diff coverage is n/a. Additional details and impacted files @@ Coverage Diff @@ ## c++14-compliant #750 +/- ## =================================================== - Coverage 90.56% 89.91% -0.65% =================================================== Files 250 90 -160 Lines 13554 6286 -7268 =================================================== - Hits 12275 5652 -6623 + Misses 1279 634 -645 Impacted Files Coverage Δ framework/src/bundle/BundleRegistry.cpp 79.60% <0.00%> (-11.97%) :arrow_down: framework/src/service/ServiceEvent.cpp 71.79% <0.00%> (-5.76%) :arrow_down: ...mework/src/service/ServiceReferenceBasePrivate.cpp 86.45% <0.00%> (-5.64%) :arrow_down: framework/src/service/ServiceRegistrationBase.cpp 77.66% <0.00%> (-4.24%) :arrow_down: framework/src/util/FrameworkEvent.cpp 61.81% <0.00%> (-3.94%) :arrow_down: framework/src/bundle/BundleResourceBuffer.cpp 67.34% <0.00%> (-3.86%) :arrow_down: framework/src/service/ServiceListeners.cpp 96.26% <0.00%> (-3.74%) :arrow_down: framework/src/util/Utils.cpp 75.58% <0.00%> (-3.32%) :arrow_down: framework/src/bundle/BundleResourceContainer.cpp 92.08% <0.00%> (-3.29%) :arrow_down: framework/src/bundle/BundleEvent.cpp 85.00% <0.00%> (-3.00%) :arrow_down: ... and 230 more
2025-04-01T04:10:18.317054
2016-05-23T13:10:45
156276939
{ "authors": [ "CedricEugeni", "CrackerakiUA" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13865", "repo": "CrackerakiUA/ngImgCropFullExtended", "url": "https://github.com/CrackerakiUA/ngImgCropFullExtended/pull/113" }
gharchive/pull-request
Fixing issue #110 @CrackerakiUA : This test if resImageQuality is set by the user and then uses this parameter to set output quality. @CedricEugeni thank you for your PR, can you recreate it with new pull from our library. Someone else have fixed those lines before you. I have merge. @CrackerakiUA : Thanks for merging this PR so quickly ! :)
2025-04-01T04:10:18.366110
2019-08-17T16:26:58
481902816
{ "authors": [ "RyderBelserion", "haeiven" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13866", "repo": "Crazy-Crew/Crazy-Enchantments", "url": "https://github.com/Crazy-Crew/Crazy-Enchantments/issues/424" }
gharchive/issue
Add a Stat trak enchant Hello, It would be an enchant who is couting each kills made a sword. The suggestion is not declined, Moved to a tracking board which will be added to the README.
2025-04-01T04:10:18.384349
2015-01-19T22:12:58
54818895
{ "authors": [ "asdaa0", "tterrag1098" ], "license": "Unlicense", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13874", "repo": "CrazyPants/EnderIO", "url": "https://github.com/CrazyPants/EnderIO/issues/1720" }
gharchive/issue
Placing ME conduit next to AE2 P2P tunnel corrupts chunk I was playing on a server and my friend decided to to use ME conduits for P2P tunnels to hide them with covers. While he was working on that he placed an ME conduit next to the P2P tunnel then the server crashed and corrupted the chunk that we were in. Reproduced in SP: http://pastebin.com/bvuFmDvW Modpack: DNS Techpack <IP_ADDRESS> EnderIO version: 1.7.10 <IP_ADDRESS>2 rv1 is no longer supported.
2025-04-01T04:10:18.386463
2021-02-19T08:46:02
811826290
{ "authors": [ "bun919tw" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13875", "repo": "CreamFi/compound-protocol", "url": "https://github.com/CreamFi/compound-protocol/pull/49" }
gharchive/pull-request
contracts/: cySUSD DO NOT MERGE. Distribute sUSD to 2 victims. No one could redeem or transfer cySUSD. Only cream multisig could liquidate cySUSD. Test: https://github.com/CreamFi/comptroller-upgrade/blob/cysusd/test/upgrade.js The code has been patched: https://etherscan.io/address/0xe9e0ad069325234035ac21bbff76b557b4d52689#code Close this.
2025-04-01T04:10:18.546142
2016-09-20T15:24:19
178092535
{ "authors": [ "Crinsane", "twilightalkalin" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13876", "repo": "Crinsane/LaravelShoppingcart", "url": "https://github.com/Crinsane/LaravelShoppingcart/issues/218" }
gharchive/issue
Quantity to 0 Hello is it possible to make the quantity set to 0 first at the first time? Have you tried? I believe it is.
2025-04-01T04:10:18.547462
2020-03-16T18:15:02
582507925
{ "authors": [ "Faks", "robbfountain" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13877", "repo": "Crinsane/LaravelShoppingcart", "url": "https://github.com/Crinsane/LaravelShoppingcart/pull/612" }
gharchive/pull-request
Wider Version Support Wider Version Support This will fail with Laravel v 6.X & 7.X as the CartItem class contains references to the array helpers
2025-04-01T04:10:18.553172
2020-03-15T15:31:36
581718539
{ "authors": [ "AnastaZIuk", "devshgraphicsprogramming" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13878", "repo": "Crisspl/IrrlichtBAW", "url": "https://github.com/Crisspl/IrrlichtBAW/pull/30" }
gharchive/pull-request
OpenEXR split & Denoiser Tonemapper status: OpenEXR split is done and tested Denoiser Tonemapper CommandLineParser is done and tested Denoiser Tonemapper example needs to be completed - @devshgraphicsprogramming it's your turn address the comment by @Crisspl and add the camera transform parameter okay, @devshgraphicsprogramming you can take control See latest review and I will close this PR closing and moving to https://github.com/buildaworldnet/IrrlichtBAW/pull/439
2025-04-01T04:10:18.573590
2021-12-13T18:27:57
1078848242
{ "authors": [ "bjgill", "tberey" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13879", "repo": "Crown-Commercial-Service/digitalmarketplace-scripts", "url": "https://github.com/Crown-Commercial-Service/digitalmarketplace-scripts/pull/747" }
gharchive/pull-request
Find active suppliers since a specified date This script was needed to find all active registered suppliers, since a given date. The script parses a date, provided by the user (if not default date of today is provided). This date is then compared against the Suppliers list, where active is true. If the date is more recent than the date the user provided, the record is captured and outputted. A count is taken and also provided. This branch can't yet be merged in, because of an outstanding question: This script outputs a file, with the results in... Perhaps I should just print it all directly to the console/terminal log, rather than out to a txt file? @tim-s-ccs the script itself works, it just seems that either: arguments = docopt(__doc__), which I am importing and initialising to grab arguments in this way arguments['--date'] it cold also be the initialisation of the data object data_api_client = DataAPIClient(get_api_endpoint_from_stage(environ["STAGE"].lower()), get_auth_token("api", environ["STAGE"].lower())) Both of these I have copied and pasted from another script, but errors are gone once these are removed (but they are needed, for the script to work) I can take a look, but a quick question on the goal before I start: are we trying to get active suppliers created since the date, or all suppliers who have been active since the date? @bjgill we are trying to get active suppliers created since the date @bjgill last push should have addressed/fixed most, if not all, of the issues you highlighted
2025-04-01T04:10:18.615238
2017-01-31T10:39:30
204258282
{ "authors": [ "Cryrivers", "tehmaestro" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13880", "repo": "Cryrivers/ember-feature-flag-solution", "url": "https://github.com/Cryrivers/ember-feature-flag-solution/issues/4" }
gharchive/issue
How can I use it with ember-cli-sass Hi, I'm using ember-cli-sass, but when I write the following code: @if $myFeature display: inline-block; It sais that $myFeature is an undefined variable. Does it have something to do with the order in which the preprocessing addon hook is called, for these two addons? If your feature flag settings are working with JavaScript and Handlebars files. You can just @import 'feature-flags'; to import all feature flag variables. Hi, thanks, I actually managed to get it working :). I couldn't get it to work in javascript however, not sure where how the code should look like. If you could post a quick example, that would be great. Also, should the configuration go to environment.js OR ember-cli-build.js ? Thanks.
2025-04-01T04:10:18.664072
2022-03-08T07:42:59
1162325686
{ "authors": [ "00Masato", "Cutwell" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13884", "repo": "Cutwell/github1s-chrome-extension", "url": "https://github.com/Cutwell/github1s-chrome-extension/issues/1" }
gharchive/issue
Options page should have fixed size. Steps to reproduce Install from https://chrome.google.com/webstore/detail/open-in-github1s/iijaagbkdohcopmmohlgfkcloefoeaoj?hl=ja open chrome://extensions/ open open-in-github1s setting page open options page change "New Tab" to "Current Tab" , and push "Save". Expected When I push "Save" , options page has fixed size. Actual When I push "Save" , options page has changed size. https://user-images.githubusercontent.com/40224377/157184454-0aa08533-9fca-4637-a33e-bc5d263cc03b.mov Environment Open in GitHub1s 1.3 Google Chrome 99.0.4844.51 Mac OS Monterey 12.2.1 Well spotted, closed by #2
2025-04-01T04:10:18.673456
2022-11-28T15:47:32
1466654148
{ "authors": [ "arnaud-lb", "romm" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13885", "repo": "CuyZ/Valinor", "url": "https://github.com/CuyZ/Valinor/issues/270" }
gharchive/issue
Support for generics in interface inference Thank you for the awesome work on this library. Valinor is super useful and makes it much easier to deal with external data in a strongly typed project. I am facing an issue since 0.11 added the following restriction: It is now mandatory to list all possible class-types that can be inferred by the mapper. This change is a step towards the library being able to deliver powerful new features such as compiling a mapper for better performance. Currently, this change makes interface inference incompatible with generic classes, as Valinor will not accept a return type like class-string<Foo<A>|Foo<B>> (rightfully, as this is not a valid type) or 'Foo<A>'|'Foo<B>'. To give some context, I'm trying to map a discriminated union, but the discriminator is on the parent node. The source looks like this: { "type": "string", "node": "..." } And the model looks like this: /** @template T */ interface RootInterface {} class Root implements RootInterface { public function __construct( /** @var T */ NodeInterface $node, ){} } interface NodeInterface {} class NodeA implements NodeInterface {} class NodeB implements NodeInterface {} Due to the position of the discriminator, I can not directly infer the type of the nested node: I must infer the type of the root node. It seems that supporting a return type that consists of a union of strings in this case would fix the issue, as I could type the mapper like this: /** @return 'Root<NodeA>'|'Root<NodeB>' */ fn (string $type) => ... Here is a full example of my use-case: https://gist.github.com/arnaud-lb/535d89874ed719265f66bc19a88b5fb0 Hi @arnaud-lb, thank you for reporting this issue; I did not have this usecase in mind when pushing 1b0ff39af650f1c5902ee930f49049042842ec08 indeed. From my POV, the following should make sense, although it is not supported by the library I certainly could make this work: /** @return class-string<Root<NodeA>|Root<NodeB>> */ function inferRootType(string $type): string { return match ($type) { 'NodeA' => 'Root<NodeA>', 'NodeB' => 'Root<NodeB>', default => throw new \Exception(), }; } But static analysis tools wont be happy about that, see PHPStan example and Psalm example; this probably makes sense as Root<NodeA> is not considered a class-string, but Root is. Do you see any other way than using string values like @return 'Root<NodeA>'|'Root<NodeB>' — even if that would mean contributing to PHPStan/Psalm? BTW big fan of your work on PHPStan generics. 🙂 Thank you :D I wish you and this project a happy new year :) But static analysis tools wont be happy about that, see PHPStan example and Psalm example; this probably makes sense as Root<NodeA> is not considered a class-string, but Root is. Agreed. I think that the class-string type would lose usefulness if class-string<Root<NodeA>> was valid, as this does not represent a class name string that PHP constructs can understand. Do you see any other way than using string values like @return 'Root<NodeA>'|'Root<NodeB>' — even if that would mean contributing to PHPStan/Psalm? Introducing a new generic-class-string<T> type would solve this issue, but it would have limited usefulness. Adding support for class-string<Root<NodeA>> would also work, if PHPStan/Psalm treated generic class-strings and non-generic ones differently (e.g. generic class-string is not valid in new), and if class-string<Root<NodeA>> was not considered a sub type of class-string. Using string values like @return 'Root<NodeA>'|'Root<NodeB>' works for me, though.
2025-04-01T04:10:18.701496
2019-08-15T03:58:08
480983581
{ "authors": [ "JayFoxRox", "PatrickvL", "RadWolfie" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13886", "repo": "Cxbx-Reloaded/XbSymbolDatabase", "url": "https://github.com/Cxbx-Reloaded/XbSymbolDatabase/issues/87" }
gharchive/issue
Search Process is Unoptimized Base on @JayFoxRox's feedback about XbSymbolDatabase. The search process is unoptimized because of not using better algorithms to gain faster scan process. Searching for XInput for example, shouldn't take longer than like.. 100ms [but it took like 10 seconds on original Xbox, when I last tried afair - which is too long] Currently, the scan process check for OV then move on to next offset, +1. If I recall correctly, it also check the reference address first. https://en.wikipedia.org/wiki/Knuth–Morris–Pratt_algorithm In my case, it is D3D8 library taking way much longer time in .text section than others. The way Dxbx scan is much faster; By using a trie data structure it doesn't need to check all symbols in the database per memory address (which makes it O(nm) cost), but instead, starting from each memory address, it walks the trie until it hits a leaf or a dead end (making this cost O(nlog(m)). afaik, trie data structure will cause the existing groups to dismantle into their own individual. This will cause harder to track what's not being used and what's currently in our database. Keep in mind, we're dealing with over 400 - 600 symbols per library by the time we're done with the database. If there's a way to keep the existing groups intact, feel free to optimize the existing macro. By using a trie data structure Dxbx.. In the respective discussion on Discord, I also recommended using an accelerator structure (although not necessarily multi-level like a tree) to search for all functions at the same time (so the XBE is only scanned as few times as necessary, and it remains in cache). The way Dxbx scans is much faster than the algorithm linked above [...] starting from each memory address, it walks the trie until it hits a leaf or a dead end. This is just a description of a trie. Please provide a link to relevant sections when talking about existing code. What are the leafs: Funtions? What are the links: XBE bytes? Which byte-index is looked at to find the next link? What happens on failure: will it go to the next byte, or can it skip bytes like Boyer-Moore search? Also, when / how does it deal with X-Refs? Multiple scans? Different tries to ensure correct prioritization? To prevent repeating myself, here a link that describes the Dxbx detection code in a bit more detail : http://dxbx-emu.com/2010/11/04/dxbx-symbol-detection/ Somehow keyword automatically marked this as closed. Reopened since it is not fully resolve.
2025-04-01T04:10:18.784739
2015-10-10T10:15:36
110785608
{ "authors": [ "Cyan4973", "xaphier" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13887", "repo": "Cyan4973/zstd", "url": "https://github.com/Cyan4973/zstd/pull/46" }
gharchive/pull-request
Remove one malloc/free from compression By making the buffer part of the zstd context structure, one malloc/free can be removed from the compression. Good catch Xavier. This was supposed to be introduced into next version of zstd, but since your patch is ready, let's use it now ;)
2025-04-01T04:10:18.810211
2024-10-10T13:18:05
2578832121
{ "authors": [ "ThomasVitale", "andreas-hilti", "jkowalleck", "nscuro" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13888", "repo": "CycloneDX/cyclonedx-cli", "url": "https://github.com/CycloneDX/cyclonedx-cli/issues/406" }
gharchive/issue
Project not distributed via Homebrew since version 0.25.1 I followed the installation documentation and got cyclonedx-cli installed on my Linux machine from Homebrew. brew install cyclonedx/cyclonedx/cyclonedx-cli The version installed is 0.25.1 even though the latest available is 0.27.1. I can see that no newer versions have been published to the Homebrew Tap since version 0.25.1: https://github.com/CycloneDX/homebrew-cyclonedx @ThomasVitale I don't know anything about Homebrew; however, I don't see anything in the release pipeline related to it. If I'm not mistaken, it requires (currently) a manual update of https://github.com/CycloneDX/homebrew-cyclonedx/blob/main/Formula/cyclonedx-cli.rb. oh, the readme actually tells about this outdated/unmaintained install path here: https://github.com/CycloneDX/cyclonedx-cli?tab=readme-ov-file#homebrew I originally set up the tap in hopes to eventually get updates to it automated, but then never came around to do it. If someone wants to have a go at it, that would be much appreciated. Otherwise, folks can simply raise a PR to get the formula updated manually, like so. Also, a few of our tools have been added to Homebrew core (AFAIK not by any of us), for example cyclonedx-gomod. i'll assign this task to me and see how we can automate all these things.
2025-04-01T04:10:18.822838
2015-04-12T18:35:56
67942311
{ "authors": [ "MartinNowak", "rainers" ], "license": "BSL-1.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13889", "repo": "D-Programming-Language/dmd", "url": "https://github.com/D-Programming-Language/dmd/pull/4585" }
gharchive/pull-request
fix testaa: destructor should not allocate https://github.com/D-Programming-Language/druntime/pull/1212 runs destructors of collected AA keys and values, but these destructors must not call back into the GC. Auto-merge toggled on
2025-04-01T04:10:18.824894
2015-09-12T17:45:00
106174104
{ "authors": [ "DmitryOlshansky", "chalucha" ], "license": "BSL-1.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13890", "repo": "D-Programming-Language/phobos", "url": "https://github.com/D-Programming-Language/phobos/pull/3650" }
gharchive/pull-request
std.encoding - optimization of windows-1252 Modified the same way as in #3626 I'm planning to add more encodings on top of this - let me know if you prefer to add it to this PR or make it separate (for example by encoding usage region). For now I kept it simple for review. GenericEncoder is modified to handle narrowed character map as with cp1252 only a small subrange is not mapped directly. Also modified the gen tool [0] to generate EncoderInstance from encoding mapping files. [0] https://github.com/chalucha/encgen LGTM Auto-merge toggled on
2025-04-01T04:10:18.835577
2017-09-27T12:51:15
260961795
{ "authors": [ "severinsimmler", "thvitt" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13893", "repo": "DARIAH-DE/Topics", "url": "https://github.com/DARIAH-DE/Topics/issues/4" }
gharchive/issue
Integrate doclist.py and meta.py Both maintain a list of documents — ideally the metadata doclist would inherit from BaseDoclist (?) and common behaviour should be merged into the base class. Using https://github.com/cophi-wue/metadata-toolbox now
2025-04-01T04:10:18.841420
2023-12-04T14:52:36
2024049691
{ "authors": [ "codecov-commenter", "mwdchang" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13894", "repo": "DARPA-ASKEM/sciml-service", "url": "https://github.com/DARPA-ASKEM/sciml-service/pull/145" }
gharchive/pull-request
ode.observables is not a required field Add guards around parsing out ode.observables field, it is not a strictly required field for petrinet-AMRs. Codecov Report All modified and coverable lines are covered by tests :white_check_mark: Comparison is base (a55c374) 77.96% compared to head (09bb5d8) 77.89%. Additional details and impacted files @@ Coverage Diff @@ ## main #145 +/- ## ========================================== - Coverage 77.96% 77.89% -0.08% ========================================== Files 3 3 Lines 463 466 +3 ========================================== + Hits 361 363 +2 - Misses 102 103 +1 :umbrella: View full report in Codecov by Sentry. :loudspeaker: Have feedback on the report? Share it here.
2025-04-01T04:10:18.898219
2024-05-19T13:42:30
2304619172
{ "authors": [ "DEstebanJV", "Lujan2408" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13896", "repo": "DEstebanJV/Backend-con-php", "url": "https://github.com/DEstebanJV/Backend-con-php/pull/1" }
gharchive/pull-request
Create file delete user New file "updatedUser.php" created for deleting an user from data base Hi, What is the change? The change is that I created a new file for deleting an user from the data base
2025-04-01T04:10:18.901157
2023-02-08T15:42:30
1576357619
{ "authors": [ "steventux", "thomasleese" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13897", "repo": "DFE-Digital/apply-for-qualified-teacher-status", "url": "https://github.com/DFE-Digital/apply-for-qualified-teacher-status/pull/1106" }
gharchive/pull-request
Assess references that have been received Trello Context and changes As an assessor that is looking at the reference that have been received about an applicants work history I need to be able to view the information that a referee has submitted so that I can decide whether to award or decline an application Adds a verify reference requests and review reference request pages. I'm not 100% sure what the happy path from verifying all references should be so for the moment this PR takes the user to the application overview. This probably isn't right. Verifying reference requests Reviewing a reference request This has been product reviewed so I'll merge this in.
2025-04-01T04:10:18.907861
2020-07-22T15:54:04
663869551
{ "authors": [ "EmmaFrith", "george-bit" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13898", "repo": "DFE-Digital/apply-for-teacher-training", "url": "https://github.com/DFE-Digital/apply-for-teacher-training/pull/2562" }
gharchive/pull-request
Rearranged and edited content on content providers view Context Editing down text on training providers available through apply for teacher training page. Changes proposed in this pull request Before: After: Guidance to review review against the design spec in the trello card https://docs.google.com/document/d/1nCmK-sAfMNPDzjdDNA0119qRtL3DysEBOFL9e5bmjvk/edit# Link to Trello card https://trello.com/c/hB66Shmz/1839-dev-content-changes-on-training-providers-available-through-apply-for-teacher-training-page Things to check [x] This code doesn't rely on migrations in the same Pull Request [x] If this code includes a migration adding or changing columns, it also backfills existing records for consistency [x] API release notes have been updated if necessary [x] New environment variables have been added to the Azure config The title has now changed on this page as a result. I wonder whether the nomenclature in en.yml and for the view itself should change to reflect this? Thanks @george-bit Am I alright to change the content slightly? (My bad) Have updated the content on the google doc here: https://docs.google.com/document/d/1nCmK-sAfMNPDzjdDNA0119qRtL3DysEBOFL9e5bmjvk/edit?usp=sharing I've changed: the text included in the Apply for teacher training hyperlink (so that it's consistent with the UCAS hyperlink) added a sentence re-iterating that candidates can only apply for 3 courses in total, as a provider complained that this message wasn't getting across Thanks @EmmaFrith I have made the suggested changes, please review. awesome looks brill! thanks @george-bit :)
2025-04-01T04:10:18.909623
2021-09-02T10:59:00
986549743
{ "authors": [ "ltello", "nigel-lowry" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13899", "repo": "DFE-Digital/get-help-with-tech", "url": "https://github.com/DFE-Digital/get-help-with-tech/pull/1989" }
gharchive/pull-request
2398 support assets link Context Changes proposed in this pull request Guidance to review Admin-BIOS Link Location (Support) We don´t have tests for the support home page to add this link to it?
2025-04-01T04:10:18.911120
2021-02-12T10:05:56
807104858
{ "authors": [ "alexbowen", "martin-bangoura" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13900", "repo": "DFE-Digital/teaching-vacancies", "url": "https://github.com/DFE-Digital/teaching-vacancies/pull/2859" }
gharchive/pull-request
[TEVA-2014] job creation required/optional revisit https://dfedigital.atlassian.net/browse/TEVA-2014 apply new required/optional appraoch to job creation and misc forms Review app deployed to https://teaching-vacancies-review-pr-2859.london.cloudapps.digital
2025-04-01T04:10:18.923323
2016-06-23T19:51:06
162008289
{ "authors": [ "DH3ALEJANDRO", "sanalpanicker" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13901", "repo": "DH3ALEJANDRO/custom-work-for-brackets", "url": "https://github.com/DH3ALEJANDRO/custom-work-for-brackets/issues/63" }
gharchive/issue
Messed up I loved your plugin, it made my work so easy. I was excited seeing your new update, but so disappointed after I installed it.Please revert back your updates I can no longer see the file icons, so file listing on the left side,my whole font and size of workspace got affected :worried: please screenshot i removed the plugin, I cant go back to see the issues. :( i think, i got it.During the installation it asked me if I want to enable the file panel and other options, i just skipped it.Where do i go back to enable it? open Console of Brackets and enter: brackets.test.CommandManager.execute("PreferencesforCustomWorkforBrackets")
2025-04-01T04:10:18.967581
2021-01-26T22:02:36
794593127
{ "authors": [ "kb1hyr", "sjamesparsonsjr" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13902", "repo": "DIYbiosphere/sphere", "url": "https://github.com/DIYbiosphere/sphere/issues/287" }
gharchive/issue
Membership request Hi! I hope this finds you well. I'm an author in the applied sciences and I have a home lab under the auspices of a biotech LLC. I'd like to be able to edit stuff here when appropriate. Always been a big fan of citizen science. p.s. yes that's my ham radio call sign. (grin) Hello kb1hyr, thank you for your interest. I'm currently working on a tool to streamline the entry process. I'll send you a link once it's complete. Hello kb1hyr, thank you for your interest. I'm currently working on a tool to streamline the entry process. I'll send you a link once it's complete. Hi! How's that tool going? I'm not a programming expert by any means, but I have some background. Maybe I can help? Maybe not too, lol, depends on what it is. But maybe! Hi Brett, the site currently has two buttons for adding content without using GitHub, it's a Google Forum. I'm currently developing a new site, if it gets approval from the committee it will be pushed live.
2025-04-01T04:10:19.011652
2020-02-10T19:41:44
562783343
{ "authors": [ "briri", "sjDCC" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13903", "repo": "DMPRoadmap/roadmap", "url": "https://github.com/DMPRoadmap/roadmap/issues/2390" }
gharchive/issue
Build out API v1 - Phase one Phase one of the API v1 changes involves setup of the basic infrastructure and a few initial endpoints. Infrastructure: Introduce JSON Web Token (JWT) security for system to system communications. Which will use a client_id and client_secret for initial auth and then a JWT in further requests. Create a new api_client table to handle JWT machine to machine security Allow access to the API from the JWT method or the v0 User api_token method Create new views that follow the RDA common metadata standard for DMPs Create views that output the RDA common metadata standard for DMPs Endpoints Build out POST /api/v1/authenticate - to authenticate the user and return a JWT Build out GET /api/v1/plans - to retrieve a list of DMPs Build out GET /api/v1/plans/[:id] - to retrieve a DMP Build out POST /api/v1/plans - to create a DMP This ticket will also provide a GET /api/v1/templates endpoint that returns the titles, descriptions, versions and owning organization names for all public templates. Will create further tickets for carrying the existing API v0 functionality into API v1 (e.g. full text API, etc.) The v1 API uses JWT authorization and allows both the existing User email + api_token or the new ApiClient (for systems that need to communicate with DMPRoadmap) client_id + client_secret. Deployed out to DMPtool stage for testing with RAMS Is this one the devs can approve please @xsrust Need a tutorial on APIs to do this myself! Closing this based on dev check
2025-04-01T04:10:19.031614
2017-06-15T16:26:26
236246265
{ "authors": [ "briri", "dsisu", "rhu001", "sjDCC", "stephaniesimms", "vyruss" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13904", "repo": "DMPRoadmap/roadmap", "url": "https://github.com/DMPRoadmap/roadmap/issues/424" }
gharchive/issue
General layout/styling issues Please record any styling/layout changes that should be made to the newly re-branded site (currently deployed on roadmap-dev and roadmap-stg) Note any color and font changes, positioning of error/informational messages, etc. Note: @briri @stephaniesimms @JEK-III and I met on 6/15 and discussed that we want to wait until more of the functionality is baked in before doing another round of graphic design changes. This round was just to set a baseline and see how we feel about a cleaner all white background. We'll be experimenting with different colors for the heading as we go. Do however, feel free to add comments about wishlist items moving forward. related to #127 which contains initial specs Agreed and that request is tied to #175. Rails has 2 message streams, one for notices and another for errors. We're not using the appropriate ones in all cases so whoever updates the placement and styling of these messages should also make sure we're using the appropriate flash[:notice] or flash[:alert] Had made comments about this before seeing this ticket. See #475 Also just spotted an example on homepage where we need more whitespace between error message and top banner. I'm happy with the banner at the top, but this does have implications for our logo so we'll need to reflect on that for our styling. We should probably also suggest unis use a white version. I wonder if we want a bottom banner to even out display. That's something I like in the Melbourne branding, see: https://dmp.research.unimelb.edu.au/en-UK/about_us @dsisu can you take a look and chip in comments here too please as you've done most of the styling work on our end. Re institutional branding and the top banner display I'll cross-reference related issue #156 @rhu001 is still working on wireframe options for this, but I think we all liked Weiwei's approach to replacing the Portage logo (or in this case roadmap) logo w/the institutional logo if admins upload one. Let us know what you think about it and/or if you always want to display the DMPonline logo somewhere in top banner? @rhu001 For reference, here's what it looks like currently after the CSS rebuild. This particular logo happens to have a lot of extra white space on the right hand side. We would need to use logos with a white background like this. Logos with transparent logos get lost on the grey bar Maybe this is a reason not to have a top banner? In general I'm for it, but if transparent logos get lost so people have to upload ones with a white background, the institutional branding is going to look a bit crummy. That was actually one of the things I had against our grey background. Logos with white surrounds looked really blocky on it. @stephaniesimms regarding institutional logos, I think the Portage approach of replacing the DMPonline / DMPTool / DMPRoadmap logo with their own one is good. Institutions would definitely like that. Only quibble is contextualisation. On Portage they have a strapline which remains. Perhaps we need to introduce something like this so it's clear what the service does when it's been customised. How about "Create, review, and share data management plans" or something of that ilk. @dsisu what are your views on all this? What do you feel our users would like? For reference here are some shots from Portage - pre login And post login, when an institutional logo has been uploaded @briri @sjDCC We could potentially do an all white background for the top header area and that would take care of the color issue. I'd like to clarify... is the University of Edinburgh logo example with the two side by side logos be a real case or were you just plopping that in to show me what that logo looked like? If so, there is a potential dimension issue as well in that space if we need to account for two side by side logos. @rhu it was just an example of what the logos would look like under the current method used to display institutional logos. There are others that do not have the extra white space but I don't have access to them all. Hi @rhu001 this is an example of what it looks like in the current model. The proposal is to have the institutional logo replace the DMPTool / DMPonline / DMProadmap logo. Currently it just sits next to it. The issue is making sure institutional logo works on banner and has context. I've suggested we keep a strapline like Portage @rhu001 to follow on comments above, if we all like the proposal to handle institutional branding by replacing the logo it's probably easier to do an all white background for the top header. this will accommodate institutional logos that tend to be color on white and save us the extra styling effort of supporting custom CSS like weiwei does. and yes to what @sjDCC said about message styling. i assume we want to distinguish between status and error msgs somehow (e.g., color) but display them in a consistent location. am happy to follow whatever you recommend. Updated the header so that the background is white and when the user is logged in it will display their institution's logo if one is present otherwise it displays the roadmap logo. We will unfortunately need to re-upload the logos because our upload process resized them to fit in the old location and they are blurry when we increase the size and display them where the roadmap logo is. Here is what it looks like with the roadmap logo on the home page. I feel like we should have a line or something that runs across the page. The explore dropdown feels like its just floating out there. And here is what it looks like with an institutional logo. This Univ. of Edinburgh logo has a lot of white space around it, there are likely better options out there for it. @briri Are the two logos in place bottom aligned with the global nav? Hi all, sorry I haven't had the time to comment on the proposed new styling, I've been away. I need time to take a look at the wireframes. As discussed at our UX meeting yesterday, we should also carry out usability testing before we roll out any changes. Just to clarify the conversation from the UX mtg on 12 Jul: usability testing is built into the MVP roadmap and planned for after the accessibility retest on 31 Jul. Styling and branding are handled independently for each of our services. This ticket pertains to generic layout and styling for the DMPRoadmap codebase, which is what others will see when they download and install the tool. I understand this. We applied our own styling to the proposed generic tool and the end-result isn’t great. I guess this is work in progress so I’m not too concerned about the DIY look of the current generic tool. The one thing that does concern me is the introduction of new elements on the page e.g. Explore. This is not a matter of styling and will have a big impact on the way people navigate the site. I’m not convinced this is more user-friendly than what we currently have and I don’t think it’s justified by accessibility either. We should not introduce changes like these without testing first. Also, accessibility and usability are not incompatible with good looks. It is important that we produce a well finished, professional looking generic tool that will entice organisations to download it and use it. I guess I’m preaching to the converted. I have other work to do now, I’ll look at the proposed wireframes next week and get back to you by Wednesday next week. Have a good weekend! #161 Copied over from issue #491 since they deal with site-wide styling changes: Remove boldface for field names in all plan creation (for multiple templates from same funder), writing, sharing, etc pages Remove colon after each field name Be consistent w/capitalization of every word for field names "Project Title" (or just first word, but we should pick one) update css to use the text '(required)' instead of '*' when the label has class="required" standardize the width(s) for entry boxes, ideally make them the same as wireframes which present 2 different sizes @sjDCC pointed out that the Page title's and descriptions are not lined up with the main content of the page. We can likely fix this sitewide via CSS We discussed how to handle institutional logos today and have a suggestion to make. We'd prefer not to lose the DMPonline one completely, but don't feel two logos work well side-by-side so in effect agree with the decision to replace this with the institutional one. What we'd propose as an addition is switching the DMPonline logo to be an icon at the start of the menu. We started to mock-up a visual of this. Will try to add a screengrab before calls. One change layout wise would be dropping the menu beneath the primary logo rather than it running adjacent to it (bottom aligned) as this would allow for different sizes (square, landscape orientation etc) of uni logos more easily @sjDCC mockup: Perfect, thanks @vyruss @sjDCC and @stephaniesimms are no longer bold and do not have a colon after them. fixed capitalization issue on profile tabs, create plan and edit project details pages. Changes are currently in PR Adding two wireframes agreed on UX call on 9th August as general layout for header / footer etc. This covers items we've been discussing under global nav and institutional branding. The wireframes marry together ideas shared by @rhu001 with some alternative suggestions based on the concerns we raised about header depth and multiple menu bars. Superseded by bootstrap work. closing out
2025-04-01T04:10:19.034182
2018-10-12T11:40:32
369515281
{ "authors": [ "Bodacious", "briri" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13905", "repo": "DMPRoadmap/roadmap", "url": "https://github.com/DMPRoadmap/roadmap/pull/1929" }
gharchive/pull-request
Fix bug preventing new sections from merging into persisted phases Fixes #1853 Changes proposed in this PR: Fixed a bug that prevented new sections on customised template from being merged onto newly created template in UpgradeCustomizationService (this was caused by trying to save a Section with Questions. The presence validation on the Question model was preventing this) failures in GTravis look like they were fixed in Sam's latest PR
2025-04-01T04:10:19.041399
2017-05-05T18:39:45
226657906
{ "authors": [ "hmillison", "timbuckley" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13906", "repo": "DNAinfo/redux-saga-test-engine", "url": "https://github.com/DNAinfo/redux-saga-test-engine/pull/21" }
gharchive/pull-request
Add mechanism to instruct test engine to throw error Handles @CoryDanielson's request in #15 Expose throwError as a faux-effect that returns an object with a key @THROW and an error message In the test engine, if we see a value that matches throwError we can instruct the generator to throw This lets consumers explicitly define that they want the generator to error out in their mappings instead of writing a value that will cause an error indirectly (like in this unit test: https://github.com/DNAinfo/redux-saga-test-engine/blob/master/tests/index.js#L336) Only thing left to do is merge master in and resolve the conflict caused by the ES5 build stuff. Specifically, we have to make sure throwError is exposed at the top level. Just updated this PR per your feedback, Cory Will merge once CI passes
2025-04-01T04:10:19.062649
2024-07-22T21:35:11
2423809713
{ "authors": [ "dblodgett-usgs", "mkoohafkan" ], "license": "CC0-1.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13907", "repo": "DOI-USGS/nhdplusTools", "url": "https://github.com/DOI-USGS/nhdplusTools/issues/397" }
gharchive/issue
XY from get_nldi_feature() can result in bad geometry from get_split_catchment() It appears that using the geometry produced by get_nldi_feature() can produce bad results compared to pairing it with get_raindrop_trace(). It can also result in different geometries, e.g. POLYGON vs MULTIPOLYGON vs GEOMETRYCOLLECTION. watershed1 = list("featureSource" = "wqp", featureID = "USGS-02471078") |> get_nldi_feature() |> st_geometry() |> get_split_catchment() |> filter(is.na(catchmentID)) |> st_cast("POLYGON") watershed2 = list("featureSource" = "wqp", featureID = "USGS-02471078") |> get_nldi_feature() |> get_raindrop_trace() |> mutate(point = map(intersection_point, function(x) st_sfc(st_point(x), crs = 4326L))) |> pull(point) |> chuck(1) |> get_split_catchment() |> filter(is.na(catchmentID)) |> st_collection_extract("POLYGON", warn = FALSE) |> st_cast("POLYGON") bind_rows( watershed1 = watershed1, watershed2 = watershed2, .id = "id" ) |> ggplot() + geom_sf() + facet_wrap(~id) Do you mind formatting this issue in terms of what you see as wrong and what your preferred behavior is? In this instance, the web service provides what it does and nhdplusTools is just a thing wrapper over the service. Ok, I think I'm struggling with differentiating between what the package is doing vs what the web service is doing. I guess this issue is another flavor of #395? Perhaps this issue as well as #395 and #392 and #396 are all resolved with documentation of the web services, or links to existing documentation of the web services. Do you mind formatting this issue in terms of what you see as wrong and what your preferred behavior is? Could you add a warning in the documentation for get_split_catchment() and/or get_nldi_feature() that the point returned by get_nldi_feature() may not fall on a flowline and be compatible with get_split_catchment(), even if it has a valid COMID identifier? And then recommend using get_raindrop_trace() for more robust results. For sure. Thanks makes sense.
2025-04-01T04:10:19.071378
2024-11-21T10:29:42
2678919900
{ "authors": [ "PeterStaar-IBM", "Upabjojr" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13908", "repo": "DS4SD/docling", "url": "https://github.com/DS4SD/docling/issues/403" }
gharchive/issue
Graphical user interface for parsed JSON? Requested feature I was wondering whether there is a GUI (graphical user interface) to view the parsing result (either the pydantic model or its JSON format). Is there an easy tool to overlap the parsed data with the original source? For example, given a PDF, is there a tool to display the parsed data on top of the source PDF or to see it side-by-side? Alternatives I haven't found any useful alternative so far. If you have suggestions, please link any useful external project :) @Upabjojr Yes, we want to add that in the short to medium term (simple app using stream-lit). This issue is a duplicate of (https://github.com/DS4SD/docling/issues/277), so I will close it for now.
2025-04-01T04:10:19.075899
2023-11-02T02:19:57
1973381816
{ "authors": [ "Keon-San", "karkir0003" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13909", "repo": "DSGT-DLP/dlp-practice", "url": "https://github.com/DSGT-DLP/dlp-practice/pull/4" }
gharchive/pull-request
Keon-part2 This PR implements all parts of Part 2 of the beginner project. A detailed video showing all parts will be attached soon Here is the video for this pull request: https://youtu.be/sJOGsS769o0 @Keon-San EXCELLENT JOB here! Looking forward to part 3! Closing this PR since we will not merge these changes into the real app
2025-04-01T04:10:19.109818
2020-02-24T16:40:22
569978439
{ "authors": [ "tdonohue" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13918", "repo": "DSpace/DSpace", "url": "https://github.com/DSpace/DSpace/pull/2690" }
gharchive/pull-request
DS-4380: Update XOAI (OAI-PMH) for full JDK11 compatibility Minor PR to update to XOAI v3.3.0, which is more compatible with JDK11. In XOAI v3.3.0, I've updated dependencies to avoid any "An illegal reflective access operation has occurred" warnings when running JDK 9 or above. See https://github.com/DSpace/xoai/releases/tag/xoai-3.3.0 This PR also makes a minor refactor to the classes which obtain the Solr configuration for dspace-oai to ensure they provide better logging & use the more correct version of the configuration oai.solr.url. I've tested both of these minor changes locally and they work well on latest master and with JDK11. At +1 and Travis approved. Merging, as this is a tiny PR & I've tested it thoroughly locally.
2025-04-01T04:10:19.114702
2019-11-21T15:06:10
526652910
{ "authors": [ "bram-atmire", "vitorsilverio" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13919", "repo": "DSpace/dspace-angular", "url": "https://github.com/DSpace/dspace-angular/pull/524" }
gharchive/pull-request
Portuguese Translation [Work in progress] Initial translation to PT [x] 404 messages [x] bitstream messages Please tell me if i need to translate everything before merge, i can keep updating this PR until there. Hi @vitorsilverio - many thanks for your translation! If you are willing to translate the entire catalog, yes, please, do it as multiple commits to this pull request & we'll finalize the review and the merge when you are done. You're also welcome to contribute it in smaller pieces, but definitely preferred if you can do it entirely in the scope of a single pull request. OK @bram-atmire i ll translate everything and ping you when i finish @bram-atmire i finished the translation I think i done something wrong when merge, a lot of files were added to my PR Closing this, the translation is PR #525 @vitorsilverio I think it's because of some changes that were applied to master in the meanwhile. I think a the owner of the pull request, you should have a button somewhere that says REBASE. Otherwise, it's also possible to close this pull request and start a new one.
2025-04-01T04:10:19.139951
2021-07-19T08:08:20
947386856
{ "authors": [ "Chew", "DV8FromTheWorld", "RedDaedalus" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13920", "repo": "DV8FromTheWorld/JDA", "url": "https://github.com/DV8FromTheWorld/JDA/pull/1736" }
gharchive/pull-request
Implement user banner & banner color Pull Request Etiquette [x] I have checked the PRs for upcoming features/bug fixes. [x] I have read the [contributing guidelines][contributing]. Changes [x] Internal code [x] Library interface (affecting end-user code) [x] Documentation Closes Issue: NaN Description Implements basic support for user banners. This is not very usable currently, missing docs, and is expected to break when the Discord API is changed shortly. Docs PR: https://github.com/discord/discord-api-docs/pull/3448 Considerations • Banner information is only sent when the user is specifically requested, a method to retrieve this might be beneficial. • The handling of unknown state from passively gathered user objects • The changes that will be made to the banner_color field are currently not known in detail. Placed in freezer until Docs PR is merged The docs PR has been merged. All of these new properties are related to a user's profile. I feel like we shouldn't add all of these top-level getters and should instead just have the POJO for Profile Just went through reviews, apologies for the delay here. The way I've implemented this, retrieveProfile returns a cached profile only when using a User directly produced by retrieving them. This prevents redundant getters, but also feels a bit confusing to me. Not sure if there's a better approach however. All review suggestions have been committed. For the future, you can apply multiple suggestions from code review...
2025-04-01T04:10:19.151408
2018-07-26T08:20:11
344741887
{ "authors": [ "DWilliames", "craigmdennis", "dlewand691", "dngrlux" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13921", "repo": "DWilliames/paddy2-beta", "url": "https://github.com/DWilliames/paddy2-beta/issues/59" }
gharchive/issue
Stack miscalculated if a masked object is contained If a group within a stack contains a larger masked element, the stacking uses the inner masked element's dimensions, rather than the mask itself. Paddy2 should ignore the dimensions of the masked element and instead calculate bounds using the mask. There are some other oddities depending if the masked element's group is at the top, middle, or bottom of the stack, and if the 'Collapse hidden views' is toggled or not. This is a little difficult to explain, but easy to see and experiment with. All of the examples attached are using the same stacking settings. Thanks for the amazing plugin! paddy2 mask stacking issue.sketch.zip I have also found this to be an issue and hope there's a fix for the next release. Thanks for this @dngrlux — this one is a pretty complex fix. Going to look into it and see what I can do. 👍 Ok. Looks like I've managed to fix this for the next release 👍 This also happens when only setting padding. I hope the fix affect all calculations.
2025-04-01T04:10:19.175044
2021-12-15T09:48:58
1080814372
{ "authors": [ "de-jcup" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13922", "repo": "Daimler/sechub", "url": "https://github.com/Daimler/sechub/issues/911" }
gharchive/issue
Provide Snippet property in region Situation Sometimes it is necessary to specify a snippet for a finding which is possible in SARIF. E.g. a secret scanner could provide here found secrets/credentials etc. Currently this is not supported by SecHub Sarif model. Wanted We want to have the region snippet possiblity Solution Inside the SARIF region object we have got an optional snippet property, see https://docs.oasis-open.org/sarif/sarif/v2.1.0/os/sarif-v2.1.0-os.html#_Toc34317698 (3.30.13) which is an `ArtifactContent? object - see https://docs.oasis-open.org/sarif/sarif/v2.1.0/os/sarif-v2.1.0-os.html#_Toc34317422 (3.3) Implement the ArtifactContent object and provide it as an optional snippet property inside region Will be done together with #606
2025-04-01T04:10:19.179478
2023-10-28T07:15:28
1966474588
{ "authors": [ "SeraphimI0", "gumbobr0t", "something-0001" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13923", "repo": "DamagingRose/Rose-Grabber", "url": "https://github.com/DamagingRose/Rose-Grabber/issues/177" }
gharchive/issue
Exe to big the built exe is way to big even when I put it at 2 mb it comes out like that Hi! The compiled stub has a static size of about 80-90MB because of the PyPi packages that are compiled into it. The file pumper does not set your file size to a specific amount, it pumps it. (Basically makes the output executable bigger.) Have a good evening, Sir! I don't really know why it comes out as 119MB tho, last time I tried I got 80MB. Must be an issue with the file pumping process calculation. Don't worry tho, we are working on a fud dropper. (Probably made in C++.) I don't really know why it comes out as 119MB tho, last time I tried I got 80MB. Must be an issue with the file pumping process calculation. Don't worry tho, we are working on a fud dropper. (Probably made in C++.) Make a C++ stealer lul. I don't really know why it comes out as 119MB tho, last time I tried I got 80MB. Must be an issue with the file pumping process calculation. Don't worry tho, we are working on a fud dropper. (Probably made in C++.) Make a C++ stealer lul. nah too hard im learning nim & cpp rn but another stealer is more experienced
2025-04-01T04:10:19.186090
2022-07-02T01:06:40
1291928121
{ "authors": [ "DamianOsipiuk", "veryard" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13924", "repo": "DamianOsipiuk/vue-query", "url": "https://github.com/DamianOsipiuk/vue-query/issues/202" }
gharchive/issue
Best Practise for working with a store? Hey, Very curious about using this package, I just have a quick question to as if there is any examples or best practices to get this package working with a store eg. Pinia or Vuex. Example of my store with an action: actions: { async fetchEmployees({ page = 1, search = "" }) { const { isLoading, data, error } = useFetch( ['employees', page, search], `/api/employees?page=${page}&search=${search}` ) this.loading = computed(() => isLoading.value); this.error = computed(() => error.value); this.employees = computed(() => data.value?.data); } } // useFetch function export function useFetch(key, url, options = {}) { return useQuery(key, () => axios(url, options) .then(res => res.data) ); } And my component looks something like: <script setup> import { useEmployeesStore } from "@/store/employees"; useHead({ title: "Manage Employees", }); const store = useEmployeesStore(); const employees = computed(() => store.employees); await store.fetchEmployees({}); </script> If I rerun the store.fetchEmployees({}) I end up getting this error: I know that you return a refetch from the useQuery function am just curious to as if you or anyone has a better approach to this. Thanks. I think the best practice would be to NOT use it with pinia, vuex or any other client state library (pushing state from vue-query to local state). You can still use pinia for your client state, that is totally fine. Vue-query does fetching, caching, deduping etc. automatically for you and it should serve as your store for async data. It will simply remove all the wiring and boilerplate you have to write. You have even dedicated devtools that should help you explore what queries reside in cache and all the state about them. You can / should abstract specific calls as a wrapper functions that would make it easier to reuse: export function useEmployeesQuery(page, search, options = {}) { return useQuery(['employees', page, search], () => axios(`/api/employees?page=${page}&search=${search}`, options) .then(res => res.data) ); } And you can call this function from whatever component you would like, and the rest will be handled by the library. If you set longer staleTime, it will not refetch until staleTime is met, even if you mount new components or remount current ones. If you mount multiple components on a page using the same query, they will send only one request. etc. etc. So instead of const store = useEmployeesStore(); const employees = computed(() => store.employees); await store.fetchEmployees({}); You should just do: const query = useEmployeesQuery(page, search) const employees = computed(() => query.data?.value); And it will fetch the data when needed. Ah awesome, I had spent a bit more time playing around and came to this conclusion. Appreciate the speedy response and great job on the package. Thanks again!
2025-04-01T04:10:19.202590
2022-03-22T18:26:53
1177145899
{ "authors": [ "RaydanOMGr", "itzJOHv", "migue802" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:13925", "repo": "DanBot-Hosting/DanBotHostingStats", "url": "https://github.com/DanBot-Hosting/DanBotHostingStats/pull/264" }
gharchive/pull-request
Trying to fix the bug to create more servers Adding a server to the used server count before the is being created, in case there is a error just remove it. useless cuz before this pr server was given once server gets created and what will happen if it can't catch the error and the bot just crashes? it will remove you a server and you cannot recover it