added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T06:38:29.869849
2021-05-11T06:12:28
886159117
{ "authors": [ "weichou1229" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5561", "repo": "edgexfoundry/edgex-go", "url": "https://github.com/edgexfoundry/edgex-go/pull/3453" }
gharchive/pull-request
feat(notifications): Implement GET /transmission/id/{id} V2 API Close: #3452 Signed-off-by: weichou<EMAIL_ADDRESS>PR Checklist Please check if your PR fulfills the following requirements: [x] Tests for the changes have been added (for bug fixes / features) [ ] Docs have been added / updated (for bug fixes / features) If your build fails due to your commit message not passing the build checks, please review the guidelines here: https://github.com/edgexfoundry/edgex-go/blob/master/.github/Contributing.md. What is the current behavior? Issue Number: #3452 What is the new behavior? Implement GET /transmission/id/{id} V2 API according to the doc https://app.swaggerhub.com/apis-docs/EdgeXFoundry1/support-notifications/2.x#/default/get_transmission_id__id_ Does this PR introduce a breaking change? [ ] Yes [x] No New Imports [ ] Yes [x] No Specific Instructions Are there any specific instructions or things that should be known prior to reviewing? Other information Rebased.
2025-04-01T06:38:29.874023
2016-07-16T10:10:09
165922811
{ "authors": [ "coveralls", "duksis", "nscyclone" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5562", "repo": "edgurgel/tentacat", "url": "https://github.com/edgurgel/tentacat/pull/91" }
gharchive/pull-request
Replace the use of the deprecated Dict module The Dict module is deprecated. Coverage remained the same at 90.135% when pulling 6f9925fd925e38535c7078934b2c3e43f3528b29 on nscyclone:fix/replace-dict into 925200ded23aa0009919a3e48a0dae1952398d63 on edgurgel:master. thank you @nscyclone
2025-04-01T06:38:29.876099
2016-10-15T20:35:38
183231569
{ "authors": [ "edinc", "munkiepus" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5563", "repo": "edinc/vagrant-jenkins", "url": "https://github.com/edinc/vagrant-jenkins/issues/1" }
gharchive/issue
Remove the Jenkins setup wizard On first provisioning of the vagrant box when running the Jenkins server the user has to go through a security setup wizard. This should be removed to further speed up the process of having the Jenkins ready. I've done this before so might be able to help, it's a case of adding a lot of jenkins config files and is quite verbose Any contributions are welcome really. This was resolved with #10. Thank you for all your contributions @munkiepus. I will add you as a contributor to the repository so that you can directly create branches and review and merge future PR-s.
2025-04-01T06:38:29.891641
2019-05-05T17:50:27
440471118
{ "authors": [ "He-Ro", "edkolev", "khalsah", "secuvim", "shrop" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5564", "repo": "edkolev/tmuxline.vim", "url": "https://github.com/edkolev/tmuxline.vim/pull/104" }
gharchive/pull-request
Adapt generated tmux config to new tmux 2.9 syntax Closes #102 Closes #100 Inspired by @secuvim in #102 I do not know if this is all that is needed, but it worked for me. This would break the backwards compatibility for users with tmux versions < 1.9. To keep the backwards compatibility you could wrap your changes in something like: let tmux_version = system("tmux -V") if (tmux_version < 'tmux 1.9') let misc_options = ... let win_options = ... else let misc_options = ... let win_options = ... elseif This is just a suggestion as I am not the maintainer of this package. @He-Ro Thanks for implementing a fix. @edkolev What is your opinion on this? Some of the presets appear to write into win_options['window-status-activity-attr'] which causes problems if you're using one of those presets: https://github.com/edkolev/tmuxline.vim/blob/c8a0295eb34bf11447779a5a203fd472147788a7/autoload/tmuxline/presets/powerline.vim#L26 https://github.com/edkolev/tmuxline.vim/blob/c8a0295eb34bf11447779a5a203fd472147788a7/autoload/tmuxline/presets/nightly_fox.vim#L15 https://github.com/edkolev/tmuxline.vim/blob/c8a0295eb34bf11447779a5a203fd472147788a7/autoload/tmuxline/presets/crosshair.vim#L23 @He-Ro amazing work on this, thanks! Could you also: git grep and change everywhere window-status-activity-attr => window-status-activity-style add yourself to the CONTRIBUTORS.md in the root of the project @khalsah good catch! This would break the backwards compatibility for users with tmux versions < 1.9. @edkolev What is your opinion on this? I'm fine with breaking compatibility for tmux <1.9 - it's been more than 5 years since 1.9 was released. Adding a conditional in the vim script would not work when you share the generated tmux conf file between machines with different versions of tmux. Added the requested changes. Also found a mention of attr in the README. Thanks for working on this. Just noting https://github.com/tmux/tmux/wiki/FAQ#how-do-i-translate--fg--bg-and--attr-options-into--style-options and curious if this was a known issue here. and curious if this was a known issue here Could you clarify the question? The link is pointing to a wiki entry about migration; the wiki entry isn't a know issue Gotcha, maybe this issue isn't related. I had to manually fix up some tmuxline generated settings to comply with the Tmux 2.9 Syntax ralated to -style and -attr. If thosed fixes have been merged in, I can try to run a new snapshot and see if that works. This PR addresses exactly this - the *-fg/bg to *-style migration. And yes, you should be able to create a snapshot which works with the lates tmux.
2025-04-01T06:38:29.954472
2022-06-16T12:49:27
1273520714
{ "authors": [ "Jerome-Liger", "edsnider" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5566", "repo": "edsnider/latestversionplugin", "url": "https://github.com/edsnider/latestversionplugin/pull/44" }
gharchive/pull-request
Workaround to repair IsLastestVersion for Android. Quick implementation from Nick Kovalsky workaround : https://stackoverflow.com/questions/72407251/how-to-get-version-number-of-application-from-play-store-using-xamarin-forms/72643625#72643625 Fixes #43 #42 . Changes proposed in this pull request: Repair plugin for Android Platform. Thank you for the PR. Please remove the second commit - a test project should not be included in this PR, just the fix. That commit is also breaking the build. Thank you. Beta nuget is available here: https://www.nuget.org/packages/Xam.Plugin.LatestVersion/2.1.1-beta.107
2025-04-01T06:38:30.028715
2020-05-18T02:20:58
619861171
{ "authors": [ "edx-webhook", "ericfab179" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5569", "repo": "edx/TinCanPython", "url": "https://github.com/edx/TinCanPython/pull/7" }
gharchive/pull-request
WIP [BD-6] OEP-18 Compliance and tests Python 3.8 Add setup.py install_requirements definition. Update travis file using new requirements. Create requirements folder. create pip_tools, base, test and travis requirements files. Add Makefile. Add makefile upgrade command. Include requirements files generated using upgrade command. Add openedx.yaml and include OEPs list. Add python 3.8 to tests. Thanks for the pull request, @ericfab179! I've created OSPR-4550 to keep track of it in JIRA. JIRA is a place for product owners to prioritize feature reviews by the engineering development teams. Feel free to add as much of the following information to the ticket: supporting documentation edx-code email threads timeline information ("this must be merged by XX date", and why that is) partner information ("this is a course on edx.org") any other information that can help Product understand the context for the PR All technical communication about the code itself will still be done via the GitHub pull request interface. As a reminder, our process documentation is here. @ericfab179 🎉 Your pull request was merged! Please take a moment to answer a two question survey so we can improve your experience in the future.
2025-04-01T06:38:30.036752
2015-05-15T18:45:35
76818565
{ "authors": [ "Nickersoft", "clintonb", "jimabramson", "rlucioni" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5570", "repo": "edx/ecommerce-api-client", "url": "https://github.com/edx/ecommerce-api-client/pull/8" }
gharchive/pull-request
API client now passes the user's full name to ecommerce This merge also requires a merge on ecommerce and edx-platform with tyler-jwt-fullname, so keep that in mind. @clintonb @rlucioni FYI: @wedaly @jimabramson Tests are required. @clintonb Will do that now Minor: Cleanup the commit message. We generally use a simple statement of the overall change as the title (e.g. Added support for passing full name via JWT). Additional information can be provided as a paragraph or list after the title. Added support for passing full name via JWT - User's full name can be passed to the API when using JWT authentication. This field is optional. - Email is no longer required for JWT authentication. Aside from the message change, :+1: . Please await a second approval from @rlucioni or @jimabramson before merging. Oh, I actually not aware that Git messages could be multiline like that. I'll change it then force a push. Thanks @clintonb. Once you get a clean build, :+1: . @Nickersoft I'm surprised to see a major version bump on this change; was expecting 0.5.0. Is there something backwards-incompatible? @jimabramson My mistake. The constructor was changed in a backwards-incompatible manner. We can make it backwards-compatible if you'd like. ah. somehow i missed that.
2025-04-01T06:38:30.056334
2021-02-17T11:57:10
810119276
{ "authors": [ "farhan-arshad-dev", "omerhabib26" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5571", "repo": "edx/edx-app-android", "url": "https://github.com/edx/edx-app-android/pull/1515" }
gharchive/pull-request
Navigate within App from dates to course component Description LEARNER-8214 Navigate within app when a link is clicked on Full Page dates screen @omerhabib26 The app still navigates to screen for the un-available course unit. https://user-images.githubusercontent.com/43750646/108349158-78f9bb00-7204-11eb-9459-e5fbb3221edc.mp4
2025-04-01T06:38:30.112083
2015-12-10T20:00:20
121565553
{ "authors": [ "clintonb", "mjfrey", "rlucioni" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5572", "repo": "edx/edx-rest-api-client", "url": "https://github.com/edx/edx-rest-api-client/pull/21" }
gharchive/pull-request
Update AUTHORS and bump version Update to AUTHORS and bump version to 1.3.0 @clintonb @rlucioni :+1: :+1:
2025-04-01T06:38:30.195930
2024-02-15T15:46:57
2136866221
{ "authors": [ "eeston", "jgoguen" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5573", "repo": "eeston/Jello", "url": "https://github.com/eeston/Jello/issues/9" }
gharchive/issue
Add manual control of Repeat and Shuffle Related to #8 is adding the ability to toggle shuffle and repeat. The play queue should have the following states: Shuffle on: randomize the order of the queue every time shuffle is turned on. Shuffle off: Version 1: It would be fine to just not change the order of the current play queue for V1 of this feature . Version 2: If playing from a defined playlist (either a playlist, album list, or a future "All Songs" view), turning shuffle off should reorder the queue according to the list display order, with playback continuing from whatever queue position holds the currently playing song. If playing from a queue modified with "Add to Queue", just leave the queue in its current order. Version 3: If the queue was created with "Play Next" or "Add to Queue", reorder the queue according to the order of whatever view it was started with and play back from whatever queue position the current song is at. Repeat off: play to the end of the current queue and stop. Repeat One: continuously play the current song, only changing if the song is manually changed by the user (by choosing a new song or pressing the Next button). Repeat All: play to the end of the current queue; if Shuffle is on randomize the queue order and start from the beginning, or if shuffle is off just go back to the beginning of the queue and continue playback. When starting playback by choosing a song or pressing Play, the Shuffle state should be whatever it was previously. When starting playback by pressing Shuffle, the Shuffle state should be turned on. Defs want to add this. At the moment playback issues are my priority. I would like to have a solid foundation before bolting on features.
2025-04-01T06:38:30.228742
2019-10-24T16:57:43
512066305
{ "authors": [ "eMaringolo", "eftomi" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5582", "repo": "eftomi/pharo-ado", "url": "https://github.com/eftomi/pharo-ado/issues/27" }
gharchive/issue
Variant type 2 (vt: 2) has to be implemented GlorpInt2Test This is solved in this PR https://github.com/tesonep/pharo-com/pull/10 PR merged, this issue can ble closed.
2025-04-01T06:38:30.253833
2024-04-08T07:34:26
2230480118
{ "authors": [ "NSystemx" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5583", "repo": "eggsy/website", "url": "https://github.com/eggsy/website/issues/56" }
gharchive/issue
I created firebase but I don't know how to create music. I created firebase but I don't know how to create music. Daily Songs with Firestore How to create firestore database for daily songs?
2025-04-01T06:38:30.260357
2020-06-08T14:50:00
634690429
{ "authors": [ "Kokiro", "mugiwarafx", "onjin" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5584", "repo": "egoist/hack", "url": "https://github.com/egoist/hack/issues/77" }
gharchive/issue
This site hackcss.egoist.moe can’t be reached Hi, This site hackcss.egoist.moe can’t be reached :( Best, Kokiro Hi @Kokiro, today I tried to restore the website with the style guidelines: https://hackcssbckp.herokuapp.com/ hope it helps! 😀 According to CNAME at github-pages the working address is hackcss.egoist.sh
2025-04-01T06:38:30.286203
2021-12-15T16:16:51
1081228264
{ "authors": [ "chandrakananandi", "femtomc", "mwillsey" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5585", "repo": "egraphs-good/egg", "url": "https://github.com/egraphs-good/egg/issues/144" }
gharchive/issue
Applying egg to SSA-based IRs Hi! What are the prospects of utilizing egg to perform rewrite-driven transformations on SSA-based IRs (?) I read this paper (whose authors appear to be collaborators with y'all): https://www.cs.cornell.edu/~ross/publications/eqsat/ where they describe E-PEG-based program representation. On the other hand, I'm working on IRs which are closer to strict SSA (where all control flow is de-sugared to basic blocks and branching) and everything is linear (e.g. blocks are vectors of instructions) vs. strictly tree-like. I'd love to find or understand how equality saturation could be applied to this representation structure -- it's likely that maybe this information can be gleaned from the above paper (but I expect that maintainers here might be able to unpack this more than I can). Hi! I use egg for a language which is mostly-SSA. In my case, things "just work" so far because I mostly care about straight line blocks. The only additional information I track which may be necessary for other SSA-based usages as well is to make sure you don't use variables before they are defined. This, in my case, was possible with a simple e-class analysis. Hi, I think @chandrakananandi is right. I don't think there is any major technical blocker stopping you from doing this today! The original eqsat paper you linked to contains a lot of good ideas on how to encode your problem into an e-graph. If I'm understanding your setting, you'll still want E-PEGs if you want to encode loops in a transparent way that you can optimize through. If you only care about optimizing one basic block (one DFG) at a time, then you don't need them. The egg paper (and tool) are mostly innovating in how equality saturation is done. Despite there not being any huge blockers, I still think it's a large and challenging task, and one that I'd like to try to tackle at some point (if I can find the time), or see someone else take a stab at! One thing that I will add: you'll have a much easier time with a tree- or dag-like IR that a linear, mutating IR. All the rewrites that you do are over trees or dags, and just overall the e-graph doesn't do a lot for you if your language is heavily sequential. If that's your case, consider building def-use chains or some other method to make things more graph-like for the e-graph. Also, this doesn't seem to be an issue, so I'm converting it to a discussion.
2025-04-01T06:38:30.295764
2023-09-08T07:35:57
1887127447
{ "authors": [ "Bastacyclop", "TrevorHansen", "mwillsey", "oflatt" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5586", "repo": "egraphs-good/extraction-gym", "url": "https://github.com/egraphs-good/extraction-gym/pull/9" }
gharchive/pull-request
An extra bottom-up recursive extractor Hi, Before I saw #8 I made a a recursive extractor, too. This code is much ugglier than #8, but is faster on some problems (for example acyclic graphs). On acyclic graphs it's able to extract in one pass through, without building the dependencies. Taking the cumulative times for all three extractors on my machine: Loaded 454 jsons. ... Cumulative time for bottom-up: 3386ms Cumulative time for bottom-up-analysis: 2483ms Cumulative time for bottom-up-recursive: 977ms bottom-up is the one currently in, bottom-up-analysis is #8, bottom-up-recursive is this PR. Is there benefit in having all three, or should the two new ones be combined? Good question! @Bastacyclop what do you think? Offhand, I'm leaning toward fewer, more distinct algorithms, so unless we can characterize that these have different advantages, I'd prefer to have just one. I feel like we should be able to combine the three bottom up versions, with some more experiments and code cleanup. Maybe by (1) doing a first pass without building dependencies; and (2) if dependencies are used, doing a bottom up analysis based on unique queues. The question would be whether using unique queues for the second stage brings performance benefits or not (I think it should: https://github.com/egraphs-good/egg/issues/239). To properly evaluate that I would like to see datasests with more costly to compute, child-dependent cost functions. PS: one thing to consider is that in egg the dependencies don't need to be computed as they are already stored in the e-graph. I would like to consider computation of dependencies (parents) as somewhat negligible, as its only linear and as @Bastacyclop says it's already there in many contexts. I'd like to preserve the "dumb" bottom up extractor as a base case. Ideally, we could consolidate the "smarter" bottom-up extractors into one (i.e., those that do not aim to do cost sharing, as a possible definition). Thoughts on that? Before I read this closely I also worked on a bottom-up extractor (#20) Sorry for duplicated work! I would also be happy with consolidating this, #20, and #8 if possible. With the recent changes to the bottom-up extractor(#20), there's now only a small time advantage for the extractor that I proposed introducing (Note these times differ from before because extra problems have been added). Currently: Cumulative time for faster-bottom-up: 2060ms [The one in #20] Cumulative time for bottom-up-recursive: 1533ms [The one in this PR]. Cumulative time for bottom-up: 4471ms Meaning that the extractor in this PR is only about 25% faster than the others, but is much uglier. However, there are some tweaks to the "faster-bottom-up" extractor which brings down its runtime to almost the same as the extractor in this PR: ################################################### faster-bottom-up vs faster-bottom-up-old extractors: ['faster-bottom-up', 'faster-bottom-up-old'] cumulative time for faster-bottom-up: 1649ms cumulative time for faster-bottom-up-old: 2060ms cumulative tree cost for faster-bottom-up:<PHONE_NUMBER>5237 cumulative tree cost for faster-bottom-up-old:<PHONE_NUMBER>5237 cumulative dag cost for faster-bottom-up: 78037 cumulative dag cost for faster-bottom-up-old: 78037 Cumulative time for faster-bottom-up: 1649ms Cumulative time for faster-bottom-up-old: 2060ms faster-bottom-up / faster-bottom-up-old geo mean tree: 1.0000 dag: 1.0000 micros: 0.8184 quantiles tree: 1.0000, 1.0000, 1.0000, 1.0000, 1.0000 dag: 1.0000, 1.0000, 1.0000, 1.0000, 1.0000 micros: 0.3611, 0.7825, 0.8227, 0.8611, 1.8333 So I've changed this PR to now just introduce some small speedups to the faster-bottom-up extractor, as well as fixing up attribution to @Bastacyclop. :+1:
2025-04-01T06:38:30.304644
2016-07-06T13:53:27
164080791
{ "authors": [ "chrisdennis", "lorban" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5587", "repo": "ehcache/ehcache3", "url": "https://github.com/ehcache/ehcache3/issues/1292" }
gharchive/issue
OffHeapChainMap thieving argument is ignored OffHeapChainMap is backed by EvictionListeningReadWriteLockedOffHeapClockCache but the latter lacks a constructor that accepts shareByThieving argument. The offheap store lib should be modified to add such constructor so that OffHeapChainMap can make use of it. Given how busted ARC is at the moment this turns out to be a non-issue. That said @AbfrmBlr is going to fixing the broken ARC implementation in clustered (that he wrote) and extending it to cover unclustered offheap and disk. He's going to run headlong in to this as a result. This will be fixed under #2215
2025-04-01T06:38:30.352575
2019-12-19T01:56:07
540026667
{ "authors": [ "ppazos", "testautomation" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5588", "repo": "ehrbase/ehrbase", "url": "https://github.com/ehrbase/ehrbase/pull/89" }
gharchive/pull-request
ehrbase/project_management#98 added ACTION examples and updated AQL test suite data @testautomation this includes the action compositions and the queries were updated too. :heavy_check_mark:
2025-04-01T06:38:30.366489
2018-02-05T12:04:36
294369745
{ "authors": [ "eidheim", "lxlenovostar" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5589", "repo": "eidheim/Simple-Web-Server", "url": "https://github.com/eidheim/Simple-Web-Server/issues/207" }
gharchive/issue
glaceful close ssl socket Hi: @eidheim void close() noexcept { error_code ec; std::unique_lock<std::mutex> lock(socket_close_mutex); // The following operations seems to be needed to run sequentially socket->lowest_layer().shutdown(asio::ip::tcp::socket::shutdown_both, ec); socket->lowest_layer().close(ec); } In this way, we directly close tcp socket, why we not close SSL socket first in HTTPS server? Thank you. Last time I studied this I came to the conclusion that calling ssl::stream::shutdown was not needed. Though I might be wrong! By the way, ssl::stream does not have a close-member function. Although, thank you for bringing this up. I'll add a couple of labels to this issue. @eidheim In a rare case, If I don't add mutex before shutdown/close, Segment error happen. Why you choose use mutex before shutdown/close?
2025-04-01T06:38:30.385946
2019-04-12T19:58:04
432721804
{ "authors": [ "alpeshjamgade", "ritzvik" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5590", "repo": "einsteinpy/einsteinpy", "url": "https://github.com/einsteinpy/einsteinpy/pull/194" }
gharchive/pull-request
Reorder parameters within functions Fixes #141 Really great contribution @alpeshjamgade . Thanks I would be reviewing in a day or two. Meanwhile could you make one change. define a private function _scr(M, c=default value, G=default value for calculating Rs. Putting units for calculating Rs seems unnecessary and slow, and may cause problems while implementing jit for the functions. Also, I guess you have made some merge commits while rebasing your branch. Try to fix that. Maybe refer https://github.com/k88hudson/git-flight-rules/ . One word of advice, try to create separate branches for fixing and developing things. The master should not ever diverge from the upstream. Believe me, I have learnt that the hard way. Really great contribution @alpeshjamgade . Thanks I would be reviewing in a day or two. Meanwhile could you make one change. define a private function _scr(M, c=default value, G=default value for calculating Rs. Putting units for calculating Rs seems unnecessary and slow, and may cause problems while implementing jit for the functions. Also, I guess you have made some merge commits while rebasing your branch. Try to fix that. Maybe refer https://github.com/k88hudson/git-flight-rules/ . One word of advice, try to create separate branches for fixing and developing things. The master should not ever diverge from the upstream. Believe me, I have learnt that the hard way. thank you you asked me to define a private function _scr. Before the schwarzschild_radius function in schwarzschid_utils.py was getting any M value as input weather it has astropy unit or not and return Rs in astropy units and also 'Rs' was the input to the most functions which was in astropy units only, but when i changed Rs with "M", i needed to make sure that "M" should be in astropy units , that means everytime i call such function i need to provide M value in astropy units otherwise it would throw an error, and we dont expect users to provide M in astropy units, it would be great if they can just give M value as input and if the given function need to calculate Rs, it would need M in astropy units then schwarzschild_radius function will convert M to astropy units, thast just what i did. please tell me if i can do any better. i think we dont need private _scr function, if we want Rs in floats we can just asked for it with command Rs.value , otherwise it will always be in astropy units. please tell me if i dont understand something. i think we dont need private _scr function, if we want Rs in floats we can just asked for it with command Rs.value , otherwise it will always be in astropy units. please tell me if i dont understand something. User always specifies things with units, but during heavy calculation internally, we don't use units. As functions like christoffels() are called thousands of time in a loop to get the trajectory it may be good to save some computation by not using astropy.units. But still wait some time, I am not sure what to do. @shreyasbapat Suggestions?? i have just created another pull request #196 for this issue , this time i didnt messed up my master. sorry for this. i did not consider on which branch i was making changes before.
2025-04-01T06:38:30.439609
2023-06-14T12:51:13
1756838022
{ "authors": [ "JonasKunz", "jackshirazi" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5591", "repo": "elastic/apm-agent-java", "url": "https://github.com/elastic/apm-agent-java/issues/3189" }
gharchive/issue
Revert workaround for micrometer bug in tests #3187 added a workaround for a micrometer bug. We should revert this workaround as soon as a micrometer version with a fix is available. fixed in https://github.com/elastic/apm-agent-java/pull/3264
2025-04-01T06:38:30.781952
2013-08-09T07:03:51
17849524
{ "authors": [ "dadoonet", "hjc1710", "jacobwgillespie", "rabidscorpio" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5592", "repo": "elastic/elasticsearch-cloud-aws", "url": "https://github.com/elastic/elasticsearch-cloud-aws/issues/31" }
gharchive/issue
Update to Elasticsearch 0.90.3 Issue reported in https://groups.google.com/d/msg/elasticsearch/bpdetIvIg5M/cMAUGwhRVf0J No problem with elasticsearch 0.90.2 and the cloud-aws 1.12.0 plugin (other than the usual node discovery flakiness), but I'm unable to get elasticsearch 0.90.3 to start: Initialization Failed ... 1) IllegalStateException[This is a proxy used to support circular references involving constructors. The object we're proxying is not constructed yet. Please wait until after injection has completed to use this object.]2) NoSuchMethodError[org.elasticsearch.discovery.zen.ZenDiscovery.<init>(Lorg/elasticsearch/common/settings/Settings;Lorg/elasticsearch/cluster/ClusterName;Lorg/elasticsearch/threadpool/ThreadPool;Lorg/elasticsearch/transport/TransportService;Lorg/elasticsearch/cluster/ClusterService;Lorg/elasticsearch/node/settings/NodeSettingsService;Lorg/elasticsearch/cluster/node/DiscoveryNodeService;Lorg/elasticsearch/discovery/zen/ping/ZenPingService;)V] Do I need to wait for a new version of cloud-aws, or is there some other problem here? My /etc/elasticsearch/elasticsearch.yml: cluster.name: foo plugin.mandatory: cloud-aws,lang-javascript cloud: aws: access_key: ******** secret_key: ******** region: us-east-1 discovery: type: ec2 ec2: ping_timeout: 15s gateway: type: s3 s3: bucket: bar For anyone (like me) who is seeing this error, my fix was to update to the correct version (as indicated by the README) - I accidentally updated Elasticsearch to 1.5.0 without updating elasticsearch-cloud-aws. @jacobwgillespie ha, thanks for pointing this out, I did the same thing! I just ran into this issue today, and it had nothing to do with mis-matched ES/Plugin versions. This error is also thrown if your config values for the cloud-aws plugin are not correct. What I had: cloud: aws: access_key: XXX secret: XXX region: us-east-1 What I needed: cloud: aws: access_key: XXX secret_key: XXX region: us-east-1 The major difference between the two of them being cloud.aws.secret in the first, non-working example; which is changed to cloud.aws.secret_key in the second, working example. So, apparently, this exact same error is what you'll get if you've completely borked your cloud-aws config. So, keep that in mind! Maybe in a future release, if you could detect a bad config state like this (access key is present, but secret key is missing) and throw an error, that would be pretty great. Alternatively, implementing it so that an error is thrown if any unknown string appears in the cloud.aws namespace would work just as well (so, an error would have been thrown because cloud.aws.secret is not a recognized and valid key), but I don't know how feasible that is to do. Just some food for thought! Hopefully someone finds this useful! @hjc1710 I think it's useful and I agree that we should better catch that kind of error. May be you would like to open an issue in elasticsearch repo now that we moved aws plugin there? Thanks! Awesome, thanks @dadoonet! I did not know that this plugin had moved and the official repo is elastic/elasticsearch now. Anyway, I opened up an issue there for this very feature, after doing a bit of rewording and thinking. If you guys need my help for that feature (for whatever reason), I'm happy to help!
2025-04-01T06:38:30.794539
2019-04-08T10:47:27
430381348
{ "authors": [ "sl0thentr0py" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5593", "repo": "elastic/elasticsearch-rails", "url": "https://github.com/elastic/elasticsearch-rails/issues/873" }
gharchive/issue
Adding search_analyzer to mapping I can't figure out how to add the search_analyzer option to the mapping block. The Mappings class only provides indexes as a way to add fields. https://www.elastic.co/guide/en/elasticsearch/reference/current/search-analyzer.html My bad, documentation parsing error from my side.
2025-04-01T06:38:30.978241
2022-12-22T15:49:57
1508117402
{ "authors": [ "Drookoo", "SashaShcherbyna", "TanguyPatte", "adrifermo", "ppatcha" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5594", "repo": "elastic/helm-charts", "url": "https://github.com/elastic/helm-charts/issues/1748" }
gharchive/issue
elasticsearch.keystore: Device or resource busy Chart version: 8.5.1 Kubernetes version: 1.23.12-gke.1600 Kubernetes provider: GKE (Google Kubernetes Engine) Helm Version: version.BuildInfo{Version:"v3.9.3", GitCommit:"414ff28d4029ae8c8b05d62aa06c7fe3dee2bc58", GitTreeState:"clean", GoVersion:"go1.19"} Describe the bug: Exception in thread "main" java.nio.file.FileSystemException: /usr/share/elasticsearch/config/elasticsearch.keystore.tmp -> /usr/share/elasticsearch/config/elasticsearch.keystore: Device or resource busy Steps to reproduce: Create the secret kubectl create secret generic elk-backup --from-file=gcs.client.elk-backup.credentials_file=./elk-backup.json To add these secrets to the keystore: keystore: - secretName: elk-backup Expected behavior: mount the secret in keystore should work Provide logs and/or server output (if relevant): Be careful to obfuscate every secrets (credentials, token, public IP, ...) that could be visible in the output before copy-pasting Exception in thread "main" java.nio.file.FileSystemException: /usr/share/elasticsearch/config/elasticsearch.keystore.tmp -> /usr/share/elasticsearch/config/elasticsearch.keystore: Device or resource busy at java.base/sun.nio.fs.UnixException.translateToIOException(UnixException.java:100) at java.base/sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:106) at java.base/sun.nio.fs.UnixCopyFile.move(UnixCopyFile.java:420) at java.base/sun.nio.fs.UnixFileSystemProvider.move(UnixFileSystemProvider.java:266) at java.base/java.nio.file.Files.move(Files.java:1430) at org.elasticsearch.common.settings.KeyStoreWrapper.save(KeyStoreWrapper.java:498) at org.elasticsearch.common.settings.KeyStoreWrapper.save(KeyStoreWrapper.java:412) at org.elasticsearch.cli.keystore.AddStringKeyStoreCommand.executeCommand(AddStringKeyStoreCommand.java:102) at org.elasticsearch.cli.keystore.BaseKeyStoreCommand.execute(BaseKeyStoreCommand.java:64) at org.elasticsearch.common.cli.EnvironmentAwareCommand.execute(EnvironmentAwareCommand.java:54) at org.elasticsearch.cli.Command.mainWithoutErrorHandling(Command.java:85) at org.elasticsearch.cli.MultiCommand.execute(MultiCommand.java:94) at org.elasticsearch.cli.Command.mainWithoutErrorHandling(Command.java:85) at org.elasticsearch.cli.Command.main(Command.java:50) at org.elasticsearch.launcher.CliToolLauncher.main(CliToolLauncher.java:64) Any additional context: I can seeing this error in pod with role master I'm not sure but maybe it's because the keystore file is mounted as a subpath in the chart - name: keystore mountPath: /usr/share/elasticsearch/config/elasticsearch.keystore subPath: elasticsearch.keystore I don't know how elasticsearch-keystore tool works but it looks like it's trying to replace elasticsearch.keystore file by elasticsearch.keystore.tmp file instead of replacing the content of it. I reproduce the issue by trying to do the following command: mv elasticsearch.keystore.tmp elasticsearch.keystore mv: cannot move 'elasticsearch.keystore.tmp' to 'elasticsearch.keystore': Device or resource busy cp command works just fine cp elasticsearch.keystore.tmp elasticsearch.keystore I see 2 possibilities to fix that: change the way elasticsearch-keystore write content in keystore file add the possibility to specify a custom path for the keystore file and mount it in is own directory to prevent the use of subpath Hi @TanguyPatte seems like there is something wrong with the current statefulset.yaml that causes this issue. FWIW deploying with this template works well for elasticsearch 8.6.2 : https://github.com/elastic/helm-charts/blob/d4e9f6bc47cf7f7ad4dfaaec102e1327d8a345e3/elasticsearch/templates/statefulset.yaml but there may be more recent iterations that may work. Chart version: 8.5.1 Kubernetes version: 1.23.12-gke.1600 Kubernetes provider: GKE (Google Kubernetes Engine) Helm Version: version.BuildInfo{Version:"v3.9.3", GitCommit:"414ff28d4029ae8c8b05d62aa06c7fe3dee2bc58", GitTreeState:"clean", GoVersion:"go1.19"} Describe the bug: Exception in thread "main" java.nio.file.FileSystemException: /usr/share/elasticsearch/config/elasticsearch.keystore.tmp -> /usr/share/elasticsearch/config/elasticsearch.keystore: Device or resource busy Steps to reproduce: Create the secret kubectl create secret generic elk-backup --from-file=gcs.client.elk-backup.credentials_file=./elk-backup.json To add these secrets to the keystore: keystore: - secretName: elk-backup Expected behavior: mount the secret in keystore should work Provide logs and/or server output (if relevant): Be careful to obfuscate every secrets (credentials, token, public IP, ...) that could be visible in the output before copy-pasting Exception in thread "main" java.nio.file.FileSystemException: /usr/share/elasticsearch/config/elasticsearch.keystore.tmp -> /usr/share/elasticsearch/config/elasticsearch.keystore: Device or resource busy at java.base/sun.nio.fs.UnixException.translateToIOException(UnixException.java:100) at java.base/sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:106) at java.base/sun.nio.fs.UnixCopyFile.move(UnixCopyFile.java:420) at java.base/sun.nio.fs.UnixFileSystemProvider.move(UnixFileSystemProvider.java:266) at java.base/java.nio.file.Files.move(Files.java:1430) at org.elasticsearch.common.settings.KeyStoreWrapper.save(KeyStoreWrapper.java:498) at org.elasticsearch.common.settings.KeyStoreWrapper.save(KeyStoreWrapper.java:412) at org.elasticsearch.cli.keystore.AddStringKeyStoreCommand.executeCommand(AddStringKeyStoreCommand.java:102) at org.elasticsearch.cli.keystore.BaseKeyStoreCommand.execute(BaseKeyStoreCommand.java:64) at org.elasticsearch.common.cli.EnvironmentAwareCommand.execute(EnvironmentAwareCommand.java:54) at org.elasticsearch.cli.Command.mainWithoutErrorHandling(Command.java:85) at org.elasticsearch.cli.MultiCommand.execute(MultiCommand.java:94) at org.elasticsearch.cli.Command.mainWithoutErrorHandling(Command.java:85) at org.elasticsearch.cli.Command.main(Command.java:50) at org.elasticsearch.launcher.CliToolLauncher.main(CliToolLauncher.java:64) Any additional context: I can seeing this error in pod with role master @SashaShcherbyna I have got the same error, Did you have a solution to resolve this issue? I can resolve this issue because I have not set ELASTIC_PASSWORD with my own credential. Just add extraEnvs: - name: ELASTIC_PASSWORD valueFrom: secretKeyRef: name: elastic-config-credentials key: password secret: enabled: false and then add the keystore just normal.
2025-04-01T06:38:32.685694
2020-10-29T16:42:53
732479122
{ "authors": [ "apmmachine", "vigneshshanmugam" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5595", "repo": "elastic/synthetics", "url": "https://github.com/elastic/synthetics/pull/112" }
gharchive/pull-request
fix: close fd after debug logs are written fix #110 runner end should be called after reset since we are logging the output inside reset Moved from nextTick to setTimeout as its called after all the micro tasks and ensure all log statements are printed before we end stream. :green_heart: Build Succeeded the below badges are clickable and redirect to their specific view in the CI or DOCS Expand to view the summary Build stats Build Cause: [Pull request #112 opened] Start Time: 2020-10-29T16:43:01.167+0000 Duration: 14 min 22 sec Test stats :test_tube: Test Results Failed 0 Passed 42 Skipped 0 Total 42
2025-04-01T06:38:32.690068
2021-01-27T13:01:00
795081257
{ "authors": [ "CLAassistant", "Ruminateer", "tizhou86" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5596", "repo": "elasticdeeplearning/edl", "url": "https://github.com/elasticdeeplearning/edl/pull/155" }
gharchive/pull-request
Fix 2 errors fix an import typo add a missing exception definition Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it. Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it. @gongweibao please review this pr, thanks~ @gongweibao please review this pr, thanks~
2025-04-01T06:38:32.703451
2020-06-21T14:04:33
642569562
{ "authors": [ "Bigsheng97", "iswyq" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5597", "repo": "elbuco1/CBAM", "url": "https://github.com/elbuco1/CBAM/issues/1" }
gharchive/issue
How to cite this work in my assignment? Thank you for your work and may I know how should I cite your works in my project? 这里已经写好了构建的
2025-04-01T06:38:32.705729
2024-01-12T23:19:43
2079872843
{ "authors": [ "coveralls", "elchininet" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5598", "repo": "elchininet/keep-texts-in-tabs", "url": "https://github.com/elchininet/keep-texts-in-tabs/pull/39" }
gharchive/pull-request
Set the docker image tag to run This pull request makes possible to run a specific Home Assistant docker image tag sending a variable to the script. If the variable is not sent, the version included in .hass/config/.HA_VERSION file will be used as the tag. Also, to generate the docker cache key, the aforementioned file will be used instead of the package.json. coverage: 100.0%. remained the same when pulling b723a4b7d1ede517f74d010d5fef0427c01cad17 on run_docker_image_tag_dynamic into ebd6f452dca960044803e08becec7c7e2efbfd39 on master.
2025-04-01T06:38:32.732620
2019-12-16T01:34:42
538141121
{ "authors": [ "aleccool213", "electerious" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5601", "repo": "electerious/Ackee", "url": "https://github.com/electerious/Ackee/issues/72" }
gharchive/issue
[Feature Request] Add Deploy to Heroku button to README.md Feature Request Hi there! Great application! I really love how simple and privacy focused it is. It would be great to have a Deploy to Heroku button, would help drive up adoption! Documentation A "Deploy to Heroku" button would be cool, but I guess that #73 should be finished first. Help is welcome! Will be part of v1.4.3.
2025-04-01T06:38:32.734729
2016-09-23T11:57:07
178858243
{ "authors": [ "electerious", "jullit31", "mhellwig" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5602", "repo": "electerious/Lychee", "url": "https://github.com/electerious/Lychee/issues/625" }
gharchive/issue
Feature request: make the auto-fullscreen behaviour configurable In PR #583, a new behaviour was introduced that would cause the picture to switch to fullscreen if the viewer didn't move his mouse for >=1 second. In my case, I need people (elderly relatives) to always see the top bar with the title. I have for now achieved that for myself by commenting out the relevant part from that PR in view.js, but it would be nicer if there was a config switch in the DB to turn the auto-fullscreen off. Notably I wouldn't have to redo this after every update and rebuild lychee. +1 for that, also being able to adjust that 1 second delay would be nice. A setting isn't planned, but we could increase the delay. I agree that 1 second might be too fast for some users. The next version will use a 2.5s delay. 1 second was too aggressive.
2025-04-01T06:38:32.739006
2024-01-22T21:13:24
2094762315
{ "authors": [ "KyleAMathews", "msfstef" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5603", "repo": "electric-sql/electric", "url": "https://github.com/electric-sql/electric/issues/868" }
gharchive/issue
Listener warning I have a view with a lot of small components that all have a small query listening for updates to their specific object — which should be a supported use case I'd assume? Or if there is an actual limit you want to warn against, set it to that. But I'm assuming the eventemitter is just set to the default (11 I think). Looking at the notifier code under src/notifiers/event.ts, we have the following: // Global singleton that all event notifiers use by default. Emitting an event // on this object will notify all subscribers in the same thread. Cross thread // notifications use the `./bridge` notifiers. const globalEmitter = new EventEmitter() // Increase the maximum number of listeners because multiple components // use this same emitter instance. globalEmitter.setMaxListeners(250) This limit was increased to 250 by @thruflo int https://github.com/electric-sql/electric/pull/377 - since this is a global emitter I find it hard to think of any reasonable way to adjust this limit for a large application while still retaining the ability to detect if listeners are being accumulated or not appropriately removed. Maybe removing this warning and checking if many listeners are being added in the same "place", or having them be identified somehow could help construct a better warning. For your particular case, would a "parent" component with view-only children work as well or would it complicate the code too much? Yeah I'm sure I could refactor — I think ideally though the subscriptions are efficient enough it's not necessary. I haven't noticed any slowdowns with the app & memory is fine — though all these listeners are to objects that are rarely changed. So what exactly is the upper limit is pretty arbitrary depending on what's updating in your app. @samwillis what's your take on this? I think it makes sense to remove the warning altogether as we explicitly use a global emitter - if we want to catch leaks we can implement our own tests or mechanisms from within our EventNotifier to catch them.
2025-04-01T06:38:32.748408
2024-08-15T10:23:54
2467801095
{ "authors": [ "KiaClouth", "TheAndrewJackson", "msfstef", "samwillis" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5604", "repo": "electric-sql/pglite", "url": "https://github.com/electric-sql/pglite/issues/208" }
gharchive/issue
PGliteWorker not working with drizzle OS: windows 11 Browser: chrome 127.0.6533.119 framework: solid-start The minimal reproduction repository: GitHub Repo Link description: I am using PGliteWorker with drizzle. I followed the instructions on https://pglite.dev/docs/multi-tab-worker to create a worker wrapper: PGliteWorker.ts worker({ async init() { const pg = await PGlite.create({...}); await pg.exec(...); const db = drizzle(pg, { schema }); const ret = await db.query.verification_token.findMany(); console.log(ret); return pg; }, }); I can see the results in the browser console. However, if I entry-client.tsx const pg = await PGliteWorker.create( new Worker(PGliteWorkerUrl, { type: "module", }) ); const db = drizzle(pg, { schema }); const ret = await db.query.verification_token.findMany(); console.log(ret); The browser console gives the following error: Uncaught DOMException: Failed to execute 'postMessage' on 'BroadcastChannel': value => value could not be cloned. at j.g (http://localhost:3001/_build/node_modules/.pnpm/@electric-sql+pglite@0.2.0/node_modules/@electric-sql/pglite/dist/worker/index.js?v=c34c2280:201:21) at j.query (http://localhost:3001/_build/node_modules/.pnpm/@electric-sql+pglite@0.2.0/node_modules/@electric-sql/pglite/dist/worker/index.js?v=c34c2280:98:56) at async PglitePreparedQuery.execute (http://localhost:3001/_build/node_modules/.vinxi/client/deps/drizzle-orm_pglite.js?v=d787a4e7:58:20) at async http://localhost:3001/_build/@fs/D:/ToramCalculator/src/entry-client.tsx:16:13 I've experienced this issue as well when trying to write to the DB with drizzle. My read queries were working. I wonder if this issue is an issue with drizzle The Failed to execute 'postMessage' on 'BroadcastChannel' error makes me think it could be our side as we use a BroadcastChannel to communicate with the worker. It's either: drizzle is trying to pass an un-cloneable object as a parameter to the query api. or we have something in our api that doesn't work with the worker and is not currently covered by the tests. or both, which is my suspicion: I'm 99% sure we will find it's the parser query option (https://pglite.dev/docs/api#query-options) and will need a bit of a refactor. The issue seems to be that we are trying to transfer ParserOptions which contains anonymous functions to the worker without proper serialization. Either the parsing is done at the calling thread (and the ParserOptions are stored and handled there) or we assume some limitations on the functions being passed and serialize them and deserialize them to run in the worker (e.g. with a toString()/eval() combination). Since these parsers might depend on various imports and what not it might be better to handle deserialization in the caller.
2025-04-01T06:38:32.775770
2024-11-12T08:07:55
2651383360
{ "authors": [ "darylgraham", "emillg", "tollercode" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5605", "repo": "electrolux-oss/infrawallet", "url": "https://github.com/electrolux-oss/infrawallet/issues/124" }
gharchive/issue
[BUG] Cost data is wrongly mapped into Budget graphs Current Situation Having this cost data in November The data is shown in the Budget Graphs in Januaray Desired Data of November should be shown in November also on Budget-Chart I just debugged this screen myself. The cost data is passed into the BudgetChart component as an array of monthly costs here: https://github.com/electrolux-oss/infrawallet/blob/main/plugins/infrawallet/src/components/Budgets/Budgets.tsx#L250 "id": "AWS", "reports": { "2023-12": 19.6492511966, "2024-01": 1939.0336325847002, "2024-02": 1610.1251975480995, "2024-03": 2149.6599309432, "2024-04": 2629.0959797089004, "2024-05": 2603.6509832689, "2024-06": 2438.023622028399, "2024-07": 2767.510390127999, "2024-08": 9312.452064800302, "2024-09": 8760.504147668702, "2024-10": 10055.072295576203, "2024-11": 5529.263631147701 } } It's then converted to a running sum here: https://github.com/electrolux-oss/infrawallet/blob/main/plugins/infrawallet/src/components/Budgets/Budgets.tsx#L91-L98 { 0: 19.6492511966, 1: 1958.6828837813002, 2: 3568.8080813294, 3: 5718.4680122726, 4: 8347.5639919815, 5: 10951.214975250401, 6: 13389.2385972788, 7: 16156.748987406798, 8: 25469.2010522071, 9: 34229.7051998758, 10: 44284.77749545201, 11: 49814.04112659971 } This is then plotted on the chart, with a constant set of x-axis titles: https://github.com/electrolux-oss/infrawallet/blob/main/plugins/infrawallet/src/components/Budgets/Budgets.tsx#L122 xAxis={[ { data: ['Jan', 'Feb', 'Mar', 'Apr', 'May', 'Jun', 'Jul', 'Aug', 'Sep', 'Oct', 'Nov', 'Dec'], scaleType: 'band', }, ]} I think that the x-axis labels just need to be dynamically generated by mapping the keys from the original report variable to the month names. @tollercode @darylgraham Can you check if version 0.2.0-20241118212048-4e21a4a works for you?
2025-04-01T06:38:32.838579
2017-05-19T16:05:20
230019792
{ "authors": [ "aliir74", "firsttracks", "ryan-christopher", "shakhassan", "zeke" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5606", "repo": "electron-userland/electron-prebuilt", "url": "https://github.com/electron-userland/electron-prebuilt/issues/254" }
gharchive/issue
Mac: Failed at the<EMAIL_ADDRESS>postinstall script 'node install.js'. Received the following message: Failed at the<EMAIL_ADDRESS>postinstall script 'node install.js'. (Exit Status 1) node.js (and I assume npm with it) were just downloaded and installed from their site. Electron version<EMAIL_ADDRESS>Operating system: macOS Sierra 10.12.4 I'm brand new to working with these types of files and installers, so I'm not sure how to correct this. Also, I need to know if I can do this as a global install, as I would not like to confront the prospect of installing this in every single project folder. In my case, npm install electron -g will do the trick. I am seeing the same error. Is there a work around for this issue? I got around this issue by looking at the link "https://docs.npmjs.com/getting-started/installing-npm-packages-locally". I tried a bunch of things and not sure exactly what fixed it but I did run "npm init" with defaults and then the "npm install mapbox-map-image-export -g" command seemed to work I've got the same error. <EMAIL_ADDRESS>postinstall: node install.js. Exit status 1 Node.js version: v8.1.4 Electron version: v1.6.11 OS: Ubuntu 16.04 @ryan-christopher I updated my node to last version and the error fixed. The electron-prebuilt repo is being retired and its code has been moved into the electron/electron repo. For the sake of historical transparency, we will leave GitHub Issues enabled on this repository, but if you are still affected by the issue reported here, please open a new issue on electron/electron repo and reference this issue from it so people can get the full context. The electron repository has a large and active contributor community, so your issue is more likely to get the attention it deserves there. Thanks!
2025-04-01T06:38:32.950412
2022-09-12T15:31:21
1370104841
{ "authors": [ "batis97", "gpronet", "lukas2", "technotip" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5607", "repo": "electron/osx-sign", "url": "https://github.com/electron/osx-sign/issues/266" }
gharchive/issue
"Unidentified developer" when opening universal app Hello, we switched from an app for x86_64 to an universal app. The app runs fine on all platforms when I sign them locally with my developer identity. However, when we go through Testflight, we are greeted with a "... cannot be opened because the developer cannot be verified" message. This only happens with the universal build, the x86_64 only one works fine. And if I force-resign the app, then it also works fine. This is how we sign: electron-osx-sign --entitlements=./entitlements/entitlements.mac.plist --entitlements-inherit=./entitlements/entitlements.mac.inherit.plist --entitlements-loginhelper=./entitlements/entitlements.mac.inherit.plist --identity=${CERTIFICATE_DEVELOPER_SHA} --keychain=signing-cert-keychain --provisioning-profile=./embedded.provisionprofile --type=distribution "${appName}" productbuild --component "${appName}" /Applications --sign "${CERTIFICATE_INSTALLER_SHA}" "app-name-${params.UPLOAD_ARTIFACT_VERSION}.pkg" Checking the app with the usual tools (codesign,pkgutil, spctl) gives no clue to any problems. Only at runtime it doesn't work. I have run out of ideas to try.. I have exactly the same problem, already spent 5 days trying different things but I've exhausted my options. Does force resigning the universal app actually work for you and you can run it in Test Flight? I couldn't even manage to do that yet. I meant that resigning locally with my developer identity makes it runnable for me, but I have not tried uploading the resigned app onto Testflight. Did you test if the same problem occurs when you attempt an app store release? Speculating if this is a problem with Testflight.. I meant that resigning locally with my developer identity makes it runnable for me, but I have not tried uploading the resigned app onto Testflight. I see, I also have no problems running the app locally when it is signed with the developer identity and notarized. However for TestFlight and the Mac Store it has to be signed either with 3rd Party Mac Developer Application or Apple Distribution certificate. And when we do this we get the above problem. Did you test if the same problem occurs when you attempt an app store release? Speculating if this is a problem with Testflight.. Yes, the app gets rejected for the same reason. I created a separate issue with more info since I am using electron-builder (not osx-sign directly) here: #https://github.com/electron-userland/electron-builder/issues/7171 I've attached a build log there which confirms signing goes well, so the problem has to be elsewhere. Have you tried regenerating the provisioning profile? Will try that next (unfortunately I don't have direct access to App Store so it's a slow process here).. after all the provisioning profile links the developer with the app.. Have you tried regenerating the provisioning profile? Yes, no difference. Transporter usually complains if your provisioning profile is not correct, so you will never reach TestFlight deployment if it was wrong. But you are welcome to try it in case I missed something. just fyi, I had another wild theory.. maybe the universal build is loading different dylibs on startup than the x86 build. That could lead it to not find something. If it doesn't find it, the loader goes to search for it. And while searching it checks some invalid directories, which are forbidden unless some entitlements are given, and therefore the signature is rejected. So i checked and the only diffrence between a working x86 app (left) and the universal app (right) was this: I think I found the reason. My app uses native modules and they need to be signed, too. For that reason, they are unpacked, using the asarUnpack config option. However, unlike with the x86 build, in the universal build I have two copies of the same node module. One in the app.asar.unpacked folder, and one in the app.asar archive itself. And only the "unpacked" one is signed. 😨 Ok I have my solution, if I set mergeASARs to false, I get a correctly signed and launching app. Albeit, it still seems that the native modules are duplicated. I would close this issue now, or leave it to you @gpronet , as I think my solution won't help you Amazing, mergeASARs: false fixes the issue for me as well. A non critical dependency of my app was using just one native module, so when I removed it, the error messages were gone as well even without mergeASARs: false. Thank you for pointing me in the right direction @lukas2. Wow, happy to hear! 🙂 Actually it clicked for me when you said in the other thread that the "Move to Bin"-deleted files are hidden in Trash. I examined mine and saw my module in unsigned form. Then I knew there had to be more than one.. :) Ok closing this. I'm still stuck with this error :-( Can you guys please post your package.json and entitlement files? Hello i still have this problem mergeASARs: false I've set it up > mergeASARs: false and the eror still show in testflight i using electron + svelkit this my electron and builder version: "electron": "^26.2.2", "electron-builder": "^24.6.4",
2025-04-01T06:38:32.982886
2022-07-14T08:48:40
1304466980
{ "authors": [ "IDoneShaveIt", "elongl" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5608", "repo": "elementary-data/elementary", "url": "https://github.com/elementary-data/elementary/pull/121" }
gharchive/pull-request
Hotfix/support backfill changes on redshift support backfill changes for alerts model on redshift LGTM! Could you please change to branch to v<IP_ADDRESS>?
2025-04-01T06:38:33.130203
2015-02-28T02:12:47
59320138
{ "authors": [ "dwightwatson", "elfet" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5609", "repo": "elfet/cherimola", "url": "https://github.com/elfet/cherimola/pull/6" }
gharchive/pull-request
Remove version constraint from the installation docs If one was to copy and paste the Composer installation command from the README when it has the ~1 version constraint they might be confused as to why they don't get all the functionality shown in the docs. Removing the version constraint will have Composer install the latest tagged version and thus they should get all the current functionality. Cool!
2025-04-01T06:38:33.143500
2019-08-07T08:40:06
477787265
{ "authors": [ "martinstz" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5612", "repo": "elgatito/plugin.video.elementum", "url": "https://github.com/elgatito/plugin.video.elementum/issues/464" }
gharchive/issue
Problem with searching Hi! I have some problems with zamunda.net. For some reason when Elementum successful login to zamunda and i search for some movie, Elementum find results with right search string: http://zamunda.net/bananas?c42=1&c25=1&c35=1&c46=1&c20=1&c19=1&c5=1&c24=1&c31=1&c28=1&search=avengers+endgame+2019&incldead=1&field=name but after that Elementum create another request with other string: https://zamunda.net:443/login.php?returnto=%2Fbananas%3Fc42%3D1%26c25%3D1%26c35%3D1%26c46%3D1%26c20%3D1%26c19%3D1%26c5%3D1%26c24%3D1%26c31%3D1%26c28%3D1%26search%3Davengers%2Bendgame%2B2019%26incldead%3D1%26field%3Dname and this string loged me out from Zamunda and finaly Elementum cannot display any results. Can you help with this? If you need user and pass for Zamunda, tell me where to send them. Thank you in advance! I fix the problem.
2025-04-01T06:38:33.156393
2024-09-22T12:28:25
2541055842
{ "authors": [ "antonsoroko", "elgatito", "olegmiercoles" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5613", "repo": "elgatito/script.elementum.burst", "url": "https://github.com/elgatito/script.elementum.burst/issues/438" }
gharchive/issue
400 Login failed on [toloka] provider When try to execute any search on toloka provider it returns no links. From log debug I can see errors of 400 bad requests and information that Request Header or Cookie Too Large. Not sure is it something about client, configuration or the provider itself. I've verified that the login/password are correct. I've also get try to lgin from postman using this form-data from config "{'username': USERNAME, 'password': PASSWORD, 'autologin': '1', 'login': 'Enter'}" and it seems to be working, at least it returns me redirect instead of bad request. Here is the part of log debug: 2024-09-22 12:49:48.502 T:4449 debug <general>: [script.elementum.burst] Searching with payload (general): {'proxy_url': '', 'internal_proxy_url': 'http://<IP_ADDRESS>:65222', 'elementum_url': 'http://<IP_ADDRESS>:65220', 'silent': False, 'skip_auth': False, 'query': 'FOO'} 2024-09-22 12:49:48.505 T:4449 warning <general>: [script.elementum.burst] Burstin' with Гуртом 2024-09-22 12:49:48.506 T:4449 warning <general>: [script.elementum.burst] No 'en' translation available... 2024-09-22 12:49:48.511 T:3816 debug <general>: ------ Window Init (DialogExtendedProgressBar.xml) ------ 2024-09-22 12:49:48.512 T:4449 debug <general>: [script.elementum.burst] Translated titles from Elementum: {'source': 'FOO', 'original': 'FOO'} 2024-09-22 12:49:48.513 T:4468 debug <general>: [script.elementum.burst] [toloka] Processing toloka with general method 2024-09-22 12:49:48.514 T:4468 debug <CSettingsManager>: requested setting (filter_music) was not found. 2024-09-22 12:49:48.515 T:4468 debug <general>: [script.elementum.burst] [toloka] General URL: https://toloka.to/tracker.php?nm=QUERYEXTRA&o=10 2024-09-22 12:49:48.517 T:4468 debug <general>: [script.elementum.burst] [toloka] execute_process for toloka with <function extract_torrents at 0x7e114938> 2024-09-22 12:49:48.518 T:4449 debug <general>: [script.elementum.burst] Timer: 0s / 27s 2024-09-22 12:49:48.527 T:4468 debug<EMAIL_ADDRESS>trying to load setting definitions from old format... 2024-09-22 12:49:48.549 T:4468 debug <general>: [script.elementum.burst] [toloka] Queries: ['{title}'] 2024-09-22 12:49:48.550 T:4468 debug <general>: [script.elementum.burst] [toloka] Extras: [''] 2024-09-22 12:49:48.551 T:4468 debug <general>: [script.elementum.burst] [toloka] Before keywords - Query: '{title}' - Extra: '' 2024-09-22 12:49:48.552 T:4468 warning <general>: [script.elementum.burst] [toloka] Falling back to original title in absence of None language title 2024-09-22 12:49:48.557 T:4468 warning <general>: [script.elementum.burst] [toloka] Using translated 'original' title 'FOO' 2024-09-22 12:49:48.559 T:4468 debug <general>: [script.elementum.burst] [toloka] After keywords - Query: '%D1%80%D0%B0%D1%82%D0%B0%D1%82%D1%83%D0%B9' - Extra: '' 2024-09-22 12:49:48.560 T:4468 debug <general>: [script.elementum.burst] - toloka query: '%D1%80%D0%B0%D1%82%D0%B0%D1%82%D1%83%D0%B9' 2024-09-22 12:49:48.561 T:4468 debug <general>: [script.elementum.burst] -- toloka url_search before token: 'https://toloka.to/tracker.php?nm=FOO&o=10' 2024-09-22 12:49:48.561 T:4468 debug <general>: [script.elementum.burst] --- toloka using POST payload: {} 2024-09-22 12:49:48.562 T:4468 debug <general>: [script.elementum.burst] ----toloka filtering with post_data: {} 2024-09-22 12:49:48.562 T:4468 debug <CSettingsManager>: requested setting (toloka_passkey) was not found. 2024-09-22 12:49:48.577 T:4468 debug <general>: [script.elementum.burst] Opening URL: b'https://toloka.to/login.php' 2024-09-22 12:49:48.770 T:4449 debug <general>: [script.elementum.burst] Timer: 0s / 27s 2024-09-22 12:49:49.029 T:4468 info <general>: Skipped 1 duplicate messages.. 2024-09-22 12:49:49.029 T:4468 debug <general>: [script.elementum.burst] Status for b'https://toloka.to/login.php' : 400 2024-09-22 12:49:49.031 T:4468 critical <general>: [script.elementum.burst] [toloka] Login failed: 400 2024-09-22 12:49:49.034 T:4468 debug <general>: [script.elementum.burst] [toloka] Failed login content: '<html>\r\n<head><title>400 Request Header Or Cookie Too Large</title></head>\r\n<body bgcolor="white">\r\n<center><h1>400 Bad Request</h1></center>\r\n<center>Request Header Or Cookie Too Large</center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>\r\n<!-- a padding to disable MSIE and Chrome friendly error page -->\r\n<!-- a padding to disable MSIE and Chrome friendly error page -->\r\n<!-- a padding to disable MSIE and Chrome friendly error page -->\r\n<!-- a padding to disable MSIE and Chrome friendly error page -->\r\n<!-- a padding to disable MSIE and Chrome friendly error page -->\r\n<!-- a padding to disable MSIE and Chrome friendly error page -->\r\n' 2024-09-22 12:49:49.047 T:4468 warning <general>: [script.elementum.burst] [toloka] >> Гуртом returned 0 results in 0.5 seconds System: Android TV Kodi version: 21.1 Elementum version: 0.1.103 Elementum burst version: 0.0.89 Any suggestions will be appreciate. @olegmiercoles Maybe it is because of long cookie. There is a file with cookies in .kodi/temp/burst/common_cookies.jar (see https://kodi.wiki/view/Kodi_data_folder for location of .kodi in your OS) It is a plain text file, you can open it and search for toloka and see if cookie is "bad" (maybe it has some garbage and it is indeed too long). Anyway - remove the line for toloka and try again. also, you can backup that file and remove the file completely. Also, in burst settings in "maintenance" tab you can remove all cookies - in case if you unable to get access to file system of your device. There is no "login_headers": field for toloka so it should not be "Request Header" issue. https://github.com/elgatito/script.elementum.burst/blob/97dcfb60aa43fa646712c6ec60fb9fc7ee80ecc0/burst/providers/providers.json#L2381 Thank you, @antonsoroko I did as you suggested: I've renamed common_cookies.jar to common_cookies_bkp.jar and it helped. Probably I will keep backup-file for some time, but later will remove it. Thank you again for assistance! @olegmiercoles I am just curious - have you tried to take a look into that file? Interesting how a line for toloka looks like. If you remove toloka_sid from that line - you can share it here (so it will not have your login cookie). Up to you of course. Sure. I did my the best to cleanup all the hashes and IDs. Hope, I didn't miss anything :) First line for toloka_data was pretty huge combo of hash+autologinid+hash+userid+hash And toloka_302_u also contained around 8.000 characters hash inside. #LWP-Cookies-2.0 Set-Cookie3: toloka_data="something-was-here-like-b%432%3b%9Bs%3d%autologinid%here-as-well-something-likeb%432%3b%9Bs%3d%userid-and-here-the-same"; path="/"; domain="toloka.to"; path_spec; secure; expires="2025-05-29 18:56:57Z"; httponly=None; version=0 Set-Cookie3: toloka_302_tt=there-was-some-numbers; path="/"; domain="toloka.to"; path_spec; secure; expires="2025-05-29 18:56:57Z"; version=0 Set-Cookie3: toloka_302_f="there-was-something-like-in-the-first-line-a%54b%33%"; path="/"; domain="toloka.to"; path_spec; secure; expires="2025-05-29 18:56:57Z"; version=0 Set-Cookie3: toloka_302_uf=there-was-some-numbers; path="/"; domain="toloka.to"; path_spec; secure; expires="2025-05-29 18:56:57Z"; version=0 Set-Cookie3: toloka_302_u="about-8.000-characters-was-here"; path="/"; domain="toloka.to"; path_spec; secure; expires="2025-05-29 18:56:57Z"; version=0 Set-Cookie3: toloka___tt=there-was-some-numbers; path="/"; domain="toloka.to"; path_spec; secure; expires="2025-03-04 19:58:37Z"; version=0 Set-Cookie3: toloka___f="there-was-something-like-in-the-first-line-a%54b%33%"; path="/"; domain="toloka.to"; path_spec; secure; expires="2025-03-04 19:58:37Z"; version=0 Set-Cookie3: toloka___uf=0; path="/"; domain="toloka.to"; path_spec; secure; expires="2025-03-04 19:58:37Z"; version=0 Set-Cookie3: toloka___u="there-was-something-like-in-the-first-line-a%54b%33%"; path="/"; domain="toloka.to"; path_spec; secure; expires="2025-03-04 19:58:37Z"; version=0 @olegmiercoles thanks! so looks like by default many popular HTTP servers have total limit for HTTP Header (and cookie is a part of header) as 8KB. so looks like in total your cookies were >8KB, thus i think we found the root cause. somehow toloka website generated such long cookie. i guess we could create a "blacklist" for cookies (e.g. ignore and do not save some cookies), although i am not sure if we really need it, since i see such issue for the first time. maybe @elgatito can add more ideas. but anyway, thanks for info. if there will be more issues like this then we can circle back to this. @antonsoroko Quick googling says there is nothing "ready" in Python's requests to control size of a request to avoid such errors. I also have problems with Toloka, even with sync, it does something (I was not debugging those), that invalidates a session everywhere, not only on Elementum/Burst side. Not sure if we should/can do something with it.
2025-04-01T06:38:33.160030
2018-04-11T23:19:27
313520137
{ "authors": [ "eliangcs", "wheelerlaw" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5614", "repo": "eliangcs/http-prompt", "url": "https://github.com/eliangcs/http-prompt/issues/142" }
gharchive/issue
Missing HTTP CONNECT method Pretty straight forward, can't seem to do an HTTP CONNECT request. For example: $ http-prompt http://example.com Version: 0.11.2 http://example.com> httpie connect http http://example.com/connect should actually be http-prompt http://example.com Version: 0.11.2 http://example.com> httpie connect http CONNECT http://example.com Addressed in #145, shipped in v1.0.0.
2025-04-01T06:38:33.166383
2016-03-18T01:20:23
141747801
{ "authors": [ "garfieldnate", "wowselim" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5615", "repo": "elibom/jogger", "url": "https://github.com/elibom/jogger/issues/19" }
gharchive/issue
archetype generation command fails I copied the command from your Wiki to start a project from the Maven archetype: mvn archetype:generate -DarchetypeGroupId=org.jogger -DarchetypeArtifactId=jogger-archetype -DarchetypeVersion=0.9.0 -DarchetypeRepository=http://repository.elibom.net/nexus/content/repositories/releases/ This was the output: mvn archetype:generate -DarchetypeGroupId=org.jogger -DarchetypeArtifactId=jogger-archetype -DarchetypeVersion=0.9.0 -DarchetypeRepository=http://repository.elibom.net/nexus/content/repositories/releases/ [INFO] Scanning for projects... [INFO] [INFO] ------------------------------------------------------------------------ [INFO] Building Maven Stub Project (No POM) 1 [INFO] ------------------------------------------------------------------------ [INFO] [INFO] >>> maven-archetype-plugin:2.4:generate (default-cli) > generate-sources @ standalone-pom >>> [INFO] [INFO] <<< maven-archetype-plugin:2.4:generate (default-cli) < generate-sources @ standalone-pom <<< [INFO] [INFO] --- maven-archetype-plugin:2.4:generate (default-cli) @ standalone-pom -- - [INFO] Generating project in Interactive mode [INFO] Archetype defined by properties Downloading: http://repository.elibom.net/nexus/content/repositories/releases/or g/jogger/jogger-archetype/0.9.0/jogger-archetype-0.9.0.jar [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 1.974 s [INFO] Finished at: 2016-03-18T10:25:32+09:00 [INFO] Final Memory: 15M/245M [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal org.apache.maven.plugins:maven-archetype-plugin:2 .4:generate (default-cli) on project standalone-pom: The desired archetype does not exist (org.jogger:jogger-archetype:0.9.0) -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e swit ch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please rea d the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureExc eption Hi @garfieldnate , I think this project is pretty much dead. In case you are looking for alternatives I suggest you take a look at either Spark or Pippo. Jooby might also be worth a try eventhough I have no experience with it. That's too bad, this was very quick to set up. It's just that the author's repo site is down. Maybe they could be moved to Maven Central if they could be found. I'll leave this issue here for future reference.
2025-04-01T06:38:33.171915
2023-04-28T01:23:42
1687749741
{ "authors": [ "gnott" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5616", "repo": "elifesciences/docmap-tools", "url": "https://github.com/elifesciences/docmap-tools/pull/15" }
gharchive/pull-request
Preprint history event data parse, update fixture. Added new parsing functions to gather preprint version event history data. The existing test fixture 2<IP_ADDRESS>2253.docmap.json is updated to reflect the latest docmap data available for it. Re issue https://github.com/elifesciences/issues/issues/7721
2025-04-01T06:38:33.190307
2024-10-16T19:15:16
2592827835
{ "authors": [ "acollings", "fred-atherden" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5617", "repo": "elifesciences/publish-reviewed-preprints-issues", "url": "https://github.com/elifesciences/publish-reviewed-preprints-issues/issues/1423" }
gharchive/issue
MSID: 99785 Version: 1 DOI: 10.1101/2<IP_ADDRESS>3526 MSID: 99785 Version: 1 Preprint DOI: https://doi.org/10.1101/2<IP_ADDRESS>3526 Step 1. Awaiting reviews Editorial to post reviews via hypothesis Useful links: DocMap: https://data-hub-api.elifesciences.org/enhanced-preprints/docmaps/v2/by-publisher/elife/get-by-manuscript-id?manuscript_id=99785 New model tracking: https://docs.google.com/spreadsheets/d/1_fHaoOy7hjyocptKtVJRijeNpUY4hBS7Ck_aVmx6ZJk/ Reviews on sciety: https://sciety.org/articles/activity/10.1101/2<IP_ADDRESS>3526 For trouble shooting (e.g. no Docmaps available): DocMap issue addressing: https://miro.com/app/board/uXjVNCwK6EI=/ Explore DataHub DocMaps API: https://lookerstudio.google.com/reporting/4c2f0368-babb-4beb-b5b3-497e7e7b0f08/page/ejphD Unmatched submissions and preprints: https://lookerstudio.google.com/u/0/reporting/9f86204f-3bf7-477c-9b18-5c5ef141bf69/page/p_gxi57ha93c Unmatched manuscripts spreadsheet: https://docs.google.com/spreadsheets/d/15QcK8w-ssB7109RQEDtFpJPZ0J5HTGxoHa_2TtpMBbg/edit#gid=1336081641 Step 2. Preview reviewed preprint Production QC content ahead of publication Instructions: QC preview: https://prod--epp.elifesciences.org/previews/99785v1 Update ticket with any problems (add blocked label) When QC OK, add QC OK label to ticket and add publication date and time to https://docs.google.com/spreadsheets/d/1amAlKvdLcaDp5W8Z8g77NmkwbMF5n_u89ArSqPMO8jg Move card to next column (At end of the day post link in #enhanced-preprint and ask for PDF to be generated) Useful links: Preprint DOI : https://doi.org/10.1101/2<IP_ADDRESS>3526 Confirm reviews returned by EPP: https://prod--epp.elifesciences.org/api/reviewed-preprints/99785/v1/reviews To update the MECA path in the docmap: https://docs.google.com/spreadsheets/d/1mctCQuNFBjSn97Lihy7_vBO6z7-N-oqyLv4clyi6zHg Step 3: Awaiting search reindex This step adds the reviewed preprint to the homepage: https://elifesciences.org The search reindex is triggered once an hour. We need the reviewed preprint to be indexed as the search application serves the journal homepage. Useful links: Jenkins pipeline to reindex search can be triggered sooner or monitored here: https://alfred.elifesciences.org/job/process/job/process-reindex-reviewed-preprints/ Step 4: Published! PDF requested Waiting for PDF to be generated Useful links: PDF tracking: https://docs.google.com/spreadsheets/d/106_XeDjmuBae7gexOTNzg60lapeqjl2aRn9DzupGyS8/ Step 5: Introduce PDF to data folder and git repo Upload PDF to relevent folder in git repo https://github.com/elifesciences/enhanced-preprints-data/ Step 6: Done! [ ] Kettle is on! Hi @acollings / @FionaBryant, please could you take a look, and if necessary tweak the assessment for this one? The authors solidly connect proteostasis ... Thanks @fred-atherden this has been fixed Many thanks! WOS query sent Confirmed Ok to proceed RE WOS. Waiting for https://sciety.org/evaluations/hypothesis:nlvGcJarEe-W4aOBdYgobQ/content to update
2025-04-01T06:38:33.206775
2021-09-07T14:27:27
990046247
{ "authors": [ "Flova", "Ledjob", "WolfgangFahl", "docsteveharris", "elimintz", "giodegas", "nielstron", "poke1024", "samiit" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5619", "repo": "elimintz/justpy", "url": "https://github.com/elimintz/justpy/issues/301" }
gharchive/issue
demjeson install cannot be done since setuptools upgrade, so neither justpy can. Today on a new deploy I did : pip3 install justpy but ends up failing because: ERROR: Cannot install justpy==0.0.5, justpy==0.0.6, justpy==0.0.7, justpy==0.0.8, justpy==0.0.9, justpy==0.1.0, justpy==0.1.1, justpy==0.1.2, justpy==0.1.3, justpy==0.1.4 and justpy==0.1.5 because these package versions have conflicting dependencies. The conflict is caused by: justpy 0.1.5 depends on demjson>=2.2.4 justpy 0.1.4 depends on demjson>=2.2.4 justpy 0.1.3 depends on demjson>=2.2.4 justpy 0.1.2 depends on demjson>=2.2.4 justpy 0.1.1 depends on demjson>=2.2.4 justpy 0.1.0 depends on demjson>=2.2.4 justpy 0.0.9 depends on demjson>=2.2.4 justpy 0.0.8 depends on demjson>=2.2.4 justpy 0.0.7 depends on demjson>=2.2.4 justpy 0.0.6 depends on demjson>=2.2.4 justpy 0.0.5 depends on demjson>=2.2.4 To fix this you could try to: 1. loosen the range of package versions you've specified 2. remove package versions to allow pip attempt to solve the dependency conflict it is an open issue at demjson repo: https://github.com/dmeranda/demjson/issues/40 Downgrading to setuptools 57.5.0 it des not work, it installs but while doing: import justpy gives: import justpy Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/data/giodegas/dev/AQ2009/myenv/lib/python3.8/site-packages/justpy/__init__.py", line 1, in <module> from .justpy import * File "/data/giodegas/dev/AQ2009/myenv/lib/python3.8/site-packages/justpy/justpy.py", line 13, in <module> from .chartcomponents import * File "/data/giodegas/dev/AQ2009/myenv/lib/python3.8/site-packages/justpy/chartcomponents.py", line 2, in <module> import demjson File "/data/giodegas/dev/AQ2009/temp/demjson/demjson.py", line 645 class json_int( (1L).__class__ ): # Have to specify base this way to satisfy 2to3 ^ SyntaxError: invalid syntax @elimintz what can I do now? Thank you for your support. Thanks for alerting me to this. I'll try to figure something out. In the meantime, it seems that if you downgrade to anything less than python 3.10 and setuptools to 57.5.0 it should work. Let me know if it doesn't I am using python 3.8.10 now, setuptools 57.5.0 with a virtual env, still with the import problem. I may try downgrading setuptools more. Ok, please let me know how it goes ok found setuptools 56.0.0 is OK! this is now my pipi listi in the virtual environment: pip list Package Version ------------------ --------- addict 2.4.0 aiofiles 0.7.0 anyio 3.3.0 asgiref 3.4.1 certifi 2021.5.30 charset-normalizer 2.0.4 click 8.0.1 demjson 2.2.4 h11 0.12.0 httpcore 0.13.6 httpx 0.19.0 idna 3.2 itsdangerous 2.0.1 Jinja2 3.0.1 justpy 0.1.5 MarkupSafe 2.0.1 pip 21.2.4 pkg_resources 0.0.0 rfc3986 1.5.0 setuptools 56.0.0 sniffio 1.2.0 starlette 0.16.0 uvicorn 0.15.0 websockets 9.1 Thanks for finding a workaround. I am now in the process of publishing a new version without using setuptools (using flit). Perhaps this will solve the issue. btw, using docker and latest python3:8 image I found this better setuptools 57.4.0: pip list Package Version ------------------ --------- addict 2.4.0 aiofiles 0.7.0 anyio 3.3.0 asgiref 3.4.1 certifi 2021.5.30 charset-normalizer 2.0.4 click 8.0.1 demjson 2.2.4 h11 0.12.0 httpcore 0.13.6 httpx 0.19.0 idna 3.2 itsdangerous 2.0.1 Jinja2 3.0.1 justpy 0.1.5 MarkupSafe 2.0.1 pip 21.2.4 rfc3986 1.5.0 setuptools 57.4.0 sniffio 1.2.0 starlette 0.16.0 uvicorn 0.15.0 websockets 9.1 wheel 0.37.0 You might want to switch to demjson3 in case you really only want to support Python 3 anymore. Thank you for the suggestion. I will make the change. I have the same problem but am unable to downgrade to Python <3.8.8 because I'm on a Mac with an M1 chip. Hi, Is there any solution yet for this problem? JustPy isn't just getting installed! Please see if you can port things to demjson3. Thanks, Sam I need to find time to release a new version with this. There is a demjson compatible package called demjson3 that solves this issue. If you want to fix it locally, change all import demjson lines to import demjson3 as demjson And you need to install demjson3. The advantage of using demjson is that it can parse correctly javascript objects where the keys do not need to be between quotes. same problem here on python 3.6.9 downgrading setuptools to 56.0.0 as giodegas stated worked for me. @elimintz Replacing demjson with demjson3 indeed fixes the issue for me. Also see PR. Version 0.2.3 is out that should this problem. Replaced demjson with demjson3. Would appreciate confirmation that this indeed is the case. Did not have time for too much testing but the changes were very limited. The changes are not reflected in the code on github yet but all I did was replace import demjson with import demjson3 as demjson in 4 places. @poke1024 @1081 @nielstron @giodegas @Ledjob @docsteveharris @Flova Some preliminary test show it is ok now. Thank you. It works fine for me too duplicate of #408
2025-04-01T06:38:33.216951
2018-09-18T10:02:36
361228128
{ "authors": [ "LAguido", "elishacloud" ], "license": "Zlib", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5620", "repo": "elishacloud/dxwrapper", "url": "https://github.com/elishacloud/dxwrapper/issues/31" }
gharchive/issue
New Worlds Hello, I've been trying to get Star Trek New Worlds to run, with DDrawCompat= 1 DSoundCtrl= 1 I do get into the game, however after a few seconds it just freezes. I'm using compatibility mode for windows 7, 16 bit color mode, disabled optimizations for full screen. I've added the logfile I got from Process Monitor: Logfile.zip I got the feeling I'm missing something really small, since it does start and you can move around for a few seconds. (Edited with new logfile, previous didn't properly include the whole game session) dxwrapper-stnw.log Here's the wrapper log when the issue occurs It looks like the game is crashing. Try setting DSoundCtrl = 0. Also try adding this line into the ini file: HandleExceptions = 0. This will allow you to see the crash. If that does not work, try using the attached updated files. This works with Star Trek Armada 1 and Star Trek Armada 2. ddraw.zip The 2 settings seemed to have worked (at least it doesn't crash in 10 seconds), I did not use the Armada files. The game needed to be launched from a FAT32 usb stick or partition to work properly. I didn't have time to test the game properly but thus far it seems ok! Closing this since the issue seems to be resolved. If the issue comes back you can reopen.
2025-04-01T06:38:33.218058
2023-12-17T17:59:36
2045314965
{ "authors": [ "JaeAeich" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5621", "repo": "elixir-cloud-aai/cloud-components", "url": "https://github.com/elixir-cloud-aai/cloud-components/pull/213" }
gharchive/pull-request
feat(wes): add service-info Description This PR creates a service info component for WES. closing this will create a PR for the same, as there have been changes in design package.
2025-04-01T06:38:33.263954
2023-06-15T22:38:00
1759636256
{ "authors": [ "josevalim", "paulo-ferraz-oliveira" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5622", "repo": "elixir-makeup/makeup", "url": "https://github.com/elixir-makeup/makeup/pull/60" }
gharchive/pull-request
Fix type definition + @type parsec_success :: {:ok, tokens, String.t(), context(), integer(), integer()} - @type parsec_success :: {:ok, tokens, String.t(), context(), {integer(), integer()}, integer()} as consumed by e.g. makeup_elixir or makeup_erlang. I take this time to also introduce dialyxir to CI, which surfaces another issue (opaqueness-related) already mentioned next to stream_data. Actions missing from this: waiting for a stream_data update and subsequent import here (potentially) updating nimble_parsec (as per acceptance of this pull request) [optional] Edit: while reading the CONTRIBUTING guide I didn't quite understand how to handle the RELEASE.md part (is this required?). Regarding the CHANGELOG.md, do you prefer I do it? Or do you, prior to release? A PR to fix the specs is welcome but we don't plan to introduce dialyzer at the moment, thank you :) Sure. I can remove that bit. Thanks. @josevalim, shall I wait for a nimble_parsec release to update this? Or are you good without it? Thanks. :green_heart: :blue_heart: :purple_heart: :yellow_heart: :heart: The goal of this update is to fix a Dialyzer -related issue with makeup_erlang and makeup_elixir (for which now I'm thinking you don't want dialyxir introduced 😄). Do you think it would make sense to tag-release it, to prevent consumers from finding a dialyxir issue.
2025-04-01T06:38:33.267066
2024-05-02T20:58:47
2276470711
{ "authors": [ "eljonny" ], "license": "Unlicense", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5623", "repo": "eljonny/TestCPP", "url": "https://github.com/eljonny/TestCPP/pull/10" }
gharchive/pull-request
Fix catching the object by value. Fixes #4 Fixed the CPPCheck issue, but this line that was changed is not covered by tests. Going to create a test for this to get the patch coverage up. That last commit is for the release build. I should have put it in the preprocessor def as an else though since iostream is already included in the debug build. Fixy fix. Fixed a bug where capturing multiple streams in the same test resulted in an Access Violation on Windows cl in Debug. PR now has sufficient patch coverage for merge.
2025-04-01T06:38:33.275215
2017-03-17T11:34:50
214978284
{ "authors": [ "bbqbaron", "julianjelfs" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5624", "repo": "elm-community/elm-datepicker", "url": "https://github.com/elm-community/elm-datepicker/issues/26" }
gharchive/issue
Cannot understand how any function that requires the Model can be used For example when I have an instance of the DatePicker type as returned by init, how can I then call getDate in order to obtain the Maybe Date value? getDate requires a Model which is not what I have. Without being able to understand how to call getDate, setDate or setFilter I really can't work out how to synchronise two date pickers into a range. Would really appreciate some help (if I am missing something) or some clarification if a fix of some kind is required. Happy to help if I can by the way. I'm glad you noticed that too! It occurred to me recently that those functions wouldn't do anyone any good. You're not missing anything; I think a (hopefully fairly simple) fix is in order to convert them to DatePicker -> instead of Model ->. I'll certainly get to it soon if no one else does. Hey, if you're happy with that change to the interface, I'm happy to have a go at making that change myself and send you a PR. I'm keen to get involved because I think there are a number of enhancements that we would like to make in the future (keyboard control, multi-month views, internationalisation etc) so it would be good to get familiar with things.
2025-04-01T06:38:33.280064
2018-09-28T10:44:44
364837135
{ "authors": [ "RalfNorthman" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5625", "repo": "elm-community/typed-svg", "url": "https://github.com/elm-community/typed-svg/issues/27" }
gharchive/issue
The type Additive have incorrect variants. The type Additive have the variants AdditiveNone | AdditiveReplace, but should have AdditiveReplace | AdditiveSum (with correction to the corresponding function in TypesToStrings.elm). Source: https://developer.mozilla.org/en-US/docs/Web/SVG/Attribute/additive Comes up when combining transform animations in this way: <rect x="10" y="10" width="40" height="20" style="stroke: #000000; fill: none;"> <animateTransform attributeName="transform" attributeType="XML" type="scale" from="1" to="3" begin="0s" dur="10s" repeatCount="indefinite" additive="sum" /> <animateTransform attributeName="transform" attributeType="XML" type="rotate" from="0 30 20" to="360 30 20" begin="0s" dur="10s" fill="freeze" repeatCount="indefinite" additive="sum" /> (http://tutorials.jenkov.com/svg/svg-animation.html) I think I can fix this.
2025-04-01T06:38:33.282393
2017-09-11T13:33:27
256704017
{ "authors": [ "evancz", "nonpop" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5626", "repo": "elm-lang/elm-compiler", "url": "https://github.com/elm-lang/elm-compiler/issues/1643" }
gharchive/issue
Compiler freezes when annotating a weird self-referential type Trying to compile the following code causes the compiler (elm 0.18.0 on Archlinux) to freeze. If I remove the type annotation from idF it will produce an error message about a weird self-referential type, but with the annotation it just gets stuck. module Test exposing (..) type alias Focus b s = { get : b -> s , update : (s -> s) -> b -> b } create : (b -> s) -> ((s -> s) -> b -> b) -> Focus b s create get update = { get = get , update = update } idF : Focus { r | id : a } a idF = create .id (\f s -> { s | id = f }) -- should be { s | id = f s.id } Development build captures it: It is not pointing out the infiniteness in an ideal way, but it does not hang at least.
2025-04-01T06:38:33.283239
2015-10-14T04:33:37
111322017
{ "authors": [ "HarleyKwyn", "evancz" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5627", "repo": "elm-lang/elm-lang.org", "url": "https://github.com/elm-lang/elm-lang.org/pull/399" }
gharchive/pull-request
Fix typo in If Expressions section Hope this will make it less confusing for people who follow in the future Nice, thanks!
2025-04-01T06:38:33.308122
2023-08-19T11:15:34
1857677467
{ "authors": [ "nothub" ], "license": "Unlicense", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5628", "repo": "elnormous/contenttype", "url": "https://github.com/elnormous/contenttype/issues/12" }
gharchive/issue
Failed parsing Firefox header I tried to parse my browsers header and ran into the invalid media type error indicating that the header is syntactically invalid. Does that mean that Firefox does not implement the header format correctly or is it an error in this project? I added a test case for the header (text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8). Result: === RUN TestGetMediaType/Firefox_header contenttype_test.go:221: Unexpected error "invalid media type" for text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8 --- FAIL: TestGetMediaType (0.00s) Ah nvm, I did not realize there is GetAcceptableMediaTypeFromHeader.
2025-04-01T06:38:33.362107
2023-10-10T20:58:01
1936236335
{ "authors": [ "KishiTheMechanic" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5629", "repo": "elsoul/skeet-solana-mobile-stack", "url": "https://github.com/elsoul/skeet-solana-mobile-stack/issues/16" }
gharchive/issue
Sign In With Solana with Solana Mobile Stack Solana Mobile Team is implementing the SIWS (Sign In With Solana) API on Mobile Wallet Adapter. (Probably on v2.0) https://github.com/solana-mobile/mobile-wallet-adapter/issues/439 After that, we can use SIWS on both of Web Apps and Solana dApps on the Saga phone. For now, SIWS is available with Next.js WebApp with Skeet. Check /webapp folder for that. https://github.com/elsoul/skeet-solana-mobile-stack/tree/main/webapp Wallet Adapter for Web is here https://github.com/elsoul/skeet-solana-mobile-stack/blob/main/webapp/src/components/providers/SolanaWalletProvider.tsx Mobile Wallet Adapter specification Version: 2.0.0-DRAFT https://solana-mobile.github.io/mobile-wallet-adapter/spec/spec.html It's done already
2025-04-01T06:38:33.369052
2016-02-20T12:39:00
135067967
{ "authors": [ "Zirak", "xiaq" ], "license": "bsd-2-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5630", "repo": "elves/elvish", "url": "https://github.com/elves/elvish/pull/149" }
gharchive/pull-request
edit: Fix crash in history listing mode When you had too few history entries to display in the terminal, entering history listing mode attempted to access a negative index and crashed. I've wanted to make two additional edits but was unsure about their possible side-effects on other parts. The first was not entering history listing mode if there was no history. Off I went to startHistoryListing, but got stuck a bit in the semantics. What happens if newHistoryListing returns an error (which makes sense in our situation)? Would the user still be in history listing mode? Perhaps it'll be best then to move the mode switch below the check. The second was in trimToLines, I badly yearned to put a if len(b.cells) < low check in the beginning, but it may introduce subtle bugs (why aren't things trimming?) if used in certain contexts, and it felt wrong to do it without more knowledge of the system. Great job on the shell, btw. It's mighty impressive. Hi! Thanks for the fix. You are correct that the mode switch should be moved down, and I prefer to leave trimToLines as it is, due to the concerns you just stated. By the way, history listing was something I didn't finish; the ultimate goal is to steal the design ptpython's history listing (#63), which allows you to scroll through the whole history and more importantly, compose a chunk of code by cherry-picking multiple entries from the history. Before trying to implement it, I observed that it has quite a lot in common with existing completion and navigation listings -- esp. wrt. the scrolling and trimming behavior -- so a good abstraction should be made to capture the common behavior. Unable to come up with a good abstraction, I didn't bother to think hard and turned to other parts of elvish instead, leaving this unfinished thing in the code :) If you find this interesting enough, you are more than welcome to contribute. Keep me informed about your progress, so that I won't rewrite the whole line editor in a midnight and ruin all your efforts.
2025-04-01T06:38:33.442053
2018-09-28T22:52:08
365065757
{ "authors": [ "PascalPrecht", "StatusSceptre", "cryptomental", "iurimatias", "subramanianv" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5631", "repo": "embark-framework/embark", "url": "https://github.com/embark-framework/embark/issues/941" }
gharchive/issue
add --nobrowser field to test codecoverage. Outline Running embark test -c will cause the browser to automatically open with the code coverage report. The goal of this task is to add --nobrowser field for users that don't want this. Acceptance Criteria running embark test -c --nobrowser should running the tests & code coverage as normal, but not open the browser. embark test -c should work as now. PR opened: https://github.com/embark-framework/embark/pull/950 @iurimatias Seems like the task is already done Thanks for pointing this out @subramanianv ! Yes this landed as https://github.com/subramanianv/embark/commit/890b46977780d3b3d0199ba8c459c102d6f85596 Closing this one. @vs77bb this issue won't allow me to payout and has been closed already, can you help? Hi @StatusSceptre it seems you need to approve me now and I will click on the 'submit' afterwards. Approved @cryptomental @StatusSceptre thank you, I submitted via GitCoin.
2025-04-01T06:38:33.476620
2022-04-23T11:30:44
1213277157
{ "authors": [ "1homebridge", "JurgenLB", "embee8" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5632", "repo": "embee8/homebridge-panasonic-ac-platform", "url": "https://github.com/embee8/homebridge-panasonic-ac-platform/issues/26" }
gharchive/issue
Status Bar and Cooling Icon I was using the previous plug-in which used to provide a cooling or heating status in the status bar of each room/level but I am not getting this function with this plug-in. I would also get a change in the colour icon of each unit based on activity from Green to Blue when Cooling and to Orange when Heating. Is this option available with this plug-in. Thanks Is this due different plugin functionality or due to a change in HomeKit, i.e. how Apple chooses to display it? Do you have a reference implementation of a plugin that achieves these UI distinctions? Hi Thank you for getting back to me. Prior to using this plug-in, I was using the Panasonic Air Conditioner plug-in by Cody1515 which has now been archived. He mentioned that your plug-in is based on the same coding, the difference being that yours uses the platform rather than individual accessories/units (trying to explain this as best as I can but my knowledge is pretty basic) His plug-in also did not initially update the status bar nor did it control the icon colour based on setting, Green for Standby, Blue for Cooling and Orange for Heating. But one of the last updates included these functions which were quite useful. So for example, when using the cooling option, the Green Circle would turn Blue and the Status Bar on the top of the Room/Level would change from Idle to Cooling and note the temperature much like the Heatmiser accessories would show the Room was Heating. I hope this clarifies the “issue”. As I am very dependent on this plug-in, I would be willing to donate but there is not such option on your plug-in. Many Thanks I’m attaching screenshots I found of the features I was trying to explain. This is the Status Bar up top that shows that the AC is on Cooling and the Temperature of the Room And this is a side by side of my Underfloor Heating on Standby (Green) and the AC on Cooling (Blue) and again showing the Temperature within the circle. I hope this helps Thanks again The results of my debugging session suggest that we might be dealing with a Homebridge or HomeKit bug. The status icon in the Home app is controlled by the CurrentHeaterCoolerState characteristic of the accessory. On this line, we set the current state to IDLE when the AC is in cooling mode and the current temperature is less than the set temperature. However, in my test the Home app displays the status of the AC as inactive. In my debugging session, I also tried setting the current state to INACTIVE instead of IDLE and I got the same result on the UI. As an additional test, I set the current state to HEATING. The Home app UI reflected this correctly (as Heating), which leaves me thinking that Homebridge or HomeKit don't distinguish between the IDLE and INACTIVE modes. @1homebridge, upon further exploration I can confirm that I actually see the "Cooling" indicator and the blue arrow in my Home app. In your previous comments, you uploaded images of how you want it to look like, but can you upload a screenshot of how it actually looks like for you right now? My previous comment alleged a bug with regards to how IDLE and INACTIVE lead to the same UI representation. It does not, however, confirm your original thesis of the indicators not being available at all. Could you help me clarify the problem statement? The overlap between the IDLE and INACTIVE statuses is addressed in this issue in the Homebridge repository. can you set Debug to "true" in your config. Restart and post the output from your log. Maybe we can resolve the issue then. My config looks like this; { "name": "Homebridge Panasonic AC Platform", "email"<EMAIL_ADDRESS> "password": "xxxxxxxxxxxxxxxxxx", "exposeOutdoorUnit": true, "debugMode": true, "platform": "Panasonic AC Platform" }, @1homebridge, is this the same type of issue as in #23 and #31? Hi, thanks for getting back to me. Yes, though with the new UI in iOS16 the Status Bar no longer indicates between Idle-Cooling-Heating states so the issue is just to obtain an indoor temperature reading from the outdoor unit as described in #23 and #31 to stop getting a 0.0* indicator. This would then allow the units to show the Blue colour indicating that they are in Cooling mode as requested in the other threads. Thanks Okay, thanks for confirming and the additional context. Will deal with it through the other open issues.
2025-04-01T06:38:33.478273
2016-07-08T18:27:19
164588776
{ "authors": [ "mdentremont" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5633", "repo": "ember-cli/broccoli-concat", "url": "https://github.com/ember-cli/broccoli-concat/pull/62" }
gharchive/pull-request
Update to latest minimatch to avoid deprecation warning Avoids npm WARN deprecated<EMAIL_ADDRESS>Please update to minimatch 3.0.2 or higher to avoid a RegExp DoS issue @stefanpenner Awesome!!
2025-04-01T06:38:33.589040
2024-03-23T07:54:15
2203758991
{ "authors": [ "bidetofevil" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5634", "repo": "embrace-io/embrace-android-sdk", "url": "https://github.com/embrace-io/embrace-android-sdk/pull/617" }
gharchive/pull-request
Convert EmbraceDeliveryServiceTest to use a real DeliveryCacheManager Goal Testing Release Notes WHAT: WHY: WHO: [!WARNING] This pull request is not mergeable via GitHub because a downstack PR is open. Once all requirements are satisfied, merge this PR as a stack on Graphite. Learn more #618 #617 👈 #616 master This stack of pull requests is managed by Graphite. Learn more about stacking. Join @bidetofevil and the rest of your teammates on Graphite Merge activity Mar 26, 3:09 AM EDT: @bidetofevil started a stack merge that includes this pull request via Graphite.
2025-04-01T06:38:33.593815
2024-06-12T21:51:31
2349740517
{ "authors": [ "bidetofevil" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5635", "repo": "embrace-io/embrace-android-sdk", "url": "https://github.com/embrace-io/embrace-android-sdk/pull/960" }
gharchive/pull-request
Add in an additional step to validate in ExecutionCoordinator Goal Testing Release Notes WHAT: WHY: WHO: #960 👈 master This stack of pull requests is managed by Graphite. Learn more about stacking. Join @bidetofevil and the rest of your teammates on Graphite Merge activity Jun 13, 2:22 AM EDT: @bidetofevil merged this pull request with Graphite.
2025-04-01T06:38:33.602139
2019-03-19T22:42:05
422986597
{ "authors": [ "emeeks", "susielu" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5636", "repo": "emeeks/react-dorling-map", "url": "https://github.com/emeeks/react-dorling-map/issues/14" }
gharchive/issue
Bug: hover data when geo is changed When hovering, the data onHover is not changed after the data is updated from a subregion to country view. Resolves in the latest version
2025-04-01T06:38:33.603403
2021-11-26T00:37:37
1063990003
{ "authors": [ "LaTonia-Mertica" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5637", "repo": "emergentworks/mentee-assessments", "url": "https://github.com/emergentworks/mentee-assessments/pull/3" }
gharchive/pull-request
La'Tonia Mertica EW Mentee Assessment (to date) as much as could complete on my own Hello Those Powering EW, Didn't see option to add reviewer(s) et cetera. Apologies if this pull request is in error in any way. Thanks for this opportunity, please stay safe.
2025-04-01T06:38:33.668882
2022-12-05T18:33:53
1477168867
{ "authors": [ "Boscop", "coderedart" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5638", "repo": "emilk/egui", "url": "https://github.com/emilk/egui/issues/2393" }
gharchive/issue
Update pure_glow example to glutin 0.30 A lot has changed in glutin 0.30, it would be nice if the pure_glow example could be updated to glutin 0.30 :) I had some code that was based on the pure_glow example before, now I'm updating it to glutin 0.30, but I'm not sure if I'm doing everything correctly with the new ways of doing things. You can find my code here: https://github.com/rust-windowing/glutin/issues/1445#issuecomment-1337903593 I would appreciate if you can let me know if it's correct. The pure_glow example could be updated similarly. @coderedart Thanks. And should I use .with_profile(GlProfile::Core) or not? :) @coderedart Thanks. And should I use .with_profile(GlProfile::Core) or not? :) always use Core, unless you are targeting really ancient hardware.
2025-04-01T06:38:33.671225
2024-05-15T08:46:55
2297249972
{ "authors": [ "emilk", "rustbasic" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5639", "repo": "emilk/egui", "url": "https://github.com/emilk/egui/pull/4498" }
gharchive/pull-request
Fix: Use default features in Image Crate Closes #4489 Related #4495 Fix: Use default features in Image Crate Because only .png is available after update #4495. Required to use JPEG, etc. No - eframe only need png support, and should not be paying for the compilation of ten other image formats
2025-04-01T06:38:33.674810
2024-02-15T06:16:35
2135733164
{ "authors": [ "keeganpotgieter" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5640", "repo": "emilkowalski/vaul", "url": "https://github.com/emilkowalski/vaul/issues/269" }
gharchive/issue
Add support for iOS PWA app bar Add the ability to match the colour of the iOS PWA application bar to the overlay colour to create a more native feel. Examples Without iOS App Bar Support With iOS App Bar Support Solution User needs these meta tags <meta name="apple-mobile-web-app-status-bar-style" content="black-translucent" /> <meta name="apple-mobile-web-app-capable" content="yes" /> Changes made in vaul/src/index.tsx to dynamically change the background colour to match overlay background colour Will also solve some of the issues with this issue https://github.com/emilkowalski/vaul/issues/259 maybe, can solve this issue #199 too It seems like #199 isn't an issue with body backgrounds.
2025-04-01T06:38:33.678981
2024-05-28T04:27:43
2320095587
{ "authors": [ "RickRyan26", "emilkowalski", "isaachinman", "max-17" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5641", "repo": "emilkowalski/vaul", "url": "https://github.com/emilkowalski/vaul/issues/360" }
gharchive/issue
How to swipe outside of the drawer to open/close? The ChatGPT mobile app has the most beautiful drawer functionality, and I'm trying to recreate that. At any time the user can side swipe the chat or drawer area to open or close the drawer and shift the entire page layout. My main hurdle is being able to swipe outside of the drawer while still being able to select what's beneath that outside swipeable area... Does my entire layout need to be inside the drawer? With only half of it looking like the drawer? And without the drawer being allowed to fully close? Any suggestion anyone reading this has would be fantastic. Thank you! check MUI swipable drawer @max-17 I haven't tried this yet but I think a horizontal CSS scroll snap on the layout solves my needs That's not supported here. The Drawer here is a Dialog meaning that it usually sits on top of other elements. If you'd want to use this Drawer for something like ChatGPT's mobile app you could position the Drawer off screen initially and add an additional drag event to the body that would translate the content. @max-17 Did you create a fork and build this functionality? I was able to achieve this quite easily: Use a controlled Vaul drawer, ie open and onOpenChange Add a react-swipeable handler to my layout, and capture specific onSwipeStart events within X pixels of the edge of the screen and set open to true
2025-04-01T06:38:33.685485
2021-08-01T16:24:30
957532127
{ "authors": [ "emilyriederer", "jonspring" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5642", "repo": "emilyriederer/data-disasters", "url": "https://github.com/emilyriederer/data-disasters/pull/1" }
gharchive/pull-request
Update data-dall.Rmd Handful of typo corrections & suggested syntax revisions. Looks like this is / will be a great resource! This is a quick test with just a handful of suggested edits. Happy to modify process if there's a better way to submit these, e.g. with separate commits (e.g. I'm not sure "consumer robotics" company is necessarily better descriptor than "e-commerce" company). Wow - thanks so much for taking the time, @jonspring ! I really appreciate these
2025-04-01T06:38:33.713659
2017-09-05T14:55:04
255306189
{ "authors": [ "NieDzejkob", "idmean" ], "license": "Artistic-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5643", "repo": "emojicode/emojicode", "url": "https://github.com/emojicode/emojicode/pull/84" }
gharchive/pull-request
Add PROJECT_BRIEF to Doxyfile Come on, why wasn't it done before? Let me know if I fucked this up, I don't have Doxygen installed. "why wasn't it done before?" Because I expect people consulting the source code documentation to already know what source they are studying...
2025-04-01T06:38:33.763412
2023-10-26T20:06:48
1964292352
{ "authors": [ "ArtiikSK", "herrardo" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5644", "repo": "empathyco/x", "url": "https://github.com/empathyco/x/issues/1342" }
gharchive/issue
query to get all elements what should I specify in the query to get all elements without filters? Thanks for asking @ArtiikSK Could you detail what are you trying to do? Are you launching a web server to try? You can add a vídeo or screenshot too if you like.
2025-04-01T06:38:33.791112
2021-07-27T15:56:08
954029439
{ "authors": [ "ateequrrahman97", "emretulek" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5645", "repo": "emretulek/jbvalidator", "url": "https://github.com/emretulek/jbvalidator/issues/15" }
gharchive/issue
How can we display Custom Error message //custom validate method validator.validator.custom = function(el, event){ if($(el).is('[name=password]') && $(el).val().length < 5){ return 'Your password is too weak.'; } } This method only return but it can not display any error on the screen. The returned string is used as error message, you can define rule and message for specific input. Okay, thanks @emretulek
2025-04-01T06:38:33.856027
2015-12-16T04:53:07
122428944
{ "authors": [ "mediavrog", "pcboy", "shawila" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5646", "repo": "en-japan/koudoku", "url": "https://github.com/en-japan/koudoku/pull/1" }
gharchive/pull-request
update stripe api version The change logs after 2015-09-08 don't have any major changes in the koudoku code. LGTM hitomi :heart:
2025-04-01T06:38:33.935786
2022-12-22T21:15:07
1508553652
{ "authors": [ "Eldar2021", "clangenb" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5647", "repo": "encointer/encointer-wallet-flutter", "url": "https://github.com/encointer/encointer-wallet-flutter/issues/927" }
gharchive/issue
Schedule notifications upon registering or claiming rewards Currently, we use the feed to schedule notifications for the registering and meetup reminders, but we can actually schedule them from within the app when we do certain actions. After registering, we can schedule two notifications, which should be reminders for the next meetup: 24 hours before the meetup time 1 hour before the meetup time The meetup time can be fetched here: https://github.com/encointer/encointer-wallet-flutter/blob/6fc236c622932d2b503f2ae5055c846dd6c7ccaa/lib/store/encointer/sub_stores/community_store/community_store.dart#L64. The value can't be null after registering, as it can only be null if we have not chosen a community. After claiming the rewards, we can set a reminder to register at the start of the registering phase. Note: We must only schedule the notification if we claim the rewards in the attesting phase, not in the registering phase, obviously. The timestamp to use is the the nextPhaseTimestamp: https://github.com/encointer/encointer-wallet-flutter/blob/6fc236c622932d2b503f2ae5055c846dd6c7ccaa/lib/store/encointer/encointer.dart#L81 Caveats: How do we ensure unique IDs that don't overlap with the IDs we assign to the notifications we get from the feed? Maybe we have to maintain an internal global counter for that. We should only schedule notifications if we are connected to the parachain on kusama, the nctr-k If we use global counter How do we ensure unique IDs that don't overlap with the IDs we assign to the notifications we get from the feed? Maybe we have to maintain an internal global counter for that. Can we use cache? I suggest generate meetupId by meetupTime void main() { final meetupTimeAfter1Day = DateTime.now().add(const Duration(days: 1)).millisecondsSinceEpoch; final meetupTimeAfter7Days = DateTime.now().add(const Duration(days: 7)).millisecondsSinceEpoch; final meetupTimeAfter10Days = DateTime.now().add(const Duration(days: 10)).millisecondsSinceEpoch; final meetupTimeAfter15Days = DateTime.now().add(const Duration(days: 15)).millisecondsSinceEpoch; final meetupTimeAfter30Days = DateTime.now().add(const Duration(days: 30)).millisecondsSinceEpoch; print(generateMeetupIdByTimeStamp(meetupTimeAfter1Day)); // 1 print(generateMeetupIdByTimeStamp(meetupTimeAfter7Days)); // 7 print(generateMeetupIdByTimeStamp(meetupTimeAfter10Days)); // 10 print(generateMeetupIdByTimeStamp(meetupTimeAfter15Days)); // 15 print(generateMeetupIdByTimeStamp(meetupTimeAfter30Days)); // 30 } int generateMeetupIdByTimeStamp(int meetupTime) { final now = DateTime.now().millisecondsSinceEpoch; int id = 0; int c = meetupTime - now; do { c -= 86400000; // 1 day = 86400000 milliseconds id++; } while (c > 0); return id; } So I think we can generate dynamic id without using any cache. Please let me know your opinion. Hi @Eldar2021, You suggestions look good! When user registers to a meetup, can he register to another meetup before that meetup ends? Yes, this is possible, but the app should not allow that because you have not yet got your reputation and you will be a newbie again. You should only be allowed to do that after the rewards have been claimed. If we use global counter How do we ensure unique IDs that don't overlap with the IDs we assign to the notifications we In general, I like your approach of using the meetup time. Why do you need to divide by 86400000, is the number too big? I think this could lead to problems where we end up with the same ID for reminders, which are very close to eachother, do you agree? Is there registration deadline time? or registration finishes when meetup starts? The registration deadline is when the registering phase is over. Remember, we have 3 phases REGISTERING > ASSIGNING > ATTESTING. As long as we are in the registering phase, the nextPhaseTimeStamp is the deadline for registering. And in the assigning phase you can already register for the meetup in the next assigning phase.
2025-04-01T06:38:33.960895
2019-06-04T16:43:42
452099739
{ "authors": [ "Shadowblitz16", "Shylie", "endlesstravel" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5648", "repo": "endlesstravel/Love2dCS", "url": "https://github.com/endlesstravel/Love2dCS/issues/75" }
gharchive/issue
Convert Image to Canvas? Is it possible to convert a image to a canvas and back so I can draw onto it? if not could you add Graphics.NewCanvas(Image Image) and Graphics.NewCanvas(ImageData ImageData)? There is no such method, but you can draw an image onto the canvas. but this requires the draw event right? I was trying to do it so it could be used anywhere See Love.Graphics.Present. This should be what you are looking for. can support be added so that images and canvases can be explicitly converted back and forth? a example here : copy the utils class into your project: https://gist.github.com/endlesstravel/027799eb772d644b0d4110284256da6a use it like : static public void Test_Issue75_ToPintImage() { ISSUE_75.Init(); var imgData = ISSUE_75.PrintImage(300, 300, () => { Graphics.SetColor(Color.LightPink); Graphics.Rectangle(DrawMode.Fill, 0, 0, 100, 100); Graphics.SetColor(Color.White); Graphics.Circle(DrawMode.Line, 100, 100, 20); }); Resource.EncodeToFile("test.png", imgData, ImageFormat.PNG); } can this be added to love2dcs as a explicit image conversion?
2025-04-01T06:38:33.963312
2022-04-27T00:44:36
1216637282
{ "authors": [ "Evernow", "istiak101" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5649", "repo": "endoflife-date/endoflife.date", "url": "https://github.com/endoflife-date/endoflife.date/pull/1102" }
gharchive/pull-request
[Fedora-Linux] Update EOL dates and command F36 was delayed, which pushed back the F34 end of life date: https://fedorapeople.org/groups/schedule/f-36/f-36-key-tasks.html F35 EOL date from: https://fedorapeople.org/groups/schedule/f-37/f-37-key-tasks.html As for the command change, it's due to lsb_release not being installed by default anymore Fedora is delayed again. EOL moved to 2022-06-07. Could be delayed more.
2025-04-01T06:38:33.971651
2017-06-27T05:08:34
238738518
{ "authors": [ "DASPRiD", "MelchiorKokernoot", "endroid", "mortezakarimi" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5650", "repo": "endroid/qr-code", "url": "https://github.com/endroid/qr-code/issues/114" }
gharchive/issue
Feature request: style dots and corners If you can add more flexible feature to style qt such as this I am very grateful!‌ :grin: Hi @mortezakarimi thank you. I will not add this myself but if someone is willing to contribute on this, that would be great. The new upstream QR code library supports that now. @DASPRiD What do you mean by "The new upstream QR code library"? @DASPRiD What do you mean by "The new upstream QR code library"? BaconQRCode, which this library is using. @DASPRiD Do you by any chance have a reference where I can learn how to do so? I can't seem to find anything about changing the style. @DASPRiD Do you by any chance have a reference where I can learn how to do so? I can't seem to find anything about changing the style. There's not really any documentation for this, but the ImageRenderer takes a RendererStyle object, which can be configured. Best to look at the source: https://github.com/Bacon/BaconQrCode/blob/master/src/Renderer/ImageRenderer.php#L26 https://github.com/Bacon/BaconQrCode/blob/master/src/Renderer/RendererStyle/RendererStyle.php Closed as this will not be implemented here.
2025-04-01T06:38:33.989565
2024-01-17T23:43:01
2087197774
{ "authors": [ "engeir" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5651", "repo": "engeir/volcano-core", "url": "https://github.com/engeir/volcano-core/pull/8" }
gharchive/pull-request
chore(main): release 0.2.3 :robot: I have created a release beep boop 0.2.3 (2024-01-17) Bug Fixes mise tasks are experimental and must be activated (4ab703b) This PR was generated with Release Please. See documentation. :robot: Release is at https://github.com/engeir/volcano-core/releases/tag/v0.2.3 :sunflower:
2025-04-01T06:38:33.997629
2016-07-03T22:38:37
163583363
{ "authors": [ "engram-design", "pixeljitsu" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5652", "repo": "engram-design/SocialPoster", "url": "https://github.com/engram-design/SocialPoster/issues/3" }
gharchive/issue
Links to social media posts It would be really handy if there were a way to access the links leading to the social media posts from within the entry. I'd like to automatically display the links as part of the craft entry. "Join the discussion on Facebook/Twitter", etc. Just to confirm, this would be from the front-end? This way, you could post the entry to Facebook, have it record the posted URL, and then being able to access in your template? Something like: {% set post = craft.socialPoster.post({ account: 'facebook', entryId: entry.id }) %} {% if post.url %} <a href="{{ post.url }}" target="_blank">Join the discussion on Facebook</a> {% endif %} Exactly. That way you could drive traffic from the website to the social media posts. Just to follow this up - this is now implemented in 1.2.0. Use the following template code: {% set posts = craft.socialPoster.posts({ element: entry }) %} {% for post in posts %} <a href="{{ post.url }}" target="_blank"> <i class="fa fa-{{ post.handle }}-square"></i> Join the discussion on {{ post.handle | capitalize }} </a> {% endfor %}
2025-04-01T06:38:34.000913
2024-12-13T11:53:32
2738220999
{ "authors": [ "ambravo", "engswee" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5653", "repo": "engswee/flashpipe", "url": "https://github.com/engswee/flashpipe/issues/40" }
gharchive/issue
Repo and Org Variables I wonder if this statement holds true: https://github.com/engswee/flashpipe/blob/687862ea650f903531ae0dfe13d1459582c95847/docs/github-actions-sync-apim.md?plain=1#L28 Github Actions offers Repo and Org variables (doc) Are you referring to something else? Thanks for highlighting this. Unfortunately, documentation often can't keep up with the speed that new features are introduced 😅 If you see the screenshot from the following page, you can see that variables were not there in the past. https://engswee.github.io/flashpipe/github-actions-sync-to-git.html I'd have to admit that I can't keep track of all the new features that are constantly being rolled out, so have definitely missed this one out. It's good to know about this, so that I can use it in my workflows and also update the documentation. I'll keep this issue open until I get around to updating the documentation 😉
2025-04-01T06:38:34.019057
2021-10-12T00:44:47
1023243522
{ "authors": [ "lognaturel", "yanokwa" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5654", "repo": "enketo/enketo-core", "url": "https://github.com/enketo/enketo-core/pull/828" }
gharchive/pull-request
Prepare 6.0.1 Stepping through the commits is the easiest way to review this because of the docs changes. Draft because we are waiting on https://github.com/enketo/openrosa-xpath-evaluator/pull/136 and then a 2.0.9 release. Verified the following: [ ] npm update [ ] npm audit fix --production [ ] npm run test [ ] npm run test-browsers [ ] npm run beautify [ ] npm run build-docs npm run test-browers has a scary Firefox failure. It'd be good to see if it fails on v5.17.6. ```Firefox 93.0 (Mac OS 10.15) merging an instance into the model when the record contains namespaced attributes, the merged result is CORRECTLY namespaced namespaces are added correctly FAILED`` https://github.com/enketo/enketo-core/pull/822 should also be merged first, right?
2025-04-01T06:38:34.031071
2021-06-15T08:11:07
921127286
{ "authors": [ "poi33" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5655", "repo": "enonic/app-office-league", "url": "https://github.com/enonic/app-office-league/issues/462" }
gharchive/issue
Site created on startup The default site is created on application start. WIth app and info-page content types, including populated fields. Should we just remove the site init code and let the user manually setup office-league with site office-league? This would still keep the init of the office-league repoes but remove the auto generated site. Remove the auto generated site. Keep the generated repo storrage.
2025-04-01T06:38:34.070386
2023-11-23T16:51:25
2008567301
{ "authors": [ "a8m", "michaelcaulley" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5656", "repo": "ent/ent", "url": "https://github.com/ent/ent/pull/3841" }
gharchive/pull-request
fix: swap the order of handling order and pred in the graph select func multi order may generate predicates, so running the order funcs first allows reusing aliases. If the predicates run first if there is a multiOrder field that generates it's own predicates, the predicates are unable to find the alias of the join and errors not finding the column on the table. Related to: https://github.com/ent/contrib/pull/559 Thanks for the contribution, @michaelcaulley 🚀
2025-04-01T06:38:34.094383
2016-02-10T07:04:26
132623609
{ "authors": [ "flyinbutrs", "gstamp", "neillturner", "stevehodgkiss", "thekindofme" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5657", "repo": "envato/stack_master", "url": "https://github.com/envato/stack_master/issues/81" }
gharchive/issue
Add support for more template generators YAML cfndsl - https://github.com/stevenjack/cfndsl Any others? It might be nice to have JSON support with comments. Would need to find a library to strip them out I guess. Another one might be just using plain old ERB? If you want JSON support with comments, technically you can just use YAML. An alternate syntax of YAML is actually just JSON + comments, so any YAML parser should be able to parse a JSON file with comments in it. [1] pry(main)> require 'yaml' => true [2] pry(main)> test = "{\n#testing\n}" => "{\n#testing\n}" [3] pry(main)> YAML.load(test) => {} So yeah, +1 for YAML input. should be pretty trivial to add, I'll fork and see. cfndsl support would be really good as it is the most popular generator of cloudformation templates that is written in ruby. it should be really easier to integrate as it is just running cfndsl commands. I also use the ppjson to pretty print the generated cloudformation code. I've built alot of examples for cfndsl at https://github.com/neillturner/cfndsl_examples and there is a utility to convert templates to cfndsl format that is very handy. Hi @neillturner we have added basic CfnDsl template support with #99. However, this dose not add support for using cfndsl variables. How would you like that supported in SM? How would you imagine a user supply values for cfndsl variables when using a cfndsl template with SM?
2025-04-01T06:38:34.097497
2017-09-27T22:16:04
261135160
{ "authors": [ "ggreenway", "htuch", "mattklein123" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5658", "repo": "envoyproxy/envoy", "url": "https://github.com/envoyproxy/envoy/issues/1761" }
gharchive/issue
Number of stats and length of stat name should be configurable Envoy currently hardcodes a limit of 16k individual stats, with a fixed maximum name length. These values should be tunable. +1 Defaults should be settable at compile-time via .bazelrc (or similar), and overridable via CLI flag at startup. @hennna It is easy to make finding unused stats O(1) with a free-list or similar. But doing the name-lookups is a little bit harder in shared memory. We'll probably need a lookup table of some kind in the shared memory. If we want to do this, I would recommend doing it as a totally independent change from the length stuff. Given how low frequency stat allocation is from shared memory the current situation is probably not that big of a deal for most people. I was going to benchmark filling up a somewhat big number of stats, maybe 1M or something, and see how long it takes. But I'm expecting that would be pretty slow: n^2 is 1 trillion operations. But I agree that it can be done in a separate change. Making the sizes tunable won't cause any degradation to existing use cases unless users opt-in to a really large stat size, and they should immediately notice some pain on startup if they choose too-large of a size.
2025-04-01T06:38:34.102156
2023-04-12T17:36:46
1664972971
{ "authors": [ "XinyiZhangAws", "mattklein123", "ytsssun" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5659", "repo": "envoyproxy/envoy", "url": "https://github.com/envoyproxy/envoy/issues/26710" }
gharchive/issue
Add Attributes connection.transport_failure_reason Title: Add Attributes connection.transport_failure_reason along with TLS connection errors Description: Describe the desired behavior, what scenario it enables and how it would be used. Purposed Attributes connection.transport_failure_reason: Currently upstream.transport_failure_reason is included in the attributes to indicate the upstream transport failure such as certificate validation failed. But there's no similar ones for downstream connection. Considering recent PR added downstream transport failure reason to access log, it can be added to attributes as well for consistency. [optional Relevant Links:] Any extra documentation required to understand the issue. https://www.envoyproxy.io/docs/envoy/latest/intro/arch_overview/advanced/attributes.html#attributes https://github.com/envoyproxy/envoy/pull/25322/files cc @kyessenov @mattklein123 could you add a "help wanted" tag to revive this issue?
2025-04-01T06:38:34.109718
2024-09-09T15:02:01
2514216849
{ "authors": [ "alyssawilk", "cainelli", "ravenblackx", "tyxia" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5660", "repo": "envoyproxy/envoy", "url": "https://github.com/envoyproxy/envoy/issues/36041" }
gharchive/issue
Test flake: //test/extensions/filters/http/ext_proc:ext_proc_integration_test https://dev.azure.com/cncf/envoy/_build/results?buildId=179508&view=logs&j=4930ecaf-18f4-5b3c-dea3-309729c3b3ae&t=573d8780-d7b9-52e3-b4e0-a89886b0b9ff&l=3840 [ RUN ] IpVersionsClientTypeDeferredProcessing/ExtProcIntegrationTest.GetAndCloseStreamWithTracing/IPv4_GoogleGrpc_WithDeferredProcessing test/extensions/filters/http/ext_proc/tracer_test_filter.cc:52: Failure Expected equality of these values: want Which is: "0" got Which is: "" grpc.status_code: 0 not found in tags: component: proxy status: canceled upstream_address: <IP_ADDRESS>:37667 upstream_cluster: ext_proc_server_0 @tyxia @cainelli Do you mind taking a look at this? Thank you! I think tracing related feature was added by your change oh sorry about that. I will take a look into it this week. @tyxia the failure is a timeout and the tracing failure is a side effect. I don't see how such simple request would take more than 5s to run. test/extensions/filters/http/ext_proc/ext_proc_integration_test.cc:281: Failure Value of: fake_upstreams_[0]->waitForHttpConnection(*dispatcher_, fake_upstream_connection_) Actual: false (Timed out waiting for new connection.) Expected: true .... test/extensions/filters/http/ext_proc/ext_proc_integration_test.cc:277: Failure Expected equality of these values: std::to_string(status_code) Which is: "200" response.headers().getStatusValue() Which is: "504" Stack trace: 0x1578018: (unknown) 0x13cb8bc: (unknown) 0x7fc368411a4d: testing::internal::HandleSehExceptionsInMethodIfSupported<>() 0x7fc3683f822e: testing::internal::HandleExceptionsInMethodIfSupported<>() 0x7fc3683dfb1d: testing::Test::Run() 0x7fc3683e060e: testing::TestInfo::Run() ... Google Test internal frames ... actually, is the timeout 5ms? should we increase it a bit? https://github.com/envoyproxy/envoy/blob/7a7df5d8887dfe673eef51ce396feab4bff9383f/test/integration/http_integration.cc#L555-L556 @cainelli Thanks for spending effort reducing the flakiness. In the past, we have test (as you linked in slack) that have larger then 5s timeout. However, that is because Please don't waitForHttpConnection with a 5s timeout if failure is expected. In your case, failure is not expected. I am not sure if the flakiness is because tracing will take a bit more time but 5s should be sufficient here. Or maybe because ext_proc_integration test has grown very big now. Do you happen to know what is the flakiness rate? if you can't repro the flake (per slack) one thing you can do is add a LogLevelSetter in that test such that CI logs more information when it flakes. then next time we see a failure you'll have more information. One thing I've found often helps if you can't repro a flake is to run stress -c 16 (or however many CPU cores) in another terminal while the test runs with --runs_per_test=n. (Flags also depending on if the flakiness is from being CPU bound or network bound or disk bound.) Thank you all for the context and tips. One thing I've found often helps if you can't repro a flake is to run stress -c 16 (or however many CPU cores) in another terminal while the test runs with --runs_per_test=n. (Flags also depending on if the flakiness is from being CPU bound or network bound or disk bound.) I did tried that with various combinations to stress during the test but did not have any luck reproducing it. if you can't repro the flake (per slack) one thing you can do is add a LogLevelSetter in that test such that CI logs more information when it flakes. then next time we see a failure you'll have more information. I will try this path moving forward (https://github.com/envoyproxy/envoy/pull/36583).
2025-04-01T06:38:34.113181
2024-10-31T09:05:26
2626347837
{ "authors": [ "nezdolik", "supercairos" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5661", "repo": "envoyproxy/envoy", "url": "https://github.com/envoyproxy/envoy/issues/36931" }
gharchive/issue
Implement a Token Introspection (RFC 7662) HTTP Filter Token Introspection (RFC 7662): Implement an HTTP Filter to verify external token Description: Allow external OAuth2/OIDC token to be valided via the Token Introspection api in order for envoy to act as a Identity Aware Proxy (IAP) Relevant Links: https://datatracker.ietf.org/doc/html/rfc7662 https://www.oauth.com/oauth2-servers/token-introspection-endpoint/ Please get familiar with our extension policy: https://github.com/envoyproxy/envoy/blob/main/EXTENSION_POLICY.md cc @tyxia @mattklein123 @TAOXUY (as oauth, jwt extension owners who may be interested in this proposal)
2025-04-01T06:38:34.132482
2019-10-17T14:39:47
508523044
{ "authors": [ "HenryYYang", "mosespx", "ramaraochavali", "saagar241290" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5662", "repo": "envoyproxy/envoy", "url": "https://github.com/envoyproxy/envoy/issues/8644" }
gharchive/issue
Redis through envoy gets high response time with redis-benchmark Hi, I’m deploying envoy redis in our environment and I’ve found out that it’s increasing the RTT of the 95 percentile requests to redis in about 7~ms, I have run redis-benchmark tool in both configurations, and through envoy the higher percents getting a much higher response time. results attached. this is the configuration file: static_resources: listeners: - address: socket_address: address: <IP_ADDRESS> port_value: 50051 filter_chains: - filters: - name: envoy.http_connection_manager config: codec_type: auto stat_prefix: ingress_http route_config: name: local_route virtual_hosts: - name: local_service domains: - "*" routes: - match: prefix: "/" route: cluster: local_service_grpc http_filters: - name: envoy.router config: {} - name: redis_nrt_listener address: socket_address: address: <IP_ADDRESS> port_value: 6379 filter_chains: - filters: - name: envoy.redis_proxy typed_config: "@type": type.googleapis.com/envoy.config.filter.network.redis_proxy.v2.RedisProxy stat_prefix: egress_redis settings: op_timeout: 0.03s enable_redirection: true enable_hashtagging: true prefix_routes: catch_all_route: cluster: redis_nrt_cluster - name: redis_vol_listener address: socket_address: address: <IP_ADDRESS> port_value: 6380 filter_chains: - filters: - name: envoy.redis_proxy typed_config: "@type": type.googleapis.com/envoy.config.filter.network.redis_proxy.v2.RedisProxy stat_prefix: egress_redis settings: op_timeout: 0.03s enable_redirection: true enable_hashtagging: true prefix_routes: catch_all_route: cluster: redis_vol_cluster clusters: - name: local_service_grpc connect_timeout: 0.250s type: logical_dns lb_policy: round_robin http2_protocol_options: {} health_checks: - timeout: 1s interval: 3s interval_jitter: 1s unhealthy_threshold: 3 healthy_threshold: 3 tcp_health_check: send: receive: [] hosts: - socket_address: address: router-us-east4-b-prod.ocddx.com port_value: 50051 - name: redis_vol_cluster connect_timeout: 1s type: strict_dns # static lb_policy: MAGLEV load_assignment: cluster_name: redis_cluster endpoints: - lb_endpoints: - endpoint: address: socket_address: address: redis-us-east4-b-prd.ocddx.com port_value: 6379 - name: redis_nrt_cluster connect_timeout: 1s type: strict_dns # static lb_policy: MAGLEV load_assignment: cluster_name: redis_cluster endpoints: - lb_endpoints: - endpoint: address: socket_address: address: redis-nrt-us-east4-b-prd.ocddx.com port_value: 6379 admin: access_log_path: "/var/log/envoy_admin_access.log" address: socket_address: address: <IP_ADDRESS> port_value: 9901 redis-benchmark running from container to the envoy sidecar: redis-benchmark -h collector_envoy -t set,get ====== SET ====== 100000 requests completed in 7.06 seconds 50 parallel clients 3 bytes payload keep alive: 1 4.92% <= 1 milliseconds 59.69% <= 2 milliseconds 73.69% <= 3 milliseconds 82.54% <= 4 milliseconds 88.11% <= 5 milliseconds 91.77% <= 6 milliseconds 94.26% <= 7 milliseconds 95.89% <= 8 milliseconds 97.18% <= 9 milliseconds 98.01% <= 10 milliseconds 98.58% <= 11 milliseconds 98.97% <= 12 milliseconds 99.22% <= 13 milliseconds 99.43% <= 14 milliseconds 99.58% <= 15 milliseconds 99.73% <= 16 milliseconds 99.78% <= 17 milliseconds 99.81% <= 18 milliseconds 99.88% <= 19 milliseconds 99.90% <= 20 milliseconds 99.90% <= 21 milliseconds 99.93% <= 22 milliseconds 99.97% <= 23 milliseconds 99.98% <= 24 milliseconds 99.98% <= 28 milliseconds 100.00% <= 29 milliseconds 14164.31 requests per second ====== GET ====== 100000 requests completed in 7.66 seconds 50 parallel clients 3 bytes payload keep alive: 1 4.18% <= 1 milliseconds 55.05% <= 2 milliseconds 71.89% <= 3 milliseconds 81.03% <= 4 milliseconds 87.10% <= 5 milliseconds 90.96% <= 6 milliseconds 93.66% <= 7 milliseconds 95.49% <= 8 milliseconds 96.44% <= 9 milliseconds 97.22% <= 10 milliseconds 97.75% <= 11 milliseconds 98.34% <= 12 milliseconds 98.83% <= 13 milliseconds 99.15% <= 14 milliseconds 99.34% <= 15 milliseconds 99.49% <= 16 milliseconds 99.59% <= 17 milliseconds 99.69% <= 18 milliseconds 99.72% <= 19 milliseconds 99.81% <= 20 milliseconds 99.87% <= 21 milliseconds 99.89% <= 22 milliseconds 99.91% <= 23 milliseconds 99.92% <= 24 milliseconds 99.93% <= 25 milliseconds 99.96% <= 26 milliseconds 99.97% <= 27 milliseconds 99.97% <= 28 milliseconds 99.98% <= 29 milliseconds 99.98% <= 34 milliseconds 99.98% <= 35 milliseconds 99.99% <= 36 milliseconds 100.00% <= 37 milliseconds 13063.36 requests per second redis-benchmark running from container to a single redis host (part of a masters cluster): bash-4.4# redis-benchmark -h <IP_ADDRESS> -t set,get ====== SET ====== 100000 requests completed in 4.95 seconds 50 parallel clients 3 bytes payload keep alive: 1 11.96% <= 1 milliseconds 94.58% <= 2 milliseconds 96.80% <= 3 milliseconds 98.02% <= 4 milliseconds 98.75% <= 5 milliseconds 99.27% <= 6 milliseconds 99.58% <= 7 milliseconds 99.75% <= 8 milliseconds 99.89% <= 9 milliseconds 99.96% <= 10 milliseconds 99.97% <= 11 milliseconds 99.97% <= 12 milliseconds 99.99% <= 13 milliseconds 100.00% <= 13 milliseconds 20197.94 requests per second ====== GET ====== 100000 requests completed in 5.07 seconds 50 parallel clients 3 bytes payload keep alive: 1 15.41% <= 1 milliseconds 90.76% <= 2 milliseconds 95.60% <= 3 milliseconds 97.79% <= 4 milliseconds 98.70% <= 5 milliseconds 99.23% <= 6 milliseconds 99.60% <= 7 milliseconds 99.76% <= 8 milliseconds 99.82% <= 9 milliseconds 99.83% <= 11 milliseconds 99.87% <= 12 milliseconds 99.89% <= 13 milliseconds 99.91% <= 14 milliseconds 99.97% <= 15 milliseconds 99.99% <= 16 milliseconds 100.00% <= 16 milliseconds 19704.43 requests per second As you can see there's a huge different between the response times, I've been trying to change some configuration for example: type to logical_dns instead of strict_dns, remove the lb_type and add max_buffer_size_before_flush and buffer_flush_timeout and even change the dns to point to only one member of the redis cluster, the same host I checked in the second test, to ensure the reliability of the redis-benchmark test. I'd be glad if someone who using redis with envoy will do the same test I did and share the results, and if someone has any recommendations to solve this response time issue @zuercher please add a BUG label, seems like its happening also in other environments Are you introducing a network hop in the envoy case in your test? Here's the result for my test against a local docker container: root@8362fe3593b4:/# redis-benchmark -h redis-server -p 7001 -t get,set ====== SET ====== 100000 requests completed in 3.07 seconds 50 parallel clients 3 bytes payload keep alive: 1 84.07% <= 1 milliseconds 99.09% <= 2 milliseconds 99.84% <= 3 milliseconds 99.93% <= 4 milliseconds 99.94% <= 5 milliseconds 99.96% <= 6 milliseconds 99.96% <= 9 milliseconds 99.96% <= 10 milliseconds 99.97% <= 11 milliseconds 99.99% <= 26 milliseconds 100.00% <= 27 milliseconds 100.00% <= 27 milliseconds 32626.43 requests per second ====== GET ====== 100000 requests completed in 3.04 seconds 50 parallel clients 3 bytes payload keep alive: 1 85.27% <= 1 milliseconds 98.79% <= 2 milliseconds 99.85% <= 3 milliseconds 99.97% <= 4 milliseconds 99.97% <= 9 milliseconds 99.99% <= 10 milliseconds 100.00% <= 11 milliseconds 32894.74 requests per second root@8362fe3593b4:/# redis-benchmark -p 6381 -t get,set ====== SET ====== 100000 requests completed in 3.78 seconds 50 parallel clients 3 bytes payload keep alive: 1 0.00% <= -32 milliseconds 0.01% <= -30 milliseconds 0.02% <= -29 milliseconds 0.03% <= -28 milliseconds 0.03% <= -26 milliseconds 0.04% <= 0 milliseconds 24.19% <= 1 milliseconds 89.52% <= 2 milliseconds 99.04% <= 3 milliseconds 99.78% <= 4 milliseconds 99.82% <= 5 milliseconds 99.84% <= 6 milliseconds 99.85% <= 7 milliseconds 99.87% <= 8 milliseconds 99.89% <= 10 milliseconds 99.94% <= 11 milliseconds 99.96% <= 12 milliseconds 99.97% <= 13 milliseconds 99.98% <= 14 milliseconds 99.98% <= 15 milliseconds 99.98% <= 16 milliseconds 99.99% <= 19 milliseconds 100.00% <= 20 milliseconds 26462.03 requests per second ====== GET ====== 100000 requests completed in 3.05 seconds 50 parallel clients 3 bytes payload keep alive: 1 28.07% <= 1 milliseconds 96.50% <= 2 milliseconds 99.51% <= 3 milliseconds 99.80% <= 4 milliseconds 99.95% <= 5 milliseconds 100.00% <= 6 milliseconds 100.00% <= 6 milliseconds 32797.64 requests per second Here's the relevant section of envoy.yaml: static_resources: listeners: - name: listener_1 address: socket_address: address: <IP_ADDRESS> port_value: 6381 filter_chains: filters: name: envoy.redis_proxy config: stat_prefix: redis_stats prefix_routes: catch_all_route: cluster: cluster_1 settings: op_timeout: 5s clusters: - name: cluster_1 connect_timeout: 0.25s lb_policy: RING_HASH hosts: - socket_address: address: redis-server port_value: 7001 type: STRICT_DNS envoy is a sidecar container on the application I ran both tests on the envoy container to get an apple to apple measurement. @mosespx i am facing the same issue. Did you got any solution or workaround? @saagar241290 no, I didn't use this solution because of this issue. please share here if you find something interesting @mosespx I tried by increasing number of connections of redis pool to 100 and it gave me a better performance. Earlier there was only a single connection. @saagar241290 when you says tried by increasing number of connections of redis pool to 100 and it gave me a better performance. Earlier there was only a single connection. Are these connections on client?
2025-04-01T06:38:34.135706
2019-11-14T14:54:41
522910625
{ "authors": [ "sriduth" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5663", "repo": "envoyproxy/envoy", "url": "https://github.com/envoyproxy/envoy/issues/9025" }
gharchive/issue
pre startup checks for windows: log if system variables need changing pre startup checks: Check os variables before startup on windows Description: As a fix for issue https://github.com/envoyproxy/envoy/issues/7130, PR #8600 introduces an interface and a posix implementation that runs platform specific checks before startup. The win32 implementation source/exe/win32/platform_checks.cc adds a no-op which should be implemented. #7130 is fixed via #9098, which just adds user documentation instead of moving checks into envoy source code.
2025-04-01T06:38:34.148945
2019-12-16T06:44:54
538217268
{ "authors": [ "alexburnos", "htuch", "kyessenov", "lambdai", "mandarjog", "mattklein123" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5664", "repo": "envoyproxy/envoy", "url": "https://github.com/envoyproxy/envoy/issues/9358" }
gharchive/issue
API Request: Add HttpFilter.instance_name HttpFilter.name is used to instantiate a specific type of filter by http_connection_manager In that sense HttpFilter.name is the className parameter. In filters like WASM and lua, a filter with the same name is deployed multiple times. This makes it difficult to identify a specific filter instance without resorting to peeking inside HttpFilter.typed_config. Add message HttpFilter { // The name of the filter to instantiate. The name must match a // :ref:`supported filter <config_http_filters>`. string name = 1 [(validate.rules).string = {min_bytes: 1}]; string instance_name = 5; // ... } Using instance_name, the filter can be identified in a meaningful way. In the following example, without instance_name=wasm.stats both filters have the same name envoy.filters.http.wasm. filters: - name: envoy.http_connection_manager typed_config: '@type': type.googleapis.com/envoy.config.filter.network.http_connection_manager.v2.HttpConnectionManager forward_client_cert_details: APPEND_FORWARD generate_request_id: true http_filters: - instance_name: wasm.stats name: envoy.filters.http.wasm config: config: root_id: stats_inbound vm_config: code: inline_string: envoy.wasm.stats runtime: envoy.wasm.runtime.null vm_id: stats_inbound - instance_name: wasm.metadata_exchange name: envoy.filters.http.wasm config: config: vm_config: code: inline_string: envoy.wasm.metadata_exchange runtime: envoy.wasm.runtime.null Here @kyessenov @rshriram My initial sniff test is that this is not generally useful, and I don't see why it's that bad to peak inside the typed_config, but happy to be swayed if there is a convincing argument. I think this would be useful for Filter Config discovery service. Having a common field across all filter configs helps in discovery and monitoring IMHO. I don't think it's possible to express that Any in the typed config must have a specific field, so pulling it up one level to filter config seems reasonable: # static config name: envoy.http.wasm config_name: wasm1 typed_config: {} # dynamic config name: envoy.http.wasm config_name: wasm1 config_source: ads: {} In the case of FCDS, we will be moving config up to a oneof, and then the config itself will have to have a name and a config source, so I think it would be covered there? Yeah, it can be done either way. But it would help us to have config name to be a peer of typed_config instead of being nested in a oneof peer. This is because we run multiple transformation passes, and having a name in the xDS helps with identifying the config/filter instance. This is consistent with the rest of xDS where each resource has a name in its proto. @htuch any thoughts on ^? I think this is useful in the context of FCDS, WASM and tooling (or control planes) that operate on opaque config (i.e. they can't peek inside). Here's an interesting thought for v3; since we will no longer have untyped Struct, and will have a world of only TypeStruct and Any, and every extension should have its own unique config proto, we should in face be able to get rid of the need to have any filter type. I.e. you don't need to write envoy.wasm, your use of the WASM config option implies that via the embedded type URL. This means that the name field could be used arbitrarily for user purposes in v3+. @htuch This would work if every filter config is unique per filter. I think there may be cases where two filters share a proto for the config. Not sure if that's something dis-allowed already. One more use case for the control plane to operate on this config in an opaque way is to be able to do partial ordering between different filters. I.e. if control plane is provided with 10 "envoy.wasm" filters, there is no meaningful way to describe a relative order between them. @htuch, just having "name" as a unique arbitrary name and rely on type for the actual type in v3 works as well. Partial order is one of the motivating use case here. Peeking inside requires specific knowledge of the filter. Changing the meaning of “name” from type_name to unique_name seems risky, though it will work. If we add instance_name / config_name field we can actually add it to both v2 and v3 For xDS v2, we can relax the constraint that the config message must match the filter name (as long as it's not a regular struct). That would allow arbitrary names without a breaking change. This instance_name which will be referenced by FCDS. Looks like in the cpp code the expression would be resource.XX_name(). Maybe calling it resource_name()? It would be good to have a v2 xDS solution here, but we need to be very clear what the semantics are if we reuse the field, i.e. it should only be possible if fully unambiguously typed configuration is otherwise present. It also might surprise some folks, as they may have built validators in their config pipeline to ensure consistency of name and config. We could also add a filter_type field that unambiguously denotes the type. If the type field is specified then name can be free form. @htuch In the linked PR, we can infer the name of the extension from the protobuf type for most cases. There are just two exceptions: Empty, which is being solved separately. Migrated APIs. I think we need this information regardless, but it seems clear that all versions of configs should be distinct. I agree about the surprise effect. Fortunately that only happens with the invalid config, e.g. some invalid config might become valid since name is not significant. I actually quite like the name/typed_config pattern. It applies in many places across the code base. @kyessenov what do you reckon the state of this issue is? I think this is resolved. If typed_config is used with the extension-specific type, name can be set to anything. I've updated the unit tests https://github.com/envoyproxy/envoy/pull/10071, https://github.com/envoyproxy/envoy/pull/10122, https://github.com/envoyproxy/envoy/pull/10130 Ack, thanks @kyessenov for the rad contribution here, closing.
2025-04-01T06:38:34.153165
2020-11-04T00:44:01
735729040
{ "authors": [ "asraa", "zasweq" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5665", "repo": "envoyproxy/envoy", "url": "https://github.com/envoyproxy/envoy/pull/13891" }
gharchive/pull-request
[fuzz] Got rid of all uninteresting call logs in health check fuzzing Signed-off-by: Zach Reyes<EMAIL_ADDRESS>Commit Message: Got rid of all uninteresting call logs in health check fuzzing Additional Description: Uninteresting call logs were slowing down my health check fuzzer in oss-fuzz. This PR gets rid of all uninteresting call logs by wrapping applicable objects in NiceMocks. However, since at the beginning of my internship I refactored the unit tests to also use fuzz tests, I put the unit test classes back in test/common/upstream:health_checker_impl_test.cc, and renamed test/common/upstream:health_checker_impl_test_utils to health_check_fuzz_test_utils. No loss in coverage over source/common/upstream/health_checker_impl.cc. Speed up to 30 exec/sec on my cloudtop instance. Risk Level: Low /assign @asraa @htuch @adisuissa Thanks! What was the before of the speed on cloudtop? Honestly Asra it was same speed, 30 exec/sec.
2025-04-01T06:38:34.156560
2021-06-13T13:14:01
919796648
{ "authors": [ "dmitri-d", "phlax" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5666", "repo": "envoyproxy/envoy", "url": "https://github.com/envoyproxy/envoy/pull/16959" }
gharchive/pull-request
docs: Use yaml build config for vars Signed-off-by: Ryan Northey<EMAIL_ADDRESS>Commit Message: docs: Use yaml build config for vars Additional Description: Another breakout from #15229 This adds a build configuration file for vars passed through to sphinx Risk Level: Testing: Docs Changes: Release Notes: Platform Specific Features: [Optional Runtime guard:] [Optional Fixes #Issue] [Optional Deprecated:] [Optional API Considerations:] I think this looks good, other than the question re: descriptor_path parameter in the validating code block. lgtm
2025-04-01T06:38:34.159873
2021-07-23T16:48:45
951740251
{ "authors": [ "keith", "moderation" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5667", "repo": "envoyproxy/envoy", "url": "https://github.com/envoyproxy/envoy/pull/17466" }
gharchive/pull-request
bazel: remove old luajit workaround According to https://luajit.org/install.html Important: this relates to LuaJIT 2.0 only — use LuaJIT 2.1 to avoid these complications. Since we have updated past 2.1 we shouldn't need these anymore which is great since it breaks on Apple Silicon https://github.com/envoyproxy/envoy/issues/16482#issuecomment-846439439 Signed-off-by: Keith Smiley<EMAIL_ADDRESS> @moderation wdyt? LGTM. I commented these lines out when I got M1 building a while back - https://github.com/envoyproxy/envoy/issues/16482#issuecomment-846439439 Removing as we don't require makes sense. MacOS CI failing however Yea I just noticed that we can probably remove them instead. Turns out I can't let the options fallthrough, hopefully green now
2025-04-01T06:38:34.166867
2021-08-03T18:47:49
959392346
{ "authors": [ "alyssawilk", "ggreenway", "rgs1" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5668", "repo": "envoyproxy/envoy", "url": "https://github.com/envoyproxy/envoy/pull/17577" }
gharchive/pull-request
conn pool: default enable runtime feature conn_pool_delete_when_idle Signed-off-by: Greg Greenway<EMAIL_ADDRESS> Commit Message: This enables the new behavior (clean up conn pools when they're idle, to avoid leaking memory in some configurations) from #17403 by default. It can still be disabled by setting runtime feature envoy.reloadable_features.conn_pool_delete_when_idle to false. Additional Description: Risk Level: Medium Testing: Test coverage was added in #17403. Docs Changes: Release Notes: Added in #17403 Platform Specific Features: [Optional Runtime guard:] [Optional Fixes #Issue] [Optional Deprecated:] [Optional API Considerations:] have we smoke tested somewhere yet? have we smoke tested somewhere yet? It's still the same code (minus one possible crash when a cluster is removed via CDS) that @rgs1 smoke tested awhile ago. have we smoke tested somewhere yet? It's still the same code (minus one possible crash when a cluster is removed via CDS) that @rgs1 smoke tested awhile ago. ... tested with the new tcp conn pool, whereas the additional crashers were with the old pool fwiw ... Ah cool, didn't realize the prior version had been canaried. Just to check my memory, the folks encountering tcp proxy crashes didn't provide additional data, and agreed they should switch back to the new pool in any case right? If so LGTM-as-long-as-you-cc-them because it's as safe as it's going to get (folks shouldn't be using the old pool without informing us the new one is problematic) @bianpengyuan FYI this change, that you reported a crash in #16948, is being reintroduced. Looking at that report again, it's very possible that it was the same crash fixed in #17522. Not enough information to know for sure, but it's a possible match, so it may be fixed. coverage test flake; unrelated: 2021-08-03T20:04:16.6349000Z test/extensions/transport_sockets/starttls/starttls_integration_test.cc:329: Failure 2021-08-03T20:04:16.6350199Z Value of: test_server_->server().listenerManager().numConnections() 2021-08-03T20:04:16.6350878Z Expected: is equal to 1 2021-08-03T20:04:16.6351556Z Actual: 0 (of type unsigned long) 2021-08-03T20:04:16.6352267Z Stack trace: 2021-08-03T20:04:16.6352891Z 0x454827: (unknown) 2021-08-03T20:04:16.6353611Z 0x7f6ad1696d96: testing::internal::HandleSehExceptionsInMethodIfSupported<>() 2021-08-03T20:04:16.6354475Z 0x7f6ad167b701: testing::internal::HandleExceptionsInMethodIfSupported<>() 2021-08-03T20:04:16.6355200Z 0x7f6ad1663042: testing::Test::Run() 2021-08-03T20:04:16.6355864Z 0x7f6ad1663b58: testing::TestInfo::Run() 2021-08-03T20:04:16.6356462Z ... Google Test internal frames ...``` /retest /retest
2025-04-01T06:38:34.169363
2018-11-26T16:51:50
384420243
{ "authors": [ "rshriram" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5669", "repo": "envoyproxy/envoy", "url": "https://github.com/envoyproxy/envoy/pull/5117" }
gharchive/pull-request
reformat dynamic metadata emitted by Mongo proxy Description: Emit metadata as map<resource, list(operations> so that it can be used in metadata matchers easily. The existing format (messages:list(structs)) is too hard to represent in metadata matchers. Risk Level: LOW Testing: Unit tests Signed-off-by: Shriram Rajagopalan<EMAIL_ADDRESS>cc @venilnoronha @dio the PR that implemented this was merged yesterday :). So users have not seen this stuff yet. So version history doesn't exist..
2025-04-01T06:38:34.176495
2023-10-14T01:05:39
1942846397
{ "authors": [ "arkodg", "shahar-h", "tmsnan" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5670", "repo": "envoyproxy/gateway", "url": "https://github.com/envoyproxy/gateway/issues/1966" }
gharchive/issue
Global downstream connection limits. Add a overload manager configuration to limit the total number of downstream connections https://www.envoyproxy.io/docs/envoy/latest/configuration/operations/overload_manager/overload_manager#limiting-active-connections It appears that the functionality is incomplete and unsuitable for production use. Should we enable it? @tmsnan im fine waiting, but atm we dont have any way of enabling overload manager @arkodg If possible, I could first add other overload manager features that are already fully supported. @tmsnan sure, imo the others like heap size based will require an API addition, putting the burden of getting it right on the user . Was trying to use this GH issue to enable sensible defaults @arkodg Maybe we can refer to Google VRP edge server configuration. `overload_manager: refresh_interval: 0.25s resource_monitors: name: "envoy.resource_monitors.fixed_heap" typed_config: "@type": type.googleapis.com/envoy.extensions.resource_monitors.fixed_heap.v3.FixedHeapConfig TODO: Tune for your system. max_heap_size_bytes:<PHONE_NUMBER> # 2 GiB actions: name: "envoy.overload_actions.shrink_heap" triggers: name: "envoy.resource_monitors.fixed_heap" threshold: value: 0.95 name: "envoy.overload_actions.stop_accepting_requests" triggers: name: "envoy.resource_monitors.fixed_heap" threshold: value: 0.98` https://www.envoyproxy.io/docs/envoy/latest/configuration/best_practices/edge#best-practices-edge Please assign me
2025-04-01T06:38:34.183618
2024-01-16T20:35:09
2084862185
{ "authors": [ "arkodg", "jaynis", "zhaohuabing" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5671", "repo": "envoyproxy/gateway", "url": "https://github.com/envoyproxy/gateway/pull/2453" }
gharchive/pull-request
feat(cors): Allowed more wildcard options A few weeks ago the allowed CORS origins have been changed from a regex to a wildcard notation (#2389). Implementation wise all kinds of wildcards are supported, however, the validation regex on the SecurityPolicy CRD limits the CORS options to hostnames prefixed with an wildcard followed by a dot, allowing all subdomains of that host. This reduces the freedom when allowing cross origins a lot compared to how it was before. This PR aims to relax the validation regex a bit to enable the following use cases: Allowing all hosts of an specific scheme (https://*) Allowing all hosts regardless of the scheme (*) Allowing all ports of a specific host (http://localhost:*) While allowing all hosts in the context of CORS might sound a bit hacky, this is sometimes required. For instance when a web service provides an API which is consumed by many third-party web applications hosted under arbitrary domains not under the control of the maintainer of aforementioned web service. In addition to that it can be very useful during application development. This is why I have added the option to allow all ports of a specific host as well. Review the new and the old validation regexes. @jaynis can sign your commits and repush ? DCO is failing @jaynis Thanks for the improvement in the wildcard host matching. The implementation looks good to me. I only have a little hesitation about the port wildcard matching. Suffix/Port wildcard matching is not a common practice for hostnames. Do you have many ports for a given hostname? Thank you for your review @zhaohuabing. I only have a little hesitation about the port wildcard matching. Suffix/Port wildcard matching is not a common practice for hostnames. In your use case, do you have many ports for a given hostname? The port range matching was solely meant to be a dev feature so that one can configure CORS for a host (e.g. localhost) regardless of the port the application runs on. But this scenario could be covered by the general wildcard as well, therefore I would also be fine with deleting it again if you think it is not required. Just let me know your preference. Thank you for your review @zhaohuabing. I only have a little hesitation about the port wildcard matching. Suffix/Port wildcard matching is not a common practice for hostnames. In your use case, do you have many ports for a given hostname? The port range matching was solely meant to be a dev feature so that one can configure CORS for a host (e.g. localhost) regardless of the port the application runs on. But this scenario could be covered by the general wildcard as well, therefore I would also be fine with deleting it again if you think it is not required. Just let me know your preference. Prefer to remove the suffix matching to keep it aligned with the common practice. Thanks.
2025-04-01T06:38:34.223972
2015-08-26T11:23:29
103241761
{ "authors": [ "kbs12e" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5672", "repo": "enyojs/moonstone", "url": "https://github.com/enyojs/moonstone/pull/2465" }
gharchive/pull-request
Enyo 2171 Add tabindex for Item to prevent accessibility timing issue Issue screen reader sometimes does not read child component's content when Item is focused. Cause Item may have components as child. However, Sometimes screen reader does not read child component's content because time to receive tabindex is late than child. To prevent this timing issue, I add tabindex to Item. Fix Add tabindex in ariaObservers. Enyo-DCO-1.1-Signed-off-by: Bongsub Kim<EMAIL_ADDRESS> I will re-create PR with latest code.
2025-04-01T06:38:34.277567
2021-09-20T10:37:52
1000814437
{ "authors": [ "Vovanella95", "dtaskai" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5673", "repo": "epam-cross-platform-lab/swagger-dart-code-generator", "url": "https://github.com/epam-cross-platform-lab/swagger-dart-code-generator/issues/245" }
gharchive/issue
[QUESTION] Path still gets generated even when told not to Path still gets generated in the swagger.dart file when specified like the following in the build.yaml file: targets: $default: sources: - lib/** - $package$ builders: chopper_generator: options: header: "//Generated code" swagger_dart_code_generator: options: input_folder: "lib/" output_folder: "lib/swagger_generated_code/" exclude_paths: - "/api/mobile/actuator/" Am I using wrong syntax or is this an actual bug? hi @dtaskai , exclude_path and include_path are regex strings. to make your path excluded, you need to add something like this: \/api\/mobile\/actuator\/ Please use Regex validator to check is you String passing or not. For example, you can use this validator: https://regex101.com/ @dtaskai , if something is not clear - please let us know The exclusion didn't work on my project even after using regex syntax, so I have tried it on the example project: Added an exclusion to /rooms swagger_dart_code_generator: options: input_folder: "lib/" output_folder: "lib/swagger_generated_code/" exclude_paths: - "\/rooms" Ran flutter build run build_runner build Then it still generated the code for /rooms @Get(path: '/rooms') Future<chopper.Response<List<Room>>> roomsGet( {@Query('id') required String? id}); Ok good let me check it @dtaskai Yep you're right. We removed it in 2+ version. Let me fix. @dtaskai Please try it on latest version. Also you can put just /rooms to exclude_path. It works. Latest version is 2.1.3+2 Works on both the example and my personal project, thank you!
2025-04-01T06:38:34.280296
2022-11-07T10:45:18
1438145424
{ "authors": [ "Vovanella95", "dfdgsdfg" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5674", "repo": "epam-cross-platform-lab/swagger-dart-code-generator", "url": "https://github.com/epam-cross-platform-lab/swagger-dart-code-generator/issues/483" }
gharchive/issue
[QUESTION] How can I config worker-pool on Chopper? Please describe a problem. The below is Chopper worker poll example. /// inspired by https://github.com/d-markey/squadron_sample/blob/main/lib/main.dart void initSquadron(String id) { Squadron.setId(id); Squadron.setLogger(ConsoleSquadronLogger()); Squadron.logLevel = SquadronLogLevel.all; Squadron.debugMode = true; } Future<void> main() async { /// initialize Squadron before using it initSquadron('worker_pool_example'); final jsonDecodeServiceWorkerPool = JsonDecodeServiceWorkerPool( // Set whatever you want here concurrencySettings: ConcurrencySettings.oneCpuThread, ); /// start the Worker Pool await jsonDecodeServiceWorkerPool.start(); /// Instantiate the JsonConverter from above final converter = JsonSerializableWorkerPoolConverter( { Resource: Resource.fromJsonFactory, }, /// make sure to provide the WorkerPool to the JsonConverter jsonDecodeServiceWorkerPool, ); /// Instantiate a ChopperClient final chopper = ChopperClient( client: client, baseUrl: 'http://localhost:8000', // bind your object factories here converter: converter, errorConverter: converter, services: [ // the generated service MyService.create(), ], /* ResponseInterceptorFunc | RequestInterceptorFunc | ResponseInterceptor | RequestInterceptor */ interceptors: [authHeader], ); /// Do stuff with myService final myService = chopper.getService<MyService>(); /// ...stuff... /// stop the Worker Pool once done jsonDecodeServiceWorkerPool.stop(); } Describe the solution you'd like How can I config worker-poll on Chopper? Hi @dfdgsdfg , Unfortunately I have no experience with WorkerPool. We just generation swagger code. If you have an idea, how to generate it - let us know.
2025-04-01T06:38:34.284128
2024-11-27T10:13:16
2697974744
{ "authors": [ "VolhaBazhkova", "YauheniyaH" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5675", "repo": "epam/ai-dial-chat", "url": "https://github.com/epam/ai-dial-chat/issues/2678" }
gharchive/issue
Background for items in context menu opened from chat header should be blue (now green) EPAM AI DIAL chat version 0.21.0 What happened? Actual: Expected: on the central part items are highlighted using blue color Example: Confidential information [X] I confirm that do not share any confidential information verified on staging successfully
2025-04-01T06:38:34.289251
2024-10-17T17:56:05
2595382305
{ "authors": [ "Derikyan" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:5676", "repo": "epam/ai-dial-chat", "url": "https://github.com/epam/ai-dial-chat/pull/2393" }
gharchive/pull-request
fix(chat): Update the phrase "Not allowed model selected. Please, change the model to proceed" (Issue #2363) Description: Update the phrase "Not allowed model selected. Please, change the model to proceed" Issues: Issue #2363 Checklist: [x] the pull request name complies with Conventional Commits [x] the pull request name starts with fix(<scope>):, feat(<scope>):, feature(<scope>):, chore(<scope>):, hotfix(<scope>): or e2e(<scope>):. If contains breaking changes then the pull request name must start with fix(<scope>)!:, feat(<scope>)!:, feature(<scope>)!:, chore(<scope>)!:, hotfix(<scope>)!: or e2e(<scope>)!: where <scope> is name of affected project: chat, chat-e2e, overlay, shared, sandbox-overlay, etc. [x] the pull request name ends with (Issue #<TICKET_ID>) (comma-separated list of issues) [x] I confirm that do not share any confidential information like API keys or any other secrets and private URLs /deploy-review /deploy-review