added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T06:38:29.869849
| 2021-05-11T06:12:28
|
886159117
|
{
"authors": [
"weichou1229"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5561",
"repo": "edgexfoundry/edgex-go",
"url": "https://github.com/edgexfoundry/edgex-go/pull/3453"
}
|
gharchive/pull-request
|
feat(notifications): Implement GET /transmission/id/{id} V2 API
Close: #3452
Signed-off-by: weichou<EMAIL_ADDRESS>PR Checklist
Please check if your PR fulfills the following requirements:
[x] Tests for the changes have been added (for bug fixes / features)
[ ] Docs have been added / updated (for bug fixes / features)
If your build fails due to your commit message not passing the build checks, please review the guidelines here: https://github.com/edgexfoundry/edgex-go/blob/master/.github/Contributing.md.
What is the current behavior?
Issue Number: #3452
What is the new behavior?
Implement GET /transmission/id/{id} V2 API according to the doc https://app.swaggerhub.com/apis-docs/EdgeXFoundry1/support-notifications/2.x#/default/get_transmission_id__id_
Does this PR introduce a breaking change?
[ ] Yes
[x] No
New Imports
[ ] Yes
[x] No
Specific Instructions
Are there any specific instructions or things that should be known prior to reviewing?
Other information
Rebased.
|
2025-04-01T06:38:29.874023
| 2016-07-16T10:10:09
|
165922811
|
{
"authors": [
"coveralls",
"duksis",
"nscyclone"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5562",
"repo": "edgurgel/tentacat",
"url": "https://github.com/edgurgel/tentacat/pull/91"
}
|
gharchive/pull-request
|
Replace the use of the deprecated Dict module
The Dict module is deprecated.
Coverage remained the same at 90.135% when pulling 6f9925fd925e38535c7078934b2c3e43f3528b29 on nscyclone:fix/replace-dict into 925200ded23aa0009919a3e48a0dae1952398d63 on edgurgel:master.
thank you @nscyclone
|
2025-04-01T06:38:29.876099
| 2016-10-15T20:35:38
|
183231569
|
{
"authors": [
"edinc",
"munkiepus"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5563",
"repo": "edinc/vagrant-jenkins",
"url": "https://github.com/edinc/vagrant-jenkins/issues/1"
}
|
gharchive/issue
|
Remove the Jenkins setup wizard
On first provisioning of the vagrant box when running the Jenkins server the user has to go through a security setup wizard. This should be removed to further speed up the process of having the Jenkins ready.
I've done this before so might be able to help, it's a case of adding a lot of jenkins config files and is quite verbose
Any contributions are welcome really.
This was resolved with #10.
Thank you for all your contributions @munkiepus. I will add you as a contributor to the repository so that you can directly create branches and review and merge future PR-s.
|
2025-04-01T06:38:29.891641
| 2019-05-05T17:50:27
|
440471118
|
{
"authors": [
"He-Ro",
"edkolev",
"khalsah",
"secuvim",
"shrop"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5564",
"repo": "edkolev/tmuxline.vim",
"url": "https://github.com/edkolev/tmuxline.vim/pull/104"
}
|
gharchive/pull-request
|
Adapt generated tmux config to new tmux 2.9 syntax
Closes #102
Closes #100
Inspired by @secuvim in #102
I do not know if this is all that is needed, but it worked for me.
This would break the backwards compatibility for users with tmux versions < 1.9.
To keep the backwards compatibility you could wrap your changes in something like:
let tmux_version = system("tmux -V")
if (tmux_version < 'tmux 1.9')
let misc_options = ...
let win_options = ...
else
let misc_options = ...
let win_options = ...
elseif
This is just a suggestion as I am not the maintainer of this package.
@He-Ro Thanks for implementing a fix.
@edkolev What is your opinion on this?
Some of the presets appear to write into win_options['window-status-activity-attr'] which causes problems if you're using one of those presets:
https://github.com/edkolev/tmuxline.vim/blob/c8a0295eb34bf11447779a5a203fd472147788a7/autoload/tmuxline/presets/powerline.vim#L26
https://github.com/edkolev/tmuxline.vim/blob/c8a0295eb34bf11447779a5a203fd472147788a7/autoload/tmuxline/presets/nightly_fox.vim#L15
https://github.com/edkolev/tmuxline.vim/blob/c8a0295eb34bf11447779a5a203fd472147788a7/autoload/tmuxline/presets/crosshair.vim#L23
@He-Ro amazing work on this, thanks! Could you also:
git grep and change everywhere window-status-activity-attr => window-status-activity-style
add yourself to the CONTRIBUTORS.md in the root of the project
@khalsah good catch!
This would break the backwards compatibility for users with tmux versions < 1.9.
@edkolev What is your opinion on this?
I'm fine with breaking compatibility for tmux <1.9 - it's been more than 5 years since 1.9 was released. Adding a conditional in the vim script would not work when you share the generated tmux conf file between machines with different versions of tmux.
Added the requested changes.
Also found a mention of attr in the README.
Thanks for working on this. Just noting https://github.com/tmux/tmux/wiki/FAQ#how-do-i-translate--fg--bg-and--attr-options-into--style-options and curious if this was a known issue here.
and curious if this was a known issue here
Could you clarify the question? The link is pointing to a wiki entry about migration; the wiki entry isn't a know issue
Gotcha, maybe this issue isn't related. I had to manually fix up some tmuxline generated settings to comply with the Tmux 2.9 Syntax ralated to -style and -attr. If thosed fixes have been merged in, I can try to run a new snapshot and see if that works.
This PR addresses exactly this - the *-fg/bg to *-style migration. And yes, you should be able to create a snapshot which works with the lates tmux.
|
2025-04-01T06:38:29.954472
| 2022-06-16T12:49:27
|
1273520714
|
{
"authors": [
"Jerome-Liger",
"edsnider"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5566",
"repo": "edsnider/latestversionplugin",
"url": "https://github.com/edsnider/latestversionplugin/pull/44"
}
|
gharchive/pull-request
|
Workaround to repair IsLastestVersion for Android.
Quick implementation from Nick Kovalsky workaround :
https://stackoverflow.com/questions/72407251/how-to-get-version-number-of-application-from-play-store-using-xamarin-forms/72643625#72643625
Fixes #43 #42 .
Changes proposed in this pull request:
Repair plugin for Android Platform.
Thank you for the PR.
Please remove the second commit - a test project should not be included in this PR, just the fix. That commit is also breaking the build.
Thank you.
Beta nuget is available here: https://www.nuget.org/packages/Xam.Plugin.LatestVersion/2.1.1-beta.107
|
2025-04-01T06:38:30.028715
| 2020-05-18T02:20:58
|
619861171
|
{
"authors": [
"edx-webhook",
"ericfab179"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5569",
"repo": "edx/TinCanPython",
"url": "https://github.com/edx/TinCanPython/pull/7"
}
|
gharchive/pull-request
|
WIP [BD-6] OEP-18 Compliance and tests Python 3.8
Add setup.py install_requirements definition.
Update travis file using new requirements.
Create requirements folder.
create pip_tools, base, test and travis requirements files.
Add Makefile.
Add makefile upgrade command.
Include requirements files generated using upgrade command.
Add openedx.yaml and include OEPs list.
Add python 3.8 to tests.
Thanks for the pull request, @ericfab179! I've created OSPR-4550 to keep track of it in JIRA. JIRA is a place for product owners to prioritize feature reviews by the engineering development teams.
Feel free to add as much of the following information to the ticket:
supporting documentation
edx-code email threads
timeline information ("this must be merged by XX date", and why that is)
partner information ("this is a course on edx.org")
any other information that can help Product understand the context for the PR
All technical communication about the code itself will still be done via the GitHub pull request interface. As a reminder, our process documentation is here.
@ericfab179 🎉 Your pull request was merged!
Please take a moment to answer a two question survey so we can improve your experience in the future.
|
2025-04-01T06:38:30.036752
| 2015-05-15T18:45:35
|
76818565
|
{
"authors": [
"Nickersoft",
"clintonb",
"jimabramson",
"rlucioni"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5570",
"repo": "edx/ecommerce-api-client",
"url": "https://github.com/edx/ecommerce-api-client/pull/8"
}
|
gharchive/pull-request
|
API client now passes the user's full name to ecommerce
This merge also requires a merge on ecommerce and edx-platform with tyler-jwt-fullname, so keep that in mind.
@clintonb @rlucioni FYI: @wedaly @jimabramson
Tests are required.
@clintonb Will do that now
Minor: Cleanup the commit message. We generally use a simple statement of the overall change as the title (e.g. Added support for passing full name via JWT). Additional information can be provided as a paragraph or list after the title.
Added support for passing full name via JWT
- User's full name can be passed to the API when using JWT authentication. This field is optional.
- Email is no longer required for JWT authentication.
Aside from the message change, :+1: . Please await a second approval from @rlucioni or @jimabramson before merging.
Oh, I actually not aware that Git messages could be multiline like that. I'll change it then force a push. Thanks @clintonb.
Once you get a clean build, :+1: .
@Nickersoft I'm surprised to see a major version bump on this change; was expecting 0.5.0. Is there something backwards-incompatible?
@jimabramson My mistake. The constructor was changed in a backwards-incompatible manner. We can make it backwards-compatible if you'd like.
ah. somehow i missed that.
|
2025-04-01T06:38:30.056334
| 2021-02-17T11:57:10
|
810119276
|
{
"authors": [
"farhan-arshad-dev",
"omerhabib26"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5571",
"repo": "edx/edx-app-android",
"url": "https://github.com/edx/edx-app-android/pull/1515"
}
|
gharchive/pull-request
|
Navigate within App from dates to course component
Description
LEARNER-8214
Navigate within app when a link is clicked on Full Page dates screen
@omerhabib26 The app still navigates to screen for the un-available course unit.
https://user-images.githubusercontent.com/43750646/108349158-78f9bb00-7204-11eb-9459-e5fbb3221edc.mp4
|
2025-04-01T06:38:30.112083
| 2015-12-10T20:00:20
|
121565553
|
{
"authors": [
"clintonb",
"mjfrey",
"rlucioni"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5572",
"repo": "edx/edx-rest-api-client",
"url": "https://github.com/edx/edx-rest-api-client/pull/21"
}
|
gharchive/pull-request
|
Update AUTHORS and bump version
Update to AUTHORS and bump version to 1.3.0
@clintonb @rlucioni
:+1:
:+1:
|
2025-04-01T06:38:30.195930
| 2024-02-15T15:46:57
|
2136866221
|
{
"authors": [
"eeston",
"jgoguen"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5573",
"repo": "eeston/Jello",
"url": "https://github.com/eeston/Jello/issues/9"
}
|
gharchive/issue
|
Add manual control of Repeat and Shuffle
Related to #8 is adding the ability to toggle shuffle and repeat. The play queue should have the following states:
Shuffle on: randomize the order of the queue every time shuffle is turned on.
Shuffle off:
Version 1: It would be fine to just not change the order of the current play queue for V1 of this feature .
Version 2: If playing from a defined playlist (either a playlist, album list, or a future "All Songs" view), turning shuffle off should reorder the queue according to the list display order, with playback continuing from whatever queue position holds the currently playing song. If playing from a queue modified with "Add to Queue", just leave the queue in its current order.
Version 3: If the queue was created with "Play Next" or "Add to Queue", reorder the queue according to the order of whatever view it was started with and play back from whatever queue position the current song is at.
Repeat off: play to the end of the current queue and stop.
Repeat One: continuously play the current song, only changing if the song is manually changed by the user (by choosing a new song or pressing the Next button).
Repeat All: play to the end of the current queue; if Shuffle is on randomize the queue order and start from the beginning, or if shuffle is off just go back to the beginning of the queue and continue playback.
When starting playback by choosing a song or pressing Play, the Shuffle state should be whatever it was previously. When starting playback by pressing Shuffle, the Shuffle state should be turned on.
Defs want to add this. At the moment playback issues are my priority. I would like to have a solid foundation before bolting on features.
|
2025-04-01T06:38:30.228742
| 2019-10-24T16:57:43
|
512066305
|
{
"authors": [
"eMaringolo",
"eftomi"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5582",
"repo": "eftomi/pharo-ado",
"url": "https://github.com/eftomi/pharo-ado/issues/27"
}
|
gharchive/issue
|
Variant type 2 (vt: 2) has to be implemented
GlorpInt2Test
This is solved in this PR https://github.com/tesonep/pharo-com/pull/10
PR merged, this issue can ble closed.
|
2025-04-01T06:38:30.253833
| 2024-04-08T07:34:26
|
2230480118
|
{
"authors": [
"NSystemx"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5583",
"repo": "eggsy/website",
"url": "https://github.com/eggsy/website/issues/56"
}
|
gharchive/issue
|
I created firebase but I don't know how to create music.
I created firebase but I don't know how to create music.
Daily Songs with Firestore
How to create firestore database for daily songs?
|
2025-04-01T06:38:30.260357
| 2020-06-08T14:50:00
|
634690429
|
{
"authors": [
"Kokiro",
"mugiwarafx",
"onjin"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5584",
"repo": "egoist/hack",
"url": "https://github.com/egoist/hack/issues/77"
}
|
gharchive/issue
|
This site hackcss.egoist.moe can’t be reached
Hi,
This site hackcss.egoist.moe can’t be reached :(
Best,
Kokiro
Hi @Kokiro,
today I tried to restore the website with the style guidelines: https://hackcssbckp.herokuapp.com/
hope it helps! 😀
According to CNAME at github-pages the working address is hackcss.egoist.sh
|
2025-04-01T06:38:30.286203
| 2021-12-15T16:16:51
|
1081228264
|
{
"authors": [
"chandrakananandi",
"femtomc",
"mwillsey"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5585",
"repo": "egraphs-good/egg",
"url": "https://github.com/egraphs-good/egg/issues/144"
}
|
gharchive/issue
|
Applying egg to SSA-based IRs
Hi! What are the prospects of utilizing egg to perform rewrite-driven transformations on SSA-based IRs (?)
I read this paper (whose authors appear to be collaborators with y'all): https://www.cs.cornell.edu/~ross/publications/eqsat/ where they describe E-PEG-based program representation.
On the other hand, I'm working on IRs which are closer to strict SSA (where all control flow is de-sugared to basic blocks and branching) and everything is linear (e.g. blocks are vectors of instructions) vs. strictly tree-like.
I'd love to find or understand how equality saturation could be applied to this representation structure -- it's likely that maybe this information can be gleaned from the above paper (but I expect that maintainers here might be able to unpack this more than I can).
Hi! I use egg for a language which is mostly-SSA. In my case, things "just work" so far because I mostly care about straight line blocks. The only additional information I track which may be necessary for other SSA-based usages as well is to make sure you don't use variables before they are defined. This, in my case, was possible with a simple e-class analysis.
Hi, I think @chandrakananandi is right. I don't think there is any major technical blocker stopping you from doing this today!
The original eqsat paper you linked to contains a lot of good ideas on how to encode your problem into an e-graph. If I'm understanding your setting, you'll still want E-PEGs if you want to encode loops in a transparent way that you can optimize through. If you only care about optimizing one basic block (one DFG) at a time, then you don't need them.
The egg paper (and tool) are mostly innovating in how equality saturation is done.
Despite there not being any huge blockers, I still think it's a large and challenging task, and one that I'd like to try to tackle at some point (if I can find the time), or see someone else take a stab at!
One thing that I will add: you'll have a much easier time with a tree- or dag-like IR that a linear, mutating IR. All the rewrites that you do are over trees or dags, and just overall the e-graph doesn't do a lot for you if your language is heavily sequential. If that's your case, consider building def-use chains or some other method to make things more graph-like for the e-graph.
Also, this doesn't seem to be an issue, so I'm converting it to a discussion.
|
2025-04-01T06:38:30.295764
| 2023-09-08T07:35:57
|
1887127447
|
{
"authors": [
"Bastacyclop",
"TrevorHansen",
"mwillsey",
"oflatt"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5586",
"repo": "egraphs-good/extraction-gym",
"url": "https://github.com/egraphs-good/extraction-gym/pull/9"
}
|
gharchive/pull-request
|
An extra bottom-up recursive extractor
Hi,
Before I saw #8 I made a a recursive extractor, too.
This code is much ugglier than #8, but is faster on some problems (for example acyclic graphs). On acyclic graphs it's able to extract in one pass through, without building the dependencies.
Taking the cumulative times for all three extractors on my machine:
Loaded 454 jsons.
...
Cumulative time for bottom-up: 3386ms
Cumulative time for bottom-up-analysis: 2483ms
Cumulative time for bottom-up-recursive: 977ms
bottom-up is the one currently in,
bottom-up-analysis is #8,
bottom-up-recursive is this PR.
Is there benefit in having all three, or should the two new ones be combined?
Good question! @Bastacyclop what do you think? Offhand, I'm leaning toward fewer, more distinct algorithms, so unless we can characterize that these have different advantages, I'd prefer to have just one.
I feel like we should be able to combine the three bottom up versions, with some more experiments and code cleanup. Maybe by (1) doing a first pass without building dependencies; and (2) if dependencies are used, doing a bottom up analysis based on unique queues. The question would be whether using unique queues for the second stage brings performance benefits or not (I think it should: https://github.com/egraphs-good/egg/issues/239). To properly evaluate that I would like to see datasests with more costly to compute, child-dependent cost functions.
PS: one thing to consider is that in egg the dependencies don't need to be computed as they are already stored in the e-graph.
I would like to consider computation of dependencies (parents) as somewhat negligible, as its only linear and as @Bastacyclop says it's already there in many contexts.
I'd like to preserve the "dumb" bottom up extractor as a base case. Ideally, we could consolidate the "smarter" bottom-up extractors into one (i.e., those that do not aim to do cost sharing, as a possible definition). Thoughts on that?
Before I read this closely I also worked on a bottom-up extractor (#20)
Sorry for duplicated work! I would also be happy with consolidating this, #20, and #8 if possible.
With the recent changes to the bottom-up extractor(#20), there's now only a small time advantage for the extractor that I proposed introducing (Note these times differ from before because extra problems have been added). Currently:
Cumulative time for faster-bottom-up: 2060ms [The one in #20]
Cumulative time for bottom-up-recursive: 1533ms [The one in this PR].
Cumulative time for bottom-up: 4471ms
Meaning that the extractor in this PR is only about 25% faster than the others, but is much uglier.
However, there are some tweaks to the "faster-bottom-up" extractor which brings down its runtime to almost the same as the extractor in this PR:
###################################################
faster-bottom-up vs faster-bottom-up-old
extractors: ['faster-bottom-up', 'faster-bottom-up-old']
cumulative time for faster-bottom-up: 1649ms
cumulative time for faster-bottom-up-old: 2060ms
cumulative tree cost for faster-bottom-up:<PHONE_NUMBER>5237
cumulative tree cost for faster-bottom-up-old:<PHONE_NUMBER>5237
cumulative dag cost for faster-bottom-up: 78037
cumulative dag cost for faster-bottom-up-old: 78037
Cumulative time for faster-bottom-up: 1649ms
Cumulative time for faster-bottom-up-old: 2060ms
faster-bottom-up / faster-bottom-up-old
geo mean
tree: 1.0000
dag: 1.0000
micros: 0.8184
quantiles
tree: 1.0000, 1.0000, 1.0000, 1.0000, 1.0000
dag: 1.0000, 1.0000, 1.0000, 1.0000, 1.0000
micros: 0.3611, 0.7825, 0.8227, 0.8611, 1.8333
So I've changed this PR to now just introduce some small speedups to the faster-bottom-up extractor, as well as fixing up attribution to @Bastacyclop.
:+1:
|
2025-04-01T06:38:30.304644
| 2016-07-06T13:53:27
|
164080791
|
{
"authors": [
"chrisdennis",
"lorban"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5587",
"repo": "ehcache/ehcache3",
"url": "https://github.com/ehcache/ehcache3/issues/1292"
}
|
gharchive/issue
|
OffHeapChainMap thieving argument is ignored
OffHeapChainMap is backed by EvictionListeningReadWriteLockedOffHeapClockCache but the latter lacks a constructor that accepts shareByThieving argument.
The offheap store lib should be modified to add such constructor so that OffHeapChainMap can make use of it.
Given how busted ARC is at the moment this turns out to be a non-issue. That said @AbfrmBlr is going to fixing the broken ARC implementation in clustered (that he wrote) and extending it to cover unclustered offheap and disk. He's going to run headlong in to this as a result.
This will be fixed under #2215
|
2025-04-01T06:38:30.352575
| 2019-12-19T01:56:07
|
540026667
|
{
"authors": [
"ppazos",
"testautomation"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5588",
"repo": "ehrbase/ehrbase",
"url": "https://github.com/ehrbase/ehrbase/pull/89"
}
|
gharchive/pull-request
|
ehrbase/project_management#98 added ACTION examples and updated AQL test suite data
@testautomation this includes the action compositions and the queries were updated too.
:heavy_check_mark:
|
2025-04-01T06:38:30.366489
| 2018-02-05T12:04:36
|
294369745
|
{
"authors": [
"eidheim",
"lxlenovostar"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5589",
"repo": "eidheim/Simple-Web-Server",
"url": "https://github.com/eidheim/Simple-Web-Server/issues/207"
}
|
gharchive/issue
|
glaceful close ssl socket
Hi: @eidheim
void close() noexcept {
error_code ec;
std::unique_lock<std::mutex> lock(socket_close_mutex); // The following operations seems to be needed to run sequentially
socket->lowest_layer().shutdown(asio::ip::tcp::socket::shutdown_both, ec);
socket->lowest_layer().close(ec);
}
In this way, we directly close tcp socket, why we not close SSL socket first in HTTPS server?
Thank you.
Last time I studied this I came to the conclusion that calling ssl::stream::shutdown was not needed. Though I might be wrong! By the way, ssl::stream does not have a close-member function.
Although, thank you for bringing this up. I'll add a couple of labels to this issue.
@eidheim In a rare case, If I don't add mutex before shutdown/close, Segment error happen. Why you choose use mutex before shutdown/close?
|
2025-04-01T06:38:30.385946
| 2019-04-12T19:58:04
|
432721804
|
{
"authors": [
"alpeshjamgade",
"ritzvik"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5590",
"repo": "einsteinpy/einsteinpy",
"url": "https://github.com/einsteinpy/einsteinpy/pull/194"
}
|
gharchive/pull-request
|
Reorder parameters within functions
Fixes #141
Really great contribution @alpeshjamgade . Thanks
I would be reviewing in a day or two. Meanwhile could you make one change.
define a private function _scr(M, c=default value, G=default value for calculating Rs. Putting units for calculating Rs seems unnecessary and slow, and may cause problems while implementing jit for the functions.
Also, I guess you have made some merge commits while rebasing your branch. Try to fix that. Maybe refer https://github.com/k88hudson/git-flight-rules/ .
One word of advice, try to create separate branches for fixing and developing things. The master should not ever diverge from the upstream. Believe me, I have learnt that the hard way.
Really great contribution @alpeshjamgade . Thanks
I would be reviewing in a day or two. Meanwhile could you make one change.
define a private function _scr(M, c=default value, G=default value for calculating Rs. Putting units for calculating Rs seems unnecessary and slow, and may cause problems while implementing jit for the functions.
Also, I guess you have made some merge commits while rebasing your branch. Try to fix that. Maybe refer https://github.com/k88hudson/git-flight-rules/ .
One word of advice, try to create separate branches for fixing and developing things. The master should not ever diverge from the upstream. Believe me, I have learnt that the hard way.
thank you
you asked me to define a private function _scr.
Before the schwarzschild_radius function in schwarzschid_utils.py was getting any M value as input weather it has astropy unit or not and return Rs in astropy units and also 'Rs' was the input to the most functions which was in astropy units only, but when i changed Rs with "M", i needed to make sure that "M" should be in astropy units , that means everytime i call such function i need to provide M value in astropy units otherwise it would throw an error, and we dont expect users to provide M in astropy units, it would be great if they can just give M value as input and if the given function need to calculate Rs, it would need M in astropy units then schwarzschild_radius function will convert M to astropy units, thast just what i did. please tell me if i can do any better.
i think we dont need private _scr function, if we want Rs in floats we can just asked for it with command Rs.value , otherwise it will always be in astropy units. please tell me if i dont understand something.
i think we dont need private _scr function, if we want Rs in floats we can just asked for it with command Rs.value , otherwise it will always be in astropy units. please tell me if i dont understand something.
User always specifies things with units, but during heavy calculation internally, we don't use units. As functions like christoffels() are called thousands of time in a loop to get the trajectory it may be good to save some computation by not using astropy.units. But still wait some time, I am not sure what to do. @shreyasbapat Suggestions??
i have just created another pull request #196 for this issue , this time i didnt messed up my master. sorry for this. i did not consider on which branch i was making changes before.
|
2025-04-01T06:38:30.439609
| 2023-06-14T12:51:13
|
1756838022
|
{
"authors": [
"JonasKunz",
"jackshirazi"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5591",
"repo": "elastic/apm-agent-java",
"url": "https://github.com/elastic/apm-agent-java/issues/3189"
}
|
gharchive/issue
|
Revert workaround for micrometer bug in tests
#3187 added a workaround for a micrometer bug. We should revert this workaround as soon as a micrometer version with a fix is available.
fixed in https://github.com/elastic/apm-agent-java/pull/3264
|
2025-04-01T06:38:30.781952
| 2013-08-09T07:03:51
|
17849524
|
{
"authors": [
"dadoonet",
"hjc1710",
"jacobwgillespie",
"rabidscorpio"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5592",
"repo": "elastic/elasticsearch-cloud-aws",
"url": "https://github.com/elastic/elasticsearch-cloud-aws/issues/31"
}
|
gharchive/issue
|
Update to Elasticsearch 0.90.3
Issue reported in https://groups.google.com/d/msg/elasticsearch/bpdetIvIg5M/cMAUGwhRVf0J
No problem with elasticsearch 0.90.2 and the cloud-aws 1.12.0 plugin (other than the usual node discovery flakiness), but I'm unable to get elasticsearch 0.90.3 to start:
Initialization Failed ...
1) IllegalStateException[This is a proxy used to support circular references involving constructors. The object we're proxying is not constructed yet. Please wait until after injection has completed to use this object.]2) NoSuchMethodError[org.elasticsearch.discovery.zen.ZenDiscovery.<init>(Lorg/elasticsearch/common/settings/Settings;Lorg/elasticsearch/cluster/ClusterName;Lorg/elasticsearch/threadpool/ThreadPool;Lorg/elasticsearch/transport/TransportService;Lorg/elasticsearch/cluster/ClusterService;Lorg/elasticsearch/node/settings/NodeSettingsService;Lorg/elasticsearch/cluster/node/DiscoveryNodeService;Lorg/elasticsearch/discovery/zen/ping/ZenPingService;)V]
Do I need to wait for a new version of cloud-aws, or is there some other problem here?
My /etc/elasticsearch/elasticsearch.yml:
cluster.name: foo
plugin.mandatory: cloud-aws,lang-javascript
cloud:
aws:
access_key: ********
secret_key: ********
region: us-east-1
discovery:
type: ec2
ec2:
ping_timeout: 15s
gateway:
type: s3
s3:
bucket: bar
For anyone (like me) who is seeing this error, my fix was to update to the correct version (as indicated by the README) - I accidentally updated Elasticsearch to 1.5.0 without updating elasticsearch-cloud-aws.
@jacobwgillespie ha, thanks for pointing this out, I did the same thing!
I just ran into this issue today, and it had nothing to do with mis-matched ES/Plugin versions. This error is also thrown if your config values for the cloud-aws plugin are not correct.
What I had:
cloud:
aws:
access_key: XXX
secret: XXX
region: us-east-1
What I needed:
cloud:
aws:
access_key: XXX
secret_key: XXX
region: us-east-1
The major difference between the two of them being cloud.aws.secret in the first, non-working example; which is changed to cloud.aws.secret_key in the second, working example.
So, apparently, this exact same error is what you'll get if you've completely borked your cloud-aws config. So, keep that in mind!
Maybe in a future release, if you could detect a bad config state like this (access key is present, but secret key is missing) and throw an error, that would be pretty great. Alternatively, implementing it so that an error is thrown if any unknown string appears in the cloud.aws namespace would work just as well (so, an error would have been thrown because cloud.aws.secret is not a recognized and valid key), but I don't know how feasible that is to do. Just some food for thought!
Hopefully someone finds this useful!
@hjc1710 I think it's useful and I agree that we should better catch that kind of error.
May be you would like to open an issue in elasticsearch repo now that we moved aws plugin there?
Thanks!
Awesome, thanks @dadoonet! I did not know that this plugin had moved and the official repo is elastic/elasticsearch now. Anyway, I opened up an issue there for this very feature, after doing a bit of rewording and thinking. If you guys need my help for that feature (for whatever reason), I'm happy to help!
|
2025-04-01T06:38:30.794539
| 2019-04-08T10:47:27
|
430381348
|
{
"authors": [
"sl0thentr0py"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5593",
"repo": "elastic/elasticsearch-rails",
"url": "https://github.com/elastic/elasticsearch-rails/issues/873"
}
|
gharchive/issue
|
Adding search_analyzer to mapping
I can't figure out how to add the search_analyzer option to the mapping block. The Mappings class only provides indexes as a way to add fields.
https://www.elastic.co/guide/en/elasticsearch/reference/current/search-analyzer.html
My bad, documentation parsing error from my side.
|
2025-04-01T06:38:30.978241
| 2022-12-22T15:49:57
|
1508117402
|
{
"authors": [
"Drookoo",
"SashaShcherbyna",
"TanguyPatte",
"adrifermo",
"ppatcha"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5594",
"repo": "elastic/helm-charts",
"url": "https://github.com/elastic/helm-charts/issues/1748"
}
|
gharchive/issue
|
elasticsearch.keystore: Device or resource busy
Chart version: 8.5.1
Kubernetes version: 1.23.12-gke.1600
Kubernetes provider: GKE (Google Kubernetes Engine)
Helm Version: version.BuildInfo{Version:"v3.9.3", GitCommit:"414ff28d4029ae8c8b05d62aa06c7fe3dee2bc58", GitTreeState:"clean", GoVersion:"go1.19"}
Describe the bug:
Exception in thread "main" java.nio.file.FileSystemException: /usr/share/elasticsearch/config/elasticsearch.keystore.tmp -> /usr/share/elasticsearch/config/elasticsearch.keystore: Device or resource busy
Steps to reproduce:
Create the secret
kubectl create secret generic elk-backup --from-file=gcs.client.elk-backup.credentials_file=./elk-backup.json
To add these secrets to the keystore:
keystore:
- secretName: elk-backup
Expected behavior:
mount the secret in keystore should work
Provide logs and/or server output (if relevant):
Be careful to obfuscate every secrets (credentials, token, public IP, ...) that could be visible in the output before copy-pasting
Exception in thread "main" java.nio.file.FileSystemException: /usr/share/elasticsearch/config/elasticsearch.keystore.tmp -> /usr/share/elasticsearch/config/elasticsearch.keystore: Device or resource busy
at java.base/sun.nio.fs.UnixException.translateToIOException(UnixException.java:100)
at java.base/sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:106)
at java.base/sun.nio.fs.UnixCopyFile.move(UnixCopyFile.java:420)
at java.base/sun.nio.fs.UnixFileSystemProvider.move(UnixFileSystemProvider.java:266)
at java.base/java.nio.file.Files.move(Files.java:1430)
at org.elasticsearch.common.settings.KeyStoreWrapper.save(KeyStoreWrapper.java:498)
at org.elasticsearch.common.settings.KeyStoreWrapper.save(KeyStoreWrapper.java:412)
at org.elasticsearch.cli.keystore.AddStringKeyStoreCommand.executeCommand(AddStringKeyStoreCommand.java:102)
at org.elasticsearch.cli.keystore.BaseKeyStoreCommand.execute(BaseKeyStoreCommand.java:64)
at org.elasticsearch.common.cli.EnvironmentAwareCommand.execute(EnvironmentAwareCommand.java:54)
at org.elasticsearch.cli.Command.mainWithoutErrorHandling(Command.java:85)
at org.elasticsearch.cli.MultiCommand.execute(MultiCommand.java:94)
at org.elasticsearch.cli.Command.mainWithoutErrorHandling(Command.java:85)
at org.elasticsearch.cli.Command.main(Command.java:50)
at org.elasticsearch.launcher.CliToolLauncher.main(CliToolLauncher.java:64)
Any additional context:
I can seeing this error in pod with role master
I'm not sure but maybe it's because the keystore file is mounted as a subpath in the chart
- name: keystore
mountPath: /usr/share/elasticsearch/config/elasticsearch.keystore
subPath: elasticsearch.keystore
I don't know how elasticsearch-keystore tool works but it looks like it's trying to replace elasticsearch.keystore file by elasticsearch.keystore.tmp file instead of replacing the content of it.
I reproduce the issue by trying to do the following command:
mv elasticsearch.keystore.tmp elasticsearch.keystore
mv: cannot move 'elasticsearch.keystore.tmp' to 'elasticsearch.keystore': Device or resource busy
cp command works just fine
cp elasticsearch.keystore.tmp elasticsearch.keystore
I see 2 possibilities to fix that:
change the way elasticsearch-keystore write content in keystore file
add the possibility to specify a custom path for the keystore file and mount it in is own directory to prevent the use of subpath
Hi @TanguyPatte
seems like there is something wrong with the current statefulset.yaml that causes this issue.
FWIW deploying with this template works well for elasticsearch 8.6.2 : https://github.com/elastic/helm-charts/blob/d4e9f6bc47cf7f7ad4dfaaec102e1327d8a345e3/elasticsearch/templates/statefulset.yaml
but there may be more recent iterations that may work.
Chart version: 8.5.1
Kubernetes version: 1.23.12-gke.1600
Kubernetes provider: GKE (Google Kubernetes Engine)
Helm Version: version.BuildInfo{Version:"v3.9.3", GitCommit:"414ff28d4029ae8c8b05d62aa06c7fe3dee2bc58", GitTreeState:"clean", GoVersion:"go1.19"}
Describe the bug: Exception in thread "main" java.nio.file.FileSystemException: /usr/share/elasticsearch/config/elasticsearch.keystore.tmp -> /usr/share/elasticsearch/config/elasticsearch.keystore: Device or resource busy
Steps to reproduce:
Create the secret
kubectl create secret generic elk-backup --from-file=gcs.client.elk-backup.credentials_file=./elk-backup.json
To add these secrets to the keystore:
keystore:
- secretName: elk-backup
Expected behavior: mount the secret in keystore should work
Provide logs and/or server output (if relevant):
Be careful to obfuscate every secrets (credentials, token, public IP, ...) that could be visible in the output before copy-pasting
Exception in thread "main" java.nio.file.FileSystemException: /usr/share/elasticsearch/config/elasticsearch.keystore.tmp -> /usr/share/elasticsearch/config/elasticsearch.keystore: Device or resource busy
at java.base/sun.nio.fs.UnixException.translateToIOException(UnixException.java:100)
at java.base/sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:106)
at java.base/sun.nio.fs.UnixCopyFile.move(UnixCopyFile.java:420)
at java.base/sun.nio.fs.UnixFileSystemProvider.move(UnixFileSystemProvider.java:266)
at java.base/java.nio.file.Files.move(Files.java:1430)
at org.elasticsearch.common.settings.KeyStoreWrapper.save(KeyStoreWrapper.java:498)
at org.elasticsearch.common.settings.KeyStoreWrapper.save(KeyStoreWrapper.java:412)
at org.elasticsearch.cli.keystore.AddStringKeyStoreCommand.executeCommand(AddStringKeyStoreCommand.java:102)
at org.elasticsearch.cli.keystore.BaseKeyStoreCommand.execute(BaseKeyStoreCommand.java:64)
at org.elasticsearch.common.cli.EnvironmentAwareCommand.execute(EnvironmentAwareCommand.java:54)
at org.elasticsearch.cli.Command.mainWithoutErrorHandling(Command.java:85)
at org.elasticsearch.cli.MultiCommand.execute(MultiCommand.java:94)
at org.elasticsearch.cli.Command.mainWithoutErrorHandling(Command.java:85)
at org.elasticsearch.cli.Command.main(Command.java:50)
at org.elasticsearch.launcher.CliToolLauncher.main(CliToolLauncher.java:64)
Any additional context: I can seeing this error in pod with role master
@SashaShcherbyna
I have got the same error, Did you have a solution to resolve this issue?
I can resolve this issue because I have not set ELASTIC_PASSWORD with my own credential.
Just add
extraEnvs:
- name: ELASTIC_PASSWORD
valueFrom:
secretKeyRef:
name: elastic-config-credentials
key: password
secret:
enabled: false
and then add the keystore just normal.
|
2025-04-01T06:38:32.685694
| 2020-10-29T16:42:53
|
732479122
|
{
"authors": [
"apmmachine",
"vigneshshanmugam"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5595",
"repo": "elastic/synthetics",
"url": "https://github.com/elastic/synthetics/pull/112"
}
|
gharchive/pull-request
|
fix: close fd after debug logs are written
fix #110
runner end should be called after reset since we are logging the output inside reset
Moved from nextTick to setTimeout as its called after all the micro tasks and ensure all log statements are printed before we end stream.
:green_heart: Build Succeeded
the below badges are clickable and redirect to their specific view in the CI or DOCS
Expand to view the summary
Build stats
Build Cause: [Pull request #112 opened]
Start Time: 2020-10-29T16:43:01.167+0000
Duration: 14 min 22 sec
Test stats :test_tube:
Test
Results
Failed
0
Passed
42
Skipped
0
Total
42
|
2025-04-01T06:38:32.690068
| 2021-01-27T13:01:00
|
795081257
|
{
"authors": [
"CLAassistant",
"Ruminateer",
"tizhou86"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5596",
"repo": "elasticdeeplearning/edl",
"url": "https://github.com/elasticdeeplearning/edl/pull/155"
}
|
gharchive/pull-request
|
Fix 2 errors
fix an import typo
add a missing exception definition
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
@gongweibao please review this pr, thanks~
@gongweibao please review this pr, thanks~
|
2025-04-01T06:38:32.703451
| 2020-06-21T14:04:33
|
642569562
|
{
"authors": [
"Bigsheng97",
"iswyq"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5597",
"repo": "elbuco1/CBAM",
"url": "https://github.com/elbuco1/CBAM/issues/1"
}
|
gharchive/issue
|
How to cite this work in my assignment?
Thank you for your work and may I know how should I cite your works in my project?
这里已经写好了构建的
|
2025-04-01T06:38:32.705729
| 2024-01-12T23:19:43
|
2079872843
|
{
"authors": [
"coveralls",
"elchininet"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5598",
"repo": "elchininet/keep-texts-in-tabs",
"url": "https://github.com/elchininet/keep-texts-in-tabs/pull/39"
}
|
gharchive/pull-request
|
Set the docker image tag to run
This pull request makes possible to run a specific Home Assistant docker image tag sending a variable to the script. If the variable is not sent, the version included in .hass/config/.HA_VERSION file will be used as the tag. Also, to generate the docker cache key, the aforementioned file will be used instead of the package.json.
coverage: 100.0%. remained the same
when pulling b723a4b7d1ede517f74d010d5fef0427c01cad17 on run_docker_image_tag_dynamic
into ebd6f452dca960044803e08becec7c7e2efbfd39 on master.
|
2025-04-01T06:38:32.732620
| 2019-12-16T01:34:42
|
538141121
|
{
"authors": [
"aleccool213",
"electerious"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5601",
"repo": "electerious/Ackee",
"url": "https://github.com/electerious/Ackee/issues/72"
}
|
gharchive/issue
|
[Feature Request] Add Deploy to Heroku button to README.md
Feature Request
Hi there! Great application! I really love how simple and privacy focused it is.
It would be great to have a Deploy to Heroku button, would help drive up adoption!
Documentation
A "Deploy to Heroku" button would be cool, but I guess that #73 should be finished first. Help is welcome!
Will be part of v1.4.3.
|
2025-04-01T06:38:32.734729
| 2016-09-23T11:57:07
|
178858243
|
{
"authors": [
"electerious",
"jullit31",
"mhellwig"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5602",
"repo": "electerious/Lychee",
"url": "https://github.com/electerious/Lychee/issues/625"
}
|
gharchive/issue
|
Feature request: make the auto-fullscreen behaviour configurable
In PR #583, a new behaviour was introduced that would cause the picture to switch to fullscreen if the viewer didn't move his mouse for >=1 second.
In my case, I need people (elderly relatives) to always see the top bar with the title. I have for now achieved that for myself by commenting out the relevant part from that PR in view.js, but it would be nicer if there was a config switch in the DB to turn the auto-fullscreen off. Notably I wouldn't have to redo this after every update and rebuild lychee.
+1 for that, also being able to adjust that 1 second delay would be nice.
A setting isn't planned, but we could increase the delay. I agree that 1 second might be too fast for some users.
The next version will use a 2.5s delay. 1 second was too aggressive.
|
2025-04-01T06:38:32.739006
| 2024-01-22T21:13:24
|
2094762315
|
{
"authors": [
"KyleAMathews",
"msfstef"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5603",
"repo": "electric-sql/electric",
"url": "https://github.com/electric-sql/electric/issues/868"
}
|
gharchive/issue
|
Listener warning
I have a view with a lot of small components that all have a small query listening for updates to their specific object — which should be a supported use case I'd assume? Or if there is an actual limit you want to warn against, set it to that. But I'm assuming the eventemitter is just set to the default (11 I think).
Looking at the notifier code under src/notifiers/event.ts, we have the following:
// Global singleton that all event notifiers use by default. Emitting an event
// on this object will notify all subscribers in the same thread. Cross thread
// notifications use the `./bridge` notifiers.
const globalEmitter = new EventEmitter()
// Increase the maximum number of listeners because multiple components
// use this same emitter instance.
globalEmitter.setMaxListeners(250)
This limit was increased to 250 by @thruflo int https://github.com/electric-sql/electric/pull/377 - since this is a global emitter I find it hard to think of any reasonable way to adjust this limit for a large application while still retaining the ability to detect if listeners are being accumulated or not appropriately removed. Maybe removing this warning and checking if many listeners are being added in the same "place", or having them be identified somehow could help construct a better warning.
For your particular case, would a "parent" component with view-only children work as well or would it complicate the code too much?
Yeah I'm sure I could refactor — I think ideally though the subscriptions are efficient enough it's not necessary.
I haven't noticed any slowdowns with the app & memory is fine — though all these listeners are to objects that are rarely changed. So what exactly is the upper limit is pretty arbitrary depending on what's updating in your app.
@samwillis what's your take on this? I think it makes sense to remove the warning altogether as we explicitly use a global emitter - if we want to catch leaks we can implement our own tests or mechanisms from within our EventNotifier to catch them.
|
2025-04-01T06:38:32.748408
| 2024-08-15T10:23:54
|
2467801095
|
{
"authors": [
"KiaClouth",
"TheAndrewJackson",
"msfstef",
"samwillis"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5604",
"repo": "electric-sql/pglite",
"url": "https://github.com/electric-sql/pglite/issues/208"
}
|
gharchive/issue
|
PGliteWorker not working with drizzle
OS: windows 11
Browser: chrome 127.0.6533.119
framework: solid-start
The minimal reproduction repository: GitHub Repo Link
description:
I am using PGliteWorker with drizzle. I followed the instructions on https://pglite.dev/docs/multi-tab-worker to create a worker wrapper:
PGliteWorker.ts
worker({ async init() { const pg = await PGlite.create({...}); await pg.exec(...); const db = drizzle(pg, { schema }); const ret = await db.query.verification_token.findMany(); console.log(ret); return pg; }, });
I can see the results in the browser console.
However, if I
entry-client.tsx
const pg = await PGliteWorker.create( new Worker(PGliteWorkerUrl, { type: "module", }) ); const db = drizzle(pg, { schema }); const ret = await db.query.verification_token.findMany(); console.log(ret);
The browser console gives the following error:
Uncaught DOMException: Failed to execute 'postMessage' on 'BroadcastChannel': value => value could not be cloned.
at j.g (http://localhost:3001/_build/node_modules/.pnpm/@electric-sql+pglite@0.2.0/node_modules/@electric-sql/pglite/dist/worker/index.js?v=c34c2280:201:21)
at j.query (http://localhost:3001/_build/node_modules/.pnpm/@electric-sql+pglite@0.2.0/node_modules/@electric-sql/pglite/dist/worker/index.js?v=c34c2280:98:56)
at async PglitePreparedQuery.execute (http://localhost:3001/_build/node_modules/.vinxi/client/deps/drizzle-orm_pglite.js?v=d787a4e7:58:20)
at async http://localhost:3001/_build/@fs/D:/ToramCalculator/src/entry-client.tsx:16:13
I've experienced this issue as well when trying to write to the DB with drizzle. My read queries were working. I wonder if this issue is an issue with drizzle
The Failed to execute 'postMessage' on 'BroadcastChannel' error makes me think it could be our side as we use a BroadcastChannel to communicate with the worker. It's either:
drizzle is trying to pass an un-cloneable object as a parameter to the query api.
or we have something in our api that doesn't work with the worker and is not currently covered by the tests.
or both, which is my suspicion:
I'm 99% sure we will find it's the parser query option (https://pglite.dev/docs/api#query-options) and will need a bit of a refactor.
The issue seems to be that we are trying to transfer ParserOptions which contains anonymous functions to the worker without proper serialization.
Either the parsing is done at the calling thread (and the ParserOptions are stored and handled there) or we assume some limitations on the functions being passed and serialize them and deserialize them to run in the worker (e.g. with a toString()/eval() combination).
Since these parsers might depend on various imports and what not it might be better to handle deserialization in the caller.
|
2025-04-01T06:38:32.775770
| 2024-11-12T08:07:55
|
2651383360
|
{
"authors": [
"darylgraham",
"emillg",
"tollercode"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5605",
"repo": "electrolux-oss/infrawallet",
"url": "https://github.com/electrolux-oss/infrawallet/issues/124"
}
|
gharchive/issue
|
[BUG] Cost data is wrongly mapped into Budget graphs
Current Situation
Having this cost data in November
The data is shown in the Budget Graphs in Januaray
Desired
Data of November should be shown in November also on Budget-Chart
I just debugged this screen myself.
The cost data is passed into the BudgetChart component as an array of monthly costs here: https://github.com/electrolux-oss/infrawallet/blob/main/plugins/infrawallet/src/components/Budgets/Budgets.tsx#L250
"id": "AWS",
"reports": {
"2023-12": 19.6492511966,
"2024-01": 1939.0336325847002,
"2024-02": 1610.1251975480995,
"2024-03": 2149.6599309432,
"2024-04": 2629.0959797089004,
"2024-05": 2603.6509832689,
"2024-06": 2438.023622028399,
"2024-07": 2767.510390127999,
"2024-08": 9312.452064800302,
"2024-09": 8760.504147668702,
"2024-10": 10055.072295576203,
"2024-11": 5529.263631147701
}
}
It's then converted to a running sum here: https://github.com/electrolux-oss/infrawallet/blob/main/plugins/infrawallet/src/components/Budgets/Budgets.tsx#L91-L98
{
0: 19.6492511966,
1: 1958.6828837813002,
2: 3568.8080813294,
3: 5718.4680122726,
4: 8347.5639919815,
5: 10951.214975250401,
6: 13389.2385972788,
7: 16156.748987406798,
8: 25469.2010522071,
9: 34229.7051998758,
10: 44284.77749545201,
11: 49814.04112659971
}
This is then plotted on the chart, with a constant set of x-axis titles: https://github.com/electrolux-oss/infrawallet/blob/main/plugins/infrawallet/src/components/Budgets/Budgets.tsx#L122
xAxis={[
{
data: ['Jan', 'Feb', 'Mar', 'Apr', 'May', 'Jun', 'Jul', 'Aug', 'Sep', 'Oct', 'Nov', 'Dec'],
scaleType: 'band',
},
]}
I think that the x-axis labels just need to be dynamically generated by mapping the keys from the original report variable to the month names.
@tollercode @darylgraham Can you check if version 0.2.0-20241118212048-4e21a4a works for you?
|
2025-04-01T06:38:32.838579
| 2017-05-19T16:05:20
|
230019792
|
{
"authors": [
"aliir74",
"firsttracks",
"ryan-christopher",
"shakhassan",
"zeke"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5606",
"repo": "electron-userland/electron-prebuilt",
"url": "https://github.com/electron-userland/electron-prebuilt/issues/254"
}
|
gharchive/issue
|
Mac: Failed at the<EMAIL_ADDRESS>postinstall script 'node install.js'.
Received the following message:
Failed at the<EMAIL_ADDRESS>postinstall script 'node install.js'. (Exit Status 1)
node.js (and I assume npm with it) were just downloaded and installed from their site.
Electron version<EMAIL_ADDRESS>Operating system: macOS Sierra 10.12.4
I'm brand new to working with these types of files and installers, so I'm not sure how to correct this.
Also, I need to know if I can do this as a global install, as I would not like to confront the prospect of installing this in every single project folder.
In my case, npm install electron -g will do the trick.
I am seeing the same error. Is there a work around for this issue?
I got around this issue by looking at the link "https://docs.npmjs.com/getting-started/installing-npm-packages-locally". I tried a bunch of things and not sure exactly what fixed it but I did run "npm init" with defaults and then the "npm install mapbox-map-image-export -g" command seemed to work
I've got the same error.
<EMAIL_ADDRESS>postinstall: node install.js. Exit status 1
Node.js version: v8.1.4
Electron version: v1.6.11
OS: Ubuntu 16.04
@ryan-christopher I updated my node to last version and the error fixed.
The electron-prebuilt repo is being retired and its code has been moved into the electron/electron repo. For the sake of historical transparency, we will leave GitHub Issues enabled on this repository, but if you are still affected by the issue reported here, please open a new issue on electron/electron repo and reference this issue from it so people can get the full context. The electron repository has a large and active contributor community, so your issue is more likely to get the attention it deserves there. Thanks!
|
2025-04-01T06:38:32.950412
| 2022-09-12T15:31:21
|
1370104841
|
{
"authors": [
"batis97",
"gpronet",
"lukas2",
"technotip"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5607",
"repo": "electron/osx-sign",
"url": "https://github.com/electron/osx-sign/issues/266"
}
|
gharchive/issue
|
"Unidentified developer" when opening universal app
Hello, we switched from an app for x86_64 to an universal app. The app runs fine on all platforms when I sign them locally with my developer identity.
However, when we go through Testflight, we are greeted with a "... cannot be opened because the developer cannot be verified" message. This only happens with the universal build, the x86_64 only one works fine. And if I force-resign the app, then it also works fine.
This is how we sign:
electron-osx-sign --entitlements=./entitlements/entitlements.mac.plist --entitlements-inherit=./entitlements/entitlements.mac.inherit.plist --entitlements-loginhelper=./entitlements/entitlements.mac.inherit.plist --identity=${CERTIFICATE_DEVELOPER_SHA} --keychain=signing-cert-keychain --provisioning-profile=./embedded.provisionprofile --type=distribution "${appName}" productbuild --component "${appName}" /Applications --sign "${CERTIFICATE_INSTALLER_SHA}" "app-name-${params.UPLOAD_ARTIFACT_VERSION}.pkg"
Checking the app with the usual tools (codesign,pkgutil, spctl) gives no clue to any problems. Only at runtime it doesn't work.
I have run out of ideas to try..
I have exactly the same problem, already spent 5 days trying different things but I've exhausted my options.
Does force resigning the universal app actually work for you and you can run it in Test Flight? I couldn't even manage to do that yet.
I meant that resigning locally with my developer identity makes it runnable for me, but I have not tried uploading the resigned app onto Testflight.
Did you test if the same problem occurs when you attempt an app store release? Speculating if this is a problem with Testflight..
I meant that resigning locally with my developer identity makes it runnable for me, but I have not tried uploading the resigned app onto Testflight.
I see, I also have no problems running the app locally when it is signed with the developer identity and notarized. However for TestFlight and the Mac Store it has to be signed either with 3rd Party Mac Developer Application or Apple Distribution certificate. And when we do this we get the above problem.
Did you test if the same problem occurs when you attempt an app store release? Speculating if this is a problem with Testflight..
Yes, the app gets rejected for the same reason. I created a separate issue with more info since I am using electron-builder (not osx-sign directly) here:
#https://github.com/electron-userland/electron-builder/issues/7171
I've attached a build log there which confirms signing goes well, so the problem has to be elsewhere.
Have you tried regenerating the provisioning profile? Will try that next (unfortunately I don't have direct access to App Store so it's a slow process here).. after all the provisioning profile links the developer with the app..
Have you tried regenerating the provisioning profile?
Yes, no difference. Transporter usually complains if your provisioning profile is not correct, so you will never reach TestFlight deployment if it was wrong. But you are welcome to try it in case I missed something.
just fyi, I had another wild theory.. maybe the universal build is loading different dylibs on startup than the x86 build. That could lead it to not find something. If it doesn't find it, the loader goes to search for it. And while searching it checks some invalid directories, which are forbidden unless some entitlements are given, and therefore the signature is rejected.
So i checked and the only diffrence between a working x86 app (left) and the universal app (right) was this:
I think I found the reason. My app uses native modules and they need to be signed, too. For that reason, they are unpacked, using the asarUnpack config option.
However, unlike with the x86 build, in the universal build I have two copies of the same node module. One in the app.asar.unpacked folder, and one in the app.asar archive itself. And only the "unpacked" one is signed. 😨
Ok I have my solution, if I set mergeASARs to false, I get a correctly signed and launching app. Albeit, it still seems that the native modules are duplicated.
I would close this issue now, or leave it to you @gpronet , as I think my solution won't help you
Amazing, mergeASARs: false fixes the issue for me as well.
A non critical dependency of my app was using just one native module, so when I removed it, the error messages were gone as well even without mergeASARs: false. Thank you for pointing me in the right direction @lukas2.
Wow, happy to hear! 🙂
Actually it clicked for me when you said in the other thread that the "Move to Bin"-deleted files are hidden in Trash. I examined mine and saw my module in unsigned form. Then I knew there had to be more than one.. :)
Ok closing this.
I'm still stuck with this error :-(
Can you guys please post your package.json and entitlement files?
Hello i still have this problem
mergeASARs: false
I've set it up > mergeASARs: false and the eror still show in testflight
i using electron + svelkit
this my electron and builder version:
"electron": "^26.2.2",
"electron-builder": "^24.6.4",
|
2025-04-01T06:38:32.982886
| 2022-07-14T08:48:40
|
1304466980
|
{
"authors": [
"IDoneShaveIt",
"elongl"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5608",
"repo": "elementary-data/elementary",
"url": "https://github.com/elementary-data/elementary/pull/121"
}
|
gharchive/pull-request
|
Hotfix/support backfill changes on redshift
support backfill changes for alerts model on redshift
LGTM!
Could you please change to branch to v<IP_ADDRESS>?
|
2025-04-01T06:38:33.130203
| 2015-02-28T02:12:47
|
59320138
|
{
"authors": [
"dwightwatson",
"elfet"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5609",
"repo": "elfet/cherimola",
"url": "https://github.com/elfet/cherimola/pull/6"
}
|
gharchive/pull-request
|
Remove version constraint from the installation docs
If one was to copy and paste the Composer installation command from the README when it has the ~1 version constraint they might be confused as to why they don't get all the functionality shown in the docs. Removing the version constraint will have Composer install the latest tagged version and thus they should get all the current functionality.
Cool!
|
2025-04-01T06:38:33.143500
| 2019-08-07T08:40:06
|
477787265
|
{
"authors": [
"martinstz"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5612",
"repo": "elgatito/plugin.video.elementum",
"url": "https://github.com/elgatito/plugin.video.elementum/issues/464"
}
|
gharchive/issue
|
Problem with searching
Hi!
I have some problems with zamunda.net. For some reason when Elementum successful login to zamunda and i search for some movie, Elementum find results with right search string: http://zamunda.net/bananas?c42=1&c25=1&c35=1&c46=1&c20=1&c19=1&c5=1&c24=1&c31=1&c28=1&search=avengers+endgame+2019&incldead=1&field=name
but after that Elementum create another request with other string: https://zamunda.net:443/login.php?returnto=%2Fbananas%3Fc42%3D1%26c25%3D1%26c35%3D1%26c46%3D1%26c20%3D1%26c19%3D1%26c5%3D1%26c24%3D1%26c31%3D1%26c28%3D1%26search%3Davengers%2Bendgame%2B2019%26incldead%3D1%26field%3Dname
and this string loged me out from Zamunda and finaly Elementum cannot display any results.
Can you help with this?
If you need user and pass for Zamunda, tell me where to send them.
Thank you in advance!
I fix the problem.
|
2025-04-01T06:38:33.156393
| 2024-09-22T12:28:25
|
2541055842
|
{
"authors": [
"antonsoroko",
"elgatito",
"olegmiercoles"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5613",
"repo": "elgatito/script.elementum.burst",
"url": "https://github.com/elgatito/script.elementum.burst/issues/438"
}
|
gharchive/issue
|
400 Login failed on [toloka] provider
When try to execute any search on toloka provider it returns no links.
From log debug I can see errors of 400 bad requests and information that Request Header or Cookie Too Large.
Not sure is it something about client, configuration or the provider itself.
I've verified that the login/password are correct.
I've also get try to lgin from postman using this form-data from config
"{'username': USERNAME, 'password': PASSWORD, 'autologin': '1', 'login': 'Enter'}"
and it seems to be working, at least it returns me redirect instead of bad request.
Here is the part of log debug:
2024-09-22 12:49:48.502 T:4449 debug <general>: [script.elementum.burst] Searching with payload (general): {'proxy_url': '', 'internal_proxy_url': 'http://<IP_ADDRESS>:65222', 'elementum_url': 'http://<IP_ADDRESS>:65220', 'silent': False, 'skip_auth': False, 'query': 'FOO'}
2024-09-22 12:49:48.505 T:4449 warning <general>: [script.elementum.burst] Burstin' with Гуртом
2024-09-22 12:49:48.506 T:4449 warning <general>: [script.elementum.burst] No 'en' translation available...
2024-09-22 12:49:48.511 T:3816 debug <general>: ------ Window Init (DialogExtendedProgressBar.xml) ------
2024-09-22 12:49:48.512 T:4449 debug <general>: [script.elementum.burst] Translated titles from Elementum: {'source': 'FOO', 'original': 'FOO'}
2024-09-22 12:49:48.513 T:4468 debug <general>: [script.elementum.burst] [toloka] Processing toloka with general method
2024-09-22 12:49:48.514 T:4468 debug <CSettingsManager>: requested setting (filter_music) was not found.
2024-09-22 12:49:48.515 T:4468 debug <general>: [script.elementum.burst] [toloka] General URL: https://toloka.to/tracker.php?nm=QUERYEXTRA&o=10
2024-09-22 12:49:48.517 T:4468 debug <general>: [script.elementum.burst] [toloka] execute_process for toloka with <function extract_torrents at 0x7e114938>
2024-09-22 12:49:48.518 T:4449 debug <general>: [script.elementum.burst] Timer: 0s / 27s
2024-09-22 12:49:48.527 T:4468 debug<EMAIL_ADDRESS>trying to load setting definitions from old format...
2024-09-22 12:49:48.549 T:4468 debug <general>: [script.elementum.burst] [toloka] Queries: ['{title}']
2024-09-22 12:49:48.550 T:4468 debug <general>: [script.elementum.burst] [toloka] Extras: ['']
2024-09-22 12:49:48.551 T:4468 debug <general>: [script.elementum.burst] [toloka] Before keywords - Query: '{title}' - Extra: ''
2024-09-22 12:49:48.552 T:4468 warning <general>: [script.elementum.burst] [toloka] Falling back to original title in absence of None language title
2024-09-22 12:49:48.557 T:4468 warning <general>: [script.elementum.burst] [toloka] Using translated 'original' title 'FOO'
2024-09-22 12:49:48.559 T:4468 debug <general>: [script.elementum.burst] [toloka] After keywords - Query: '%D1%80%D0%B0%D1%82%D0%B0%D1%82%D1%83%D0%B9' - Extra: ''
2024-09-22 12:49:48.560 T:4468 debug <general>: [script.elementum.burst] - toloka query: '%D1%80%D0%B0%D1%82%D0%B0%D1%82%D1%83%D0%B9'
2024-09-22 12:49:48.561 T:4468 debug <general>: [script.elementum.burst] -- toloka url_search before token: 'https://toloka.to/tracker.php?nm=FOO&o=10'
2024-09-22 12:49:48.561 T:4468 debug <general>: [script.elementum.burst] --- toloka using POST payload: {}
2024-09-22 12:49:48.562 T:4468 debug <general>: [script.elementum.burst] ----toloka filtering with post_data: {}
2024-09-22 12:49:48.562 T:4468 debug <CSettingsManager>: requested setting (toloka_passkey) was not found.
2024-09-22 12:49:48.577 T:4468 debug <general>: [script.elementum.burst] Opening URL: b'https://toloka.to/login.php'
2024-09-22 12:49:48.770 T:4449 debug <general>: [script.elementum.burst] Timer: 0s / 27s
2024-09-22 12:49:49.029 T:4468 info <general>: Skipped 1 duplicate messages..
2024-09-22 12:49:49.029 T:4468 debug <general>: [script.elementum.burst] Status for b'https://toloka.to/login.php' : 400
2024-09-22 12:49:49.031 T:4468 critical <general>: [script.elementum.burst] [toloka] Login failed: 400
2024-09-22 12:49:49.034 T:4468 debug <general>: [script.elementum.burst] [toloka] Failed login content: '<html>\r\n<head><title>400 Request Header Or Cookie Too Large</title></head>\r\n<body bgcolor="white">\r\n<center><h1>400 Bad Request</h1></center>\r\n<center>Request Header Or Cookie Too Large</center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>\r\n<!-- a padding to disable MSIE and Chrome friendly error page -->\r\n<!-- a padding to disable MSIE and Chrome friendly error page -->\r\n<!-- a padding to disable MSIE and Chrome friendly error page -->\r\n<!-- a padding to disable MSIE and Chrome friendly error page -->\r\n<!-- a padding to disable MSIE and Chrome friendly error page -->\r\n<!-- a padding to disable MSIE and Chrome friendly error page -->\r\n'
2024-09-22 12:49:49.047 T:4468 warning <general>: [script.elementum.burst] [toloka] >> Гуртом returned 0 results in 0.5 seconds
System: Android TV
Kodi version: 21.1
Elementum version: 0.1.103
Elementum burst version: 0.0.89
Any suggestions will be appreciate.
@olegmiercoles
Maybe it is because of long cookie.
There is a file with cookies in .kodi/temp/burst/common_cookies.jar (see https://kodi.wiki/view/Kodi_data_folder for location of .kodi in your OS)
It is a plain text file, you can open it and search for toloka and see if cookie is "bad" (maybe it has some garbage and it is indeed too long).
Anyway - remove the line for toloka and try again. also, you can backup that file and remove the file completely.
Also, in burst settings in "maintenance" tab you can remove all cookies - in case if you unable to get access to file system of your device.
There is no "login_headers": field for toloka so it should not be "Request Header" issue.
https://github.com/elgatito/script.elementum.burst/blob/97dcfb60aa43fa646712c6ec60fb9fc7ee80ecc0/burst/providers/providers.json#L2381
Thank you, @antonsoroko
I did as you suggested: I've renamed common_cookies.jar to common_cookies_bkp.jar and it helped.
Probably I will keep backup-file for some time, but later will remove it.
Thank you again for assistance!
@olegmiercoles I am just curious - have you tried to take a look into that file? Interesting how a line for toloka looks like. If you remove toloka_sid from that line - you can share it here (so it will not have your login cookie).
Up to you of course.
Sure. I did my the best to cleanup all the hashes and IDs. Hope, I didn't miss anything :)
First line for toloka_data was pretty huge combo of hash+autologinid+hash+userid+hash
And toloka_302_u also contained around 8.000 characters hash inside.
#LWP-Cookies-2.0
Set-Cookie3: toloka_data="something-was-here-like-b%432%3b%9Bs%3d%autologinid%here-as-well-something-likeb%432%3b%9Bs%3d%userid-and-here-the-same"; path="/"; domain="toloka.to"; path_spec; secure; expires="2025-05-29 18:56:57Z"; httponly=None; version=0
Set-Cookie3: toloka_302_tt=there-was-some-numbers; path="/"; domain="toloka.to"; path_spec; secure; expires="2025-05-29 18:56:57Z"; version=0
Set-Cookie3: toloka_302_f="there-was-something-like-in-the-first-line-a%54b%33%"; path="/"; domain="toloka.to"; path_spec; secure; expires="2025-05-29 18:56:57Z"; version=0
Set-Cookie3: toloka_302_uf=there-was-some-numbers; path="/"; domain="toloka.to"; path_spec; secure; expires="2025-05-29 18:56:57Z"; version=0
Set-Cookie3: toloka_302_u="about-8.000-characters-was-here"; path="/"; domain="toloka.to"; path_spec; secure; expires="2025-05-29 18:56:57Z"; version=0
Set-Cookie3: toloka___tt=there-was-some-numbers; path="/"; domain="toloka.to"; path_spec; secure; expires="2025-03-04 19:58:37Z"; version=0
Set-Cookie3: toloka___f="there-was-something-like-in-the-first-line-a%54b%33%"; path="/"; domain="toloka.to"; path_spec; secure; expires="2025-03-04 19:58:37Z"; version=0
Set-Cookie3: toloka___uf=0; path="/"; domain="toloka.to"; path_spec; secure; expires="2025-03-04 19:58:37Z"; version=0
Set-Cookie3: toloka___u="there-was-something-like-in-the-first-line-a%54b%33%"; path="/"; domain="toloka.to"; path_spec; secure; expires="2025-03-04 19:58:37Z"; version=0
@olegmiercoles thanks!
so looks like by default many popular HTTP servers have total limit for HTTP Header (and cookie is a part of header) as 8KB.
so looks like in total your cookies were >8KB, thus i think we found the root cause.
somehow toloka website generated such long cookie.
i guess we could create a "blacklist" for cookies (e.g. ignore and do not save some cookies), although i am not sure if we really need it, since i see such issue for the first time. maybe @elgatito can add more ideas.
but anyway, thanks for info.
if there will be more issues like this then we can circle back to this.
@antonsoroko Quick googling says there is nothing "ready" in Python's requests to control size of a request to avoid such errors.
I also have problems with Toloka, even with sync, it does something (I was not debugging those), that invalidates a session everywhere, not only on Elementum/Burst side.
Not sure if we should/can do something with it.
|
2025-04-01T06:38:33.160030
| 2018-04-11T23:19:27
|
313520137
|
{
"authors": [
"eliangcs",
"wheelerlaw"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5614",
"repo": "eliangcs/http-prompt",
"url": "https://github.com/eliangcs/http-prompt/issues/142"
}
|
gharchive/issue
|
Missing HTTP CONNECT method
Pretty straight forward, can't seem to do an HTTP CONNECT request.
For example:
$ http-prompt http://example.com
Version: 0.11.2
http://example.com> httpie connect
http http://example.com/connect
should actually be
http-prompt http://example.com
Version: 0.11.2
http://example.com> httpie connect
http CONNECT http://example.com
Addressed in #145, shipped in v1.0.0.
|
2025-04-01T06:38:33.166383
| 2016-03-18T01:20:23
|
141747801
|
{
"authors": [
"garfieldnate",
"wowselim"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5615",
"repo": "elibom/jogger",
"url": "https://github.com/elibom/jogger/issues/19"
}
|
gharchive/issue
|
archetype generation command fails
I copied the command from your Wiki to start a project from the Maven archetype:
mvn archetype:generate -DarchetypeGroupId=org.jogger -DarchetypeArtifactId=jogger-archetype -DarchetypeVersion=0.9.0 -DarchetypeRepository=http://repository.elibom.net/nexus/content/repositories/releases/
This was the output:
mvn archetype:generate -DarchetypeGroupId=org.jogger -DarchetypeArtifactId=jogger-archetype -DarchetypeVersion=0.9.0 -DarchetypeRepository=http://repository.elibom.net/nexus/content/repositories/releases/
[INFO] Scanning for projects...
[INFO]
[INFO] ------------------------------------------------------------------------
[INFO] Building Maven Stub Project (No POM) 1
[INFO] ------------------------------------------------------------------------
[INFO]
[INFO] >>> maven-archetype-plugin:2.4:generate (default-cli) > generate-sources
@ standalone-pom >>>
[INFO]
[INFO] <<< maven-archetype-plugin:2.4:generate (default-cli) < generate-sources
@ standalone-pom <<<
[INFO]
[INFO] --- maven-archetype-plugin:2.4:generate (default-cli) @ standalone-pom --
-
[INFO] Generating project in Interactive mode
[INFO] Archetype defined by properties
Downloading: http://repository.elibom.net/nexus/content/repositories/releases/or
g/jogger/jogger-archetype/0.9.0/jogger-archetype-0.9.0.jar
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 1.974 s
[INFO] Finished at: 2016-03-18T10:25:32+09:00
[INFO] Final Memory: 15M/245M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-archetype-plugin:2
.4:generate (default-cli) on project standalone-pom: The desired archetype does
not exist (org.jogger:jogger-archetype:0.9.0) -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e swit
ch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please rea
d the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureExc
eption
Hi @garfieldnate ,
I think this project is pretty much dead. In case you are looking for alternatives I suggest you take a look at either Spark or Pippo. Jooby might also be worth a try eventhough I have no experience with it.
That's too bad, this was very quick to set up. It's just that the author's repo site is down. Maybe they could be moved to Maven Central if they could be found. I'll leave this issue here for future reference.
|
2025-04-01T06:38:33.171915
| 2023-04-28T01:23:42
|
1687749741
|
{
"authors": [
"gnott"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5616",
"repo": "elifesciences/docmap-tools",
"url": "https://github.com/elifesciences/docmap-tools/pull/15"
}
|
gharchive/pull-request
|
Preprint history event data parse, update fixture.
Added new parsing functions to gather preprint version event history data.
The existing test fixture 2<IP_ADDRESS>2253.docmap.json is updated to reflect the latest docmap data available for it.
Re issue https://github.com/elifesciences/issues/issues/7721
|
2025-04-01T06:38:33.190307
| 2024-10-16T19:15:16
|
2592827835
|
{
"authors": [
"acollings",
"fred-atherden"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5617",
"repo": "elifesciences/publish-reviewed-preprints-issues",
"url": "https://github.com/elifesciences/publish-reviewed-preprints-issues/issues/1423"
}
|
gharchive/issue
|
MSID: 99785 Version: 1 DOI: 10.1101/2<IP_ADDRESS>3526
MSID: 99785
Version: 1
Preprint DOI: https://doi.org/10.1101/2<IP_ADDRESS>3526
Step 1. Awaiting reviews
Editorial to post reviews via hypothesis
Useful links:
DocMap: https://data-hub-api.elifesciences.org/enhanced-preprints/docmaps/v2/by-publisher/elife/get-by-manuscript-id?manuscript_id=99785
New model tracking: https://docs.google.com/spreadsheets/d/1_fHaoOy7hjyocptKtVJRijeNpUY4hBS7Ck_aVmx6ZJk/
Reviews on sciety: https://sciety.org/articles/activity/10.1101/2<IP_ADDRESS>3526
For trouble shooting (e.g. no Docmaps available):
DocMap issue addressing: https://miro.com/app/board/uXjVNCwK6EI=/
Explore DataHub DocMaps API: https://lookerstudio.google.com/reporting/4c2f0368-babb-4beb-b5b3-497e7e7b0f08/page/ejphD
Unmatched submissions and preprints: https://lookerstudio.google.com/u/0/reporting/9f86204f-3bf7-477c-9b18-5c5ef141bf69/page/p_gxi57ha93c
Unmatched manuscripts spreadsheet: https://docs.google.com/spreadsheets/d/15QcK8w-ssB7109RQEDtFpJPZ0J5HTGxoHa_2TtpMBbg/edit#gid=1336081641
Step 2. Preview reviewed preprint
Production QC content ahead of publication
Instructions:
QC preview: https://prod--epp.elifesciences.org/previews/99785v1
Update ticket with any problems (add blocked label)
When QC OK, add QC OK label to ticket and add publication date and time to https://docs.google.com/spreadsheets/d/1amAlKvdLcaDp5W8Z8g77NmkwbMF5n_u89ArSqPMO8jg
Move card to next column
(At end of the day post link in #enhanced-preprint and ask for PDF to be generated)
Useful links:
Preprint DOI : https://doi.org/10.1101/2<IP_ADDRESS>3526
Confirm reviews returned by EPP: https://prod--epp.elifesciences.org/api/reviewed-preprints/99785/v1/reviews
To update the MECA path in the docmap: https://docs.google.com/spreadsheets/d/1mctCQuNFBjSn97Lihy7_vBO6z7-N-oqyLv4clyi6zHg
Step 3: Awaiting search reindex
This step adds the reviewed preprint to the homepage: https://elifesciences.org
The search reindex is triggered once an hour. We need the reviewed preprint to be indexed as the search application serves the journal homepage.
Useful links:
Jenkins pipeline to reindex search can be triggered sooner or monitored here: https://alfred.elifesciences.org/job/process/job/process-reindex-reviewed-preprints/
Step 4: Published! PDF requested
Waiting for PDF to be generated
Useful links:
PDF tracking: https://docs.google.com/spreadsheets/d/106_XeDjmuBae7gexOTNzg60lapeqjl2aRn9DzupGyS8/
Step 5: Introduce PDF to data folder and git repo
Upload PDF to relevent folder in git repo https://github.com/elifesciences/enhanced-preprints-data/
Step 6: Done!
[ ] Kettle is on!
Hi @acollings / @FionaBryant, please could you take a look, and if necessary tweak the assessment for this one?
The authors solidly connect proteostasis ...
Thanks @fred-atherden this has been fixed
Many thanks!
WOS query sent
Confirmed Ok to proceed RE WOS.
Waiting for https://sciety.org/evaluations/hypothesis:nlvGcJarEe-W4aOBdYgobQ/content to update
|
2025-04-01T06:38:33.206775
| 2021-09-07T14:27:27
|
990046247
|
{
"authors": [
"Flova",
"Ledjob",
"WolfgangFahl",
"docsteveharris",
"elimintz",
"giodegas",
"nielstron",
"poke1024",
"samiit"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5619",
"repo": "elimintz/justpy",
"url": "https://github.com/elimintz/justpy/issues/301"
}
|
gharchive/issue
|
demjeson install cannot be done since setuptools upgrade, so neither justpy can.
Today on a new deploy I did :
pip3 install justpy
but ends up failing because:
ERROR: Cannot install justpy==0.0.5, justpy==0.0.6, justpy==0.0.7, justpy==0.0.8, justpy==0.0.9, justpy==0.1.0, justpy==0.1.1, justpy==0.1.2, justpy==0.1.3, justpy==0.1.4 and justpy==0.1.5 because these package versions have conflicting dependencies.
The conflict is caused by:
justpy 0.1.5 depends on demjson>=2.2.4
justpy 0.1.4 depends on demjson>=2.2.4
justpy 0.1.3 depends on demjson>=2.2.4
justpy 0.1.2 depends on demjson>=2.2.4
justpy 0.1.1 depends on demjson>=2.2.4
justpy 0.1.0 depends on demjson>=2.2.4
justpy 0.0.9 depends on demjson>=2.2.4
justpy 0.0.8 depends on demjson>=2.2.4
justpy 0.0.7 depends on demjson>=2.2.4
justpy 0.0.6 depends on demjson>=2.2.4
justpy 0.0.5 depends on demjson>=2.2.4
To fix this you could try to:
1. loosen the range of package versions you've specified
2. remove package versions to allow pip attempt to solve the dependency conflict
it is an open issue at demjson repo:
https://github.com/dmeranda/demjson/issues/40
Downgrading to setuptools 57.5.0 it des not work, it installs but while doing:
import justpy
gives:
import justpy
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/data/giodegas/dev/AQ2009/myenv/lib/python3.8/site-packages/justpy/__init__.py", line 1, in <module>
from .justpy import *
File "/data/giodegas/dev/AQ2009/myenv/lib/python3.8/site-packages/justpy/justpy.py", line 13, in <module>
from .chartcomponents import *
File "/data/giodegas/dev/AQ2009/myenv/lib/python3.8/site-packages/justpy/chartcomponents.py", line 2, in <module>
import demjson
File "/data/giodegas/dev/AQ2009/temp/demjson/demjson.py", line 645
class json_int( (1L).__class__ ): # Have to specify base this way to satisfy 2to3
^
SyntaxError: invalid syntax
@elimintz what can I do now?
Thank you for your support.
Thanks for alerting me to this. I'll try to figure something out. In the meantime, it seems that if you downgrade to anything less than python 3.10 and setuptools to 57.5.0 it should work. Let me know if it doesn't
I am using python 3.8.10 now, setuptools 57.5.0 with a virtual env, still with the import problem.
I may try downgrading setuptools more.
Ok, please let me know how it goes
ok found setuptools 56.0.0 is OK!
this is now my pipi listi in the virtual environment:
pip list
Package Version
------------------ ---------
addict 2.4.0
aiofiles 0.7.0
anyio 3.3.0
asgiref 3.4.1
certifi 2021.5.30
charset-normalizer 2.0.4
click 8.0.1
demjson 2.2.4
h11 0.12.0
httpcore 0.13.6
httpx 0.19.0
idna 3.2
itsdangerous 2.0.1
Jinja2 3.0.1
justpy 0.1.5
MarkupSafe 2.0.1
pip 21.2.4
pkg_resources 0.0.0
rfc3986 1.5.0
setuptools 56.0.0
sniffio 1.2.0
starlette 0.16.0
uvicorn 0.15.0
websockets 9.1
Thanks for finding a workaround.
I am now in the process of publishing a new version without using setuptools (using flit). Perhaps this will solve the issue.
btw, using docker and latest python3:8 image I found this better setuptools 57.4.0:
pip list
Package Version
------------------ ---------
addict 2.4.0
aiofiles 0.7.0
anyio 3.3.0
asgiref 3.4.1
certifi 2021.5.30
charset-normalizer 2.0.4
click 8.0.1
demjson 2.2.4
h11 0.12.0
httpcore 0.13.6
httpx 0.19.0
idna 3.2
itsdangerous 2.0.1
Jinja2 3.0.1
justpy 0.1.5
MarkupSafe 2.0.1
pip 21.2.4
rfc3986 1.5.0
setuptools 57.4.0
sniffio 1.2.0
starlette 0.16.0
uvicorn 0.15.0
websockets 9.1
wheel 0.37.0
You might want to switch to demjson3 in case you really only want to support Python 3 anymore.
Thank you for the suggestion. I will make the change.
I have the same problem but am unable to downgrade to Python <3.8.8 because I'm on a Mac with an M1 chip.
Hi,
Is there any solution yet for this problem? JustPy isn't just getting installed!
Please see if you can port things to demjson3.
Thanks,
Sam
I need to find time to release a new version with this. There is a demjson compatible package called demjson3 that solves this issue. If you want to fix it locally, change all import demjson lines to import demjson3 as demjson
And you need to install demjson3.
The advantage of using demjson is that it can parse correctly javascript objects where the keys do not need to be between quotes.
same problem here on python 3.6.9
downgrading setuptools to 56.0.0 as giodegas stated worked for me.
@elimintz Replacing demjson with demjson3 indeed fixes the issue for me. Also see PR.
Version 0.2.3 is out that should this problem. Replaced demjson with demjson3.
Would appreciate confirmation that this indeed is the case. Did not have time for too much testing but the changes were very limited. The changes are not reflected in the code on github yet but all I did was replace import demjson with import demjson3 as demjson in 4 places.
@poke1024 @1081 @nielstron @giodegas @Ledjob @docsteveharris @Flova
Some preliminary test show it is ok now. Thank you.
It works fine for me too
duplicate of #408
|
2025-04-01T06:38:33.216951
| 2018-09-18T10:02:36
|
361228128
|
{
"authors": [
"LAguido",
"elishacloud"
],
"license": "Zlib",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5620",
"repo": "elishacloud/dxwrapper",
"url": "https://github.com/elishacloud/dxwrapper/issues/31"
}
|
gharchive/issue
|
New Worlds
Hello, I've been trying to get Star Trek New Worlds to run, with
DDrawCompat= 1
DSoundCtrl= 1
I do get into the game, however after a few seconds it just freezes.
I'm using compatibility mode for windows 7, 16 bit color mode, disabled optimizations for full screen.
I've added the logfile I got from Process Monitor:
Logfile.zip
I got the feeling I'm missing something really small, since it does start and you can move around for a few seconds.
(Edited with new logfile, previous didn't properly include the whole game session)
dxwrapper-stnw.log
Here's the wrapper log when the issue occurs
It looks like the game is crashing. Try setting DSoundCtrl = 0. Also try adding this line into the ini file: HandleExceptions = 0. This will allow you to see the crash.
If that does not work, try using the attached updated files. This works with Star Trek Armada 1 and Star Trek Armada 2.
ddraw.zip
The 2 settings seemed to have worked (at least it doesn't crash in 10 seconds), I did not use the Armada files. The game needed to be launched from a FAT32 usb stick or partition to work properly. I didn't have time to test the game properly but thus far it seems ok!
Closing this since the issue seems to be resolved. If the issue comes back you can reopen.
|
2025-04-01T06:38:33.218058
| 2023-12-17T17:59:36
|
2045314965
|
{
"authors": [
"JaeAeich"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5621",
"repo": "elixir-cloud-aai/cloud-components",
"url": "https://github.com/elixir-cloud-aai/cloud-components/pull/213"
}
|
gharchive/pull-request
|
feat(wes): add service-info
Description
This PR creates a service info component for WES.
closing this will create a PR for the same, as there have been changes in design package.
|
2025-04-01T06:38:33.263954
| 2023-06-15T22:38:00
|
1759636256
|
{
"authors": [
"josevalim",
"paulo-ferraz-oliveira"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5622",
"repo": "elixir-makeup/makeup",
"url": "https://github.com/elixir-makeup/makeup/pull/60"
}
|
gharchive/pull-request
|
Fix type definition
+ @type parsec_success :: {:ok, tokens, String.t(), context(), integer(), integer()}
- @type parsec_success :: {:ok, tokens, String.t(), context(), {integer(), integer()}, integer()}
as consumed by e.g. makeup_elixir or makeup_erlang.
I take this time to also introduce dialyxir to CI, which surfaces another issue (opaqueness-related) already mentioned next to stream_data.
Actions missing from this:
waiting for a stream_data update and subsequent import here
(potentially) updating nimble_parsec (as per acceptance of this pull request) [optional]
Edit: while reading the CONTRIBUTING guide I didn't quite understand how to handle the RELEASE.md part (is this required?). Regarding the CHANGELOG.md, do you prefer I do it? Or do you, prior to release?
A PR to fix the specs is welcome but we don't plan to introduce dialyzer at the moment, thank you :)
Sure. I can remove that bit. Thanks.
@josevalim, shall I wait for a nimble_parsec release to update this? Or are you good without it? Thanks.
:green_heart: :blue_heart: :purple_heart: :yellow_heart: :heart:
The goal of this update is to fix a Dialyzer -related issue with makeup_erlang and makeup_elixir (for which now I'm thinking you don't want dialyxir introduced 😄). Do you think it would make sense to tag-release it, to prevent consumers from finding a dialyxir issue.
|
2025-04-01T06:38:33.267066
| 2024-05-02T20:58:47
|
2276470711
|
{
"authors": [
"eljonny"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5623",
"repo": "eljonny/TestCPP",
"url": "https://github.com/eljonny/TestCPP/pull/10"
}
|
gharchive/pull-request
|
Fix catching the object by value.
Fixes #4
Fixed the CPPCheck issue, but this line that was changed is not covered by tests.
Going to create a test for this to get the patch coverage up.
That last commit is for the release build. I should have put it in the preprocessor def as an else though since iostream is already included in the debug build.
Fixy fix.
Fixed a bug where capturing multiple streams in the same test resulted in an Access Violation on Windows cl in Debug.
PR now has sufficient patch coverage for merge.
|
2025-04-01T06:38:33.275215
| 2017-03-17T11:34:50
|
214978284
|
{
"authors": [
"bbqbaron",
"julianjelfs"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5624",
"repo": "elm-community/elm-datepicker",
"url": "https://github.com/elm-community/elm-datepicker/issues/26"
}
|
gharchive/issue
|
Cannot understand how any function that requires the Model can be used
For example when I have an instance of the DatePicker type as returned by init, how can I then call getDate in order to obtain the Maybe Date value? getDate requires a Model which is not what I have.
Without being able to understand how to call getDate, setDate or setFilter I really can't work out how to synchronise two date pickers into a range.
Would really appreciate some help (if I am missing something) or some clarification if a fix of some kind is required.
Happy to help if I can by the way.
I'm glad you noticed that too! It occurred to me recently that those functions wouldn't do anyone any good. You're not missing anything; I think a (hopefully fairly simple) fix is in order to convert them to DatePicker -> instead of Model ->. I'll certainly get to it soon if no one else does.
Hey, if you're happy with that change to the interface, I'm happy to have a go at making that change myself and send you a PR. I'm keen to get involved because I think there are a number of enhancements that we would like to make in the future (keyboard control, multi-month views, internationalisation etc) so it would be good to get familiar with things.
|
2025-04-01T06:38:33.280064
| 2018-09-28T10:44:44
|
364837135
|
{
"authors": [
"RalfNorthman"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5625",
"repo": "elm-community/typed-svg",
"url": "https://github.com/elm-community/typed-svg/issues/27"
}
|
gharchive/issue
|
The type Additive have incorrect variants.
The type Additive have the variants AdditiveNone | AdditiveReplace, but should have AdditiveReplace | AdditiveSum (with correction to the corresponding function in TypesToStrings.elm).
Source: https://developer.mozilla.org/en-US/docs/Web/SVG/Attribute/additive
Comes up when combining transform animations in this way:
<rect x="10" y="10" width="40" height="20"
style="stroke: #000000; fill: none;">
<animateTransform attributeName="transform" attributeType="XML"
type="scale"
from="1" to="3"
begin="0s" dur="10s"
repeatCount="indefinite"
additive="sum"
/>
<animateTransform attributeName="transform" attributeType="XML"
type="rotate"
from="0 30 20" to="360 30 20"
begin="0s" dur="10s"
fill="freeze"
repeatCount="indefinite"
additive="sum"
/>
(http://tutorials.jenkov.com/svg/svg-animation.html)
I think I can fix this.
|
2025-04-01T06:38:33.282393
| 2017-09-11T13:33:27
|
256704017
|
{
"authors": [
"evancz",
"nonpop"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5626",
"repo": "elm-lang/elm-compiler",
"url": "https://github.com/elm-lang/elm-compiler/issues/1643"
}
|
gharchive/issue
|
Compiler freezes when annotating a weird self-referential type
Trying to compile the following code causes the compiler (elm 0.18.0 on Archlinux) to freeze. If I remove the type annotation from idF it will produce an error message about a weird self-referential type, but with the annotation it just gets stuck.
module Test exposing (..)
type alias Focus b s =
{ get : b -> s
, update : (s -> s) -> b -> b
}
create : (b -> s) -> ((s -> s) -> b -> b) -> Focus b s
create get update =
{ get = get
, update = update
}
idF : Focus { r | id : a } a
idF =
create .id (\f s -> { s | id = f }) -- should be { s | id = f s.id }
Development build captures it:
It is not pointing out the infiniteness in an ideal way, but it does not hang at least.
|
2025-04-01T06:38:33.283239
| 2015-10-14T04:33:37
|
111322017
|
{
"authors": [
"HarleyKwyn",
"evancz"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5627",
"repo": "elm-lang/elm-lang.org",
"url": "https://github.com/elm-lang/elm-lang.org/pull/399"
}
|
gharchive/pull-request
|
Fix typo in If Expressions section
Hope this will make it less confusing for people who follow in the future
Nice, thanks!
|
2025-04-01T06:38:33.308122
| 2023-08-19T11:15:34
|
1857677467
|
{
"authors": [
"nothub"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5628",
"repo": "elnormous/contenttype",
"url": "https://github.com/elnormous/contenttype/issues/12"
}
|
gharchive/issue
|
Failed parsing Firefox header
I tried to parse my browsers header and ran into the invalid media type error indicating that the header is syntactically invalid.
Does that mean that Firefox does not implement the header format correctly or is it an error in this project?
I added a test case for the header (text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8).
Result:
=== RUN TestGetMediaType/Firefox_header
contenttype_test.go:221: Unexpected error "invalid media type" for text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8
--- FAIL: TestGetMediaType (0.00s)
Ah nvm, I did not realize there is GetAcceptableMediaTypeFromHeader.
|
2025-04-01T06:38:33.362107
| 2023-10-10T20:58:01
|
1936236335
|
{
"authors": [
"KishiTheMechanic"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5629",
"repo": "elsoul/skeet-solana-mobile-stack",
"url": "https://github.com/elsoul/skeet-solana-mobile-stack/issues/16"
}
|
gharchive/issue
|
Sign In With Solana with Solana Mobile Stack
Solana Mobile Team is implementing the SIWS (Sign In With Solana) API on Mobile Wallet Adapter. (Probably on v2.0)
https://github.com/solana-mobile/mobile-wallet-adapter/issues/439
After that, we can use SIWS on both of Web Apps and Solana dApps on the Saga phone.
For now, SIWS is available with Next.js WebApp with Skeet.
Check /webapp folder for that.
https://github.com/elsoul/skeet-solana-mobile-stack/tree/main/webapp
Wallet Adapter for Web is here
https://github.com/elsoul/skeet-solana-mobile-stack/blob/main/webapp/src/components/providers/SolanaWalletProvider.tsx
Mobile Wallet Adapter specification
Version: 2.0.0-DRAFT
https://solana-mobile.github.io/mobile-wallet-adapter/spec/spec.html
It's done already
|
2025-04-01T06:38:33.369052
| 2016-02-20T12:39:00
|
135067967
|
{
"authors": [
"Zirak",
"xiaq"
],
"license": "bsd-2-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5630",
"repo": "elves/elvish",
"url": "https://github.com/elves/elvish/pull/149"
}
|
gharchive/pull-request
|
edit: Fix crash in history listing mode
When you had too few history entries to display in the terminal,
entering history listing mode attempted to access a negative index and
crashed.
I've wanted to make two additional edits but was unsure about their possible side-effects on other parts.
The first was not entering history listing mode if there was no history. Off I went to startHistoryListing, but got stuck a bit in the semantics. What happens if newHistoryListing returns an error (which makes sense in our situation)? Would the user still be in history listing mode? Perhaps it'll be best then to move the mode switch below the check.
The second was in trimToLines, I badly yearned to put a if len(b.cells) < low check in the beginning, but it may introduce subtle bugs (why aren't things trimming?) if used in certain contexts, and it felt wrong to do it without more knowledge of the system.
Great job on the shell, btw. It's mighty impressive.
Hi! Thanks for the fix. You are correct that the mode switch should be moved down, and I prefer to leave trimToLines as it is, due to the concerns you just stated.
By the way, history listing was something I didn't finish; the ultimate goal is to steal the design ptpython's history listing (#63), which allows you to scroll through the whole history and more importantly, compose a chunk of code by cherry-picking multiple entries from the history. Before trying to implement it, I observed that it has quite a lot in common with existing completion and navigation listings -- esp. wrt. the scrolling and trimming behavior -- so a good abstraction should be made to capture the common behavior. Unable to come up with a good abstraction, I didn't bother to think hard and turned to other parts of elvish instead, leaving this unfinished thing in the code :)
If you find this interesting enough, you are more than welcome to contribute. Keep me informed about your progress, so that I won't rewrite the whole line editor in a midnight and ruin all your efforts.
|
2025-04-01T06:38:33.442053
| 2018-09-28T22:52:08
|
365065757
|
{
"authors": [
"PascalPrecht",
"StatusSceptre",
"cryptomental",
"iurimatias",
"subramanianv"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5631",
"repo": "embark-framework/embark",
"url": "https://github.com/embark-framework/embark/issues/941"
}
|
gharchive/issue
|
add --nobrowser field to test codecoverage.
Outline
Running embark test -c will cause the browser to automatically open with the code coverage report. The goal of this task is to add --nobrowser field for users that don't want this.
Acceptance Criteria
running embark test -c --nobrowser should running the tests & code coverage as normal, but not open the browser.
embark test -c should work as now.
PR opened: https://github.com/embark-framework/embark/pull/950
@iurimatias Seems like the task is already done
Thanks for pointing this out @subramanianv !
Yes this landed as https://github.com/subramanianv/embark/commit/890b46977780d3b3d0199ba8c459c102d6f85596
Closing this one.
@vs77bb this issue won't allow me to payout and has been closed already, can you help?
Hi @StatusSceptre it seems you need to approve me now and I will click on the 'submit' afterwards.
Approved @cryptomental
@StatusSceptre thank you, I submitted via GitCoin.
|
2025-04-01T06:38:33.476620
| 2022-04-23T11:30:44
|
1213277157
|
{
"authors": [
"1homebridge",
"JurgenLB",
"embee8"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5632",
"repo": "embee8/homebridge-panasonic-ac-platform",
"url": "https://github.com/embee8/homebridge-panasonic-ac-platform/issues/26"
}
|
gharchive/issue
|
Status Bar and Cooling Icon
I was using the previous plug-in which used to provide a cooling or heating status in the status bar of each room/level but I am not getting this function with this plug-in. I would also get a change in the colour icon of each unit based on activity from Green to Blue when Cooling and to Orange when Heating. Is this option available with this plug-in.
Thanks
Is this due different plugin functionality or due to a change in HomeKit, i.e. how Apple chooses to display it? Do you have a reference implementation of a plugin that achieves these UI distinctions?
Hi
Thank you for getting back to me. Prior to using this plug-in, I was using the Panasonic Air Conditioner plug-in by Cody1515 which has now been archived. He mentioned that your plug-in is based on the same coding, the difference being that yours uses the platform rather than individual accessories/units (trying to explain this as best as I can but my knowledge is pretty basic) His plug-in also did not initially update the status bar nor did it control the icon colour based on setting, Green for Standby, Blue for Cooling and Orange for Heating. But one of the last updates included these functions which were quite useful. So for example, when using the cooling option, the Green Circle would turn Blue and the Status Bar on the top of the Room/Level would change from Idle to Cooling and note the temperature much like the Heatmiser accessories would show the Room was Heating. I hope this clarifies the “issue”. As I am very dependent on this plug-in, I would be willing to donate but there is not such option on your plug-in.
Many Thanks
I’m attaching screenshots I found of the features I was trying to explain.
This is the Status Bar up top that shows that the AC is on Cooling and the Temperature of the Room
And this is a side by side of my Underfloor Heating on Standby (Green) and the AC on Cooling (Blue) and again showing the Temperature within the circle.
I hope this helps
Thanks again
The results of my debugging session suggest that we might be dealing with a Homebridge or HomeKit bug.
The status icon in the Home app is controlled by the CurrentHeaterCoolerState characteristic of the accessory.
On this line, we set the current state to IDLE when the AC is in cooling mode and the current temperature is less than the set temperature. However, in my test the Home app displays the status of the AC as inactive.
In my debugging session, I also tried setting the current state to INACTIVE instead of IDLE and I got the same result on the UI.
As an additional test, I set the current state to HEATING. The Home app UI reflected this correctly (as Heating), which leaves me thinking that Homebridge or HomeKit don't distinguish between the IDLE and INACTIVE modes.
@1homebridge, upon further exploration I can confirm that I actually see the "Cooling" indicator and the blue arrow in my Home app.
In your previous comments, you uploaded images of how you want it to look like, but can you upload a screenshot of how it actually looks like for you right now?
My previous comment alleged a bug with regards to how IDLE and INACTIVE lead to the same UI representation. It does not, however, confirm your original thesis of the indicators not being available at all. Could you help me clarify the problem statement?
The overlap between the IDLE and INACTIVE statuses is addressed in this issue in the Homebridge repository.
can you set Debug to "true" in your config. Restart and post the output from your log.
Maybe we can resolve the issue then.
My config looks like this;
{ "name": "Homebridge Panasonic AC Platform", "email"<EMAIL_ADDRESS> "password": "xxxxxxxxxxxxxxxxxx", "exposeOutdoorUnit": true, "debugMode": true, "platform": "Panasonic AC Platform" },
@1homebridge, is this the same type of issue as in #23 and #31?
Hi, thanks for getting back to me.
Yes, though with the new UI in iOS16 the Status Bar no longer indicates between Idle-Cooling-Heating states so the issue is just to obtain an indoor temperature reading from the outdoor unit as described in #23 and #31 to stop getting a 0.0* indicator. This would then allow the units to show the Blue colour indicating that they are in Cooling mode as requested in the other threads.
Thanks
Okay, thanks for confirming and the additional context. Will deal with it through the other open issues.
|
2025-04-01T06:38:33.478273
| 2016-07-08T18:27:19
|
164588776
|
{
"authors": [
"mdentremont"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5633",
"repo": "ember-cli/broccoli-concat",
"url": "https://github.com/ember-cli/broccoli-concat/pull/62"
}
|
gharchive/pull-request
|
Update to latest minimatch to avoid deprecation warning
Avoids npm WARN deprecated<EMAIL_ADDRESS>Please update to minimatch 3.0.2 or higher to avoid a RegExp DoS issue
@stefanpenner Awesome!!
|
2025-04-01T06:38:33.589040
| 2024-03-23T07:54:15
|
2203758991
|
{
"authors": [
"bidetofevil"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5634",
"repo": "embrace-io/embrace-android-sdk",
"url": "https://github.com/embrace-io/embrace-android-sdk/pull/617"
}
|
gharchive/pull-request
|
Convert EmbraceDeliveryServiceTest to use a real DeliveryCacheManager
Goal
Testing
Release Notes
WHAT:
WHY:
WHO:
[!WARNING]
This pull request is not mergeable via GitHub because a downstack PR is open. Once all requirements are satisfied, merge this PR as a stack on Graphite.
Learn more
#618
#617 👈
#616
master
This stack of pull requests is managed by Graphite. Learn more about stacking.
Join @bidetofevil and the rest of your teammates on Graphite
Merge activity
Mar 26, 3:09 AM EDT: @bidetofevil started a stack merge that includes this pull request via Graphite.
|
2025-04-01T06:38:33.593815
| 2024-06-12T21:51:31
|
2349740517
|
{
"authors": [
"bidetofevil"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5635",
"repo": "embrace-io/embrace-android-sdk",
"url": "https://github.com/embrace-io/embrace-android-sdk/pull/960"
}
|
gharchive/pull-request
|
Add in an additional step to validate in ExecutionCoordinator
Goal
Testing
Release Notes
WHAT:
WHY:
WHO:
#960 👈
master
This stack of pull requests is managed by Graphite. Learn more about stacking.
Join @bidetofevil and the rest of your teammates on Graphite
Merge activity
Jun 13, 2:22 AM EDT: @bidetofevil merged this pull request with Graphite.
|
2025-04-01T06:38:33.602139
| 2019-03-19T22:42:05
|
422986597
|
{
"authors": [
"emeeks",
"susielu"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5636",
"repo": "emeeks/react-dorling-map",
"url": "https://github.com/emeeks/react-dorling-map/issues/14"
}
|
gharchive/issue
|
Bug: hover data when geo is changed
When hovering, the data onHover is not changed after the data is updated from a subregion to country view.
Resolves in the latest version
|
2025-04-01T06:38:33.603403
| 2021-11-26T00:37:37
|
1063990003
|
{
"authors": [
"LaTonia-Mertica"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5637",
"repo": "emergentworks/mentee-assessments",
"url": "https://github.com/emergentworks/mentee-assessments/pull/3"
}
|
gharchive/pull-request
|
La'Tonia Mertica EW Mentee Assessment (to date)
as much as could complete on my own
Hello Those Powering EW,
Didn't see option to add reviewer(s) et cetera. Apologies if this pull request is in error in any way.
Thanks for this opportunity, please stay safe.
|
2025-04-01T06:38:33.668882
| 2022-12-05T18:33:53
|
1477168867
|
{
"authors": [
"Boscop",
"coderedart"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5638",
"repo": "emilk/egui",
"url": "https://github.com/emilk/egui/issues/2393"
}
|
gharchive/issue
|
Update pure_glow example to glutin 0.30
A lot has changed in glutin 0.30, it would be nice if the pure_glow example could be updated to glutin 0.30 :)
I had some code that was based on the pure_glow example before, now I'm updating it to glutin 0.30, but I'm not sure if I'm doing everything correctly with the new ways of doing things.
You can find my code here: https://github.com/rust-windowing/glutin/issues/1445#issuecomment-1337903593
I would appreciate if you can let me know if it's correct. The pure_glow example could be updated similarly.
@coderedart Thanks. And should I use .with_profile(GlProfile::Core) or not? :)
@coderedart Thanks. And should I use .with_profile(GlProfile::Core) or not? :)
always use Core, unless you are targeting really ancient hardware.
|
2025-04-01T06:38:33.671225
| 2024-05-15T08:46:55
|
2297249972
|
{
"authors": [
"emilk",
"rustbasic"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5639",
"repo": "emilk/egui",
"url": "https://github.com/emilk/egui/pull/4498"
}
|
gharchive/pull-request
|
Fix: Use default features in Image Crate
Closes #4489
Related #4495
Fix: Use default features in Image Crate
Because only .png is available after update #4495.
Required to use JPEG, etc.
No - eframe only need png support, and should not be paying for the compilation of ten other image formats
|
2025-04-01T06:38:33.674810
| 2024-02-15T06:16:35
|
2135733164
|
{
"authors": [
"keeganpotgieter"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5640",
"repo": "emilkowalski/vaul",
"url": "https://github.com/emilkowalski/vaul/issues/269"
}
|
gharchive/issue
|
Add support for iOS PWA app bar
Add the ability to match the colour of the iOS PWA application bar to the overlay colour to create a more native feel.
Examples
Without iOS App Bar Support
With iOS App Bar Support
Solution
User needs these meta tags
<meta name="apple-mobile-web-app-status-bar-style" content="black-translucent" />
<meta name="apple-mobile-web-app-capable" content="yes" />
Changes made in vaul/src/index.tsx to dynamically change the background colour to match overlay background colour
Will also solve some of the issues with this issue https://github.com/emilkowalski/vaul/issues/259
maybe, can solve this issue #199 too
It seems like #199 isn't an issue with body backgrounds.
|
2025-04-01T06:38:33.678981
| 2024-05-28T04:27:43
|
2320095587
|
{
"authors": [
"RickRyan26",
"emilkowalski",
"isaachinman",
"max-17"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5641",
"repo": "emilkowalski/vaul",
"url": "https://github.com/emilkowalski/vaul/issues/360"
}
|
gharchive/issue
|
How to swipe outside of the drawer to open/close?
The ChatGPT mobile app has the most beautiful drawer functionality, and I'm trying to recreate that.
At any time the user can side swipe the chat or drawer area to open or close the drawer and shift the entire page layout.
My main hurdle is being able to swipe outside of the drawer while still being able to select what's beneath that outside swipeable area...
Does my entire layout need to be inside the drawer? With only half of it looking like the drawer? And without the drawer being allowed to fully close?
Any suggestion anyone reading this has would be fantastic.
Thank you!
check MUI swipable drawer
@max-17 I haven't tried this yet but I think a horizontal CSS scroll snap on the layout solves my needs
That's not supported here. The Drawer here is a Dialog meaning that it usually sits on top of other elements.
If you'd want to use this Drawer for something like ChatGPT's mobile app you could position the Drawer off screen initially and add an additional drag event to the body that would translate the content.
@max-17 Did you create a fork and build this functionality?
I was able to achieve this quite easily:
Use a controlled Vaul drawer, ie open and onOpenChange
Add a react-swipeable handler to my layout, and capture specific onSwipeStart events within X pixels of the edge of the screen and set open to true
|
2025-04-01T06:38:33.685485
| 2021-08-01T16:24:30
|
957532127
|
{
"authors": [
"emilyriederer",
"jonspring"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5642",
"repo": "emilyriederer/data-disasters",
"url": "https://github.com/emilyriederer/data-disasters/pull/1"
}
|
gharchive/pull-request
|
Update data-dall.Rmd
Handful of typo corrections & suggested syntax revisions. Looks like this is / will be a great resource!
This is a quick test with just a handful of suggested edits. Happy to modify process if there's a better way to submit these, e.g. with separate commits (e.g. I'm not sure "consumer robotics" company is necessarily better descriptor than "e-commerce" company).
Wow - thanks so much for taking the time, @jonspring ! I really appreciate these
|
2025-04-01T06:38:33.713659
| 2017-09-05T14:55:04
|
255306189
|
{
"authors": [
"NieDzejkob",
"idmean"
],
"license": "Artistic-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5643",
"repo": "emojicode/emojicode",
"url": "https://github.com/emojicode/emojicode/pull/84"
}
|
gharchive/pull-request
|
Add PROJECT_BRIEF to Doxyfile
Come on, why wasn't it done before?
Let me know if I fucked this up, I don't have Doxygen installed.
"why wasn't it done before?" Because I expect people consulting the source code documentation to already know what source they are studying...
|
2025-04-01T06:38:33.763412
| 2023-10-26T20:06:48
|
1964292352
|
{
"authors": [
"ArtiikSK",
"herrardo"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5644",
"repo": "empathyco/x",
"url": "https://github.com/empathyco/x/issues/1342"
}
|
gharchive/issue
|
query to get all elements
what should I specify in the query to get all elements without filters?
Thanks for asking @ArtiikSK
Could you detail what are you trying to do?
Are you launching a web server to try?
You can add a vídeo or screenshot too if you like.
|
2025-04-01T06:38:33.791112
| 2021-07-27T15:56:08
|
954029439
|
{
"authors": [
"ateequrrahman97",
"emretulek"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5645",
"repo": "emretulek/jbvalidator",
"url": "https://github.com/emretulek/jbvalidator/issues/15"
}
|
gharchive/issue
|
How can we display Custom Error message
//custom validate method
validator.validator.custom = function(el, event){ if($(el).is('[name=password]') && $(el).val().length < 5){ return 'Your password is too weak.'; } }
This method only return but it can not display any error on the screen.
The returned string is used as error message, you can define rule and message for specific input.
Okay, thanks @emretulek
|
2025-04-01T06:38:33.856027
| 2015-12-16T04:53:07
|
122428944
|
{
"authors": [
"mediavrog",
"pcboy",
"shawila"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5646",
"repo": "en-japan/koudoku",
"url": "https://github.com/en-japan/koudoku/pull/1"
}
|
gharchive/pull-request
|
update stripe api version
The change logs after 2015-09-08 don't have any major changes in the koudoku code.
LGTM
hitomi :heart:
|
2025-04-01T06:38:33.935786
| 2022-12-22T21:15:07
|
1508553652
|
{
"authors": [
"Eldar2021",
"clangenb"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5647",
"repo": "encointer/encointer-wallet-flutter",
"url": "https://github.com/encointer/encointer-wallet-flutter/issues/927"
}
|
gharchive/issue
|
Schedule notifications upon registering or claiming rewards
Currently, we use the feed to schedule notifications for the registering and meetup reminders, but we can actually schedule them from within the app when we do certain actions.
After registering, we can schedule two notifications, which should be reminders for the next meetup:
24 hours before the meetup time
1 hour before the meetup time
The meetup time can be fetched here: https://github.com/encointer/encointer-wallet-flutter/blob/6fc236c622932d2b503f2ae5055c846dd6c7ccaa/lib/store/encointer/sub_stores/community_store/community_store.dart#L64. The value can't be null after registering, as it can only be null if we have not chosen a community.
After claiming the rewards, we can set a reminder to register at the start of the registering phase. Note: We must only schedule the notification if we claim the rewards in the attesting phase, not in the registering phase, obviously.
The timestamp to use is the the nextPhaseTimestamp: https://github.com/encointer/encointer-wallet-flutter/blob/6fc236c622932d2b503f2ae5055c846dd6c7ccaa/lib/store/encointer/encointer.dart#L81
Caveats:
How do we ensure unique IDs that don't overlap with the IDs we assign to the notifications we get from the feed? Maybe we have to maintain an internal global counter for that.
We should only schedule notifications if we are connected to the parachain on kusama, the nctr-k
If we use global counter How do we ensure unique IDs that don't overlap with the IDs we assign to the notifications we get from the feed? Maybe we have to maintain an internal global counter for that.
Can we use cache?
I suggest generate meetupId by meetupTime
void main() {
final meetupTimeAfter1Day = DateTime.now().add(const Duration(days: 1)).millisecondsSinceEpoch;
final meetupTimeAfter7Days = DateTime.now().add(const Duration(days: 7)).millisecondsSinceEpoch;
final meetupTimeAfter10Days = DateTime.now().add(const Duration(days: 10)).millisecondsSinceEpoch;
final meetupTimeAfter15Days = DateTime.now().add(const Duration(days: 15)).millisecondsSinceEpoch;
final meetupTimeAfter30Days = DateTime.now().add(const Duration(days: 30)).millisecondsSinceEpoch;
print(generateMeetupIdByTimeStamp(meetupTimeAfter1Day)); // 1
print(generateMeetupIdByTimeStamp(meetupTimeAfter7Days)); // 7
print(generateMeetupIdByTimeStamp(meetupTimeAfter10Days)); // 10
print(generateMeetupIdByTimeStamp(meetupTimeAfter15Days)); // 15
print(generateMeetupIdByTimeStamp(meetupTimeAfter30Days)); // 30
}
int generateMeetupIdByTimeStamp(int meetupTime) {
final now = DateTime.now().millisecondsSinceEpoch;
int id = 0;
int c = meetupTime - now;
do {
c -= 86400000; // 1 day = 86400000 milliseconds
id++;
} while (c > 0);
return id;
}
So I think we can generate dynamic id without using any cache. Please let me know your opinion.
Hi @Eldar2021,
You suggestions look good!
When user registers to a meetup, can he register to another meetup before that meetup ends?
Yes, this is possible, but the app should not allow that because you have not yet got your reputation and you will be a newbie again. You should only be allowed to do that after the rewards have been claimed.
If we use global counter How do we ensure unique IDs that don't overlap with the IDs we assign to the notifications we
In general, I like your approach of using the meetup time. Why do you need to divide by 86400000, is the number too big? I think this could lead to problems where we end up with the same ID for reminders, which are very close to eachother, do you agree?
Is there registration deadline time? or registration finishes when meetup starts?
The registration deadline is when the registering phase is over. Remember, we have 3 phases REGISTERING > ASSIGNING > ATTESTING. As long as we are in the registering phase, the nextPhaseTimeStamp is the deadline for registering. And in the assigning phase you can already register for the meetup in the next assigning phase.
|
2025-04-01T06:38:33.960895
| 2019-06-04T16:43:42
|
452099739
|
{
"authors": [
"Shadowblitz16",
"Shylie",
"endlesstravel"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5648",
"repo": "endlesstravel/Love2dCS",
"url": "https://github.com/endlesstravel/Love2dCS/issues/75"
}
|
gharchive/issue
|
Convert Image to Canvas?
Is it possible to convert a image to a canvas and back so I can draw onto it?
if not could you add Graphics.NewCanvas(Image Image) and Graphics.NewCanvas(ImageData ImageData)?
There is no such method, but you can draw an image onto the canvas.
but this requires the draw event right?
I was trying to do it so it could be used anywhere
See Love.Graphics.Present. This should be what you are looking for.
can support be added so that images and canvases can be explicitly converted back and forth?
a example here :
copy the utils class into your project:
https://gist.github.com/endlesstravel/027799eb772d644b0d4110284256da6a
use it like :
static public void Test_Issue75_ToPintImage()
{
ISSUE_75.Init();
var imgData = ISSUE_75.PrintImage(300, 300, () =>
{
Graphics.SetColor(Color.LightPink);
Graphics.Rectangle(DrawMode.Fill, 0, 0, 100, 100);
Graphics.SetColor(Color.White);
Graphics.Circle(DrawMode.Line, 100, 100, 20);
});
Resource.EncodeToFile("test.png", imgData, ImageFormat.PNG);
}
can this be added to love2dcs as a explicit image conversion?
|
2025-04-01T06:38:33.963312
| 2022-04-27T00:44:36
|
1216637282
|
{
"authors": [
"Evernow",
"istiak101"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5649",
"repo": "endoflife-date/endoflife.date",
"url": "https://github.com/endoflife-date/endoflife.date/pull/1102"
}
|
gharchive/pull-request
|
[Fedora-Linux] Update EOL dates and command
F36 was delayed, which pushed back the F34 end of life date: https://fedorapeople.org/groups/schedule/f-36/f-36-key-tasks.html
F35 EOL date from: https://fedorapeople.org/groups/schedule/f-37/f-37-key-tasks.html
As for the command change, it's due to lsb_release not being installed by default anymore
Fedora is delayed again. EOL moved to 2022-06-07. Could be delayed more.
|
2025-04-01T06:38:33.971651
| 2017-06-27T05:08:34
|
238738518
|
{
"authors": [
"DASPRiD",
"MelchiorKokernoot",
"endroid",
"mortezakarimi"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5650",
"repo": "endroid/qr-code",
"url": "https://github.com/endroid/qr-code/issues/114"
}
|
gharchive/issue
|
Feature request: style dots and corners
If you can add more flexible feature to style qt such as this
I am very grateful! :grin:
Hi @mortezakarimi thank you. I will not add this myself but if someone is willing to contribute on this, that would be great.
The new upstream QR code library supports that now.
@DASPRiD What do you mean by "The new upstream QR code library"?
@DASPRiD What do you mean by "The new upstream QR code library"?
BaconQRCode, which this library is using.
@DASPRiD Do you by any chance have a reference where I can learn how to do so?
I can't seem to find anything about changing the style.
@DASPRiD Do you by any chance have a reference where I can learn how to do so?
I can't seem to find anything about changing the style.
There's not really any documentation for this, but the ImageRenderer takes a RendererStyle object, which can be configured. Best to look at the source:
https://github.com/Bacon/BaconQrCode/blob/master/src/Renderer/ImageRenderer.php#L26
https://github.com/Bacon/BaconQrCode/blob/master/src/Renderer/RendererStyle/RendererStyle.php
Closed as this will not be implemented here.
|
2025-04-01T06:38:33.989565
| 2024-01-17T23:43:01
|
2087197774
|
{
"authors": [
"engeir"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5651",
"repo": "engeir/volcano-core",
"url": "https://github.com/engeir/volcano-core/pull/8"
}
|
gharchive/pull-request
|
chore(main): release 0.2.3
:robot: I have created a release beep boop
0.2.3 (2024-01-17)
Bug Fixes
mise tasks are experimental and must be activated (4ab703b)
This PR was generated with Release Please. See documentation.
:robot: Release is at https://github.com/engeir/volcano-core/releases/tag/v0.2.3 :sunflower:
|
2025-04-01T06:38:33.997629
| 2016-07-03T22:38:37
|
163583363
|
{
"authors": [
"engram-design",
"pixeljitsu"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5652",
"repo": "engram-design/SocialPoster",
"url": "https://github.com/engram-design/SocialPoster/issues/3"
}
|
gharchive/issue
|
Links to social media posts
It would be really handy if there were a way to access the links leading to the social media posts from within the entry. I'd like to automatically display the links as part of the craft entry. "Join the discussion on Facebook/Twitter", etc.
Just to confirm, this would be from the front-end? This way, you could post the entry to Facebook, have it record the posted URL, and then being able to access in your template?
Something like:
{% set post = craft.socialPoster.post({ account: 'facebook', entryId: entry.id }) %}
{% if post.url %}
<a href="{{ post.url }}" target="_blank">Join the discussion on Facebook</a>
{% endif %}
Exactly. That way you could drive traffic from the website to the social media posts.
Just to follow this up - this is now implemented in 1.2.0. Use the following template code:
{% set posts = craft.socialPoster.posts({ element: entry }) %}
{% for post in posts %}
<a href="{{ post.url }}" target="_blank">
<i class="fa fa-{{ post.handle }}-square"></i> Join the discussion on {{ post.handle | capitalize }}
</a>
{% endfor %}
|
2025-04-01T06:38:34.000913
| 2024-12-13T11:53:32
|
2738220999
|
{
"authors": [
"ambravo",
"engswee"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5653",
"repo": "engswee/flashpipe",
"url": "https://github.com/engswee/flashpipe/issues/40"
}
|
gharchive/issue
|
Repo and Org Variables
I wonder if this statement holds true:
https://github.com/engswee/flashpipe/blob/687862ea650f903531ae0dfe13d1459582c95847/docs/github-actions-sync-apim.md?plain=1#L28
Github Actions offers Repo and Org variables (doc)
Are you referring to something else?
Thanks for highlighting this. Unfortunately, documentation often can't keep up with the speed that new features are introduced 😅
If you see the screenshot from the following page, you can see that variables were not there in the past.
https://engswee.github.io/flashpipe/github-actions-sync-to-git.html
I'd have to admit that I can't keep track of all the new features that are constantly being rolled out, so have definitely missed this one out. It's good to know about this, so that I can use it in my workflows and also update the documentation.
I'll keep this issue open until I get around to updating the documentation 😉
|
2025-04-01T06:38:34.019057
| 2021-10-12T00:44:47
|
1023243522
|
{
"authors": [
"lognaturel",
"yanokwa"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5654",
"repo": "enketo/enketo-core",
"url": "https://github.com/enketo/enketo-core/pull/828"
}
|
gharchive/pull-request
|
Prepare 6.0.1
Stepping through the commits is the easiest way to review this because of the docs changes.
Draft because we are waiting on https://github.com/enketo/openrosa-xpath-evaluator/pull/136 and then a 2.0.9 release.
Verified the following:
[ ] npm update
[ ] npm audit fix --production
[ ] npm run test
[ ] npm run test-browsers
[ ] npm run beautify
[ ] npm run build-docs
npm run test-browers has a scary Firefox failure. It'd be good to see if it fails on v5.17.6.
```Firefox 93.0 (Mac OS 10.15) merging an instance into the model when the record contains namespaced attributes, the merged result is CORRECTLY namespaced namespaces are added correctly FAILED``
https://github.com/enketo/enketo-core/pull/822 should also be merged first, right?
|
2025-04-01T06:38:34.031071
| 2021-06-15T08:11:07
|
921127286
|
{
"authors": [
"poi33"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5655",
"repo": "enonic/app-office-league",
"url": "https://github.com/enonic/app-office-league/issues/462"
}
|
gharchive/issue
|
Site created on startup
The default site is created on application start. WIth app and info-page content types, including populated fields.
Should we just remove the site init code and let the user manually setup office-league with site office-league?
This would still keep the init of the office-league repoes but remove the auto generated site.
Remove the auto generated site. Keep the generated repo storrage.
|
2025-04-01T06:38:34.070386
| 2023-11-23T16:51:25
|
2008567301
|
{
"authors": [
"a8m",
"michaelcaulley"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5656",
"repo": "ent/ent",
"url": "https://github.com/ent/ent/pull/3841"
}
|
gharchive/pull-request
|
fix: swap the order of handling order and pred in the graph select func
multi order may generate predicates, so running the order funcs first allows reusing aliases.
If the predicates run first if there is a multiOrder field that generates it's own predicates, the predicates are unable to find the alias of the join and errors not finding the column on the table.
Related to: https://github.com/ent/contrib/pull/559
Thanks for the contribution, @michaelcaulley 🚀
|
2025-04-01T06:38:34.094383
| 2016-02-10T07:04:26
|
132623609
|
{
"authors": [
"flyinbutrs",
"gstamp",
"neillturner",
"stevehodgkiss",
"thekindofme"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5657",
"repo": "envato/stack_master",
"url": "https://github.com/envato/stack_master/issues/81"
}
|
gharchive/issue
|
Add support for more template generators
YAML
cfndsl - https://github.com/stevenjack/cfndsl
Any others?
It might be nice to have JSON support with comments. Would need to find a library to strip them out I guess.
Another one might be just using plain old ERB?
If you want JSON support with comments, technically you can just use YAML. An alternate syntax of YAML is actually just JSON + comments, so any YAML parser should be able to parse a JSON file with comments in it.
[1] pry(main)> require 'yaml'
=> true
[2] pry(main)> test = "{\n#testing\n}"
=> "{\n#testing\n}"
[3] pry(main)> YAML.load(test)
=> {}
So yeah, +1 for YAML input. should be pretty trivial to add, I'll fork and see.
cfndsl support would be really good as it is the most popular generator of cloudformation templates that is written in ruby. it should be really easier to integrate as it is just running cfndsl commands.
I also use the ppjson to pretty print the generated cloudformation code.
I've built alot of examples for cfndsl at https://github.com/neillturner/cfndsl_examples
and there is a utility to convert templates to cfndsl format that is very handy.
Hi @neillturner we have added basic CfnDsl template support with #99. However, this dose not add support for using cfndsl variables. How would you like that supported in SM? How would you imagine a user supply values for cfndsl variables when using a cfndsl template with SM?
|
2025-04-01T06:38:34.097497
| 2017-09-27T22:16:04
|
261135160
|
{
"authors": [
"ggreenway",
"htuch",
"mattklein123"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5658",
"repo": "envoyproxy/envoy",
"url": "https://github.com/envoyproxy/envoy/issues/1761"
}
|
gharchive/issue
|
Number of stats and length of stat name should be configurable
Envoy currently hardcodes a limit of 16k individual stats, with a fixed maximum name length. These values should be tunable.
+1
Defaults should be settable at compile-time via .bazelrc (or similar), and overridable via CLI flag at startup.
@hennna
It is easy to make finding unused stats O(1) with a free-list or similar. But doing the name-lookups is a little bit harder in shared memory. We'll probably need a lookup table of some kind in the shared memory.
If we want to do this, I would recommend doing it as a totally independent change from the length stuff. Given how low frequency stat allocation is from shared memory the current situation is probably not that big of a deal for most people.
I was going to benchmark filling up a somewhat big number of stats, maybe 1M or something, and see how long it takes. But I'm expecting that would be pretty slow: n^2 is 1 trillion operations.
But I agree that it can be done in a separate change. Making the sizes tunable won't cause any degradation to existing use cases unless users opt-in to a really large stat size, and they should immediately notice some pain on startup if they choose too-large of a size.
|
2025-04-01T06:38:34.102156
| 2023-04-12T17:36:46
|
1664972971
|
{
"authors": [
"XinyiZhangAws",
"mattklein123",
"ytsssun"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5659",
"repo": "envoyproxy/envoy",
"url": "https://github.com/envoyproxy/envoy/issues/26710"
}
|
gharchive/issue
|
Add Attributes connection.transport_failure_reason
Title: Add Attributes connection.transport_failure_reason along with TLS connection errors
Description:
Describe the desired behavior, what scenario it enables and how it
would be used.
Purposed Attributes
connection.transport_failure_reason:
Currently upstream.transport_failure_reason is included in the attributes to indicate the upstream transport failure such as certificate validation failed. But there's no similar ones for downstream connection. Considering recent PR added downstream transport failure reason to access log, it can be added to attributes as well for consistency.
[optional Relevant Links:]
Any extra documentation required to understand the issue.
https://www.envoyproxy.io/docs/envoy/latest/intro/arch_overview/advanced/attributes.html#attributes
https://github.com/envoyproxy/envoy/pull/25322/files
cc @kyessenov
@mattklein123 could you add a "help wanted" tag to revive this issue?
|
2025-04-01T06:38:34.109718
| 2024-09-09T15:02:01
|
2514216849
|
{
"authors": [
"alyssawilk",
"cainelli",
"ravenblackx",
"tyxia"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5660",
"repo": "envoyproxy/envoy",
"url": "https://github.com/envoyproxy/envoy/issues/36041"
}
|
gharchive/issue
|
Test flake: //test/extensions/filters/http/ext_proc:ext_proc_integration_test
https://dev.azure.com/cncf/envoy/_build/results?buildId=179508&view=logs&j=4930ecaf-18f4-5b3c-dea3-309729c3b3ae&t=573d8780-d7b9-52e3-b4e0-a89886b0b9ff&l=3840
[ RUN ] IpVersionsClientTypeDeferredProcessing/ExtProcIntegrationTest.GetAndCloseStreamWithTracing/IPv4_GoogleGrpc_WithDeferredProcessing
test/extensions/filters/http/ext_proc/tracer_test_filter.cc:52: Failure
Expected equality of these values:
want
Which is: "0"
got
Which is: ""
grpc.status_code: 0 not found in tags:
component: proxy
status: canceled
upstream_address: <IP_ADDRESS>:37667
upstream_cluster: ext_proc_server_0
@tyxia
@cainelli Do you mind taking a look at this? Thank you!
I think tracing related feature was added by your change
oh sorry about that. I will take a look into it this week.
@tyxia the failure is a timeout and the tracing failure is a side effect. I don't see how such simple request would take more than 5s to run.
test/extensions/filters/http/ext_proc/ext_proc_integration_test.cc:281: Failure
Value of: fake_upstreams_[0]->waitForHttpConnection(*dispatcher_, fake_upstream_connection_)
Actual: false (Timed out waiting for new connection.)
Expected: true
....
test/extensions/filters/http/ext_proc/ext_proc_integration_test.cc:277: Failure
Expected equality of these values:
std::to_string(status_code)
Which is: "200"
response.headers().getStatusValue()
Which is: "504"
Stack trace:
0x1578018: (unknown)
0x13cb8bc: (unknown)
0x7fc368411a4d: testing::internal::HandleSehExceptionsInMethodIfSupported<>()
0x7fc3683f822e: testing::internal::HandleExceptionsInMethodIfSupported<>()
0x7fc3683dfb1d: testing::Test::Run()
0x7fc3683e060e: testing::TestInfo::Run()
... Google Test internal frames ...
actually, is the timeout 5ms? should we increase it a bit?
https://github.com/envoyproxy/envoy/blob/7a7df5d8887dfe673eef51ce396feab4bff9383f/test/integration/http_integration.cc#L555-L556
@cainelli Thanks for spending effort reducing the flakiness.
In the past, we have test (as you linked in slack) that have larger then 5s timeout. However, that is because Please don't waitForHttpConnection with a 5s timeout if failure is expected. In your case, failure is not expected.
I am not sure if the flakiness is because tracing will take a bit more time but 5s should be sufficient here. Or maybe because ext_proc_integration test has grown very big now. Do you happen to know what is the flakiness rate?
if you can't repro the flake (per slack) one thing you can do is add a LogLevelSetter in that test such that CI logs more information when it flakes. then next time we see a failure you'll have more information.
One thing I've found often helps if you can't repro a flake is to run stress -c 16 (or however many CPU cores) in another terminal while the test runs with --runs_per_test=n. (Flags also depending on if the flakiness is from being CPU bound or network bound or disk bound.)
Thank you all for the context and tips.
One thing I've found often helps if you can't repro a flake is to run stress -c 16 (or however many CPU cores) in another terminal while the test runs with --runs_per_test=n. (Flags also depending on if the flakiness is from being CPU bound or network bound or disk bound.)
I did tried that with various combinations to stress during the test but did not have any luck reproducing it.
if you can't repro the flake (per slack) one thing you can do is add a LogLevelSetter in that test such that CI logs more information when it flakes. then next time we see a failure you'll have more information.
I will try this path moving forward (https://github.com/envoyproxy/envoy/pull/36583).
|
2025-04-01T06:38:34.113181
| 2024-10-31T09:05:26
|
2626347837
|
{
"authors": [
"nezdolik",
"supercairos"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5661",
"repo": "envoyproxy/envoy",
"url": "https://github.com/envoyproxy/envoy/issues/36931"
}
|
gharchive/issue
|
Implement a Token Introspection (RFC 7662) HTTP Filter
Token Introspection (RFC 7662): Implement an HTTP Filter to verify external token
Description:
Allow external OAuth2/OIDC token to be valided via the Token Introspection api in order for envoy to act as a Identity Aware Proxy (IAP)
Relevant Links:
https://datatracker.ietf.org/doc/html/rfc7662
https://www.oauth.com/oauth2-servers/token-introspection-endpoint/
Please get familiar with our extension policy: https://github.com/envoyproxy/envoy/blob/main/EXTENSION_POLICY.md
cc @tyxia @mattklein123 @TAOXUY (as oauth, jwt extension owners who may be interested in this proposal)
|
2025-04-01T06:38:34.132482
| 2019-10-17T14:39:47
|
508523044
|
{
"authors": [
"HenryYYang",
"mosespx",
"ramaraochavali",
"saagar241290"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5662",
"repo": "envoyproxy/envoy",
"url": "https://github.com/envoyproxy/envoy/issues/8644"
}
|
gharchive/issue
|
Redis through envoy gets high response time with redis-benchmark
Hi,
I’m deploying envoy redis in our environment and I’ve found out that it’s increasing the RTT of the 95 percentile requests to redis in about 7~ms, I have run redis-benchmark tool in both configurations, and through envoy the higher percents getting a much higher response time. results attached.
this is the configuration file:
static_resources:
listeners:
- address:
socket_address:
address: <IP_ADDRESS>
port_value: 50051
filter_chains:
- filters:
- name: envoy.http_connection_manager
config:
codec_type: auto
stat_prefix: ingress_http
route_config:
name: local_route
virtual_hosts:
- name: local_service
domains:
- "*"
routes:
- match:
prefix: "/"
route:
cluster: local_service_grpc
http_filters:
- name: envoy.router
config: {}
- name: redis_nrt_listener
address:
socket_address:
address: <IP_ADDRESS>
port_value: 6379
filter_chains:
- filters:
- name: envoy.redis_proxy
typed_config:
"@type": type.googleapis.com/envoy.config.filter.network.redis_proxy.v2.RedisProxy
stat_prefix: egress_redis
settings:
op_timeout: 0.03s
enable_redirection: true
enable_hashtagging: true
prefix_routes:
catch_all_route:
cluster: redis_nrt_cluster
- name: redis_vol_listener
address:
socket_address:
address: <IP_ADDRESS>
port_value: 6380
filter_chains:
- filters:
- name: envoy.redis_proxy
typed_config:
"@type": type.googleapis.com/envoy.config.filter.network.redis_proxy.v2.RedisProxy
stat_prefix: egress_redis
settings:
op_timeout: 0.03s
enable_redirection: true
enable_hashtagging: true
prefix_routes:
catch_all_route:
cluster: redis_vol_cluster
clusters:
- name: local_service_grpc
connect_timeout: 0.250s
type: logical_dns
lb_policy: round_robin
http2_protocol_options: {}
health_checks:
- timeout: 1s
interval: 3s
interval_jitter: 1s
unhealthy_threshold: 3
healthy_threshold: 3
tcp_health_check:
send:
receive: []
hosts:
- socket_address:
address: router-us-east4-b-prod.ocddx.com
port_value: 50051
- name: redis_vol_cluster
connect_timeout: 1s
type: strict_dns # static
lb_policy: MAGLEV
load_assignment:
cluster_name: redis_cluster
endpoints:
- lb_endpoints:
- endpoint:
address:
socket_address:
address: redis-us-east4-b-prd.ocddx.com
port_value: 6379
- name: redis_nrt_cluster
connect_timeout: 1s
type: strict_dns # static
lb_policy: MAGLEV
load_assignment:
cluster_name: redis_cluster
endpoints:
- lb_endpoints:
- endpoint:
address:
socket_address:
address: redis-nrt-us-east4-b-prd.ocddx.com
port_value: 6379
admin:
access_log_path: "/var/log/envoy_admin_access.log"
address:
socket_address:
address: <IP_ADDRESS>
port_value: 9901
redis-benchmark running from container to the envoy sidecar:
redis-benchmark -h collector_envoy -t set,get
====== SET ======
100000 requests completed in 7.06 seconds
50 parallel clients
3 bytes payload
keep alive: 1
4.92% <= 1 milliseconds
59.69% <= 2 milliseconds
73.69% <= 3 milliseconds
82.54% <= 4 milliseconds
88.11% <= 5 milliseconds
91.77% <= 6 milliseconds
94.26% <= 7 milliseconds
95.89% <= 8 milliseconds
97.18% <= 9 milliseconds
98.01% <= 10 milliseconds
98.58% <= 11 milliseconds
98.97% <= 12 milliseconds
99.22% <= 13 milliseconds
99.43% <= 14 milliseconds
99.58% <= 15 milliseconds
99.73% <= 16 milliseconds
99.78% <= 17 milliseconds
99.81% <= 18 milliseconds
99.88% <= 19 milliseconds
99.90% <= 20 milliseconds
99.90% <= 21 milliseconds
99.93% <= 22 milliseconds
99.97% <= 23 milliseconds
99.98% <= 24 milliseconds
99.98% <= 28 milliseconds
100.00% <= 29 milliseconds
14164.31 requests per second
====== GET ======
100000 requests completed in 7.66 seconds
50 parallel clients
3 bytes payload
keep alive: 1
4.18% <= 1 milliseconds
55.05% <= 2 milliseconds
71.89% <= 3 milliseconds
81.03% <= 4 milliseconds
87.10% <= 5 milliseconds
90.96% <= 6 milliseconds
93.66% <= 7 milliseconds
95.49% <= 8 milliseconds
96.44% <= 9 milliseconds
97.22% <= 10 milliseconds
97.75% <= 11 milliseconds
98.34% <= 12 milliseconds
98.83% <= 13 milliseconds
99.15% <= 14 milliseconds
99.34% <= 15 milliseconds
99.49% <= 16 milliseconds
99.59% <= 17 milliseconds
99.69% <= 18 milliseconds
99.72% <= 19 milliseconds
99.81% <= 20 milliseconds
99.87% <= 21 milliseconds
99.89% <= 22 milliseconds
99.91% <= 23 milliseconds
99.92% <= 24 milliseconds
99.93% <= 25 milliseconds
99.96% <= 26 milliseconds
99.97% <= 27 milliseconds
99.97% <= 28 milliseconds
99.98% <= 29 milliseconds
99.98% <= 34 milliseconds
99.98% <= 35 milliseconds
99.99% <= 36 milliseconds
100.00% <= 37 milliseconds
13063.36 requests per second
redis-benchmark running from container to a single redis host (part of a masters cluster):
bash-4.4# redis-benchmark -h <IP_ADDRESS> -t set,get
====== SET ======
100000 requests completed in 4.95 seconds
50 parallel clients
3 bytes payload
keep alive: 1
11.96% <= 1 milliseconds
94.58% <= 2 milliseconds
96.80% <= 3 milliseconds
98.02% <= 4 milliseconds
98.75% <= 5 milliseconds
99.27% <= 6 milliseconds
99.58% <= 7 milliseconds
99.75% <= 8 milliseconds
99.89% <= 9 milliseconds
99.96% <= 10 milliseconds
99.97% <= 11 milliseconds
99.97% <= 12 milliseconds
99.99% <= 13 milliseconds
100.00% <= 13 milliseconds
20197.94 requests per second
====== GET ======
100000 requests completed in 5.07 seconds
50 parallel clients
3 bytes payload
keep alive: 1
15.41% <= 1 milliseconds
90.76% <= 2 milliseconds
95.60% <= 3 milliseconds
97.79% <= 4 milliseconds
98.70% <= 5 milliseconds
99.23% <= 6 milliseconds
99.60% <= 7 milliseconds
99.76% <= 8 milliseconds
99.82% <= 9 milliseconds
99.83% <= 11 milliseconds
99.87% <= 12 milliseconds
99.89% <= 13 milliseconds
99.91% <= 14 milliseconds
99.97% <= 15 milliseconds
99.99% <= 16 milliseconds
100.00% <= 16 milliseconds
19704.43 requests per second
As you can see there's a huge different between the response times, I've been trying to change some configuration for example: type to logical_dns instead of strict_dns, remove the lb_type and add max_buffer_size_before_flush and buffer_flush_timeout and even change the dns to point to only one member of the redis cluster, the same host I checked in the second test, to ensure the reliability of the redis-benchmark test.
I'd be glad if someone who using redis with envoy will do the same test I did and share the results, and if someone has any recommendations to solve this response time issue
@zuercher please add a BUG label, seems like its happening also in other environments
Are you introducing a network hop in the envoy case in your test?
Here's the result for my test against a local docker container:
root@8362fe3593b4:/# redis-benchmark -h redis-server -p 7001 -t get,set
====== SET ======
100000 requests completed in 3.07 seconds
50 parallel clients
3 bytes payload
keep alive: 1
84.07% <= 1 milliseconds
99.09% <= 2 milliseconds
99.84% <= 3 milliseconds
99.93% <= 4 milliseconds
99.94% <= 5 milliseconds
99.96% <= 6 milliseconds
99.96% <= 9 milliseconds
99.96% <= 10 milliseconds
99.97% <= 11 milliseconds
99.99% <= 26 milliseconds
100.00% <= 27 milliseconds
100.00% <= 27 milliseconds
32626.43 requests per second
====== GET ======
100000 requests completed in 3.04 seconds
50 parallel clients
3 bytes payload
keep alive: 1
85.27% <= 1 milliseconds
98.79% <= 2 milliseconds
99.85% <= 3 milliseconds
99.97% <= 4 milliseconds
99.97% <= 9 milliseconds
99.99% <= 10 milliseconds
100.00% <= 11 milliseconds
32894.74 requests per second
root@8362fe3593b4:/# redis-benchmark -p 6381 -t get,set
====== SET ======
100000 requests completed in 3.78 seconds
50 parallel clients
3 bytes payload
keep alive: 1
0.00% <= -32 milliseconds
0.01% <= -30 milliseconds
0.02% <= -29 milliseconds
0.03% <= -28 milliseconds
0.03% <= -26 milliseconds
0.04% <= 0 milliseconds
24.19% <= 1 milliseconds
89.52% <= 2 milliseconds
99.04% <= 3 milliseconds
99.78% <= 4 milliseconds
99.82% <= 5 milliseconds
99.84% <= 6 milliseconds
99.85% <= 7 milliseconds
99.87% <= 8 milliseconds
99.89% <= 10 milliseconds
99.94% <= 11 milliseconds
99.96% <= 12 milliseconds
99.97% <= 13 milliseconds
99.98% <= 14 milliseconds
99.98% <= 15 milliseconds
99.98% <= 16 milliseconds
99.99% <= 19 milliseconds
100.00% <= 20 milliseconds
26462.03 requests per second
====== GET ======
100000 requests completed in 3.05 seconds
50 parallel clients
3 bytes payload
keep alive: 1
28.07% <= 1 milliseconds
96.50% <= 2 milliseconds
99.51% <= 3 milliseconds
99.80% <= 4 milliseconds
99.95% <= 5 milliseconds
100.00% <= 6 milliseconds
100.00% <= 6 milliseconds
32797.64 requests per second
Here's the relevant section of envoy.yaml:
static_resources:
listeners:
- name: listener_1
address:
socket_address:
address: <IP_ADDRESS>
port_value: 6381
filter_chains:
filters:
name: envoy.redis_proxy
config:
stat_prefix: redis_stats
prefix_routes:
catch_all_route:
cluster: cluster_1
settings:
op_timeout: 5s
clusters:
- name: cluster_1
connect_timeout: 0.25s
lb_policy: RING_HASH
hosts:
- socket_address:
address: redis-server
port_value: 7001
type: STRICT_DNS
envoy is a sidecar container on the application
I ran both tests on the envoy container to get an apple to apple measurement.
@mosespx i am facing the same issue. Did you got any solution or workaround?
@saagar241290 no, I didn't use this solution because of this issue.
please share here if you find something interesting
@mosespx I tried by increasing number of connections of redis pool to 100 and it gave me a better performance. Earlier there was only a single connection.
@saagar241290 when you says
tried by increasing number of connections of redis pool to 100 and it gave me a better performance. Earlier there was only a single connection.
Are these connections on client?
|
2025-04-01T06:38:34.135706
| 2019-11-14T14:54:41
|
522910625
|
{
"authors": [
"sriduth"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5663",
"repo": "envoyproxy/envoy",
"url": "https://github.com/envoyproxy/envoy/issues/9025"
}
|
gharchive/issue
|
pre startup checks for windows: log if system variables need changing
pre startup checks: Check os variables before startup on windows
Description:
As a fix for issue https://github.com/envoyproxy/envoy/issues/7130, PR #8600 introduces an interface and a posix implementation that runs platform specific checks before startup.
The win32 implementation source/exe/win32/platform_checks.cc adds a no-op which should be implemented.
#7130 is fixed via #9098, which just adds user documentation instead of moving checks into envoy source code.
|
2025-04-01T06:38:34.148945
| 2019-12-16T06:44:54
|
538217268
|
{
"authors": [
"alexburnos",
"htuch",
"kyessenov",
"lambdai",
"mandarjog",
"mattklein123"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5664",
"repo": "envoyproxy/envoy",
"url": "https://github.com/envoyproxy/envoy/issues/9358"
}
|
gharchive/issue
|
API Request: Add HttpFilter.instance_name
HttpFilter.name is used to instantiate a specific type of filter by http_connection_manager
In that sense HttpFilter.name is the className parameter.
In filters like WASM and lua, a filter with the same name is deployed multiple times.
This makes it difficult to identify a specific filter instance without resorting to peeking inside HttpFilter.typed_config.
Add
message HttpFilter {
// The name of the filter to instantiate. The name must match a
// :ref:`supported filter <config_http_filters>`.
string name = 1 [(validate.rules).string = {min_bytes: 1}];
string instance_name = 5;
// ...
}
Using instance_name, the filter can be identified in a meaningful way.
In the following example, without instance_name=wasm.stats both filters have the same name envoy.filters.http.wasm.
filters:
- name: envoy.http_connection_manager
typed_config:
'@type': type.googleapis.com/envoy.config.filter.network.http_connection_manager.v2.HttpConnectionManager
forward_client_cert_details: APPEND_FORWARD
generate_request_id: true
http_filters:
- instance_name: wasm.stats
name: envoy.filters.http.wasm
config:
config:
root_id: stats_inbound
vm_config:
code:
inline_string: envoy.wasm.stats
runtime: envoy.wasm.runtime.null
vm_id: stats_inbound
- instance_name: wasm.metadata_exchange
name: envoy.filters.http.wasm
config:
config:
vm_config:
code:
inline_string: envoy.wasm.metadata_exchange
runtime: envoy.wasm.runtime.null
Here
@kyessenov @rshriram
My initial sniff test is that this is not generally useful, and I don't see why it's that bad to peak inside the typed_config, but happy to be swayed if there is a convincing argument.
I think this would be useful for Filter Config discovery service. Having a common field across all filter configs helps in discovery and monitoring IMHO. I don't think it's possible to express that Any in the typed config must have a specific field, so pulling it up one level to filter config seems reasonable:
# static config
name: envoy.http.wasm
config_name: wasm1
typed_config: {}
# dynamic config
name: envoy.http.wasm
config_name: wasm1
config_source:
ads: {}
In the case of FCDS, we will be moving config up to a oneof, and then the config itself will have to have a name and a config source, so I think it would be covered there?
Yeah, it can be done either way. But it would help us to have config name to be a peer of typed_config instead of being nested in a oneof peer. This is because we run multiple transformation passes, and having a name in the xDS helps with identifying the config/filter instance. This is consistent with the rest of xDS where each resource has a name in its proto.
@htuch any thoughts on ^?
I think this is useful in the context of FCDS, WASM and tooling (or control planes) that operate on opaque config (i.e. they can't peek inside).
Here's an interesting thought for v3; since we will no longer have untyped Struct, and will have a world of only TypeStruct and Any, and every extension should have its own unique config proto, we should in face be able to get rid of the need to have any filter type. I.e. you don't need to write envoy.wasm, your use of the WASM config option implies that via the embedded type URL. This means that the name field could be used arbitrarily for user purposes in v3+.
@htuch This would work if every filter config is unique per filter. I think there may be cases where two filters share a proto for the config. Not sure if that's something dis-allowed already.
One more use case for the control plane to operate on this config in an opaque way is to be able to do partial ordering between different filters. I.e. if control plane is provided with 10 "envoy.wasm" filters, there is no meaningful way to describe a relative order between them.
@htuch, just having "name" as a unique arbitrary name and rely on type for the actual type in v3 works as well.
Partial order is one of the motivating use case here. Peeking inside requires specific knowledge of the filter.
Changing the meaning of “name” from type_name to unique_name seems risky, though it will work.
If we add instance_name / config_name field we can actually add it to both v2 and v3
For xDS v2, we can relax the constraint that the config message must match the filter name (as long as it's not a regular struct). That would allow arbitrary names without a breaking change.
This instance_name which will be referenced by FCDS.
Looks like in the cpp code the expression would be resource.XX_name(). Maybe calling it resource_name()?
It would be good to have a v2 xDS solution here, but we need to be very clear what the semantics are if we reuse the field, i.e. it should only be possible if fully unambiguously typed configuration is otherwise present. It also might surprise some folks, as they may have built validators in their config pipeline to ensure consistency of name and config.
We could also add a filter_type field that unambiguously denotes the type. If the type field is specified then name can be free form.
@htuch In the linked PR, we can infer the name of the extension from the protobuf type for most cases. There are just two exceptions:
Empty, which is being solved separately.
Migrated APIs. I think we need this information regardless, but it seems clear that all versions of configs should be distinct.
I agree about the surprise effect. Fortunately that only happens with the invalid config, e.g. some invalid config might become valid since name is not significant.
I actually quite like the name/typed_config pattern. It applies in many places across the code base.
@kyessenov what do you reckon the state of this issue is?
I think this is resolved. If typed_config is used with the extension-specific type, name can be set to anything. I've updated the unit tests https://github.com/envoyproxy/envoy/pull/10071, https://github.com/envoyproxy/envoy/pull/10122, https://github.com/envoyproxy/envoy/pull/10130
Ack, thanks @kyessenov for the rad contribution here, closing.
|
2025-04-01T06:38:34.153165
| 2020-11-04T00:44:01
|
735729040
|
{
"authors": [
"asraa",
"zasweq"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5665",
"repo": "envoyproxy/envoy",
"url": "https://github.com/envoyproxy/envoy/pull/13891"
}
|
gharchive/pull-request
|
[fuzz] Got rid of all uninteresting call logs in health check fuzzing
Signed-off-by: Zach Reyes<EMAIL_ADDRESS>Commit Message: Got rid of all uninteresting call logs in health check fuzzing
Additional Description: Uninteresting call logs were slowing down my health check fuzzer in oss-fuzz. This PR gets rid of all uninteresting call logs by wrapping applicable objects in NiceMocks. However, since at the beginning of my internship I refactored the unit tests to also use fuzz tests, I put the unit test classes back in test/common/upstream:health_checker_impl_test.cc, and renamed test/common/upstream:health_checker_impl_test_utils to health_check_fuzz_test_utils.
No loss in coverage over source/common/upstream/health_checker_impl.cc. Speed up to 30 exec/sec on my cloudtop instance.
Risk Level: Low
/assign @asraa @htuch @adisuissa
Thanks! What was the before of the speed on cloudtop?
Honestly Asra it was same speed, 30 exec/sec.
|
2025-04-01T06:38:34.156560
| 2021-06-13T13:14:01
|
919796648
|
{
"authors": [
"dmitri-d",
"phlax"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5666",
"repo": "envoyproxy/envoy",
"url": "https://github.com/envoyproxy/envoy/pull/16959"
}
|
gharchive/pull-request
|
docs: Use yaml build config for vars
Signed-off-by: Ryan Northey<EMAIL_ADDRESS>Commit Message: docs: Use yaml build config for vars
Additional Description:
Another breakout from #15229
This adds a build configuration file for vars passed through to sphinx
Risk Level:
Testing:
Docs Changes:
Release Notes:
Platform Specific Features:
[Optional Runtime guard:]
[Optional Fixes #Issue]
[Optional Deprecated:]
[Optional API Considerations:]
I think this looks good, other than the question re: descriptor_path parameter in the validating code block.
lgtm
|
2025-04-01T06:38:34.159873
| 2021-07-23T16:48:45
|
951740251
|
{
"authors": [
"keith",
"moderation"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5667",
"repo": "envoyproxy/envoy",
"url": "https://github.com/envoyproxy/envoy/pull/17466"
}
|
gharchive/pull-request
|
bazel: remove old luajit workaround
According to https://luajit.org/install.html
Important: this relates to LuaJIT 2.0 only — use LuaJIT 2.1 to avoid these complications.
Since we have updated past 2.1 we shouldn't need these anymore which is
great since it breaks on Apple Silicon
https://github.com/envoyproxy/envoy/issues/16482#issuecomment-846439439
Signed-off-by: Keith Smiley<EMAIL_ADDRESS>
@moderation wdyt?
LGTM. I commented these lines out when I got M1 building a while back - https://github.com/envoyproxy/envoy/issues/16482#issuecomment-846439439
Removing as we don't require makes sense. MacOS CI failing however
Yea I just noticed that we can probably remove them instead. Turns out I can't let the options fallthrough, hopefully green now
|
2025-04-01T06:38:34.166867
| 2021-08-03T18:47:49
|
959392346
|
{
"authors": [
"alyssawilk",
"ggreenway",
"rgs1"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5668",
"repo": "envoyproxy/envoy",
"url": "https://github.com/envoyproxy/envoy/pull/17577"
}
|
gharchive/pull-request
|
conn pool: default enable runtime feature conn_pool_delete_when_idle
Signed-off-by: Greg Greenway<EMAIL_ADDRESS>
Commit Message: This enables the new behavior (clean up conn pools when they're idle, to avoid leaking memory in some configurations) from #17403 by default. It can still be disabled by setting runtime feature envoy.reloadable_features.conn_pool_delete_when_idle to false.
Additional Description:
Risk Level: Medium
Testing: Test coverage was added in #17403.
Docs Changes:
Release Notes: Added in #17403
Platform Specific Features:
[Optional Runtime guard:]
[Optional Fixes #Issue]
[Optional Deprecated:]
[Optional API Considerations:]
have we smoke tested somewhere yet?
have we smoke tested somewhere yet?
It's still the same code (minus one possible crash when a cluster is removed via CDS) that @rgs1 smoke tested awhile ago.
have we smoke tested somewhere yet?
It's still the same code (minus one possible crash when a cluster is removed via CDS) that @rgs1 smoke tested awhile ago.
... tested with the new tcp conn pool, whereas the additional crashers were with the old pool fwiw ...
Ah cool, didn't realize the prior version had been canaried.
Just to check my memory, the folks encountering tcp proxy crashes didn't provide additional data, and agreed they should switch back to the new pool in any case right? If so LGTM-as-long-as-you-cc-them because it's as safe as it's going to get (folks shouldn't be using the old pool without informing us the new one is problematic)
@bianpengyuan FYI this change, that you reported a crash in #16948, is being reintroduced.
Looking at that report again, it's very possible that it was the same crash fixed in #17522. Not enough information to know for sure, but it's a possible match, so it may be fixed.
coverage test flake; unrelated:
2021-08-03T20:04:16.6349000Z test/extensions/transport_sockets/starttls/starttls_integration_test.cc:329: Failure
2021-08-03T20:04:16.6350199Z Value of: test_server_->server().listenerManager().numConnections()
2021-08-03T20:04:16.6350878Z Expected: is equal to 1
2021-08-03T20:04:16.6351556Z Actual: 0 (of type unsigned long)
2021-08-03T20:04:16.6352267Z Stack trace:
2021-08-03T20:04:16.6352891Z 0x454827: (unknown)
2021-08-03T20:04:16.6353611Z 0x7f6ad1696d96: testing::internal::HandleSehExceptionsInMethodIfSupported<>()
2021-08-03T20:04:16.6354475Z 0x7f6ad167b701: testing::internal::HandleExceptionsInMethodIfSupported<>()
2021-08-03T20:04:16.6355200Z 0x7f6ad1663042: testing::Test::Run()
2021-08-03T20:04:16.6355864Z 0x7f6ad1663b58: testing::TestInfo::Run()
2021-08-03T20:04:16.6356462Z ... Google Test internal frames ...```
/retest
/retest
|
2025-04-01T06:38:34.169363
| 2018-11-26T16:51:50
|
384420243
|
{
"authors": [
"rshriram"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5669",
"repo": "envoyproxy/envoy",
"url": "https://github.com/envoyproxy/envoy/pull/5117"
}
|
gharchive/pull-request
|
reformat dynamic metadata emitted by Mongo proxy
Description: Emit metadata as map<resource, list(operations> so that it can be used in metadata matchers easily. The existing format (messages:list(structs)) is too hard to represent in metadata matchers.
Risk Level: LOW
Testing: Unit tests
Signed-off-by: Shriram Rajagopalan<EMAIL_ADDRESS>cc @venilnoronha
@dio the PR that implemented this was merged yesterday :). So users have not seen this stuff yet. So version history doesn't exist..
|
2025-04-01T06:38:34.176495
| 2023-10-14T01:05:39
|
1942846397
|
{
"authors": [
"arkodg",
"shahar-h",
"tmsnan"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5670",
"repo": "envoyproxy/gateway",
"url": "https://github.com/envoyproxy/gateway/issues/1966"
}
|
gharchive/issue
|
Global downstream connection limits.
Add a overload manager configuration to limit the total number of downstream connections
https://www.envoyproxy.io/docs/envoy/latest/configuration/operations/overload_manager/overload_manager#limiting-active-connections
It appears that the functionality is incomplete and unsuitable for production use. Should we enable it?
@tmsnan im fine waiting, but atm we dont have any way of enabling overload manager
@arkodg If possible, I could first add other overload manager features that are already fully supported.
@tmsnan sure, imo the others like heap size based will require an API addition, putting the burden of getting it right on the user .
Was trying to use this GH issue to enable sensible defaults
@arkodg Maybe we can refer to Google VRP edge server configuration.
`overload_manager:
refresh_interval: 0.25s
resource_monitors:
name: "envoy.resource_monitors.fixed_heap"
typed_config:
"@type": type.googleapis.com/envoy.extensions.resource_monitors.fixed_heap.v3.FixedHeapConfig
TODO: Tune for your system.
max_heap_size_bytes:<PHONE_NUMBER> # 2 GiB
actions:
name: "envoy.overload_actions.shrink_heap"
triggers:
name: "envoy.resource_monitors.fixed_heap"
threshold:
value: 0.95
name: "envoy.overload_actions.stop_accepting_requests"
triggers:
name: "envoy.resource_monitors.fixed_heap"
threshold:
value: 0.98`
https://www.envoyproxy.io/docs/envoy/latest/configuration/best_practices/edge#best-practices-edge
Please assign me
|
2025-04-01T06:38:34.183618
| 2024-01-16T20:35:09
|
2084862185
|
{
"authors": [
"arkodg",
"jaynis",
"zhaohuabing"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5671",
"repo": "envoyproxy/gateway",
"url": "https://github.com/envoyproxy/gateway/pull/2453"
}
|
gharchive/pull-request
|
feat(cors): Allowed more wildcard options
A few weeks ago the allowed CORS origins have been changed from a regex to a wildcard notation (#2389). Implementation wise all kinds of wildcards are supported, however, the validation regex on the SecurityPolicy CRD limits the CORS options to hostnames prefixed with an wildcard followed by a dot, allowing all subdomains of that host. This reduces the freedom when allowing cross origins a lot compared to how it was before. This PR aims to relax the validation regex a bit to enable the following use cases:
Allowing all hosts of an specific scheme (https://*)
Allowing all hosts regardless of the scheme (*)
Allowing all ports of a specific host (http://localhost:*)
While allowing all hosts in the context of CORS might sound a bit hacky, this is sometimes required. For instance when a web service provides an API which is consumed by many third-party web applications hosted under arbitrary domains not under the control of the maintainer of aforementioned web service. In addition to that it can be very useful during application development. This is why I have added the option to allow all ports of a specific host as well.
Review the new and the old validation regexes.
@jaynis can sign your commits and repush ? DCO is failing
@jaynis Thanks for the improvement in the wildcard host matching. The implementation looks good to me.
I only have a little hesitation about the port wildcard matching. Suffix/Port wildcard matching is not a common practice for hostnames. Do you have many ports for a given hostname?
Thank you for your review @zhaohuabing.
I only have a little hesitation about the port wildcard matching. Suffix/Port wildcard matching is not a common practice for hostnames. In your use case, do you have many ports for a given hostname?
The port range matching was solely meant to be a dev feature so that one can configure CORS for a host (e.g. localhost) regardless of the port the application runs on. But this scenario could be covered by the general wildcard as well, therefore I would also be fine with deleting it again if you think it is not required. Just let me know your preference.
Thank you for your review @zhaohuabing.
I only have a little hesitation about the port wildcard matching. Suffix/Port wildcard matching is not a common practice for hostnames. In your use case, do you have many ports for a given hostname?
The port range matching was solely meant to be a dev feature so that one can configure CORS for a host (e.g. localhost) regardless of the port the application runs on. But this scenario could be covered by the general wildcard as well, therefore I would also be fine with deleting it again if you think it is not required. Just let me know your preference.
Prefer to remove the suffix matching to keep it aligned with the common practice. Thanks.
|
2025-04-01T06:38:34.223972
| 2015-08-26T11:23:29
|
103241761
|
{
"authors": [
"kbs12e"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5672",
"repo": "enyojs/moonstone",
"url": "https://github.com/enyojs/moonstone/pull/2465"
}
|
gharchive/pull-request
|
Enyo 2171 Add tabindex for Item to prevent accessibility timing issue
Issue
screen reader sometimes does not read child component's content when Item is focused.
Cause
Item may have components as child. However, Sometimes screen reader does not read child component's content because time to receive tabindex is late than child.
To prevent this timing issue, I add tabindex to Item.
Fix
Add tabindex in ariaObservers.
Enyo-DCO-1.1-Signed-off-by: Bongsub Kim<EMAIL_ADDRESS>
I will re-create PR with latest code.
|
2025-04-01T06:38:34.277567
| 2021-09-20T10:37:52
|
1000814437
|
{
"authors": [
"Vovanella95",
"dtaskai"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5673",
"repo": "epam-cross-platform-lab/swagger-dart-code-generator",
"url": "https://github.com/epam-cross-platform-lab/swagger-dart-code-generator/issues/245"
}
|
gharchive/issue
|
[QUESTION] Path still gets generated even when told not to
Path still gets generated in the swagger.dart file when specified like the following in the build.yaml file:
targets:
$default:
sources:
- lib/**
- $package$
builders:
chopper_generator:
options:
header: "//Generated code"
swagger_dart_code_generator:
options:
input_folder: "lib/"
output_folder: "lib/swagger_generated_code/"
exclude_paths:
- "/api/mobile/actuator/"
Am I using wrong syntax or is this an actual bug?
hi @dtaskai , exclude_path and include_path are regex strings. to make your path excluded, you need to add something like this:
\/api\/mobile\/actuator\/
Please use Regex validator to check is you String passing or not. For example, you can use this validator: https://regex101.com/
@dtaskai , if something is not clear - please let us know
The exclusion didn't work on my project even after using regex syntax, so I have tried it on the example project:
Added an exclusion to /rooms
swagger_dart_code_generator: options: input_folder: "lib/" output_folder: "lib/swagger_generated_code/" exclude_paths: - "\/rooms"
Ran flutter build run build_runner build
Then it still generated the code for /rooms
@Get(path: '/rooms')
Future<chopper.Response<List<Room>>> roomsGet(
{@Query('id') required String? id});
Ok good let me check it
@dtaskai Yep you're right. We removed it in 2+ version. Let me fix.
@dtaskai Please try it on latest version. Also you can put just /rooms to exclude_path. It works.
Latest version is 2.1.3+2
Works on both the example and my personal project, thank you!
|
2025-04-01T06:38:34.280296
| 2022-11-07T10:45:18
|
1438145424
|
{
"authors": [
"Vovanella95",
"dfdgsdfg"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5674",
"repo": "epam-cross-platform-lab/swagger-dart-code-generator",
"url": "https://github.com/epam-cross-platform-lab/swagger-dart-code-generator/issues/483"
}
|
gharchive/issue
|
[QUESTION] How can I config worker-pool on Chopper?
Please describe a problem.
The below is Chopper worker poll example.
/// inspired by https://github.com/d-markey/squadron_sample/blob/main/lib/main.dart
void initSquadron(String id) {
Squadron.setId(id);
Squadron.setLogger(ConsoleSquadronLogger());
Squadron.logLevel = SquadronLogLevel.all;
Squadron.debugMode = true;
}
Future<void> main() async {
/// initialize Squadron before using it
initSquadron('worker_pool_example');
final jsonDecodeServiceWorkerPool = JsonDecodeServiceWorkerPool(
// Set whatever you want here
concurrencySettings: ConcurrencySettings.oneCpuThread,
);
/// start the Worker Pool
await jsonDecodeServiceWorkerPool.start();
/// Instantiate the JsonConverter from above
final converter = JsonSerializableWorkerPoolConverter(
{
Resource: Resource.fromJsonFactory,
},
/// make sure to provide the WorkerPool to the JsonConverter
jsonDecodeServiceWorkerPool,
);
/// Instantiate a ChopperClient
final chopper = ChopperClient(
client: client,
baseUrl: 'http://localhost:8000',
// bind your object factories here
converter: converter,
errorConverter: converter,
services: [
// the generated service
MyService.create(),
],
/* ResponseInterceptorFunc | RequestInterceptorFunc | ResponseInterceptor | RequestInterceptor */
interceptors: [authHeader],
);
/// Do stuff with myService
final myService = chopper.getService<MyService>();
/// ...stuff...
/// stop the Worker Pool once done
jsonDecodeServiceWorkerPool.stop();
}
Describe the solution you'd like
How can I config worker-poll on Chopper?
Hi @dfdgsdfg , Unfortunately I have no experience with WorkerPool. We just generation swagger code. If you have an idea, how to generate it - let us know.
|
2025-04-01T06:38:34.284128
| 2024-11-27T10:13:16
|
2697974744
|
{
"authors": [
"VolhaBazhkova",
"YauheniyaH"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5675",
"repo": "epam/ai-dial-chat",
"url": "https://github.com/epam/ai-dial-chat/issues/2678"
}
|
gharchive/issue
|
Background for items in context menu opened from chat header should be blue (now green)
EPAM AI DIAL chat version
0.21.0
What happened?
Actual:
Expected: on the central part items are highlighted using blue color
Example:
Confidential information
[X] I confirm that do not share any confidential information
verified on staging successfully
|
2025-04-01T06:38:34.289251
| 2024-10-17T17:56:05
|
2595382305
|
{
"authors": [
"Derikyan"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:5676",
"repo": "epam/ai-dial-chat",
"url": "https://github.com/epam/ai-dial-chat/pull/2393"
}
|
gharchive/pull-request
|
fix(chat): Update the phrase "Not allowed model selected. Please, change the model to proceed" (Issue #2363)
Description:
Update the phrase "Not allowed model selected. Please, change the model to proceed"
Issues:
Issue #2363
Checklist:
[x] the pull request name complies with Conventional Commits
[x] the pull request name starts with fix(<scope>):, feat(<scope>):, feature(<scope>):, chore(<scope>):, hotfix(<scope>): or e2e(<scope>):. If contains breaking changes then the pull request name must start with fix(<scope>)!:, feat(<scope>)!:, feature(<scope>)!:, chore(<scope>)!:, hotfix(<scope>)!: or e2e(<scope>)!: where <scope> is name of affected project: chat, chat-e2e, overlay, shared, sandbox-overlay, etc.
[x] the pull request name ends with (Issue #<TICKET_ID>) (comma-separated list of issues)
[x] I confirm that do not share any confidential information like API keys or any other secrets and private URLs
/deploy-review
/deploy-review
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.