added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T06:40:41.023217
| 2016-12-15T11:33:15
|
195782945
|
{
"authors": [
"dram",
"infradig"
],
"license": "bsd-2-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11462",
"repo": "trealla-lang/trealla",
"url": "https://github.com/trealla-lang/trealla/issues/54"
}
|
gharchive/issue
|
Redundant parentheses cause memory corruption
Following code cause memory corruption:
main :- (foo, (bar(baz))), halt.
Yep,
main :- (bar(baz)), halt.
is enough.
Not sure if related, following code does not parse correctly:
main :- write(a), nl, (write(b), nl), halt.
Listing:
Trealla v0.1alpha, 64-bits (int=128-bits), Dec 16 2016
main :-
write(a),
nl(write(b),nl),
halt.
According to valgrind's result, those lines trealla.c#L2058-L2063 are suspicious.
After I commented them out, memory corruption disappeared.
Yes, very suspect. Can't imagine what it is supposed to be doing.
Perhaps it was supposed to be NLIST_POP_FRONT, not NLIST_FRONT.
Anyway, it still doesn't parse correctly.
Now your test case will cause an infinite loop:
main :- (bar(baz)), halt.
Now for this:
main :- write(a), nl, (write(b), nl), halt.
Seems ok now?
Cool! Fix confirmed in devel branch. I'll close this issue.
|
2025-04-01T06:40:41.024431
| 2022-04-08T10:05:53
|
1197117811
|
{
"authors": [
"UWN"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11463",
"repo": "trealla-prolog/trealla",
"url": "https://github.com/trealla-prolog/trealla/issues/520"
}
|
gharchive/issue
|
Unification segv
?- A=[A|A],B=[A|B],A=[A|B], false.
Segmentation fault (core dumped)
(This one blocks a lot)
I can only repeat that this blocks a lot, I just reencountered exactly this error.
|
2025-04-01T06:40:41.070132
| 2020-11-30T14:28:14
|
753524358
|
{
"authors": [
"arielshaqed"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11464",
"repo": "treeverse/lakeFS",
"url": "https://github.com/treeverse/lakeFS/pull/979"
}
|
gharchive/pull-request
|
Add "lakectl fs upload --recursive" flag
Behaves a lot like s3 cp --recursive: uploads files using their path relative from the local directory.
Example
Suppose you have these files:
/tmp/a/foo
/tmp/a/b/bar
/tmp/a/b/c/d/e/xyzzy
Then lakectl upload --recursive --source /tmp/a/ lakefs://repo@master/top/ will create these objects:
lakefs://repo@master/top/foo
lakefs://repo@master/top/b/bar
lakefs://repo@master/top/b/c/d/e/xyzzy
Alternative solution for https://lakefs.slack.com/archives/C016726JLJW/p1606735311131500.
Alternative to #977, probably don't want both...
Thanks! Pulling this one, throwing #977 away.
|
2025-04-01T06:40:41.077971
| 2022-10-23T22:49:26
|
1419965042
|
{
"authors": [
"alexheiner",
"christopherkindl",
"mitrotasios",
"tonyxiao"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11465",
"repo": "tremorlabs/tremor",
"url": "https://github.com/tremorlabs/tremor/issues/145"
}
|
gharchive/issue
|
Search bar
I think it would be useful to have a search bar component to filter rows of a table/list or other data. I've seen your examples of using a MultiSelectbox box for filtering, but I think the search functionality would be helpful too.
Hey @alexheiner, thanks for the comment! Did you have the chance to check out the SelectBox component already, which essentially provides a search-box to filter among the available options? Or are you looking for something different? Feel free to elaborate ๐
I actually didn't notice the ability to search in the SelectBox, thanks for pointing that out. Sorry I don't know how I missed that. I was originally thinking of a more general purpose table/data search, such as searching for a specific address, phone number, etc in a table. But I thInk the SelectBox could probably be used to do something similar.
@alexheiner Let us know if a simple search bar (including further features, such as auto complete) still would be more appealing to your use cases! :)
Sounds good! I do think it could be a useful feature, but it could be something that is added down the line in a future release. Thanks for getting back to me about it! I think this is a really cool project, I'm excited to keep using it!
SearchBox does no appear to be documented in the list of components here through. https://www.tremor.so/docs/components
|
2025-04-01T06:40:41.099019
| 2015-02-25T16:31:06
|
58928164
|
{
"authors": [
"trentrichardson",
"zackperdue"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11466",
"repo": "trentrichardson/jQuery-Impromptu",
"url": "https://github.com/trentrichardson/jQuery-Impromptu/issues/57"
}
|
gharchive/issue
|
Can I use my own form?
like:
$.prompt({
html: '<form action="/user/settings" method="post"......',
})
Impromptu wraps the all of the states into a single form, so doing so would create a form inside a form, which will cause issues.
|
2025-04-01T06:40:41.109098
| 2016-07-28T21:27:55
|
168201123
|
{
"authors": [
"SpainTrain",
"ransford",
"willjschmitt"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11467",
"repo": "treyhunner/django-simple-history",
"url": "https://github.com/treyhunner/django-simple-history/issues/230"
}
|
gharchive/issue
|
Query previous/next historical entry
One of the main use cases I've found with django-simple-history has been to get the previous historical entry to calculate deltas in values.
Issue #201 talks about the Django utility functions, which allows for getting the last value by a field. I think it would be great to add a previous property to the historical model that grabbed the get_previous_by_history_date that reflected the same tracked instance automatically. I feel like a next and previous method fits perfectly within the historical records context.
As a dependent though, comparing versions between points in time would be excellent, too, but I'll submit a second issue based on response to this.
What is the best way, given a historical entry, to get the "previous" entry?
In my use case, thing and thing.history.most_recent() have the same field values in all cases, except they don't compare identically because thing._state != thing.history.most_recent()._state, and id(thing) != id(thing.history.most_recent()).
:white_check_mark: https://github.com/treyhunner/django-simple-history/pull/365
|
2025-04-01T06:40:41.138540
| 2023-10-02T21:26:43
|
1922749330
|
{
"authors": [
"matt-aitken",
"neo773"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11468",
"repo": "triggerdotdev/trigger.dev",
"url": "https://github.com/triggerdotdev/trigger.dev/pull/530"
}
|
gharchive/pull-request
|
fix: #526 CLI warn if using @trigger.dev/react package with Next.js pages directory
Closes #526
โ
Checklist
[x] I have followed every step in the contributing guide
[x] The PR title follows the convention.
[x] I ran and tested the code works
Screenshots
This is a bit stale now
|
2025-04-01T06:40:46.193574
| 2022-08-09T18:44:25
|
1333651515
|
{
"authors": [
"codecov-commenter",
"elysee15"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11469",
"repo": "trisacrypto/directory",
"url": "https://github.com/trisacrypto/directory/pull/699"
}
|
gharchive/pull-request
|
sc-7976 GDS User UI : Remove blue border from css custom-styled button
Scope of changes
sc-7976 GDS User UI : Remove blue border from css custom-styled button
Type of change
[ ] bug fix
[ ] new feature
[ ] documentation
[ ] other (describe)
Acceptance criteria
Describe how reviewers can test this change to be sure that it works correctly. Add a checklist if possible
Author checklist
[x] I have manually tested the change and/or added automation in the form of unit tests or integration tests
[ ] I have updated the dependencies list
[ ] I have recompiled and included new protocol buffers to reflect changes I made
[ ] I have added new test fixtures as needed to support added tests
[x] Check this box if a reviewer can merge this pull request after approval (leave it unchecked if you want to do it yourself)
[x] I have moved the associated Shortcut story to "Ready for Review"
Reviewer(s) checklist
[ ] Any new user-facing content that has been added for this PR has been QA'ed to ensure correct grammar, spelling, and understandability.
Codecov Report
Merging #699 (916900f) into main (c9e02fb) will decrease coverage by 5.80%.
The diff coverage is n/a.
@@ Coverage Diff @@
## main #699 +/- ##
==========================================
- Coverage 43.89% 38.08% -5.81%
==========================================
Files 580 155 -425
Lines 17363 2434 -14929
Branches 1273 618 -655
==========================================
- Hits 7621 927 -6694
+ Misses 8557 1330 -7227
+ Partials 1185 177 -1008
Impacted Files
Coverage ฮ
...eb/gds-user-ui/src/components/FileUpload/index.tsx
...i/src/components/NeedsAttention/AttentionAlert.tsx
web/gds-user-ui/src/theme/components/Button.ts
...github.com/trisacrypto/directory/pkg/gds/backup.go
...ficateManagement/CertificateManagement.stories.tsx
...isacrypto/directory/pkg/utils/logger/middleware.go
web/gds-user-ui/src/locales/fr/messages.js
web/gds-user-ui/src/components/NotFound/index.tsx
...om/trisacrypto/directory/pkg/gds/client/profile.go
...-user-ui/src/components/Navbar/Landing/NavItem.tsx
... and 415 more
Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here.
|
2025-04-01T06:40:46.204365
| 2023-10-06T10:59:16
|
1929918655
|
{
"authors": [
"Yashrocky96",
"kareemmahlees",
"tristanisham"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11470",
"repo": "tristanisham/logysia",
"url": "https://github.com/tristanisham/logysia/issues/4"
}
|
gharchive/issue
|
Output showing in s (Seconds) instead of micro or milli seconds And Output not in readable format
Added an Image to show the output
In the first you can see it's showing "s" which is not correct
And in the second log it's not in a readable format
Request:-
Add a UTC timestamp as a configurable option maybe because we can stream these logs or access from Docker logs we would be able to confirm with a timestamp
Hi ๐ , it this is accepted I would like to work on this, could you assign it to me ๐!
Okay, I will take a look at it. The timing got messed up in the last update from a bug I didn't catch.
I could include that. Maybe a few other logging options too like JSON for structured loggers?
@tristanisham
Logger does look good and this will help a lot in future.
You can provide more options, also you can maybe as well use local timestamp so that people developing in their own timezone will be able to relate by the time easier. You can enable this as an option.
What do you have in mind for Structured loggers?
Also are you working on adding testing?
@Yashrocky96 could you make a pull request for the errant seconds label?
As for logging, I'll add some options to enable logging in UTC and support for the TZ variable.
@tristanisham I have raised a PR, Please review and let me know
#5 is almost done. One more change and it's good to merge @Yashrocky96
@tristanisham I have updated the PR
Also, Please check this out if you can.
Story:- When there's an error during execution of onError for example during development you are giving a wrong JSON and you get an error that the input is Failed to parse JSON.
In this scenario, onError hook is executing twice and logging the information twice.
For now, I am fairly new with Elysia and do not understand why that is happening.
@tristanisham Please close this issue and would I not join as a Contributor here?
@tristanisham
Hi,
I am on vacation and will review these changes when I return in
November.
Tristan Isham
On October 23, 2023, Tristan Isham @.***> wrote:
@tristanisham https://github.com/tristanisham
โ
Reply to this email directly, view it on GitHub
<https://github.com/tristanisham/logysia/issues/4#issuecomment-
1774702710>, or unsubscribe
<https://github.com/notifications/unsubscribe-
auth/AFQNWUQIXLSLPUOWAMZVWC3YAYUTJAVCNFSM6AAAAAA5VUOQ4GVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTONZUG4YDENZRGA>.
You are receiving this because you were mentioned.Message ID:
@.***>
Hi Tristanisham,
Hope that you are back from your vacations, Would you please add me as a contributor and do a patch release please?
Just send them here? Is there a pull request?
Could you please send them here?
What do I send and how do I send, This is my first Open-Source Contribution @tristanisham . Please guide me. And thanks.
You have already merged my changes from my forked repository
Hi @tristanisham Please do a patch release and add me as a contributor to the package, if my application does use this package.
@tristanisham Please do a patch release
|
2025-04-01T06:40:46.226621
| 2023-02-06T12:29:40
|
1572484324
|
{
"authors": [
"ayusharma",
"matsko"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11471",
"repo": "trivago/prettier-plugin-sort-imports",
"url": "https://github.com/trivago/prettier-plugin-sort-imports/issues/206"
}
|
gharchive/issue
|
Do not call yarn directly in the package.json pre commands
Your Environment
Prettier version: 2.8.3
node version [12.x.x, 14.x.x]: 18.13
package manager<EMAIL_ADDRESS>IDE: VSCode
Describe the bug
When installed via pnpm install, pnpm expresses that the command fails due to yarn missing.
Install a fresh nvm version (like 18.13) and run npm install -g pnpm. Then create a new project add @trivago/prettier-plugin-sort-imports as a dep. Run install and see the error.
Expected behavior
Should work without the need to have a global yarn package manager.
Any ideas?
hello?
Please use require while loading the plugin if you use it pnpm. Something like this https://github.com/verdaccio/verdaccio/blob/c9bd22dcaa2f0d919338fb387ae261992f0e1116/prettier.config.js#L10
Feel free to reopen if you still face the issue.
|
2025-04-01T06:40:46.231198
| 2021-06-09T16:04:24
|
916374962
|
{
"authors": [
"jlchan",
"ranocha",
"sloede"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11472",
"repo": "trixi-framework/Trixi.jl",
"url": "https://github.com/trixi-framework/Trixi.jl/issues/633"
}
|
gharchive/issue
|
Rename basis part of DG to operator or something similar?
This came up in https://github.com/trixi-framework/Trixi.jl/pull/617#discussion_r644467670. The name basis might not fit very well for FD-SBP methods...
@andrewwinters5000 suggested "spectral approximation", but that doesn't really fit for FD methods, I think. The "approximation" part is probably okay, though - "approximation space"? But that's also quite long compared to "basis"...
I like the approximation-related name. One could also use a name related to the reference element/domain since all quantities are defined there?
One could also use a name related to the reference element/domain since all quantities are defined there?
Like what?
I was thinking ReferenceElementData/ReferenceDomainData after deal.ii's FiniteElementData and GridAP's ReferenceFE, but those names are pretty long.
Hm. So it would be something like LobattoLegendreReferenceElementData? That is quite long...
Yeah, that's definitely too long. Maybe ReferenceElementData would be better as an abstract type, something like
LobattoLegendreElement <: AbstractReferenceElementData{RealT}.
Triage: Not necessary
|
2025-04-01T06:40:46.234176
| 2022-03-27T22:13:21
|
1182680642
|
{
"authors": [
"basarozcan",
"gabrielbryk",
"rinchen"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11473",
"repo": "trntv/sshed",
"url": "https://github.com/trntv/sshed/issues/16"
}
|
gharchive/issue
|
Brew Tap is broken
~ brew install trntv/sshed/sshed 1 [15:45:25]
==> Tapping trntv/sshed
Cloning into '/usr/local/Homebrew/Library/Taps/trntv/homebrew-sshed'...
remote: Enumerating objects: 6, done.
remote: Counting objects: 100% (6/6), done.
remote: Compressing objects: 100% (4/4), done.
remote: Total 6 (delta 2), reused 0 (delta 0), pack-reused 0
Receiving objects: 100% (6/6), done.
Resolving deltas: 100% (2/2), done.
Error: Invalid formula: /usr/local/Homebrew/Library/Taps/trntv/homebrew-sshed/sshed.rb
sshed: Calling bottle :unneeded is disabled! There is no replacement.
Please report this issue to the trntv/sshed tap (not Homebrew/brew or Homebrew/core):
/usr/local/Homebrew/Library/Taps/trntv/homebrew-sshed/sshed.rb:9
Error: Cannot tap trntv/sshed: invalid syntax in tap!
Same for me with M1 Mac Pro
Same here
|
2025-04-01T06:40:46.318547
| 2015-03-30T03:50:55
|
65128708
|
{
"authors": [
"jpsfs",
"kozilla",
"troolee"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11479",
"repo": "troolee/gridstack.js",
"url": "https://github.com/troolee/gridstack.js/issues/111"
}
|
gharchive/issue
|
multiple instances of gridster
Hi Troolee,
I am trying to create 2 separate instances of gridster on one page how ever i have found some difficulties as they are using the same naming conventions. I have tried to initialize gridster with options (item_class ,placeholder_class) how ever I couldnt see any differences. Could you advise pls if gridster support multiple instances on one project and if yes ;) how correctly initialize each instance. I am trying :
for second instance I am trying :
var options = {
cell_height: 10,
vertical_margin: 10,
item_class: "my_iteam_class",
placeholder_class: "my_placeholder_class"
};
$('.grid-stack.test').gridstack(options);
Thanks for advance...
I'm also having problems with this.
@kozilla did you managed to put this to work?
@troolee Any suggestion?
Please check 2 grids demo http://troolee.github.io/gridstack.js/demo/two.html
|
2025-04-01T06:40:46.336477
| 2017-05-22T18:18:16
|
230477453
|
{
"authors": [
"1sourcecontrol",
"bleehu",
"trowl223"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11480",
"repo": "trowl223/Compound_X",
"url": "https://github.com/trowl223/Compound_X/pull/107"
}
|
gharchive/pull-request
|
nerfing pin/multi-pin
Per issue #67 Attempt to balance marksman pin/multi-pin abilities.
Nerfed the bonus accuracy. Redesigned other things.
I tried to reword some text to remove vagueness.
What do you think of the following changes:
"If the target is fired at by you" -> "If you hit the selected target"
The effect should only trigger when a target is hit.
"The Pin/Multi-Pin ability can be activated by a pinned target up to 3 times per Marksman player's turn cycle" -> "This ability allows a marksman to react to a pinned target once per turn."
This version of the feat only allows one target to be pinned, so it doesn't makes sense that the only pinned target would trigger the marksman's reaction shot more than once per turn.
Also, "Only 1 target may be pinned by the marksman at a time." should be added to the first iteration of this feat.
"You may select and pin up to PER/2 (round down) targets at a time (you have to fire at a target to pin it).+30% to hit a Pinned target (instead of +10%). Only usable with long-rifles. " -> "You may monitor and react to up to PER/2 (round down) pinned targets at a time. +20% to hit a pinned target. Only usable with Long Rifles."
Cleared up some phrasing. A target is only pinned when hit, that is defined in the first section of the feat so we don't have to mention that again. The original +20% replaces any other bonus, this is the intended behavior of all class feats. if class X has a level 5 feat that says "+15% to do Y", and the level 10 variant says "+30% to do Y" that means the total bonus to performing Y is 30%. This phrasing is fairly ambiguous and I apologize for that. Maybe we should change all second-tier class feats so that bonuses are written as "The bonus for doing Y increases to +30%". That is fairly explicit.
Are we also adding in that this ability takes actions off of the marksman's turn, or leaving it as-is? I'm fine with either decision, I thought'd I'd bring it up since we'd discussed it in the past.
What do you think of the following changes:
"If the target is fired at by you" -> "If you hit the selected target"
The effect should only trigger when a target is hit.
Why should the effect only trigger when a target is hit?
My reasoning was that I didn't want to straight nerf marksman, so it gives them an advantage this way while other ways get nerfed. Also realism, 'pinning' sounds analogous to 'suppressing' a target, and that is more about psychological warfare than physical damage. I suppose you may be considering the first shot to be a maiming shot? Also, most long rifles are bolt action, making it much harder to pin targets if it's an on-hit effect (especially in regards to multi-pin).
"The Pin/Multi-Pin ability can be activated by a pinned target up to 3 times per Marksman player's turn cycle" -> "This ability allows a marksman to react to a pinned target once per turn."
This version of the feat only allows one target to be pinned, so it doesn't makes sense that the only pinned target would trigger the marksman's reaction shot more than once per turn.
Why doesn't it make sense? In my mind, the target could be darting from cover to cover on their turn. If I was trying to kill someone, I would be shooting every chance I get.
Also, "Only 1 target may be pinned by the marksman at a time." should be added to the first iteration of this feat.
"If the target dies you may select a new target." already says that, but we can make it more explicit if you prefer.
"You may select and pin up to PER/2 (round down) targets at a time (you have to fire at a target to pin it).+30% to hit a Pinned target (instead of +10%). Only usable with long-rifles. " -> "You may monitor and react to up to PER/2 (round down) pinned targets at a time. +20% to hit a pinned target. Only usable with Long Rifles."
Cleared up some phrasing. A target is only pinned when hit, that is defined in the first section of the feat so we don't have to mention that again. The original +20% replaces any other bonus, this is the intended behavior of all class feats. if class X has a level 5 feat that says "+15% to do Y", and the level 10 variant says "+30% to do Y" that means the total bonus to performing Y is 30%. This phrasing is fairly ambiguous and I apologize for that. Maybe we should change all second-tier class feats so that bonuses are written as "The bonus for doing Y increases to +30%". That is fairly explicit.
Ok.
Are we also adding in that this ability takes actions off of the marksman's turn, or leaving it as-is? I'm fine with either decision, I thought'd I'd bring it up since we'd discussed it in the past.
I put the following in, "The Pin/Multi-Pin ability can be activated by a pinned target up to 3 times per Marksman player's turn cycle" for that reason. I figured that it would be ok to grant the marksman the ability to react without extra expense.
I think this is good discussion!
I felt pinned by Alex's nerf monstrosity at nerf wars in past even when not hit by his shots.
If the trigger is the target leaving cover, how would it be possible for the marksman's ability to trigger more than once per turn per target? That also doesn't make sense to me.
We should make that much more explicit. Players sometimes create loopholes when they don't quite understand what's written. We should avoid attempting to communicate with nuance.
Whichever way we communicate it, we need to make it very clear.
I still don't understand how a target could proc 3 sniper reaction shots in a single turn.
@bleehu,
2. If a target re-enters cover and leaves it again.
Example:, I am pinned and hiding behind light cover in a hallway, I run ahead to then next cover in the hallway, proc'ing marksman Pin. Then I leave from that cover, to a third cover. Proc'ing marksman pin a 2nd time.
5. I tried to make it about multi-pin as well as 'single'-pin in one swoop. The way I meant it to be written is that the Marksman player cannot shoot more than 3 reaction shots per his full turn cycle, no matter how many people he is pinning.
That makes a little more sense. =)
@trowl223, fyi I won't do (don't want to do) any more work on this until you respond to the recent post(s) =)
(I have plenty of other things to do, so take your time)
So sorry! working on this now.
Makes sense.
I don't think so. The marksman is attuned to only the precise locations of the target he has pinned. if one leaves cover and is not hit by the reaction shot, the marksman has lost track of them temporarily and must readjust. The time frames seem to be in favor of the marksman too much. three shots in the span of their enemy's turn while they've already had a full turn is like granting them extra actions.
The situation will not always be that someone is moving in a straight line down a hallway towards you as in your given example. In that instance, it may be up to the DM to allow that if the Marksman has a semi-auto rifle since they don't have to readjust, but in a different example this may not be the case. What if the Marksman was providing cover to their squad from a bell-tower positioned 200 meters from combat, and they have to watch a 300m spread of land in front of them? A pinned target moving side-to-side form the Marksman's POV would require considerable time for the marksman to readjust if they miss the first shot as the target leaves cover.
My thoughts exactly. An ability of this caliber needs to be pretty explicit.
I still don't think that they should receive an extended amount of extra firing time. At the second tier, I think that with multiple targets the Marksman can have multiple reaction shots (as per my original edit) but at the first tier we're talking about a potential Level 5 ability, its too much out of turn action IMO.
@trowl223, sounds good.
Please review my latest commit
|
2025-04-01T06:40:46.390837
| 2023-07-26T14:44:25
|
1822581803
|
{
"authors": [
"Qubad786",
"bugclerk"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11481",
"repo": "truenas/ixdiagnose",
"url": "https://github.com/truenas/ixdiagnose/pull/29"
}
|
gharchive/pull-request
|
NAS-123236 / 23.10 / Allow waiting for jobs to complete
Context
A new parameter has been added to MiddlewareCommand to allow waiting for a job to complete so we get the job's endpoints output and not the job number.
Jira URL: https://ixsystems.atlassian.net/browse/NAS-123236
|
2025-04-01T06:40:46.474204
| 2024-05-02T00:03:51
|
2274370803
|
{
"authors": [
"mmatheson"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11482",
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/12951"
}
|
gharchive/pull-request
|
broomwood, folks
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[b,f]
/trunk merge
|
2025-04-01T06:40:46.476100
| 2024-05-04T01:24:16
|
2278645807
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11483",
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/14364"
}
|
gharchive/pull-request
|
favillous, gigantostracous
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[f,g]
/trunk merge
|
2025-04-01T06:40:46.477689
| 2024-05-05T12:03:28
|
2279494188
|
{
"authors": [
"epes"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11484",
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/15406"
}
|
gharchive/pull-request
|
fostered, crosses
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[c,f]
/trunk merge
|
2025-04-01T06:40:46.479230
| 2024-07-02T19:05:48
|
2386947651
|
{
"authors": [
"mmatheson"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11485",
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/16514"
}
|
gharchive/pull-request
|
drilling, enters
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[d,e]
/trunk merge
|
2025-04-01T06:40:46.480647
| 2024-07-04T02:17:30
|
2389784460
|
{
"authors": [
"epes"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11486",
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/17863"
}
|
gharchive/pull-request
|
aerometry, dominion
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[a,d]
/trunk merge
|
2025-04-01T06:40:46.482222
| 2024-07-06T23:08:15
|
2393776690
|
{
"authors": [
"mmatheson"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11487",
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/21343"
}
|
gharchive/pull-request
|
goats, exactingness
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[e,g]
/trunk merge
|
2025-04-01T06:40:46.483846
| 2024-07-09T17:09:01
|
2398739126
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11488",
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/22916"
}
|
gharchive/pull-request
|
galvanoplastical, descanting
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[d,g]
/trunk merge
|
2025-04-01T06:40:46.485610
| 2024-07-10T19:09:31
|
2401468170
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11489",
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/23420"
}
|
gharchive/pull-request
|
culturization
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[c]
/trunk merge
|
2025-04-01T06:40:46.488655
| 2024-07-20T07:07:52
|
2420639699
|
{
"authors": [
"mmatheson"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11491",
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/27175"
}
|
gharchive/pull-request
|
fiscalization, accounters
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[a,f]
/trunk merge
|
2025-04-01T06:40:46.490229
| 2024-07-20T21:04:17
|
2421055504
|
{
"authors": [
"mmatheson"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11492",
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/27497"
}
|
gharchive/pull-request
|
dovecot, fuddledness
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[d,f]
/trunk merge
|
2025-04-01T06:40:46.491931
| 2024-07-21T17:08:02
|
2421525688
|
{
"authors": [
"mmatheson"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11493",
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/28145"
}
|
gharchive/pull-request
|
atypicality, bractlets
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[a,b]
/trunk merge
|
2025-04-01T06:40:46.493536
| 2024-07-24T04:04:17
|
2426547392
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11494",
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/30714"
}
|
gharchive/pull-request
|
fireplough
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[f]
/trunk merge
|
2025-04-01T06:40:46.495352
| 2024-07-25T04:05:33
|
2428935389
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11495",
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/31916"
}
|
gharchive/pull-request
|
emesidae, chromatophoric
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[c,e]
/trunk merge
|
2025-04-01T06:40:46.498570
| 2024-07-30T08:05:57
|
2437184287
|
{
"authors": [
"mmatheson"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11497",
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/37980"
}
|
gharchive/pull-request
|
gendarmes, conceivability
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[c,g]
/trunk merge
|
2025-04-01T06:40:46.501695
| 2024-04-24T20:02:08
|
2262087802
|
{
"authors": [
"epes"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11499",
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/7819"
}
|
gharchive/pull-request
|
equimultiple, frequence
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 100
sleep for: 300s
close stale after: 4 hours
[pullrequest]
requests per hour: 0
deps=[e,f]
/trunk merge
|
2025-04-01T06:40:46.502422
| 2024-04-11T07:17:00
|
2237065894
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11500",
"repo": "trunk-io/mergequeue-staging",
"url": "https://github.com/trunk-io/mergequeue-staging/pull/870"
}
|
gharchive/pull-request
|
aasvogel, disarray (logical-conflict)
This pull request was generated by the 'mq' tool
/trunk merge
|
2025-04-01T06:40:46.504012
| 2024-07-20T14:05:52
|
2420912047
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11501",
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/108428"
}
|
gharchive/pull-request
|
gurdy, blinking
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 600s
close stale after: 24 hours
[pullrequest]
requests per hour: 20
deps=[b,g]
/trunk merge
|
2025-04-01T06:40:46.514543
| 2024-05-08T07:59:22
|
2284951764
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11502",
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/47745"
}
|
gharchive/pull-request
|
frangula, bureaucratical
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 2100s
close stale after: 24 hours
[pullrequest]
requests per hour: 100
deps=[b,f]
/trunk merge
|
2025-04-01T06:40:46.516157
| 2024-05-09T13:53:08
|
2287747019
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11503",
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/48534"
}
|
gharchive/pull-request
|
appaloosa, flavobacterium
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 2100s
close stale after: 24 hours
[pullrequest]
requests per hour: 100
deps=[a,f]
/trunk merge
|
2025-04-01T06:40:46.517760
| 2024-05-10T19:26:42
|
2290289390
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11504",
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/49262"
}
|
gharchive/pull-request
|
charcoalist
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 2100s
close stale after: 24 hours
[pullrequest]
requests per hour: 100
deps=[c]
/trunk merge
|
2025-04-01T06:40:46.519344
| 2024-05-13T10:25:53
|
2292419265
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11505",
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/51089"
}
|
gharchive/pull-request
|
chaton, agruif
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 2100s
close stale after: 24 hours
[pullrequest]
requests per hour: 100
deps=[a,c]
/trunk merge
|
2025-04-01T06:40:46.522588
| 2024-05-30T14:48:47
|
2325866079
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11507",
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/61964"
}
|
gharchive/pull-request
|
emprizes, glitch
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 2100s
close stale after: 24 hours
[pullrequest]
requests per hour: 100
deps=[e,g]
/trunk merge
|
2025-04-01T06:40:46.524424
| 2024-06-01T13:31:36
|
2329148117
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11508",
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/63536"
}
|
gharchive/pull-request
|
gallivorous, consentaneously
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 2100s
close stale after: 24 hours
[pullrequest]
requests per hour: 100
deps=[c,g]
/trunk merge
|
2025-04-01T06:40:46.526038
| 2024-06-05T19:12:17
|
2336647040
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11509",
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/66950"
}
|
gharchive/pull-request
|
dissyllabify, foredoes
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 600s
close stale after: 24 hours
[pullrequest]
requests per hour: 20
deps=[d,f]
/trunk merge
|
2025-04-01T06:40:46.527606
| 2024-06-07T12:58:19
|
2340395494
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11510",
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/68596"
}
|
gharchive/pull-request
|
botong, apodous
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 600s
close stale after: 24 hours
[pullrequest]
requests per hour: 20
deps=[a,b]
/trunk merge
|
2025-04-01T06:40:46.529221
| 2024-06-09T13:58:01
|
2342272863
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11511",
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/70536"
}
|
gharchive/pull-request
|
betattered
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 600s
close stale after: 24 hours
[pullrequest]
requests per hour: 20
deps=[b]
/trunk merge
|
2025-04-01T06:40:46.530834
| 2024-06-10T02:25:43
|
2342699236
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11512",
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/71026"
}
|
gharchive/pull-request
|
excommunicable
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 600s
close stale after: 24 hours
[pullrequest]
requests per hour: 20
deps=[e]
/trunk merge
|
2025-04-01T06:40:46.532612
| 2024-06-13T06:43:06
|
2350291037
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11513",
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/74010"
}
|
gharchive/pull-request
|
azotise, exploit
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 600s
close stale after: 24 hours
[pullrequest]
requests per hour: 20
deps=[a,e]
/trunk merge
|
2025-04-01T06:40:46.534453
| 2024-06-13T18:02:23
|
2351757096
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11514",
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/74460"
}
|
gharchive/pull-request
|
belauds, dicyclopentadienyliron
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 600s
close stale after: 24 hours
[pullrequest]
requests per hour: 20
deps=[b,d]
/trunk merge
|
2025-04-01T06:40:46.536026
| 2024-06-13T22:48:58
|
2352179224
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11515",
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/74649"
}
|
gharchive/pull-request
|
franklinite, clarigation
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 600s
close stale after: 24 hours
[pullrequest]
requests per hour: 20
deps=[c,f]
/trunk merge
|
2025-04-01T06:40:46.542754
| 2024-06-27T14:29:16
|
2378328176
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11519",
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/87642"
}
|
gharchive/pull-request
|
aurivorous, calombigas
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 600s
close stale after: 24 hours
[pullrequest]
requests per hour: 20
deps=[a,c]
/trunk merge
|
2025-04-01T06:40:46.544629
| 2024-06-29T00:08:04
|
2381336349
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11520",
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/88935"
}
|
gharchive/pull-request
|
biacuminate, erythroblastosis
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 600s
close stale after: 24 hours
[pullrequest]
requests per hour: 20
deps=[b,e]
/trunk merge
|
2025-04-01T06:40:46.547876
| 2024-07-01T13:56:04
|
2383931989
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11522",
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/91376"
}
|
gharchive/pull-request
|
fleysome, aversive
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 600s
close stale after: 24 hours
[pullrequest]
requests per hour: 20
deps=[a,f]
/trunk merge
|
2025-04-01T06:40:46.552784
| 2024-07-10T15:22:26
|
2401061316
|
{
"authors": [
"EliSchleifer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11525",
"repo": "trunk-io/mergequeue",
"url": "https://github.com/trunk-io/mergequeue/pull/99126"
}
|
gharchive/pull-request
|
fuddles, ejectives
This pull request was generated by the 'mq' tool
[test]
flake rate: 0.1
logical conflict every: 1000
sleep for: 600s
close stale after: 24 hours
[pullrequest]
requests per hour: 20
deps=[e,f]
/trunk merge
|
2025-04-01T06:40:46.558932
| 2022-12-15T21:59:43
|
1499169368
|
{
"authors": [
"TylerJang27"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11526",
"repo": "trunk-io/plugins",
"url": "https://github.com/trunk-io/plugins/pull/68"
}
|
gharchive/pull-request
|
Improve testing workflow
Changes targeted by this PR:
Avoid snapshot assertions failing when it's a new version that doesn't have a snapshot. This includes removing the requirement to specify output files, and now snapshots are always autogenerated. See comment in linterCheckTest() for specifics on this design. This decision was rooted in a couple factors, namely:
As identified in #65, the lack of a snapshot for a specific version was causing unnecessary failures.
Linters can still change their messages, outputs, etc (as sqlfluff did yesterday). This is something that we don't want to cause too much friction, but we still want to call it out as a noticeable change. If this happens during someone's workflow (should still be relatively infrequent), now simply run the specific command in tests/readme.md and validate expected output. We may want to highlight this specific command better or with more visibility.
The JSON assertions and snapshot code were a bit redundant and unnecessarily complex. This should simplify and be flexible enough for the near future. If we want any additional assertions, we can do so via the postCheck hook.
Make tests run in parallel. This also simplifies the setup for the config_check_test.ts. It's a tad redundant, but its presence makes triaging problems a bit easier.
Attempt to reduce flakiness (the parallelism appears to reduce flakiness to about 1-3%--it'd be nice to have analytics on this). I'll continue investigating alternatives for dealing with the daemon lifetime.
Average test time right now locally with these 4 tests is about 10 seconds in total :)
They're about a minute in total on CI because of the need to download and install linters.
Remaining todos (unordered):
Investigate daemon lifetime changes
Modify sandboxing to shadow tree from root
Use debugging library in testing
Add contributing.md
Add option for specifying a range of linter versions (ideally for nightlies)
Migrate rest of definitions
Addressing all comments in a follow-up to unblock folks with testing failures
|
2025-04-01T06:40:46.560073
| 2022-07-29T22:18:20
|
1322718691
|
{
"authors": [
"dauphin3"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11527",
"repo": "trustgraph/trustgraph",
"url": "https://github.com/trustgraph/trustgraph/issues/34"
}
|
gharchive/issue
|
fix data import paths in weav/instance.py
files currently in weav folder but need to figure out how to properly import using ././ etc while opening json and pandas read_csv
wrong repo
|
2025-04-01T06:40:46.562505
| 2024-11-20T16:31:55
|
2676450468
|
{
"authors": [
"dejanb",
"jcrossley3"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11528",
"repo": "trustification/trustify",
"url": "https://github.com/trustification/trustify/issues/1030"
}
|
gharchive/issue
|
Status should probably be a single field rather than an array
This was an observation in #1006 that I think warrants its own issue...
The response from /api/v1/vulnerability/{id} contains a field advisories.sboms.status which is an array and not a single field. That technically means that an a Single advisory can say that an SBOM has the status affected and not_affected at the same time. E.g.
{
"advisories": [
{
"identifier": "Advisory1",
"sboms": [
{
"id": "Sbom1",
"status": ["affected", "not_affected"]
// other fields
},
]
}
]
}
This structure forces the client to come up with some kind of algorithm to parse the data and solve conflicts between multiple advisory.sboms.status values which might contradict each other.
This should be fixed by #1098. The status field is replaced with purl_statuses that contains list of purls with the specific status. As advisory can refer to different packages with different statuses, this should remain the array
|
2025-04-01T06:40:46.571510
| 2023-06-13T10:12:41
|
1754530990
|
{
"authors": [
"RobGeada"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11529",
"repo": "trustyai-explainability/trustyai-explainability",
"url": "https://github.com/trustyai-explainability/trustyai-explainability/pull/211"
}
|
gharchive/pull-request
|
Adds Checkout to main-branch action
... because otherwise there's nothing to build from
Many thanks for submitting your Pull Request :heart:!
Please make sure that your PR meets the following requirements:
[ ] You have read the contributors guide
[ ] Pull Request title is properly formatted: FAI-XYZ Subject
[ ] Pull Request title contains the target branch if not targeting main: [0.3.x] FAI-XYZ Subject
[ ] Pull Request contains link to the JIRA issue
[ ] Pull Request contains link to any dependent or related Pull Request
[ ] Pull Request contains description of the issue
[ ] Pull Request does not include fixes for issues other than the main ticket
\approve
\lgtm
|
2025-04-01T06:40:46.573922
| 2017-04-01T15:43:45
|
218703852
|
{
"authors": [
"trvinh"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11530",
"repo": "trvinh/phyloprofile",
"url": "https://github.com/trvinh/phyloprofile/issues/7"
}
|
gharchive/issue
|
problem with online version
Somehow the Detailed plot of the online version doesn't work the same as the offline version. The order of sequences is incorrect :(
it works now, still don't know why =))
|
2025-04-01T06:40:46.596185
| 2022-02-13T11:07:40
|
1135668352
|
{
"authors": [
"KODIKAS-NL",
"Romakita",
"coveralls"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11531",
"repo": "tsedio/tsed",
"url": "https://github.com/tsedio/tsed/pull/1762"
}
|
gharchive/pull-request
|
fix(json-mapper): fix serialization of additional properties
Serialization of additional properties if @additionalProperties(true) decorator is set on Model
Information
Type
Breaking change
Fix
No
I expected the addition of the @additionalProperties decorator to a Model to also allow serialization of any additional properties but it didn't. This changes fixes this.
Pull Request Test Coverage Report for Build<PHONE_NUMBER>
9 of 9 (100.0%) changed or added relevant lines in 1 file are covered.
No unchanged relevant lines lost coverage.
Overall coverage increased (+0.007%) to 95.688%
Totals
Change from base Build<PHONE_NUMBER>:
0.007%
Covered Lines:
4037
Relevant Lines:
4050
๐ - Coveralls
Hello @KODIKAS-NL
Thanks for this first PR ;)
Just minor changes!
See you
Romain
:tada: This PR is included in version 6.102.1 :tada:
The release is available on:
v6.102.1
GitHub release
Your semantic-release bot :package::rocket:
|
2025-04-01T06:40:46.613422
| 2022-05-22T08:12:39
|
1244179115
|
{
"authors": [
"DavyMorgan",
"p501501",
"zwx980624"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11532",
"repo": "tsinghua-fib-lab/CLSR",
"url": "https://github.com/tsinghua-fib-lab/CLSR/issues/9"
}
|
gharchive/issue
|
About your trained model
Could you please provide your trained model?
Thank you so much.
ไฝ ๅฅฝ๏ผไฝ ่ฎญ็ปๅบๆทๅฎๆฐๆฎ้ไบๅ๏ผ้ๅธธๆ่ฐข
ไฝ ่ฝๆไพไฝ ่ฎญ็ปๆ็ด ็ๆจกๅๅ๏ผ ๅคชๆ่ฐขไบใ
1
Thanks for your attention on our work. We have released the pretrained model. Please check pretrained-model-evaluation.
|
2025-04-01T06:40:46.623514
| 2019-03-25T15:37:13
|
424971079
|
{
"authors": [
"dmayle",
"tsl0922"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11533",
"repo": "tsl0922/ttyd",
"url": "https://github.com/tsl0922/ttyd/pull/167"
}
|
gharchive/pull-request
|
Use SIGCHLD for graceful child cleanup without hanging
We can't guarantee that child processes exit immediately. Rather than hanging, we handle the SIGCHLD we receive to ensure that we wait on a process when it's ready to be waited upon.
e73a8e85878ddd5e60075b8b6daca5d52c6dbd06
|
2025-04-01T06:40:46.624865
| 2020-07-07T20:35:12
|
652621830
|
{
"authors": [
"rexim"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11534",
"repo": "tsoding/cm",
"url": "https://github.com/tsoding/cm/issues/106"
}
|
gharchive/issue
|
Programs the produce their output too fast block cm
$ cargo run yes
This is because we are "flushing" the input until we hit "EAGAIN" (I don't remember if it's actually EAGAIN, but it's besides the point). So in cause of yes it never finishes "flushing". We can try to fix that by limiting amount of lines we "flush" to unblock the UI.
It also rises the problem of potentially infinite output that my consume all the memory.
|
2025-04-01T06:40:46.694763
| 2023-02-21T02:58:48
|
1592700215
|
{
"authors": [
"eth-p",
"tth05"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11535",
"repo": "tth05/obsidian-completr",
"url": "https://github.com/tth05/obsidian-completr/pull/54"
}
|
gharchive/pull-request
|
Use Callout Manager plugin as source for callouts.
This pull requests adds support for the Callout Manager API, allowing Callout Manager to provide a list of callouts (and their icons/colors) to use for callout completions. It introduces a new drop-down setting to change the source for callout suggestions.
By default, Completr's callouts.json file will be used.
If Callout Manager is installed, a new option labelled "Callout Manager" will be available.
Selecting this option will have Completr suggest the callouts detected by Callout Manager instead.
Screenshot:
(the `deprecated` callout is a custom callout)
Sorry for the delay. I've been pretty inactive on here.
|
2025-04-01T06:40:46.767359
| 2018-04-20T23:28:17
|
316439142
|
{
"authors": [
"prestone"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11536",
"repo": "tucnak/telebot",
"url": "https://github.com/tucnak/telebot/issues/134"
}
|
gharchive/issue
|
Message proxy to bot
I have a bot. I want to make a chat inside the bot. Admins can send messages
Is it possible to send *Message from user to all admins?
Its like proxy, user send message to bot, bot send it to other users.
Thank you
yep its call forward)
|
2025-04-01T06:40:46.774410
| 2022-07-12T12:14:34
|
1301977493
|
{
"authors": [
"ahueck",
"coveralls"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11537",
"repo": "tudasc/TypeART",
"url": "https://github.com/tudasc/TypeART/pull/131"
}
|
gharchive/pull-request
|
Release candidate 1.8.1
Fix issue with phpmap git tag, set it to current relese 1.35
Remove AD CI config for safeptr, see #64
Pull Request Test Coverage Report for Build<PHONE_NUMBER>
0 of 0 changed or added relevant lines in 0 files are covered.
No unchanged relevant lines lost coverage.
Overall coverage increased (+0.06%) to 89.921%
Totals
Change from base Build<PHONE_NUMBER>:
0.06%
Covered Lines:
3203
Relevant Lines:
3562
๐ - Coveralls
|
2025-04-01T06:40:46.777229
| 2016-09-12T18:23:36
|
176447351
|
{
"authors": [
"brunoalr",
"sampollard",
"wlngai"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11538",
"repo": "tudelft-atlarge/graphalytics",
"url": "https://github.com/tudelft-atlarge/graphalytics/issues/109"
}
|
gharchive/issue
|
Can not download graphs
The link to http://atlarge.ewi.tudelft.nl/graphalytics/ is broken.
Our server is currently down due to an issue in the data center. Currently we are working on migrating data back from our mirror site. Our apologies for the inconvenience, and we will provide you an update on this within a week.
Thanks for the answer! Hope the issue is solved soon.
@wlngai any updates?
Now I can access the mirror ๐
Thanks for reporting the issue. All dataset should be made accessible by end of today.
Looks like the server is down again (at http://atlarge.ewi.tudelft.nl/graphalytics/)
|
2025-04-01T06:40:46.866714
| 2023-06-20T18:21:56
|
1765917037
|
{
"authors": [
"allenbenz",
"turboderp"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11539",
"repo": "turboderp/exllama",
"url": "https://github.com/turboderp/exllama/pull/84"
}
|
gharchive/pull-request
|
Minor import time output suppression for windows
These outputs are problematic when piping output to other programs.
"where" writes some unhelpful stuff to stderr/stdout "INFO: Could not find files for the given pattern(s). so this quiets that.
Put the "injected compiler path" behind the verbose flag.
The failure to find cl.exe message is now written to stderr to avoid having it unintentionally redirected.
Seems reasonable.
|
2025-04-01T06:40:46.891724
| 2022-10-26T17:54:14
|
1424474654
|
{
"authors": [
"cbruno10",
"prosinger"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11540",
"repo": "turbot/steampipe-plugin-uptimerobot",
"url": "https://github.com/turbot/steampipe-plugin-uptimerobot/issues/11"
}
|
gharchive/issue
|
Issues around and <create_date_time> columns
Describe the bug
When I do
select create_date_time,ssl from uptimerobot_monitor where type = 1;
I get
+----------------------+--------+
| create_date_time | ssl |
+----------------------+--------+
| 1970-01-18T04:55:16Z | |
| 1970-01-19T22:28:28Z | |
| 1970-01-20T05:56:19Z | |
It doesn't seem to read the creation date and if SSL check is active for the monitor. This happnes of course also without filtering out the create_date_time and ssl columns
Steampipe version (steampipe -v)
Example: v0.3.0
Plugin version (steampipe plugin list)
Example: v0.5.0
To reproduce
select create_date_time,ssl from uptimerobot_monitor where type = 1;
Expected behavior
I expected to get creation date/time (time seems to be right) and if the SSL check on the monitor is enabled or not (0/1 for example)
Additional context
Add any other context about the problem here.
Hey @prosinger , thanks for opening this issue!
I was able to reproduce the issue with the create_date_time column, and pushed a fixed to the main branch (this has not been released yet in a new version).
For SSL, I was unable to reproduce since SSL appears to be a paid feature (or at least I misunderstood where the ssl column data should come from), but I've created a branch called add-raw-data-monitor that adds a new column called raw to the uptimerobot_monitor table that should help us debug.
If you get a chance, can you please:
Clone this repo and switch to the add-raw-data-monitor branch
Run make to locally build the plugin
Then run select raw from uptimerobot_monitor and share the results here (with any sensitive info removed)
This column will contain all of the response data we're getting back from the API, so it can help us determine if we're looking in the right field when populating the ssl column.
FYI, this branch also contains the create_date_time column fix, so please let me know if you still don't see the correct date time showing up.
Thanks!
Ok, give me some time for this please
I tried to compile it but I'm having dependency issues all the time (right now I'm missing redis-dev). I'll need some more time for this, over the weekend or after
@prosinger What OS are you working on? If you'd like to share the error(s) you're receiving, I'd be happy to try and help debug them. Thanks!
@prosinger What OS are you working on? If you'd like to share the error(s) you're receiving, I'd be happy to try and help debug them. Thanks!
Ubuntu 20 LTS. I'll follow up on this soon. Sorry for the delay
I managed to compile it on another PC. Apparently the droplet I was tryin on lacks RAM.
So, now I get:
Error: 'api_key' must be set in the connection configuration. Edit your connection configuration file and then restart Steampipe (SQLSTATE HV000)
although the connection is there, and it's read-only
...although this sounds more like it's not being able to find the connection config file, while it's actually there
can you check which config file is it actually expecting (where, name)? It's uptimerobot.spc in the config dir
is one entry enough? (I also have the problem here that it lists only 50 checks)
| {"alert_contacts":null,"create_datetime":1636104766,"friendly_name":"":"","http_username":"","id":789645145,"interval":60,"is_group_main":0,"keyword_case_type":null,"keyword_type":null,"keyword_value":"","logs":null,"monitor_group":0,"port":"","ssl":null,"status":2,"sub_type":"","timeout":30,"type":1,"url":"https://akamai-rap1.somedonain.com/image31.jpg"}
Thanks @prosinger , so looking at the response data, I can see ssl is null. Is SSL enabled for that monitor? If so, can you please try running the cURL command (replacing enterYourAPIKeyHere):
curl -X POST -H "Content-Type: application/x-www-form-urlencoded" -H "Cache-Control: no-cache" -d 'api_key=enterYourAPIKeyHere&format=json&logs=1' "https://api.uptimerobot.com/v2/getMonitors"
And see if you get back the correct SSL information for each monitor? If you can find one with SSL disabled and one with it enabled, that will be helpful.
This will help determine if it's the SDK or the API returning null for the SSL property.
Thanks!
@cbruno10 yeah Monitor SSL is enabled. This is what I actually notices - you get SSL for all rows in the table
and woah, is there a way to isolate the ecact monitor with the curl? I get s loy of non-wrapped data when I do your curl format
@prosinger Yes, you can, the full API documentation is available at https://uptimerobot.com/api/. For the getMonitors method, you can specify a specific one like monitors=15830-32696-83920.
Here it is:
{"stat":"ok","pagination":{"offset":0,"limit":50,"total":1},"monitors":[{"id":789645145,"friendly_name":"Akamai latency endpoint","url":"https://akamai-rap1.somedomain.com/image31.jpg","type":1,"sub_type":"","keyword_type":null,"keyword_case_type":null,"keyword_value":"","port":"","interval":60,"timeout":30,"status":2,"create_datetime":1636104766,"logs":[{"id":2321806411,"type":2,"datetime":1636562549,"duration":31522323,"reason":{"code":"200","detail":"OK"}},{"id":2320797641,"type":1,"datetime":1636530598,"duration":31951,"reason":{"code":"300001","detail":"The SSL certificate is expired."}},{"id":2307464418,"type":2,"datetime":1636101202,"duration":429396,"reason":{"code":"200","detail":"OK"}},{"id":2307463957,"type":98,"datetime":1636101166,"duration":36,"reason":{"code":"98","detail":"Monitor started"}}],"lastLogTypeBeforeStartDate":{}}]}
ssl
optional (defines if SSL certificate info for each monitor will be returned)
so when I add sss=1 into parameters, I get this:
{"stat":"ok","pagination":{"offset":0,"limit":50,"total":1},"monitors":[{"id":789645145,"friendly_name":"Akamai latency endpoint","url":"https://akamai-rap1.somedomain.com/image31.jpg","type":1,"sub_type":"","keyword_type":null,"keyword_case_type":null,"keyword_value":"","port":"","interval":60,"timeout":30,"status":2,"create_datetime":1636104766,"ssl":{"brand":"Let's Encrypt","product":"R3","expires":1674722671,"last_check":"2022-11-10","ignore_errors":0,"disable_notifications":1},"logs":[{"id":2321806411,"type":2,"datetime":1636562549,"duration":31522799,"reason":{"code":"200","detail":"OK"}},{"id":2320797641,"type":1,"datetime":1636530598,"duration":31951,"reason":{"code":"300001","detail":"The SSL certificate is expired."}},{"id":2307464418,"type":2,"datetime":1636101202,"duration":429396,"reason":{"code":"200","detail":"OK"}},{"id":2307463957,"type":98,"datetime":1636101166,"duration":36,"reason":{"code":"98","detail":"Monitor started"}}],"lastLogTypeBeforeStartDate":{}}]}
Thanks for helping with the debugging @prosinger ! I've released a fix for both of these columns in v0.0.3, if you're still seeing issues, please re-open this issue and let us know.
|
2025-04-01T06:40:46.900293
| 2017-02-02T03:14:55
|
204772138
|
{
"authors": [
"lawlite19",
"michaelyma12"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11541",
"repo": "turi-code/python-libffm",
"url": "https://github.com/turi-code/python-libffm/issues/10"
}
|
gharchive/issue
|
Error in Make process: 'No such file or directory'
I am running the quick start procedures on Windows terminal. After I do 'cd ffm' and then 'make', I receive the following error:
'fatal error: parallel/winpthreadsll.h: No such file or directory'
I attached a screenshot of the full output below.
I have the same error. How did you solve it?
|
2025-04-01T06:40:46.909637
| 2017-12-04T00:05:02
|
278836699
|
{
"authors": [
"Elektordi",
"turmoni"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11542",
"repo": "turmoni/temp-probe-exporter",
"url": "https://github.com/turmoni/temp-probe-exporter/pull/1"
}
|
gharchive/pull-request
|
Temp values from 1w have not padding zeros.
Prevents value freeze for temperatures below 10ยฐ (like now in France).
Thanks, my only 1w temperature probe is indoors so it's always been >10. Do you know if it works with negative temperatures? I've been unable to verify that yet.
I'll see as soon as it's freezing outside. ;)
Just got one today:
fb ff 4b 46 7f ff 0c 10 06 : crc=06 YES
fb ff 4b 46 7f ff 0c 10 06 t=-312
And it seems to work:
onewire_temperature_c{location="exterieur"} -0.312
Nice, thanks!
(Meanwhile I've been in sub-zero temperatures in two countries in the time it took your area to drop to the negatives, I have clearly made some bad decisions somewhere.)
|
2025-04-01T06:40:46.911775
| 2024-12-08T03:54:00
|
2724936406
|
{
"authors": [
"JamesChenX",
"weishirongzhen"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11543",
"repo": "turms-im/turms",
"url": "https://github.com/turms-im/turms/issues/1588"
}
|
gharchive/issue
|
what flutter sdk version use in chat demo ?
Looks like current flutter stable version 3.24.5 does not has 'spacing' params on Column or Row
but in .fvmrc it shows use 3.24.5
I will test and try out the project setup workflow from scratch this weekend, so please wait.
I will @ you once done.
|
2025-04-01T06:40:46.988882
| 2016-11-26T18:20:01
|
191819296
|
{
"authors": [
"mdo",
"viktor-yakubiv"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11547",
"repo": "twbs/bootstrap",
"url": "https://github.com/twbs/bootstrap/issues/21210"
}
|
gharchive/issue
|
.btn-outline- caption text colors
For .btn-primary (and others) was made an option to set color of caption text on _variables.scss
But for all .btn-outline-* caption text in active state is white, even if black is better to look. It would be good feature add similar option for .btn-outline- or use colors from $btn-*-color in button-outline-variant mixin, or use sass functions for checking color lightness.
Outline buttons are specifically designed to share the same color for color and border-color. No planned changes here for making that any different.
Yeah, in normal state. But in hover and active states buttons change its background, and on different backgrounds we have only white text. That's I mean.
|
2025-04-01T06:40:46.991872
| 2017-09-30T07:34:07
|
261825808
|
{
"authors": [
"Johann-S",
"josh-byster",
"mayank-mittal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11548",
"repo": "twbs/bootstrap",
"url": "https://github.com/twbs/bootstrap/issues/24180"
}
|
gharchive/issue
|
Navbar not working with current dist?
With the current distribution bootstrap.min.css, none of the navbar examples on the docs show up or function even remotely correctly. Tried both the CDN and current file in the dist folder. I haven't had this problem before.
Example Gist
Actually you are using the latest distribution of bootstrap but not the latest version of the code. Actually bootstrap 4 has a different structure all together
For you navbar to work use this link
And for future references if you want to bootstrap v4, you have to look for https://v4-alpha.getbootstrap.com in the url.
Don't forget to close the issue.
The new URL of our latest release documentation : https://getbootstrap.com/docs/4.0/getting-started/introduction/
|
2025-04-01T06:40:46.992796
| 2021-03-24T14:50:13
|
839804727
|
{
"authors": [
"XhmikosR",
"crdo"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11549",
"repo": "twbs/bootstrap",
"url": "https://github.com/twbs/bootstrap/issues/33462"
}
|
gharchive/issue
|
V5 Offcanvas - Optional sizes
Similarly as it is in the Modal component it would be great to have optional sizes for the Offcanvas component (Extra Large, Large, Small).
Duplicate of #33387
|
2025-04-01T06:40:47.000997
| 2017-01-13T20:43:38
|
200726291
|
{
"authors": [
"Johann-S",
"XhmikosR",
"bardiharborow",
"mdo",
"vanduynslagerp"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11550",
"repo": "twbs/bootstrap",
"url": "https://github.com/twbs/bootstrap/pull/21715"
}
|
gharchive/pull-request
|
JS Mangle private properties => reduce minified JS by 11.9% (4.8% gzipped)
All the private attribute (variables or function) in Bootstrap JS modules starts with _ so we can easily configure Uglify to mandle them and produce a small minified JS file.
Minified size
Minified gain
Minified + zip size
Minified + zip gain
Before
46,653b
-
12,648b
-
After
41,059b
11.9%
12,063b
4.8%
In addition, this size reductions cumulate with #21711
I'm all for this in principle, but there's a problem, which is that --mangle-props does not exist in UglifyJS v2.4.10 (the latest version on npm). I've asked the UglifyJS team why they aren't publishing new versions to npm, and I'll update this issue when I hear back.
This needs review and conflicts resolved before we could merge it now. @Johann-S @XhmikosR can you take a look and let me know if we should proceed with this PR or something else?
@vanduynslagerp: how safe is that?
BTW @vanduynslagerp you would need to move the change in build/uglifyjs.config.json
@vanduynslagerp: how safe is that?
It's safe as long as you don't prefix your public class methods and properties with _.
And it provides a benefit as long as you do prefix your private class methods and properties with _.
It was the case when I opened this PR.
I didn't contribute to Bootstrap for a while and I didn't follow the recent updates, so I can't tell about the current safety/efficiency of this modification with the current code base.
Feel free to re-use or adapt the modification or close the PR if the code has changed and doesn't follow the convention of prefixing private methods and properties with _.
@Johann-S: what do you think about this? If you agree we should make a new branch and test this ourselves.
Yes sure, our minified file size will grow with the removal off jQuery from our dependencies so optimizations are welcome here ๐
|
2025-04-01T06:40:47.006182
| 2021-11-04T13:00:48
|
1044737689
|
{
"authors": [
"ErinvanderVeen",
"Infinisil"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11551",
"repo": "tweag/haskell-fido2",
"url": "https://github.com/tweag/haskell-fido2/issues/40"
}
|
gharchive/issue
|
Implement apple attestation format
This is used by iPhone's TouchID, see https://www.w3.org/TR/webauthn-2/#sctn-apple-anonymous-attestation
We will eventually need https://www.apple.com/certificateauthority/private/
Closed with #41
|
2025-04-01T06:40:47.027583
| 2023-01-13T23:31:33
|
1533032616
|
{
"authors": [
"Gelbpunkt",
"MrGlockenspiel",
"vilgotf"
],
"license": "ISC",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11552",
"repo": "twilight-rs/twilight",
"url": "https://github.com/twilight-rs/twilight/pull/2060"
}
|
gharchive/pull-request
|
Fix dependency name in example
Very minor README change
We mention futures in alot of examples, the book, and also depend upon it in dev-deps of some packages. I think we should change all instances to futures-util at once if we want to do this. It should nearly just be a search and replace
Newer iterations of the README have a completely different example code that uses neither of the futures crates. Closing this stalled PR.
|
2025-04-01T06:40:47.030860
| 2024-12-26T01:11:50
|
2759154993
|
{
"authors": [
"Erk-",
"QuangBoo"
],
"license": "ISC",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11553",
"repo": "twilight-rs/twilight",
"url": "https://github.com/twilight-rs/twilight/pull/2398"
}
|
gharchive/pull-request
|
feat(http): add get voice state endpoints
Discord supports endpoints that allow to fetch voice state of the current user or another user in a guild.
https://github.com/discord/discord-api-docs/pull/7061
missing entry in request.rs to parse the url into the route again, looking good otherwise
I believe that you mean the tests in routing.rs. I just added
missing entry in request.rs to parse the url into the route again, looking good otherwise
I think it already is there:
https://github.com/twilight-rs/twilight/blob/4896c94a818d1f736838a8cb42f81a583603842e/twilight-http-ratelimiting/src/request.rs#L449
I don't think we need one that has the user id as well since it is probably not ratelimited per user.
|
2025-04-01T06:40:47.032385
| 2023-10-25T19:56:40
|
1962129541
|
{
"authors": [
"brandon-lango",
"dremin"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11554",
"repo": "twilio-professional-services/flex-project-template",
"url": "https://github.com/twilio-professional-services/flex-project-template/issues/402"
}
|
gharchive/issue
|
Update GitHub actions to Node 20
Node 16 has reached end of life. GitHub is planning to migrate all actions to Node 20 by Spring 2024.
https://github.blog/changelog/2023-09-22-github-actions-transitioning-from-node-16-to-node-20/
Thanks! This should now be taken care of.
|
2025-04-01T06:40:47.033243
| 2021-12-14T00:02:50
|
1079155619
|
{
"authors": [
"timrozum"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11555",
"repo": "twilio/twilio-live-interactive-audio",
"url": "https://github.com/twilio/twilio-live-interactive-audio/pull/11"
}
|
gharchive/pull-request
|
Set media processor max duration to 30 minutes
Set this to 30 minutes since the new default is only 5 minutes.
I think CI has not been setup since moving to the monorepo so that is why CI fails.
|
2025-04-01T06:40:47.042987
| 2020-11-06T19:56:32
|
738006794
|
{
"authors": [
"idallas",
"timmydoza"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11556",
"repo": "twilio/twilio-video-app-react",
"url": "https://github.com/twilio/twilio-video-app-react/issues/356"
}
|
gharchive/issue
|
Unable to change audio inputs using Firefox
Describe the bug
Changing audio inputs does not work in firefox due to a MediaStreamError
To Reproduce
Steps to reproduce the behavior:
Connect to app with a computer that has more than one audio input
Put in your name and room name and click Continue button
Click on Settings button
Select Audio and Video Settings from the Settings menu
Try to change the Audio Input
The audio input doesn't change, and there is an error in the console.log.
Expected behavior
The audio input should change to what was selected.
Environment (please complete the following information):
OS: macOS 10.15.7 and Windows 10
Browser: Firefox 82
App Version: 0.2.1
SDK Version: 2.8.0-beta.2
Node.js version: 12.18.4
Additional context
Console.log errors:
2020-11-06 19:52:04.776Z | ERROR in [LocalAudioTrack #3: {7b6cc5f0-df45-c840-bb3d-cf660da5580d}]: Failed to re-acquire the MediaStreamTrack:
MediaStreamError { name: "NotReadableError", message: "Concurrent mic process limit.", constraint: "", stack: "" }
Object { deviceId: {โฆ} }
Uncaught (in promise)
MediaStreamError { name: "NotReadableError", message: "Concurrent mic process limit.", constraint: "", stack: "" }
Thanks for the issue @idallas!
Unfortunately, I'm not able to reproduce this issue. I tried switching audio devices in Firefox 82 and it seemed to work fine.
I have seen the Concurrent mic process limit error before. Firefox won't let you start a new audio stream if an existing stream (or clone of a stream) is still active. We take care to stop the audio tracks that are cloned in the AudioLevelIndicator component.
I wonder if you would still encounter the error if you commented out this line of code and then tried to switch devices before you join a room. If this works for you, then it means that there is still some issue in the AudioLevelIndicator component.
I'm also curious about the devices that are connected to your machine. What device are you switching away from when you see this error? I wonder if there is some kind of problem stopping the track (audioTrack.stop() must be called on the existing track before a new one is started).
Sorry I couldn't find the cause of your error, but maybe some of these ideas will yield us some clues. Let me know what you think. Thanks!
Thanks for taking a look! I was able to reproduce it on a third computer (another macOS Catalina) by just adding the virtual ZoomAudioDevice, used for sharing computer audio over Zoom. Maybe it only happens when virtual audio devices are present? I have multiple virtual audio devices on my Windows machine as well, but not the Zoom audio device. I'll get a test going with only physical audio devices and see if it still happens.
For extra context, the virtual audio devices I currently see on my Windows setup are an Oculus Virtual Audio Device (part of the Oculus Rift software), Steam Streaming Microphone, Virtual Desktop Audio, and Bigscreen Audio Stream. Those might all be VR software related.
Ok, this was my mistake and the issue is resolved. I thought I had the latest version of your code, but I downloaded it again just to be sure and now I can't reproduce the issue. Sorry to use up your time!
Not a problem @idallas - I'm happy to hear you got it working!
|
2025-04-01T06:40:47.071138
| 2020-12-07T01:52:35
|
758091345
|
{
"authors": [
"alexbibi",
"jftaylor"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11557",
"repo": "twintproject/twint",
"url": "https://github.com/twintproject/twint/issues/1060"
}
|
gharchive/issue
|
Issue regarding the number of tweets returned
Issue Template
Please use this template!
Hello All, thank you so much to anyone contributing to this project you guys are amazing :)
Initial Check
If the issue is a request please specify that it is a request in the title (Example: [REQUEST] more features). If this is a question regarding 'twint' please specify that it's a question in the title (Example: [QUESTION] What is x?). Please only submit issues related to 'twint'. Thanks.
Make sure you've checked the following:
[] Python version is 3.6; OK
[] Updated Twint with pip3 install --user --upgrade -e git+https://github.com/twintproject/twint.git@origin/master#egg=twint; OK
[] I have searched the issues and there are no duplicates of this issue/question/request. Done
Command Ran
twint -s $CB --since 2017-01-01 --until 2017-01-31
results:
[!] No more data! Scraping will stop now.
found 0 deleted tweets in this search.
Some other queries return results (witn other keyword) but the number of results do not seem to match how many there is on twitter
Tried using VPN to see if more results would be found but no effect (tried 3 different IP)
When I look at :
https://twitter.com/search?q=%24CB until%3A2017-01-31 since%3A2017-01-01&src=typed_query&f=live
I see a lot of results few hundreds at least
Description of Issue
I noticed that I get a lot less results when compared to an advanced search in twitter
Environment Details
Using Windows 10, Pycharm 2020.3, same problem in terminal and in code
Without quotes, $CB is probably not searching for what you want toto be searching for.
Did you try twint -s '$CB' --since 2017-01-01 --until 2017-01-31
twint -s '$CB' --since 2017-01-01 --until 2017-01-31
826188387942014976 2017-01-30 14:00:36 -0800 #Earnings Tomorrow-P2 $CB $CHRW $EA $EQR $FBHS $FFIC $FICO $GFF $ILMN $KTCC $MANH $MSL Free stock app https://t.co/S59zN0R1wP #iTunes
thanks for the reply, I tried your suggestion: twint -s '$CB' --since 2017-01-01 --until 2017-01-31
but got the exact same results:
[!] No more data! Scraping will stop now.
found 0 deleted tweets in this search.
After several other tests, I noticed that specifying a language was actually solving the issue, all results were discovered.
Thank you for the help,
|
2025-04-01T06:40:47.079462
| 2022-11-30T16:06:08
|
1469877678
|
{
"authors": [
"sadilekivan",
"twistedfall"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11558",
"repo": "twistedfall/opencv-rust",
"url": "https://github.com/twistedfall/opencv-rust/issues/396"
}
|
gharchive/issue
|
Cross-compiling for aarch64, fatal error: limits: No such file or directory
I have a working build when compiling for x86_64, but my final intention is to cross-compile for a RaspberryPi.
Fedora 37 (nushell)
sudo dnf install opencv-devel, which installed the following:
/home/ivanใdnf list installed | find opencv 11/30/2022 04:54:15 PM
โญโโโโฌโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโฎ
โ 0 โ opencv.x86_64 4.6.0-7.fc37 @fedora โ
โ 1 โ opencv-contrib.x86_64 4.6.0-7.fc37 @fedora โ
โ 2 โ opencv-core.x86_64 4.6.0-7.fc37 @fedora โ
โ 3 โ opencv-devel.x86_64 4.6.0-7.fc37 @fedora โ
โฐโโโโดโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโฏ
/home/ivanใrustc --version 11/30/2022 04:54:19 PM
rustc 1.65.0 (897e37553 2022-11-02)
/home/ivan/code/recosใlet-env RUST_BACKTRACE = "full" ; cargo build -vv | save log => log.gz
Anything obvious I missed? I feel like I have all the dependencies I need.
I would guess that your cross-compile toolchain is not set up/installed properly. I have been successfully cross-building for RPi Zero although from the Gentoo system (using https://wiki.gentoo.org/wiki/Crossdev). This specific error message has already been reported several times, there is an issue for it: https://github.com/twistedfall/opencv-rust/issues/322 Maybe you can find a workaround there?
|
2025-04-01T06:40:47.093838
| 2015-02-09T21:05:38
|
57089664
|
{
"authors": [
"caniszczyk",
"ianoc",
"johnynek",
"sriramkrishnan",
"sritchie"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11559",
"repo": "twitter/algebird",
"url": "https://github.com/twitter/algebird/pull/414"
}
|
gharchive/pull-request
|
Add coverageAggregate
We need this to get better, but still not accurate, results due to: https://github.com/scoverage/sbt-scoverage/issues/87
lgtm, merge when green
This is good but we fail when we upload results to coveralls:
https://travis-ci.org/twitter/algebird/jobs/50114430#L1321
Sorry my bad, git foo on cmd line broke stuff and closed all of these
It looks like the tests take too long to run when doing the coverage tracking.
Maybe we can turn down the number of trials in the CMS tests. They seem to be the slow ones in my check:
https://github.com/twitter/algebird/blob/develop/algebird-test/src/test/scala/com/twitter/algebird/CountMinSketchTest.scala#L267
Maybe we can turn down the number of instances on some of these?
I think it is actually working now: https://coveralls.io/github/twitter/algebird
/cc @sriramkrishnan
@ianoc I guess this is green now, and we can start showing code coverage. Merge?
Nice. Does it also work if the tests are in a separate package now?
https://github.com/scoverage/sbt-scoverage/issues/87 is still open.
The code coverage values don't look right do they? It also reports on that website that develop was tested/updated last night. But we've made no changes there.
Looking at the results on that coverage page it seems to report only 28 files are covered... none of which are from algebird-core from what i can see.
about develop: I think that is travis test of the push.
It looks like aggregate testing is indeed broken still. I guess we need a later sbt-coverage plugin, but it appears the coveralls support is lagging, and does not yet support sbt-coverage 1.3.3.
I guess algebird is a uniquely bad case here because all the tests are in another package.
We've been trying this for a year or more.
Closing, since we now have codecov.
|
2025-04-01T06:40:47.101266
| 2021-03-19T18:02:57
|
836242093
|
{
"authors": [
"codecov-io",
"scala-steward"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11560",
"repo": "twitter/bijection",
"url": "https://github.com/twitter/bijection/pull/492"
}
|
gharchive/pull-request
|
Update guava to 30.1.1-jre
Updates com.google.guava:guava from 30.1-jre to 30.1.1-jre.
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Configure Scala Steward for your repository with a .scala-steward.conf file.
Have a fantastic day writing Scala!
Ignore future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "com.google.guava", artifactId = "guava" } ]
labels: library-update
Codecov Report
Merging #492 (cc00539) into develop (47b29ec) will not change coverage.
The diff coverage is n/a.
@@ Coverage Diff @@
## develop #492 +/- ##
========================================
Coverage 77.25% 77.25%
========================================
Files 46 46
Lines 1051 1051
Branches 30 30
========================================
Hits 812 812
Misses 239 239
Continue to review full report at Codecov.
Legend - Click here to learn more
ฮ = absolute <relative> (impact), รธ = not affected, ? = missing data
Powered by Codecov. Last update 47b29ec...cc00539. Read the comment docs.
|
2025-04-01T06:40:47.116641
| 2011-09-29T20:12:36
|
1779868
|
{
"authors": [
"OttoAllmendinger",
"austegard",
"bezmenov",
"chrido",
"dlo",
"dyve",
"etaque",
"fat",
"j-hernandez",
"jakeasmith",
"marcopotento",
"markdotto",
"michaelmior",
"nahuelange",
"ouranos",
"pokonski",
"powder96",
"santiquiss",
"santuxus"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11561",
"repo": "twitter/bootstrap",
"url": "https://github.com/twitter/bootstrap/issues/341"
}
|
gharchive/issue
|
Add autocomplete
It would be great if Bootstrap supported "autocomplete" like for a search field.
I think the idea behind these requests is to stop having to use jquery ui :P
this is what we have planned for 2.0: http://cl.ly/B7Mf
Has someone already started on this?
Haven't started - gotta get 1.4 out first :)
I'm about to implement this for a client--any design guidelines? Or are you going to do with internally @ Twitter?
I'd be interested in seeing what you came up with.
My approach has been minimal options, offload as much to css as possible (don't set styles in js), keep it lightweight, less is more.
My favorite is https://github.com/jstayton/jquery-marcopolo
I've yet to find a jquery plugin that
Does autocomplete
Provides a dropdown menu
Looks good
Wraps an existing <select> tag
FlexBox comes close, but it would be fantastic if bootstrap provided something in that direction.
FYI, http://harvesthq.github.com/chosen/ integrates well with bootstrap styles, and match @OttoAllmendinger requirements.
@etaque thanks a lot, that is precisely what I was looking for.
@etaque, now that's pretty!
Chosen + Bootstrap + A little Aristo = Sexy Autocompletes. Bonus points for Chosen + [Isotope] (http://isotope.metafizzy.co/) autocomplete filtering+sorting.
Chosen doesn't seem to support fetching items via AJAX which is one of the great things about jQuery UI autocomplete.
@pokonski, I'd be happy, if you create matching styles for Chosen, because it's default styles and styles of Bootstrap are different little bit.
PS: I'm using [http://driversworld.us/app/link?id=5141] your progress bar styles and I'm very satisfied with them. Thanks.
It would be better to create matching styles, but it is not really necessary. Chosen's default style differs from Bootstrap's just a little bit. I have combined Bootstrap and Chosen on this page http://driversworld.us/app/adddrivers/ , and as you can see, it looks OK.
I needed styles for autocomplete too. As long as it's not implemented in bootstrap, I think that JQuery autocomplete is a quite nice solution: http://jqueryui.com/demos/autocomplete/. JQuery UI has it's own stylesheets (https://github.com/jquery/jquery-ui/tree/master/themes/base) and in case of autocomplete combined with bootstrap css it doesn't look bad. All jquery css element classes start with prefix 'ui-', so there are no conflicts.
jQuery UI is very bloated and complicated compared to Bootstrap. Also it's too invasive to change the default styles. For example jQuery UI tabs force certain styles like rounded corners, by adding classes. And there's butload of those :(
@pokonski It would be possible to write a theme for jQuery UI which mirrors Bootstrap. But I agree that's it's really bloated anyway and using both jQuery UI and Bootstrap is overkill.
@michaelmior - thank you for the link!
Very happy to say work has started on this for Bootstrap 2.0 and the early version is now in 2.0-wip. We'll finish it off for launch as soon as possible.
Nice! As autocomplete author (https://github.com/dyve/jquery-autocomplete) I'm looking forward to see where this leads. Also, KUDOS to Chosen, if you don't need remote data that is pretty awesome. I'm gonna use it!
haven't tried it myself, but there is also a chosen which could be used with remote data, http://github.com/meltingice/ajax-chosen
Is there a specific reason that this doesn't support remote data? If not, I'd love to submit a patch for it.
Instead of explicitly storing remote data, it's probably easier and more flexible to allow the source to be specified as a callback which returns the results to use given the currently entered text. Then you can implement remote data or anything else you want on top.
+1 for being able to specify a source callback
I found this thread specifically when looking to see if anyone had implemented a remote source for this...
Is that, or using a callback, in the works anywhere?
@austegard: I've made some modifications to the html generated by marcopolo to add boostrap styles.
Still work in progress but it looks nice so far: https://github.com/ouranos/jquery-marcopolo/tree/bootstrap
See https://github.com/twitter/bootstrap/pull/1983 and https://github.com/twitter/bootstrap/issues/1336
@ouranos - it does look good. We're trying out marcopolo now...
|
2025-04-01T06:40:47.127775
| 2015-11-09T04:04:15
|
115787120
|
{
"authors": [
"elruwen"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11562",
"repo": "twitter/cloudhopper-smpp",
"url": "https://github.com/twitter/cloudhopper-smpp/issues/117"
}
|
gharchive/issue
|
Threads get stuck at DefaultSmppClient.createConnectedChannel
Aloha!
We have occasionally Threads stuck. All the thread dumps show the following:
"Sender Heartbeat 1" prio=10 tid=0x00007f0b50feb800 nid=0x7fa in Object.wait() [0x00007f0b849e2000]
java.lang.Thread.State: WAITING (on object monitor)
at java.lang.Object.wait(Native Method)
- waiting on <0x00000000caa999b0> (a org.jboss.netty.channel.DefaultChannelFuture)
at java.lang.Object.wait(Object.java:503)
at org.jboss.netty.channel.DefaultChannelFuture.awaitUninterruptibly(DefaultChannelFuture.java:259)
- locked <0x00000000caa999b0> (a org.jboss.netty.channel.DefaultChannelFuture)
at com.cloudhopper.smpp.impl.DefaultSmppClient.createConnectedChannel(DefaultSmppClient.java:286)
at com.cloudhopper.smpp.impl.DefaultSmppClient.doOpen(DefaultSmppClient.java:224)
at com.cloudhopper.smpp.impl.DefaultSmppClient.bind(DefaultSmppClient.java:193)
When our connection drops, we unbind the SmppSession (if isBound()). The we use client.bind() again. The issue occurs sometimes, I couldn't reproduce it yet. It happens with different providers.
Maybe it makes sense as a workaround not to wait forever in com.cloudhopper.smpp.impl.DefaultSmppClient#createConnectedChannel
// attempt to connect to the remote system
ChannelFuture connectFuture = this.clientBootstrap.connect(socketAddr);
// wait until the connection is made successfully
// boolean timeout = !connectFuture.await(connectTimeoutMillis);
// BAD: using .await(timeout)
// see http://netty.io/3.9/api/org/jboss/netty/channel/ChannelFuture.html
connectFuture.awaitUninterruptibly();
//assert connectFuture.isDone();
I am using version: 5.0.8
Cheers
Ruwen
I wanted to move this bug report to the new repo, but I can't raise issues there. Can anyone fix it?
|
2025-04-01T06:40:47.129180
| 2015-03-23T17:09:51
|
63777691
|
{
"authors": [
"jjlauer",
"pgoergler"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11563",
"repo": "twitter/cloudhopper-smpp",
"url": "https://github.com/twitter/cloudhopper-smpp/pull/85"
}
|
gharchive/pull-request
|
Add host address in SmppServerConfiguration
Sometimes you don't want to bind on localhost nor <IP_ADDRESS>
This patch allow to bind DefaultSmppServer to a specific address.
Hey @pgoergler this is a useful PR. Nice work. Not sure when we'll do the next release to maven central -- I'm assuming you could use this pretty quickly?
|
2025-04-01T06:40:47.132123
| 2016-06-10T23:54:28
|
159743691
|
{
"authors": [
"billonahill",
"maosongfu",
"objmagic",
"taishi8117"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11564",
"repo": "twitter/heron",
"url": "https://github.com/twitter/heron/pull/902"
}
|
gharchive/pull-request
|
Additions to getting-started-troubleshooting.md
This is an updated addition to getting-started-troubleshooting.md. Please ignore the previous pull requests #900 and #901 (already closed).
Added fixation for an issue with resolving localhost.
@billonahill fixed.
๐
๐
๐
@maosongfu The sample error output is already written.
|
2025-04-01T06:40:47.188818
| 2015-12-27T19:53:37
|
123994199
|
{
"authors": [
"CaptainElliott",
"Eudaldca",
"Pablete1234",
"TiTo-418",
"dentmaged",
"twizmwazin",
"xFracTV"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11565",
"repo": "twizmwazin/CardinalPGM",
"url": "https://github.com/twizmwazin/CardinalPGM/pull/950"
}
|
gharchive/pull-request
|
Implement fancy tab
The feature is fully working except that it's missing skins, everything else works, even tho i would like to remove the code from the header module, and create it's own module, also some events that update it may be missing
PS: any help implementing skins will be nice
/team alias doesn't change the name in the tab
There are others bugs like this one http://imgur.com/09Eygy1
@TiTo-418 i already fixed the team join updating
@Pablete1234 Minor, but some tab UUID's match real player's UUID.
how should i go arround it?
This should be ready to use in all servers, but may have issues if the server has +70 players (even tho it shouldn't)
Only known bug atm is that hats don't show up (skins show up, but not the overlay)
Great job @Pablete1234! :clap:
@Pablete1234 look at this: http://dentmaged.github.io/docs/sportbukkit/org/bukkit/Skin.html
@xFracTV already known bug, and i have a patch for it, i'll push it later
@xFracTV fixed
Made a few notes on formatting, this is looking really great though.
when you use " ". Team name does not appear
I knew you could "/team alias blue \u00A74Red" and it would use unicode values, didn't think quotes would give problems
@twizmwazin what is the best way i could escape that?
L329: IChatBaseComponent.ChatSerializer.a("{text:\"" + displayName + "\"}")
means:
{text:" + whatever + "}
I think you figured it out. Alternately, we could prohibit certain chars from being used in team names
I don't think we need to prohibit any characters, i only need to escape the characters that go into the text, my question is, what is the best method to escape a string?
I don't think we need to prohibit any characters, i only need to escape the characters that go into the text, my question is, what is the best method to escape a string?
Backslashes are the only way to escape strings
I know that, but what method/library/util/whatever does it with all chararters that need it?
@Pablete1234 try out apache commons https://commons.apache.org/proper/commons-lang/javadocs/api-2.6/org/apache/commons/lang/StringEscapeUtils.html#escapeJava(java.lang.String)
moved on to fancy tab V2 #956
|
2025-04-01T06:40:47.190149
| 2023-06-27T16:58:54
|
1777386730
|
{
"authors": [
"davelebbing",
"twocanoes"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11566",
"repo": "twocanoes/xcreds",
"url": "https://github.com/twocanoes/xcreds/issues/83"
}
|
gharchive/issue
|
Refactor refresh
The current implementation of refresh interval could be improved to consider system sleep. Instead of setting a timer interval to check for refresh this could set a timer to check system time. Then if the system sleeps for some duration the intended refresh behavior could be maintained.
i looked and i had already done this. i now added in the time it will check again in the menu as of 5077
|
2025-04-01T06:40:47.193246
| 2020-02-24T22:19:58
|
570183124
|
{
"authors": [
"connorhsm"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11567",
"repo": "twohoursonelife/dictator",
"url": "https://github.com/twohoursonelife/dictator/issues/9"
}
|
gharchive/issue
|
Su command for moderators
Allow moderators to run a command as the user they tag instead of wasting time repeatedly instructuing them how to run a command.
We need to understand how this will work with passing context to the command.
Won't add this. Instead we should consider separate commands for mods to quickly dump helpful information.
Won't add this. Instead we should consider separate commands for mods to quickly dump helpful information.
|
2025-04-01T06:40:47.196127
| 2016-04-26T07:15:08
|
151060348
|
{
"authors": [
"kpaharelau",
"scottdraves"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11568",
"repo": "twosigma/beaker-notebook",
"url": "https://github.com/twosigma/beaker-notebook/pull/4024"
}
|
gharchive/pull-request
|
groovy output control fails #4003
actually it works fine, but the documentation (groovy-examples.bkr) is wrong it should say ...useOutputPanel = true
FIXED
do not also change the output display logic.
|
2025-04-01T06:40:47.196929
| 2016-09-08T10:18:10
|
175717054
|
{
"authors": [
"scottdraves",
"sluhtoja"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11569",
"repo": "twosigma/beaker-notebook",
"url": "https://github.com/twosigma/beaker-notebook/pull/4698"
}
|
gharchive/pull-request
|
3203 render tex errors in red instead of throwing
somehow got it to work
does not fix the blue lambda. please follow the new plan i wrote yesterday on the issue.
|
2025-04-01T06:40:47.198082
| 2017-10-27T16:39:00
|
269167098
|
{
"authors": [
"lovasko"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11570",
"repo": "twosigma/mbeat",
"url": "https://github.com/twosigma/mbeat/issues/28"
}
|
gharchive/issue
|
Add logging of the PID
both mpub and msub processes should report their process ID in the DEBUG notification level
done in 28b32fd
|
2025-04-01T06:40:47.237732
| 2017-09-28T08:25:00
|
261233085
|
{
"authors": [
"CasdDesnDR",
"baiyancheng20",
"sriprabhar",
"twtygqyy",
"yuanshuai220"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11571",
"repo": "twtygqyy/pytorch-LapSRN",
"url": "https://github.com/twtygqyy/pytorch-LapSRN/issues/5"
}
|
gharchive/issue
|
Question about data augmentation
Thanks for your code, it helps me a lot. But I have some questions about data augmentation. In the generate_train_lap_pry.m, you only used downsizing to make more training data. While in the paper, the author augments the training data in three ways, scaling, rotation and flipping. Your performance is better than the paper, but your training data only has 7488 examples. I'm confused about it.
@yuanshuai220 Hi๏ผ I am reproducing the paper result recently. I get the datasets size of (11712, 1, 32, 32). After 200 epochs, I get average psnr 31.32 on Set5 for 4X. After server test, I find that the training datasets play a important roles in resluts. The more richer training datasets is, the better result you will get. Meanwhile data augmentation is also important, you can add scaling, rotation and flipping function in generate_train_lap_pry.m script by yourself.
@CasdDesnDR I agree with you. If the trianing data is not enough, the nerual network will overfit with the training set. So the performance on test set is not good. I will add rotation and flipping in the generate_train_lap_pry.m
@yuanshuai220 @CasdDesnDR Please refer https://github.com/twtygqyy/pytorch-SRResNet/blob/master/data/generate_train_srresnet.m for adding flipping and rotation
@twtygqyy Hi, Thank you for sharing your code. I want to know why you convert RGB images into YCbCr colour space and only use the Y channel information. How about the results directly using all RGB channels?
Hi @baiyancheng20, I followed the LapSRN paper for the implementation. Actually, you can check https://github.com/twtygqyy/pytorch-SRResNet which I used RGB image as inputs.
@twtygqyy Hi, Thank you for sharing your LAPSRN code. I took your pytorch code from Git-hub and executed it. It works only for grayscale images.
I modified the lapsrn.py to extend support for RGB color images. Then I took just one color image from Urban100 dataset. On this image, I performed augmentations as given in your matlab code (generate_train_lap_pry.m). I got around 165 color image patches of size 128x128. Using these image patches (h5 file ) I trained the network for 100 epochs.
For testing I modified the test.py for color images, and gave as input,
a 32x32 cropped image from the original training image. The results are very poor. I'm not sure where I'm going wrong. I have attached modified codes and my results. I request that I may be kindly given technical advise as to how I can proceed further to get the correct results.
sourcefiles.zip
@sriprabhar Hi, I understand that you tried to overfit the network on a small dataset. How is the loss looks like in your training? Did it converge well?
Thanks for your response.
I took the building image (attached) and took several overlapping patches.
Training trial 1
stride = 64 and
number of patched = 15x11 patches of size 128x128.
Convergence noticed. Attached Plot no 1.
Trained for 100 epochs
Training trial 2
stride = 16 and
number of patches were 41x57, each of size 128x128.
Convergence noticed. Attached Plot no 2.
Trained for 5 epochs using training trial 1 model as pre-trained model.
Test image
Image of size 32x32 cropped from building image
Image patch taken from building image used for training. This image is downsampled to 32x32.
I'm not sure how to solve, if its an overfitting problem. Please have a look at the attachments.
Hi,
Also, for training, we have to create dataset in hdf5 format using matlab code. For creating h5 file using patches of single image, the file size is huge. (for 165 color patches, h5 file size is around 500 MB for 57x41 patches the file size is around 3GB)
If I have a folder containing around 50 images of size 1080x1080 and if I run the matlab code for RGB color images, the system hangs.
I'm not sure if I'm following the correct methods for dataset creation and training. Thanks for any kind of help/suggestions
@sriprabhar Hi, I think the way how you generate the h5 is correct, while you will probably get to many small patches out of 1080x1080 because of the huge size. (3GB is not that big, TBH : ) )
A quick way to solve this is to change the stride when you run the matlab code for generation.
The best way to solve this is to generate multiple h5 files, create a new generator with the folder contains h5 files as input, and fetch data from one h5 at a time.
@sriprabhar also the result you plotted makes sense to me. Cause the image you tested might not be the exactly same one as you used in training. Grab one image from h5 which you used for training, and see if the result looks better.
Thank you for your response, I will give the training patch and try.
Also, one more doubt is that, the LapSRN works for Y component alone, we combined the bicubic interpolated cb and cr and merged with the Lap-srn super resolved y component. The results were good. If y component is sufficient for training and for PSNR measurement, then I would like to know why we have to train for RGB images (like in SRResNet)
Hi @sriprabhar you can have a look at section 5.1 in this paper Fast and Accurate Image Super-Resolution Using A Combined Loss. They compared the difference between training with Y and RGB for SR.
|
2025-04-01T06:40:47.243610
| 2022-04-16T18:56:58
|
1206192766
|
{
"authors": [
"SebastienGllmt",
"scarmuega"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11572",
"repo": "txpipe/pallas",
"url": "https://github.com/txpipe/pallas/issues/88"
}
|
gharchive/issue
|
AuxiliaryData encoding is not using uint keys
The line here uses the type transaction_metadata: &KeyValuePairs<Metadatum, Metadatum> which doesn't match the Cardano binary spec. While Metadatum uses int, the spec has a special type for top-level keys called transaction_metadatum_label which uses uint instead.
Concretely, you can see in this transaction that the metadata representation used by Oura ends up being 82a21b12127f810d7dcee28264594554491907e53b08ef38d198e051b4a080 which gives the following cbor.me representation:
82 # array(2)
A2 # map(2)
1B 12127F810D7DCEE2 # unsigned(1302243434517352162)
82 # array(2)
64 # text(4)
59455449 # "YETI"
19 07E5 # unsigned(2021)
3B 08EF38D198E051B4 # negative(643795744601428404)
A0 # map(0)
80 # array(0)
As you can see, instead of the expected<PHONE_NUMBER>9108123211 key, we instead get -643795744601428404
You can see, for example, Cardano Multiplatform Lib uses the special label type here
thanks @SebastienGllmt. Nice catch, the metadatum label should indeed be a uint instead of the current representation.
Regardless of the required change, is it possible that you're using an old version of Oura / Pallas to generate the CBOR? The current version of the lib represents INT metadatum using a minicbor struct that allows up to 2^64, which should be able to accommodate the mentioned value.
I've downloaded the block and decoded it. The rust representation of the auxiliary data shows me:
[
(
Int(
Int {
neg: false,
val:<PHONE_NUMBER>517352162,
},
),
Array(
Def(
[
Text(
"YETI",
),
Int(
Int {
neg: false,
val: 2021,
},
),
],
),
),
),
(
Int(
Int {
neg: false,
val:<PHONE_NUMBER>9108123211,
},
),
Map(
Def(
[],
),
),
),
],
I'm using Oura 1.3
Here are all the steps more concretely:
let event = input.recv()?;
let data = event.data.clone();
match data {
EventData::Block(block_record) => {
let alonzo::BlockWrapper(_, block) = alonzo::BlockWrapper::decode_fragment(payload);
let aux_data = alonzo_block
.auxiliary_data_set
.iter()
.find(|(index, _)| *index as usize == idx);
let auxiliary_data = aux_data.map(|(_, a)| {
let auxiliary_data_payload = a.encode_fragment().unwrap();
println!("{:?}", auxiliary_data_payload);
}
}
}
|
2025-04-01T06:40:47.491221
| 2023-04-28T13:22:51
|
1688553753
|
{
"authors": [
"bakerac4",
"dfreeman"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11573",
"repo": "typed-ember/glint",
"url": "https://github.com/typed-ember/glint/issues/566"
}
|
gharchive/issue
|
Autocompletion of args in vscode is incorrect
When I try to use a gts file, it seems like if there are other autocompletions it wont show the autocompletions for my component args. However when I get to a more specific argument (past those other autocompletions) it will show, and then if you delete characters to get it back to where it wasn't showing before it will still show (and not show the other autocompletions).
I suspect this could be a env issue with my vscode but would love for someone to confirm this for me.
Here is what I am talking about with pictures.
And just to confirm, this is only in .gts filesโyou don't see this behavior in .hbs ones?
Thats correct. I just made a test
file with a normal template. Here is an attached picture.
|
2025-04-01T06:40:47.502131
| 2023-02-27T18:29:59
|
1601733606
|
{
"authors": [
"BennyMcBenBen",
"armanbilge"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11574",
"repo": "typelevel/sbt-tpolecat",
"url": "https://github.com/typelevel/sbt-tpolecat/issues/139"
}
|
gharchive/issue
|
Warn unused is reporting on XML literals
If I am using scala-xml literals, then I am seeing compile warnings starting with Scala version 2.13.9.
Example:
import scala.xml.Elem
case class ToDoItem(name: String) {
def toXml: Elem = {
<to-do-item>
<name>{name}</name>
</to-do-item>
}
}
Compile errors:
[error] ToDoItem.scala:5:5: unused value of type scala.xml.NodeBuffer (add `: Unit` to discard silently)
[error] <to-do-item>
[error] ^
[error] /ToDoItem.scala:6:7: unused value of type scala.xml.NodeBuffer (add `: Unit` to discard silently)
[error] <name>{name}</name>
Perhaps this is a bug report for the Scala compiler? :)
You are probably right. I'll open an issue with scala-xml
Closing because this is covered under https://github.com/scala/bug/issues/12658
|
2025-04-01T06:40:47.504747
| 2022-07-22T02:20:10
|
1314052075
|
{
"authors": [
"jducoeur",
"rossabaker"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11575",
"repo": "typelevel/typelevel.github.com",
"url": "https://github.com/typelevel/typelevel.github.com/pull/362"
}
|
gharchive/pull-request
|
Restore "core" projects to front page
We can quibble about which these are, but this matches the status quo.
See #347.
Yep. I'll want us to think about what belongs here, but getting back to status quo is the right move IMO.
I'm merging what I perceive to be non-controversial ones aggressively. Unless someone says whoa, this one's getting merged in the morning.
|
2025-04-01T06:40:47.505624
| 2015-06-30T06:02:50
|
91994980
|
{
"authors": [
"CodaFi"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11576",
"repo": "typelift/SwiftCheck",
"url": "https://github.com/typelift/SwiftCheck/pull/57"
}
|
gharchive/pull-request
|
Stop hogging the incredibly generic identifier "State"
Quits hogging all the cool identifiers.
:boat:
|
2025-04-01T06:40:47.508089
| 2020-05-09T10:56:54
|
615149524
|
{
"authors": [
"Tekl",
"trendschau"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11577",
"repo": "typemill/typemill",
"url": "https://github.com/typemill/typemill/issues/151"
}
|
gharchive/issue
|
Drag & Drop reordering seems to be broken
In Safari the dragging of a page in the sidebar just behaves like dragging a normal link. In Firefox Typemill v<IP_ADDRESS> does show the initial drag, but dropping or reordering has no effect.
In Firefox 68 the reordering works well, in Firefox 76 it doesn't.
Hmm, I use FF 76 in Linux and Windows and it works well, you probably use iOS? That is hard for me to test, I can check if I still have access to browserstack...
I'm using macOS 10.15. In iOS Firefox is just another Webkit browser without the Gecko engine.
But I'm so sorry ๐. This is no bug. It's caused by the Browser-extension StopTheMadness which I forgot to disable for Typemill. Now it works even in Safari.
That is good to hear !!! ๐
|
2025-04-01T06:40:47.546929
| 2015-05-31T08:38:13
|
83040464
|
{
"authors": [
"coconilu",
"ryantemple"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11578",
"repo": "typicode/json-server",
"url": "https://github.com/typicode/json-server/issues/107"
}
|
gharchive/issue
|
How can I delete or put or push or patch using similar SQL's where?
I have read the total README.DOC
and can do successfully the example.
But I have a problem ,how can I delete the post with its title instead of its id.
example:
{
"posts": [
{
"id": 1,
"title": "json-server",
"author": "typicode"
},
{
"title": "rest",
"author": "Alpha",
"id": 5
},
{
"course": "restful",
"teacher": "Mr.Alpha",
"id": 6
},
{
"id": 7,
"title": "hello kit",
"author": "alpha",
"titles": "hello kit",
"titless": "hello kit"
},
{
"title": "hello world",
"id": 8
}
],
"comments": [
{
"id": 1,
"body": "some comment",
"postId": 1
}
]
}
Can I delete the post which title is hello world ?
I know I can do it with DELETE /posts/8.
Thanks for the trouble.
You'd need to find the post(s) with the specific title you are looking for using the "q" parameter in the collection
GET /posts?q="hello world"
Then grab the IDs from the response for the DELETE call. As far as I know there is no way to go straight to the title (or any other attribute).
A little disappointment,but thanks for your answer. :)
|
2025-04-01T06:40:47.564072
| 2023-11-15T20:30:29
|
1995518276
|
{
"authors": [
"laurmaedje",
"mpmdean"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11579",
"repo": "typst/hayagriva",
"url": "https://github.com/typst/hayagriva/pull/82"
}
|
gharchive/pull-request
|
fix volume reference
This implements a fix suggested entirely by @zepinglee
The issue, which is described in more detail below, is that hayagriva fails to find the volume when it is stored in the articles parent.
https://github.com/typst/typst/issues/2643
Thanks!
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.