added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T06:40:37.992073
| 2024-06-10T14:03:44
|
2344015458
|
{
"authors": [
"iwpnd",
"tidwall",
"undeadcat"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11346",
"repo": "tidwall/tile38",
"url": "https://github.com/tidwall/tile38/pull/745"
}
|
gharchive/pull-request
|
Feature/set circle support
Please do not open a pull request without first filing an issue and/or discussing the feature directly with the project maintainer.
Please ensure you adhere to every item in this list
[ ] This PR was pre-approved by the project maintainer
Raised discussion on Slack: https://tile38.slack.com/archives/C0M6QN8VC/p1718027811980879
[x] I have self-reviewed the code
[x] I have added all necessary tests
[ ] Fix tests, test for concentric circles contains fails.
Depends on: https://github.com/tidwall/geojson/pull/27
Describe your changes
Adds support for SET ... CIRCLE.
<IP_ADDRESS>:9851> SET cities sf POINT 37.769092 -122.451505
OK
<IP_ADDRESS>:9851> SET cities 10_km_from_sf CIRCLE 37.769092 -122.451505 10000
OK
<IP_ADDRESS>:9851> SET cities 100_km_from_sf CIRCLE 37.769092 -122.451505 100000
OK
<IP_ADDRESS>:9851> GET cities 10_km_from_sf
"{\"type\":\"Feature\",\"geometry\":{\"type\":\"Point\",\"coordinates\":[-122.451505,37.769092]},\"properties\":{\"type\":\"Circle\",\"radius\":10000,\"radius_units\":\"m\"}}"
<IP_ADDRESS>:9851> INTERSECTS cities POINT 37.769092 -122.451505
1) (integer) 0
2) 1) 1) "100_km_from_sf"
2) "{\"type\":\"Feature\",\"geometry\":{\"type\":\"Point\",\"coordinates\":[-122.451505,37.769092]},\"properties\":{\"type\":\"Circle\",\"radius\":100000,\"radius_units\":\"m\"}}"
2) 1) "10_km_from_sf"
2) "{\"type\":\"Feature\",\"geometry\":{\"type\":\"Point\",\"coordinates\":[-122.451505,37.769092]},\"properties\":{\"type\":\"Circle\",\"radius\":10000,\"radius_units\":\"m\"}}"
3) 1) "sf"
2) "{\"type\":\"Point\",\"coordinates\":[-122.451505,37.769092]}"
<IP_ADDRESS>:9851> INTERSECTS cities IDS POINT 37.769092 -122.451505
1) (integer) 0
2) 1) "100_km_from_sf"
2) "10_km_from_sf"
3) "sf"
<IP_ADDRESS>:9851> INTERSECTS cities IDS POINT 37.554478 -122.020158
1) (integer) 0
2) 1) "100_km_from_sf"
This is based on PR https://github.com/tidwall/tile38/pull/649, but with some differences:
https://github.com/tidwall/tile38/pull/649 no longer compiles
this uses tidwall/geojsons Circle type, while the previous PR stored the polygon representation of the circle.
Issue number and link
Pull request require a prior issue with discussion.
Include the issue number of link here.
Duplicate of #649
The problem with a circle type as a geometry stored in a collection is that there is no GeoJSON compatible way to represent it without losing precision on output. I'm not opposed with the idea of a SET ... CIRCLE ... command but I would like to somehow solve the compatibility issue.
|
2025-04-01T06:40:38.061335
| 2023-06-13T08:42:39
|
1754356887
|
{
"authors": [
"aohsato",
"manato"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11348",
"repo": "tier4/edge-auto-jetson",
"url": "https://github.com/tier4/edge-auto-jetson/pull/26"
}
|
gharchive/pull-request
|
feat(andible): update environmental setup
Related Links
https://github.com/tier4/edge-auto/pull/4
Description
Update ansible roles to set up edge.auto_jetson environment.
Review Procedure
Please run setup-dev-env.sh to set up the environment.
Remarks
This PR contains the netplan role, which fixes the IP address for eth1. If you are using this network interface for ssh connection, running ansilbe-playbook may cause ssh disconnection.
Pre-Review Checklist for the PR Author
PR Author should check the checkboxes below when creating the PR.
[x] Assign PR to reviewer
Checklist for the PR Reviewer
Reviewers should check the checkboxes below before approval.
[ ] Commits are properly organized and messages are according to the guideline
[ ] (Optional) Unit tests have been written for new behavior
[ ] PR title describes the changes
Post-Review Checklist for the PR Author
PR Author should check the checkboxes below before merging.
[ ] All open points are addressed and tracked via issues or tickets
CI Checks
Build and test for PR: Required to pass before the merge.
@manato How about PR status?
@aohsato As well as other PR, I'll open this PR because I guess it's ready.
Please note that I set the default camera configuration to be:
FSYNC trigger mode
fixed exposure time (11ms)
in this PR.
Please note that I set the default camera configuration to be...
@manato Could you add this explain to our tutorials? https://github.com/tier4/edge-auto/pull/3
|
2025-04-01T06:40:38.088692
| 2022-06-19T04:14:39
|
1275948277
|
{
"authors": [
"caseydavenport",
"tmjd",
"vikas027"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11349",
"repo": "tigera/operator",
"url": "https://github.com/tigera/operator/issues/2026"
}
|
gharchive/issue
|
Question - How do I install Calico add-on in an EKS cluster?
I am looking to install Calico add-on in an EKS cluster along with AWS VPC CNI as suggested in the official calico documentation which points to the AWS official documentation, which further states that the recommended way of installing Calico on EKS is by using this operator.
I am not sure how to do the same with the operator.
What I have tried :)
look in the certification courses but that too has the old way of installing it
Alternatives
Helm Installation helm install calico projectcalico/tigera-operator --version v3.23.1
Yeah it isn't clear how to do this.
I think you should be able to follow the Calico quickstart directions and just ensure you properly configure the custom-resources.yaml for the AWS CNI plugin (after you install the AWS CNI plugin).
The Amazon EKS documentation has steps for installing Calico listed here: https://docs.aws.amazon.com/eks/latest/userguide/calico.html#calico-install
Looks like it could use an update to the latest version, though.
The Amazon EKS documentation has steps for installing Calico listed here: https://docs.aws.amazon.com/eks/latest/userguide/calico.html#calico-install
Looks like it could use an update to the latest version, though.
Yes, this is the link I have in my query as well. It does not state how to install using the operator.
@vikas027 I see that the doc is not very clear, but the instructions listed under the "Helm" tab on that page are the correct instructions to install the tigera-operator (using helm), very similar to the instructions we have here: https://projectcalico.docs.tigera.io/getting-started/kubernetes/helm
We should clean up the language on the Amazon doc so that it's less confusing, but the content there is correct.
|
2025-04-01T06:40:38.106387
| 2023-02-28T19:37:44
|
1603692095
|
{
"authors": [
"codecov-commenter",
"efirs"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11350",
"repo": "tigrisdata/tigris",
"url": "https://github.com/tigrisdata/tigris/pull/882"
}
|
gharchive/pull-request
|
Enable quota metrics by default
Describe your changes
How best to test these changes
Issue ticket number and link
Codecov Report
Patch coverage: 100.00% and project coverage change: -27.71 :warning:
Comparison is base (c0eb3c0) 57.56% compared to head (b071362) 29.85%.
:exclamation: Current head b071362 differs from pull request most recent head 93343b8. Consider uploading reports for the commit 93343b8 to get more accurate results
:mega: This organization is not using Codecov’s GitHub App Integration. We recommend you install it so Codecov can continue to function properly for your repositories. Learn more
Additional details and impacted files
@@ Coverage Diff @@
## main #882 +/- ##
===========================================
- Coverage 57.56% 29.85% -27.71%
===========================================
Files 175 175
Lines 19569 19568 -1
===========================================
- Hits 11265 5843 -5422
- Misses 7349 13042 +5693
+ Partials 955 683 -272
Impacted Files
Coverage Δ
server/metrics/metrics.go
83.69% <100.00%> (-7.71%)
:arrow_down:
server/search/factory.go
0.00% <0.00%> (-100.00%)
:arrow_down:
lib/container/priority_queue.go
0.00% <0.00%> (-100.00%)
:arrow_down:
server/quota/namespace.go
0.00% <0.00%> (-98.37%)
:arrow_down:
schema/lang/go.go
0.00% <0.00%> (-90.63%)
:arrow_down:
schema/lang/java.go
0.00% <0.00%> (-90.63%)
:arrow_down:
schema/lang/typescript.go
0.00% <0.00%> (-90.63%)
:arrow_down:
server/search/sorted_hits.go
0.00% <0.00%> (-89.10%)
:arrow_down:
server/search/facets.go
0.00% <0.00%> (-88.60%)
:arrow_down:
schema/inference.go
0.00% <0.00%> (-85.06%)
:arrow_down:
... and 111 more
Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here.
:umbrella: View full report at Codecov.
:loudspeaker: Do you have feedback about the report comment? Let us know in this issue.
|
2025-04-01T06:40:38.108712
| 2023-03-06T13:44:15
|
1611461086
|
{
"authors": [
"garrensmith",
"pboros"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11351",
"repo": "tigrisdata/tigris",
"url": "https://github.com/tigrisdata/tigris/pull/902"
}
|
gharchive/pull-request
|
fix: only log unexpected errors in match
Describe your changes
We logging an error that is acceptable. There will be cases when the field we looking for does not exist in the
document or in the json timestamp object when we do a query. In those cases rather check first the field does not exist and return false. Only if it is an unexpected error do we log it.
How best to test these changes
Test should pass
Issue ticket number and link
I verified that this indeed fixes the logging issue.
|
2025-04-01T06:40:38.118179
| 2021-10-31T07:56:38
|
1040402347
|
{
"authors": [
"apatait",
"tikhop"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11352",
"repo": "tikhop/TPInAppReceipt",
"url": "https://github.com/tikhop/TPInAppReceipt/issues/94"
}
|
gharchive/issue
|
Objective C example code
Hi
This is not an issue per se, but I could not find a way to communicate this and I thought it'll also help others in similar boat; hence posting here.
Do you have an example of using TPInAppReceipt library in an objective C project? Sample code on how to instantiate the class and call various methods would be great. Or if you have Objective C class documentation, that'd be great too.
I am not familiar as much with Swift, and my project is in Objective C hence the request.
Thank you in advance!
Hi @apatait,
Unfortunately, I don't have an example, but you can use it in objc project.
Here you can find how to install it:
https://github.com/tikhop/TPInAppReceipt/blob/master/Documentation/UseInObjCProject.md
Here is a similar question:
https://github.com/tikhop/TPInAppReceipt/issues/84
Some old example that probably doesn't work anymore.
https://gist.github.com/tikhop/7d1215bbcea7fe8a3be3ce5ef25e15ee
Anyway, it must be pretty easy to use the library in objective c project just by checking swift examples from readme file
|
2025-04-01T06:40:38.122164
| 2023-06-20T04:36:35
|
1764620095
|
{
"authors": [
"cfzjywxk",
"zyguan"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11353",
"repo": "tikv/client-go",
"url": "https://github.com/tikv/client-go/pull/847"
}
|
gharchive/pull-request
|
lock_resolver: let resolve lock timeout properly
fix https://github.com/pingcap/tidb/issues/44822 , after this PR tidb can pass the test.
I think it's better to merge PRs affecting critical path with 2-3 approvals.
|
2025-04-01T06:40:38.138974
| 2023-09-21T06:48:16
|
1906235826
|
{
"authors": [
"Yui-Song"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11354",
"repo": "tikv/tikv",
"url": "https://github.com/tikv/tikv/issues/15653"
}
|
gharchive/issue
|
TableFullScan gets slower resulting in a 5% performance regression in TPCDS
Bug Report
What version of TiKV are you using?
good commit: https://github.com/tikv/tikv/pull/15523
bad commit: https://github.com/tikv/tikv/pull/15522
What operating system and CPU are you using?
Steps to reproduce
deploy a tidb cluster with 1 tidb + 3 tikv with the good and the bad commit
run TPCDS 50g
What did you expect?
no performance regression
What did happened?
Q88 elapsed time: 34s -> 38.9s
Q28 elapsed time: 24.8s->28.3s
exection_plan.zip
/severity critical
It was caused by https://github.com/tikv/tikv/pull/15584
/remove may-affects-5.3
/remove may-affects-5.4
/remove may-affects-6.1
/remove may-affects-6.5
/remove may-affects-7.1
/unlabel may-affects-5.3
/remove-label may-affects-5.3
/remove-label may-affects-5.4
/remove-label may-affects-6.1
/remove-label may-affects-6.5
/remove-label may-affects-7.1
|
2025-04-01T06:40:38.147331
| 2021-12-23T07:52:57
|
1087465884
|
{
"authors": [
"windtalker",
"youjiali1995"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11355",
"repo": "tikv/tikv",
"url": "https://github.com/tikv/tikv/pull/11708"
}
|
gharchive/pull-request
|
coprocessor: not return rows when there is no input for simple aggregation
What is changed and how it works?
Close pingcap/tidb#30923
What's Changed:
For SimpleAggregation does not return row if there is no input rows
Related changes
PR to update pingcap/docs/pingcap/docs-cn:
PR to update pingcap/tidb-ansible:
Need to cherry-pick to the release branch
Check List
Tests
Unit test
Manual test (add detailed scripts or steps below)
Side effects
Performance regression
Consumes more CPU
Consumes more MEM
Breaking backward compatibility
Release note
Fix wrong `any_value` result when there are regions returning empty result
/label needs-cherry-pick-5.3
/label needs-cherry-pick-5.2
/label needs-cherry-pick-5.1
/label needs-cherry-pick-5.0
/label needs-cherry-pick-4.0
/rebuild
pingcap/tidb#30923
TiDB v3.1.2/v3.0.20/v2.1.19, all of them have the same issue.
BTW seems that TiKV is keeping this behavior since v2.0.0, so this might be a bug exists for a lot of versions. Better to check v3.0 and v2.1 as well.
Tested in TiDB v3.1.2/v3.0.20/v2.1.19, all of them have the same issue.
/label needs-cherry-pick-3.1
/label needs-cherry-pick-3.0
/label needs-cherry-pick-2.1
/merge
@breeswish can you help merge it?
/merge
/merge
|
2025-04-01T06:40:38.153205
| 2024-09-06T07:07:45
|
2509688354
|
{
"authors": [
"overvenus",
"wuhuizuo"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11356",
"repo": "tikv/tikv",
"url": "https://github.com/tikv/tikv/pull/17490"
}
|
gharchive/pull-request
|
ci: update owners aliases
What is changed and how it works?
Issue Number: Close #17004
What's Changed:
Related changes
[ ] PR to update pingcap/docs/pingcap/docs-cn:
[ ] Need to cherry-pick to the release branch
Check List
Tests
[ ] Unit test
[ ] Integration test
[ ] Manual test (add detailed scripts or steps below)
[x] No code
Side effects
[ ] Performance regression: Consumes more CPU
[ ] Performance regression: Consumes more Memory
[ ] Breaking backward compatibility
Release note
None
/cc @yudongusa @benmeadowcroft @zhangjinpeng87
/cc @overvenus @LykxSassinator
/ok-to-test
/retest
/retest
flaky unit test
/cc cfzjywxk
|
2025-04-01T06:40:38.160772
| 2020-06-03T10:11:27
|
629869531
|
{
"authors": [
"youjiali1995"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11357",
"repo": "tikv/tikv",
"url": "https://github.com/tikv/tikv/pull/8009"
}
|
gharchive/pull-request
|
test: stabilize lock_manager tests
Signed-off-by: youjiali1995<EMAIL_ADDRESS>
What problem does this PR solve?
Issue Number: close #xxx
Problem Summary:
What is changed and how it works?
Proposal: xxx
What's Changed:
Related changes
PR to update pingcap/docs/pingcap/docs-cn:
PR to update pingcap/tidb-ansible:
Need to cherry-pick to the release branch
Check List
Tests
Unit test
Integration test
Manual test (add detailed scripts or steps below)
No code
Side effects
Performance regression
Consumes more CPU
Consumes more MEM
Breaking backward compatibility
Release note
No release note
/test
/test
/run-all-tests
/rebuild
/test
/test
/test
/test
/test
/test
/test
/test
|
2025-04-01T06:40:38.170083
| 2016-07-23T22:53:10
|
167207515
|
{
"authors": [
"das-keyboard",
"fenryxo",
"lonlazer"
],
"license": "bsd-2-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11358",
"repo": "tiliado/nuvola-app-amazon-cloud-player",
"url": "https://github.com/tiliado/nuvola-app-amazon-cloud-player/issues/17"
}
|
gharchive/issue
|
Nuvola does not recognise the current song
Nuvola does not recognise the current song. It shows in the sidebar only (null). So neither the lyrics plugin nor the desktop notifications are working.
I installed the Version 5.1 of the Plugin and Version 3.1.0 of Nuvola via the Arch User Repository.
Hello @lonlazer. Thanks for taking the time to report this issue. I can confirm it.
@SteVwonder, could you take care of it? Thanks!
I wonder if there is still someone taking care of updating this service/app.
Can someone give us some information about it?
@fenryxo @SteVwonder
Hello @das-keyboard. This script is currently unmaintained as the maintainer @SteVwonder seems to have lost interest in it. I might take over it next month if there is any spare time.
I've adopted this abandoned script and released version 5.2 with fixes.
-- Jiri Janousek, Nuvola Player developer • My work on Nuvola Player and user support are funded by Nuvola Patrons. :heart: Make a donation to keep the project going. Thank you.
|
2025-04-01T06:40:38.202503
| 2024-03-03T07:51:40
|
2165235175
|
{
"authors": [
"zaneduffield"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11359",
"repo": "tim-harding/soa-rs",
"url": "https://github.com/tim-harding/soa-rs/issues/6"
}
|
gharchive/issue
|
Compile error from soa-rs-derive: can't leak private type
Using version 0.3.1 of soa-rs I cannot compile anything because a private type is being leaked from soa-rs-derive.
For me, to reproduce the issue it's as simple as
cargo init test_soa_rs;
cd test_soa_rs;
cargo add soa-rs;
cargo check;
I get the following error message
Compiling soa-rs-derive v0.3.1
error[E0446]: private type `SoaDerive` in public interface
--> C:\Users\zane\.cargo\registry\src\index.crates.io-6f17d22bba15001f\soa-rs-derive-0.3.1\src\fields.rs:9:1
|
9 | / pub fn fields_struct(
10 | | ident: Ident,
11 | | vis: Visibility,
12 | | fields: Punctuated<Field, Comma>,
13 | | kind: FieldKind,
14 | | soa_derive: SoaDerive,
15 | | ) -> Result<TokenStream, syn::Error> {
| |____________________________________^ can't leak private type
|
::: C:\Users\zane\.cargo\registry\src\index.crates.io-6f17d22bba15001f\soa-rs-derive-0.3.1\src\lib.rs:75:1
|
75 | struct SoaDerive {
| ---------------- `SoaDerive` declared as private
For more information about this error, try `rustc --explain E0446`.
error: could not compile `soa-rs-derive` (lib) due to previous error
I upgraded to Rust 1.76 from 1.74 and I no longer have the issue.
|
2025-04-01T06:40:38.208627
| 2021-11-18T21:05:58
|
1057799645
|
{
"authors": [
"BjoernPetersen",
"torbencarstens"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11360",
"repo": "tim-hat-die-hand-an-der-maus/preliminary-frontend",
"url": "https://github.com/tim-hat-die-hand-an-der-maus/preliminary-frontend/issues/12"
}
|
gharchive/issue
|
Remove size from image url
movie.imdb.coverUrl.rsplit("@", 1)
lazy loading
Bump
Could you prioritize this?
This is blocking important processes! (My sense of aesthetics)
Real talk: shouldn't this happen in the imdb-resolver?
I actually want to do this in the frontend since the load-times will go down. The removal of the actual size can happen in imdb-resolver.
I don't understand, what do you want to do where? (Keep in mind that I refuse to have a look at the frontend code)
imdb-resolver removes the size from the cover url, but then we'd load the full size image in the frontend which isn't necessary since we'll never display a full size cover.
That's why I want to add the actual dimensions (which are set in the frontend anyways) so we have fewer resources to load -> better loading times/less bandwidth waste/no resizing necessary.
So does imdb-resolver already remove the size from the URL, or do we need to create an issue for that?
https://github.com/tim-hat-die-hand-an-der-maus/imdb-resolver/pull/6
I actually need the image size ratio or original width and height.
Maybe we should convert coverUrl to
cover: {
"url": "<String>",
"width": <int>,
"height": <int>
}
or replace width/height with ratio: <float>
what do you think?
I think ratio is probably the best solution
additionally:
_V1_SY150_CR0,0,101,150_ no idea what CR does, doesn't seem to affect the output (visually at least)
SY150 is the height of the image, S can apparently be replaced with U, I couldn't find any differences in those two.
Other characters seem to just ignore the size and show the given width/height based from the top left.
If Y is replaced by X it takes the width for stretching.
|
2025-04-01T06:40:38.214915
| 2023-08-28T18:47:29
|
1870280901
|
{
"authors": [
"SlothUS",
"fcom000",
"tim-hub"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11361",
"repo": "tim-hub/obsidian-bible-reference",
"url": "https://github.com/tim-hub/obsidian-bible-reference/issues/117"
}
|
gharchive/issue
|
not working on windows
the plugin simply does not work, it is active so I can run the command "Bible reference: verse lookup", but when I write inside the verse there is always no answer, nothing. Also if I write --John1:1 etc I never had any output. Tried also without firewall no difference.
Is there some extra plugin I need to run it?
I tried it also in android phone, same problem, nothing happens...
If on windows, can I try open developer console to see
what logs is there?
this is the toggle for enable the developer console.
If you can get some logs, that will be very helpful. I am very curious why this could happen as well.
Also getting the same issue. Here's what my console produces -
Also getting the same issue. Here's what my console produces -
Hi @SlothUS
Thanks for this.
@fcom000 @SlothUS
Can you try to switch the Bible Version provider by Bolls life? in the settings.
This is a CORS issue, which is a server side issue, I know this will happen one day, just don't know it happens so fast.
CORS issue, which it blocks the 3rd part API consume to consume its API.
This is why it failed.
To fix this, need to host a server by my self.
For now I will suggest you to use bolls life versions.
this should be fixed in 1.9.0 by switching the default version, please re-open this if not.
thanks very much, it works with the Bolls life. I was wondering if it is easy to host a Bible Api offline in my computer so even without internet the plugin works (and there will be no update issues?
|
2025-04-01T06:40:38.233379
| 2019-07-23T18:43:24
|
471865221
|
{
"authors": [
"gustavolanzas"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11362",
"repo": "timber/timber",
"url": "https://github.com/timber/timber/issues/2048"
}
|
gharchive/issue
|
wordpress filtered content not available in post object?
I'm not getting the post `content` as expected anymore.
on a single page, i'm using this in the page.php:
`$context = Timber::get_context();
$context['post']= new Timber\Post();'
All my templates that use {{ post.content }} stopped working. When I dump the post object, I don't see the 'content' key as being set in the object.
Expected behavior
{{ post.content }} should output wordpress filtered content (with wpautop, shortcodes, and oembed filters applied.
Actual behavior
Timber\Post Object
(
[ImageClass] => Timber\Image
[PostClass] => Timber\Post
[TermClass] => Timber\Term
[object_type] => post
[custom] => Array
(
[_edit_lock] =><PHONE_NUMBER>:79
[_edit_last] => 79
[_wp_page_template] => default
)
[_content:protected] =>
[_permalink:protected] =>
[_next:protected] => Array
(
)
[_prev:protected] => Array
(
)
[_css_class:protected] =>
[id] => 15061
[ID] => 15061
[post_author] => 79
[post_content] => this is the content
[post_date] => 2019-07-23 09:57:18
[post_excerpt] =>
[post_parent] => 0
[post_status] => publish
[post_title] => test page
[post_type] => page
[slug] => test-page
[__type:protected] =>
[_edit_lock] =><PHONE_NUMBER>:79
[_edit_last] => 79
[_wp_page_template] => default
[post_date_gmt] => 2019-07-23 16:57:18
[comment_status] => closed
[ping_status] => closed
[post_password] =>
[post_name] => test-page
[to_ping] =>
[pinged] =>
[post_modified] => 2019-07-23 09:57:18
[post_modified_gmt] => 2019-07-23 16:57:18
[post_content_filtered] =>
[guid] => http://localhost:8080/library/?page_id=15061
[menu_order] => 0
[post_mime_type] =>
[comment_count] => 0
[filter] => raw
[status] => publish
)
Steps to reproduce behavior
page.php:
$context = Timber::get_context();
$context['post']= new Timber\Post();
Timber::render( array( 'page.twig' ), $context );
page.twig:
{{ post.content }}
or
{{ post | print_r }}
What version of WordPress, PHP and Timber are you using?
PHP 7.1.3, Wordpress 5.2.2, Timber 1.9.2
How did you install Timber? (for example, from GitHub, Composer/Packagist, WP.org?)
Latest from Wordpress plugins
Sorry - discovered some changes to a custom plugin that were breaking this. Not a timber issue.
|
2025-04-01T06:40:38.267557
| 2021-02-07T10:55:30
|
802924213
|
{
"authors": [
"denouche",
"timdorr"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11363",
"repo": "timdorr/tesla-api",
"url": "https://github.com/timdorr/tesla-api/issues/299"
}
|
gharchive/issue
|
auth.tesla.cn ?
Hello
One of my users reported me that he cannot login anymore since I updated my app on the new SSO login.
He also told me that on his side the login page is:
https://auth.tesla.cn/oauth2/v1/authorize?client_id=teslaweb-cn&response_type=code&scope=openid email profile&redirect_uri=https%3A//www.tesla.cn/openid-connect/generic&state=Ht7mUM_AjhcOj5gNaU_xHLUqy-v9TljYQOtnLUghodY&locale=zh-CN
When I asked him to login on Tesla.com he told me that this does not works. On my side, I cannot login to auth.tesla.cn.
So there would be multiple SSO servers not sharing their users databases?
Have you encountered this case?
Last thing I tested on my side, when I try to load: https://auth.tesla.cn/oauth2/v3/authorize?client_id=ownerapi&redirect_uri=https%3A%2F%2Fauth.tesla.cn%2Fvoid%2Fcallback&response_type=code&scope=openid email offline_access&state=xxx
I have an "invalid-redirect-uri" error.
With an account created on tesla.cn I succeed to login using this URL:
https://auth.tesla.cn/oauth2/v3/authorize?client_id=ownerapi&redirect_uri=https%3A%2F%2Fauth.tesla.com%2Fvoid%2Fcallback&response_type=code&scope=openid email offline_access&state=xxxxxx
So the SSO server seems to be on auth.tesla.cn and the redirect_uri seems to be common between all SSO servers (would be logical because the Android app is the same everywhere)
Last thing, the /authorize page seems to be different on auth.tesla.cn and auth.tesla.com , there is one more step on auth.tesla.cn (the password field is hidden when you load the page and you have to click on "Next" to see it).
The first POST to /authorize is sent with the following body:
_csrf: azertyuiopzertyuiop
_phase: identity
transaction_id: azerty
cancel:
identity: user@email
It answer with a 200 and load the /authorize page again, with the password field this time.
Then I post the form with the password and the body contains:
_csrf: qsdfghjklqsdfghjkl
_phase: authenticate
_process: 1
transaction_id: azerty
cancel:
identity: user@email
credential: userPassword
And this time I have the 302 response with the code in the location header.
So it seems there is an additional "identity" phase on the CN login page!
Ok so by adding the extra POST /authorize with the "identity" phase I succeed to obtain an SSO token.
But...
After getting it I cannot obtain an owner API token using the step 4.
The answer is:
Response{protocol=h2, code=404, message=, url=https://owner-api.teslamotors.com/oauth/token} {"response":null,"error":"https://mobilesvc.teslamotors.com:443/accounts/GetUidByFederationId => {:code=>404, :message=>\"Could not get account by federation id\"}","error_description":""}
login_hint query parameter on first GET /authorize seems to be the way to get the good environment, see here:
https://github.com/adriankumpf/teslamate/issues/1297
I'm guessing they distribute a different app build in China, because there are references to the tesla.cn domain in there, but I don't see any code path to actually have them set as the SSO URL or other related variables. So, there may not be that much to glean from the build in the global app stores. Given I'm not a citizen of China, I don't really have much to offer here, unfortunately.
I think the last error I have when trying to exchange my bearer an owner API token is because my account in CN is not associated to any vehicle.
So I sent an alpha version of my app with the login process using the login_hint query parameter, following the redirects on step 1 GET and reuse the Host value of the last redirect.
I will keep you informed.
Ok so I got a feedback from the CN user I sent the alpha version of my app, and this is working for him now!
So I guess the error I got Could not get account by federation id with the CN account I created for test purpose was because I have no vehicle associated with this account.
To summarize, in order to be compliant with all Tesla SSO servers, the GET at step 1 should include a login_hint query parameter with the email as value, and then 2 cases:
you get a 200 response
you get a 303 response redirecting you to another SSO server (for example auth.tesla.cn)
And the rest of the process (step 2 and step 3) should be executed on this SSO server.
So in my case I just followed the redirect if any, then get the request Host, and use this new value as Tesla SSO Environment.
A big thanks to @cliffchen and @adriankumpf for finding a way to handle this!
Documented in #309. Feel free to submit a PR if you want to add anything else.
Thanks! And thanks for this documentation and this repository!!
|
2025-04-01T06:40:38.270014
| 2022-05-05T03:30:17
|
1226163816
|
{
"authors": [
"insaindesign",
"timdorr"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11364",
"repo": "timdorr/tesla-api",
"url": "https://github.com/timdorr/tesla-api/issues/572"
}
|
gharchive/issue
|
Powerwall and energy endpoints
Between this (timdorr/tesla-api) and joeblau/teslaapi.io, there is pretty good coverage of the API. timdorr/tesla-api has a lot more updated information on the auth endpoints, but joeblau/teslaapi.io has much more on the powerwall and energy sites. Can we make these one source of truth?
We've got a start with #580!
|
2025-04-01T06:40:38.296638
| 2018-01-24T12:22:52
|
291189262
|
{
"authors": [
"vpp"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11365",
"repo": "timeoff-management/application",
"url": "https://github.com/timeoff-management/application/issues/232"
}
|
gharchive/issue
|
Black out period on department basis
Would be VERY helpful if we could block out a section of the
calendar in a different colour and put a message on that months
team view or a legend of the colours on it so that it says that
people should not book leave on those dates - we have a period in
the year where we don't allow leave between certain dates
(although we need the option to still book it if its approved by
directors) - just a colour indication I think would be good. So
I would set a date period to be in red, showing that employees
shouldn't book leave on those dates.
+1 as in #283
|
2025-04-01T06:40:38.298876
| 2024-02-02T06:48:13
|
2114217803
|
{
"authors": [
"jovezhong",
"zliang-min"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11366",
"repo": "timeplus-io/proton",
"url": "https://github.com/timeplus-io/proton/pull/550"
}
|
gharchive/pull-request
|
enhancement: input_format_skip_unknown_fields default to true for external streams
PR checklist:
Did you run ClangFormat ?
Did you separate headers to a different section in existing community code base ?
Did you surround proton: starts/ends for new code in existing community code base ?
Please write user-readable short description of the changes:
Closes #549 .
(Jove Github Bot) added it to the current sprint.
(Jove Github Bot) moved this ticket out of the GitHub project(up to 1200 tickets for one project).
|
2025-04-01T06:40:38.300927
| 2020-09-01T05:47:22
|
689833655
|
{
"authors": [
"drpebcak",
"feikesteenbergen"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11367",
"repo": "timescale/timescaledb-kubernetes",
"url": "https://github.com/timescale/timescaledb-kubernetes/pull/216"
}
|
gharchive/pull-request
|
use named port for pgbouncer
The services(s) call for targetPort: pgbouncer but the sts does not currently define that name. This leads to the endpoint not getting computed, and it causes issues with service loadbalancers.
Thanks for the PR, this is part of https://github.com/timescale/timescaledb-kubernetes/pull/218 now (including this commit).
We cannot merge this PR directly, as this commit is now a bugfix on top of 0.7.0, whereas master has diverged a lot already.
|
2025-04-01T06:40:38.373646
| 2015-08-09T21:32:34
|
99930854
|
{
"authors": [
"scriptingosx",
"timsutton"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11368",
"repo": "timsutton/mcxToProfile",
"url": "https://github.com/timsutton/mcxToProfile/pull/11"
}
|
gharchive/pull-request
|
added '--defaults' option that gets data with the defaults command
This adds a --defaults option as an alternative to --plist and --dsobject. Basically this will run the defaults read command and create the mobileconfig from that output. That way you do not have to worry where the actual plist is stored and wether cfprefsd has already updated. There is an extra option --currentHost which will enable the -currentHost flag when running the defaults command.
Example:
./mcxToProfile.py --defaults com.apple.screensaver --identifier com.example.screensaver --currentHost --manage Once
I like the idea! However, since we have Python and Foundation at our fingertips, we could do this without shelling out to Python and use CFPreferences natively. I'm also not sure if the output from defaults properly distinguishes between an integer 0 and a Boolean false. I've never tried to throw the defaults output at plistlib directly as you did here.
I don't know if there's a convenient way to do the equivalent of defaults read with a single API call but there is a method to list all keys, and then each value can be retrieved. Then you'd still get things in a nice dict structure that should map to a plist.
You're right, it wasn't that hard at all.
I have also added flags (or a placeholder) to access the 'anyUser' and 'anyApplication' (Global) domain. This is how you get at the different files:
Application
User
Host
File Location
app_id
current
current
~/Library/Preferences/ByHost/app_id.host_id.plist
app_id
current
any
~/Library/Preferences/app_id.plist
app_id
any
current
/Library/Preferences/app_id.plist
app_id
any
any
???
any
current
current
~/Library/Preferences/ByHost/.GlobalPrefences.host_id.plist
any
current
any
~/Library/Preferences/.GlobalPrefences.plist
any
any
current
/Library/Preferences/.GlobalPreferences.plist
any
any
any
???
I wish I knew what the ??? are. I thought it might read forced profiles if present, but that does not seem to work. Setting those parameter always seems to return empty dicts. Maybe with MCX? If there really is no (current) application for this combinations, it might make sense to assume --currentHost when --anyUser is set.
Examples:
./mcxToProfile.py --defaults com.apple.screensaver --identifier ScreenSaver --currentHost
reads ~/Library/Preferences/ByHost/com.apple.screensaver.host_id.plist
./mcxToProfile.py --defaults com.apple.screensaver --identifier ScreenSaver
reads ~/Library/Preferences/com.apple.screensaver.plist
./mcxToProfile.py --defaults com.apple.loginwindow --identifier loginwindow --anyUser --currentHost
reads /Library/Preferences/com.apple.loginwindow.plist
./mcxToProfile.py --defaults ManagedInstalls --identifier MunkiSettings --currentHost --anyUser
reads `/Library/Preferences/ManagedInstalls.plist
./mcxToProfile.py --defaults NSGlobalDomain --identifier GlobalPrefs --currentHost
reads ~/Library/Preferences/ByHost/.GlobalPrefences.host_id.plist
./mcxToProfile.py --defaults NSGlobalDomain --identifier GlobalPrefs
reads ~/Library/Preferences/.GlobalPrefences.plist
./mcxToProfile.py --defaults NSGlobalDomain --identifier GlobalPrefs --anyUser --currentHost
reads /Library/Preferences/.GlobalPreferences.plist
I think the any-host / any-user combination may have been used in the old days of network-mounted storage and user data. I think it's reasonable enough to make these options for --defaults work similarly to how the defaults command works.
Thanks for adding this!
Oh, and this was just merged: https://github.com/timsutton/mcxToProfile/commit/cd1314f31cec9fe8b874b6727c0e998097269153
Thinking again about the behaviour when --currentHost and --anyUser are both set, it doesn't generate the kind of "scope" we want.
Say I want to capture the settings that are effectively stored in /L/P/com.apple.loginwindow:
I need to type this:
./mcxToProfile.py --defaults com.apple.loginwindow --currentHost --anyUser --identifier Loginwindow
But this gives me something like:
<key>PayloadContent</key>
<dict>
<key>com.apple.loginwindow.ByHost</key>
<dict>
<key>Forced</key>
<array>
<dict>
<key>mcx_preference_settings</key>
<dict>
<key>Hide500Users</key>
<true/>
</dict>
</dict>
</array>
</dict>
</dict>
Notice the ".ByHost" at the end of the domain - except that this is not actually a ByHost pref. I'm starting to wonder if we should instead try and make it behave analogous to defaults, and support the --currentHost flag only. To read from the any-user domain, we could simply take the /L/P/com.whatever.domain syntax that defaults uses, and not put the .ByHost for the domain.
I kind of like the extra behavior, we could also fix the logic to add the "ByHost" to the preference domain:
https://github.com/timsutton/mcxToProfile/pull/12
Thanks, I just wonder what the extra behavior of --anyUser gains us if it only gives expected results when combined with --currentHost. But in practice, it's going to be less common that you want to pull preferences from the any-user domain anyway. We can just document this in an example.
|
2025-04-01T06:40:38.382976
| 2021-06-18T13:04:03
|
924884557
|
{
"authors": [
"timtorChen"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11369",
"repo": "timtorChen/pi-cluster",
"url": "https://github.com/timtorChen/pi-cluster/pull/154"
}
|
gharchive/pull-request
|
(bot) Update ghcr.io/fluxcd/helm-controller Docker tag to v0.11.1
This PR contains the following updates:
Package
Update
Change
ghcr.io/fluxcd/helm-controller
minor
v0.10.1 -> v0.11.1
Release Notes
fluxcd/helm-controller
v0.11.1
Compare Source
CHANGELOG
v0.11.0
Compare Source
CHANGELOG
Configuration
📅 Schedule: At any time (no schedule defined).
🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied.
♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox.
🔕 Ignore: Close this PR and you won't be reminded about this update again.
[ ] If you want to rebase/retry this PR, check this box.
This PR has been generated by Renovate Bot.
Renovate Ignore Notification
As this PR has been closed unmerged, Renovate will now ignore this update (v0.11.1). You will still receive a PR once a newer version is released, so if you wish to permanently ignore this dependency, please add it to the ignoreDeps array of your renovate config.
If this PR was closed by mistake or you changed your mind, you can simply rename this PR and you will soon get a fresh replacement PR opened.
|
2025-04-01T06:40:38.412269
| 2021-02-12T07:31:42
|
807018179
|
{
"authors": [
"atifsyedali",
"r00dY",
"spbyrne"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11370",
"repo": "tinacms/tinacms",
"url": "https://github.com/tinacms/tinacms/issues/1743"
}
|
gharchive/issue
|
StyledFocusRing has width:100%;height:100% which can break CSS layout
Hey guys,
Awesome job with Tina, couldn't love your project more!!!
Just one issue I noticed. react-tina-inline adds StyledFocusRing to the DOM. It's all fine but by making it width: 100%; height: 100% it can break CSS layout. I'm talking about this line of code:
https://github.com/tinacms/tinacms/blob/a22063fdf2e10f873df9bd401f5d224c73ef6729/packages/react-tinacms-inline/src/styles/focus-ring.tsx#L85-L86
Imagine you have an InlineGroup or InlineBlock with content that has some "intrinsic natural width" which is not 100%. Like a button with width: 200px. Imagine that the parent of this element is display: flex. Flex parent makes a use of natural width of its children during layout (uses width as flex-basis if the latter is missing). However, if you add width: 100% between parent and child (StyledFocusRing) then layout breaks, because flex parent thinks that its child is 100% width. Basically any kind of layout that makes a use of content size of its children will break if we add Tina container in between.
In my opinion StyledFocusRing shouldn't have neither width nor height. Its size should be determined by the content and it should be as non-intrusive as possible. (can't wait for display: contents to have better support btw).
I know this change could break backward compatibility in some projects. Maybe we could add a flag? I'd be happy to do a PR, but curious about rationale behind the current behaviour.
Just with some quick testing it seems that this was done for components that use relative or absolute positioning within their parent, which is probably the minority compared to intrinsic sizing. I think possibly just accepting a class string and passing it to the focus ring, or only setting width/height when some prop is set could solve the issue for now. This component is used by the inline group and inline block components, and they already accept focus ring options which could accommodate this change.
Currently for us, certain types of elements that expect direct children to have certain CSS constraints ends up with incorrect layouts while editing. For example, a FlexGroup that has FlexItems requires FlexItems to have flex properties.
It would be great if we can just pass a className to be appended to the StyledFocusRing elements. This way, if I have an element like FlexGroup and another set called FlexItems, then I can give classNames to StyledFocusRing elements that would make them behave like FlexItems.
In the long term, it would be great to not even have feedback elements injected. For example, notion.so adds separate feedback elements that are positioned absolutely for mouse move and dnd effects.
Also, thanks for an amazing project!
|
2025-04-01T06:40:39.479052
| 2022-07-22T12:28:42
|
1314946708
|
{
"authors": [
"codecov-commenter",
"martins0n"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11371",
"repo": "tinkoff-ai/etna",
"url": "https://github.com/tinkoff-ai/etna/pull/813"
}
|
gharchive/pull-request
|
[BUG] Raise errors in models.nn if they can't make in-sample and some cases out-sample predictions
Before submitting (must do checklist)
[x] Did you read the contribution guide?
[x] Did you update the docs? We use Numpy format for all the methods and classes.
[x] Did you write any new necessary tests?
[x] Did you update the CHANGELOG?
Proposed Changes
Closing issues
closes #787
Codecov Report
Merging #813 (2cc0ddd) into master (41fcce4) will decrease coverage by 34.59%.
The diff coverage is 70.00%.
@@ Coverage Diff @@
## master #813 +/- ##
===========================================
- Coverage 84.05% 49.46% -34.60%
===========================================
Files 126 126
Lines 7220 7236 +16
===========================================
- Hits 6069 3579 -2490
- Misses 1151 3657 +2506
Impacted Files
Coverage Δ
etna/models/nn/deepar.py
82.95% <70.00%> (-17.05%)
:arrow_down:
etna/models/nn/tft.py
71.69% <70.00%> (-28.31%)
:arrow_down:
etna/commands/__init__.py
0.00% <0.00%> (-100.00%)
:arrow_down:
etna/commands/backtest_command.py
0.00% <0.00%> (-97.06%)
:arrow_down:
etna/commands/forecast_command.py
0.00% <0.00%> (-94.88%)
:arrow_down:
etna/commands/__main__.py
0.00% <0.00%> (-87.50%)
:arrow_down:
etna/models/utils.py
14.28% <0.00%> (-85.72%)
:arrow_down:
etna/commands/resolvers.py
0.00% <0.00%> (-80.00%)
:arrow_down:
etna/analysis/outliers/density_outliers.py
22.44% <0.00%> (-75.52%)
:arrow_down:
etna/datasets/datasets_generation.py
27.02% <0.00%> (-72.98%)
:arrow_down:
... and 77 more
:mega: Codecov can now indicate which changes are the most critical in Pull Requests. Learn more
|
2025-04-01T06:40:39.481597
| 2018-11-12T15:54:25
|
379839322
|
{
"authors": [
"Scottmar93",
"tinosulzer"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11372",
"repo": "tinosulzer/PyBaMM",
"url": "https://github.com/tinosulzer/PyBaMM/issues/9"
}
|
gharchive/issue
|
Add mesh and operators for spherical domain
Summary
Add (dimensionless?) spherical domain to mesh.py, independently of the existing Cartesian domain
Add grad and div operators (Finite Volumes) for the spherical domain
Mesh shall be unchanged. Will use if domain is "negative particle" or "positive particle" to determine when spherical operators should be used.
Closed by #142
|
2025-04-01T06:40:39.517500
| 2024-02-22T11:11:57
|
2148816076
|
{
"authors": [
"geohot"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11373",
"repo": "tinygrad/tinygrad",
"url": "https://github.com/tinygrad/tinygrad/pull/3475"
}
|
gharchive/pull-request
|
move all reduces to the end in lazy
This is just an experiment.
I don't like how the reduce LazyOp doesn't specify the actual reduce operation. Neither new_shape nor axis list expresses it fully, but rather depends on what shape the input buffer has.
One idea is to have the arg just be the number of elements that are being reduced and move them to the end, but I'm not sure this is expressive enough for multireduce.
Oh I think this actually might work! The final permute isn't really a permute, it's a reshape of only 1s.
This is also already happening in the Linearizer
|
2025-04-01T06:40:39.530446
| 2021-04-03T10:48:45
|
849632155
|
{
"authors": [
"CLAassistant",
"Mickls",
"leonwanghui"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11374",
"repo": "tinyms-ai/tinyms",
"url": "https://github.com/tinyms-ai/tinyms/pull/66"
}
|
gharchive/pull-request
|
fix: start_server不支持windows启动flask的问题
What type of PR is this?
enhancement
What does this PR do / why do we need it:
The start_server method in tinyms does not support running in the windows environment. It is a huge blow to the friends who use the windows environment to learn tinyms.
We have also seen that tinyms stated in the introduction that it supports the operation of the windows system, but it is a very uncomfortable thing that it cannot be used as one of its features.
Which issue(s) this PR fixes:
Fixes #65
Special notes for your reviewers:
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.Dragonfruit_JC seems not to be a GitHub user. You need a GitHub account to be able to sign the CLA. If you have already a GitHub account, please add the email address used for this commit to your account.You have signed the CLA already but the status is still pending? Let us recheck it.
在FlaskServer类中,linux_run_server方法中,有一行循环监听信号的代码,我给了一些注释,这个地方是我存疑的地方,希望你们看完谨慎合并
@lyd911 Please take a look, thanks!
|
2025-04-01T06:40:39.540473
| 2021-02-27T22:57:54
|
818050014
|
{
"authors": [
"drewroberts"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11375",
"repo": "tipoff/laravel-google-api",
"url": "https://github.com/tipoff/laravel-google-api/issues/8"
}
|
gharchive/issue
|
GMB Accounts
Multiple Accounts need to be present to implement data pulls through the Google My Business API (#7) for Locations (https://github.com/tipoff/locations/issues/42). This is also mentioned in this issue:
https://github.com/tipoff/locations/issues/50
https://developers.google.com/my-business/reference/rest/v4/accounts
https://support.google.com/business/answer/6085339?ref_topic=6085325
|
2025-04-01T06:40:39.568938
| 2024-02-06T21:31:26
|
2121699475
|
{
"authors": [
"KeisukeYamashita",
"jackton1"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11376",
"repo": "tj-actions/changed-files",
"url": "https://github.com/tj-actions/changed-files/pull/1918"
}
|
gharchive/pull-request
|
fix: broken link to matrix example workflow
Why
Because the link was broken, I got 404.
It should be linked to this file → https://github.com/tj-actions/changed-files/blob/main/.github/workflows/matrix-example.yml
@all-contributors please add @KeisukeYamashita for docs
|
2025-04-01T06:40:39.575426
| 2022-05-26T02:32:15
|
1249014101
|
{
"authors": [
"jackton1",
"janzzen-tp"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11377",
"repo": "tj-actions/eslint-changed-files",
"url": "https://github.com/tj-actions/eslint-changed-files/issues/632"
}
|
gharchive/issue
|
[BUG] - Warning: Unexpected input(s) 'extensions',
Is there an existing issue for this?
[X] I have searched the existing issues
Does this issue exist in the latest version?
[X] I'm using the latest release
Describe the bug?
To eslint-changed-files maintainers,
First off, i would like to extend gratitude to the maintainers for creating this amazing workflow to automate the linting process.
This workflow has become a great help to our team to ensure code readability and code quality.
However, we found that there were some breaking changes from @v10 so we updated the package to @v11
after updating the latest v11, we have started to receive a warning that is blocking the github actions in the workflow
please see the following snippet:
14s
##[debug]Evaluating condition for step: 'Run ESLint on changed files'
##[debug]Evaluating: success()
##[debug]Evaluating success:
##[debug]=> true
##[debug]Result: true
##[debug]Starting: Run ESLint on changed files
##[debug]Register post job cleanup for action: tj-actions/eslint-changed-files@v[1](https://github.com/fible/fible-app/runs/6603050516?check_suite_focus=true#step:6:1)1
##[debug]Loading inputs
##[debug]Evaluating: github.token
##[debug]Evaluating Index:
##[debug]..Evaluating github:
##[debug]..=> Object
##[debug]..Evaluating String:
##[debug]..=> 'token'
##[debug]=> '***'
##[debug]Result: '***'
Warning: Unexpected input(s) 'extensions', valid inputs are ['token', 'all_files', 'config_path', 'ignore_path', 'file_extensions', 'extra_args', 'skip_annotations', 'level', 'reporter', 'filter_mode', 'fail_on_error']
##[debug]Loading env
Run tj-actions/eslint-changed-files@v[11](https://github.com/fible/fible-app/runs/6603050516?check_suite_focus=true#step:6:11)
with:
config_path: .eslintrc.json
extensions: js,jsx
extra_args: --quiet
token: ***
all_files: false
file_extensions: **/*.ts
**/*.tsx
**/*.js
**/*.jsx
With that said, would there be something missing the configuration that's causing this issue? Thanks for accommodating me.
To Reproduce
This is how the github action is written
name: Run Linter
on:
pull_request:
branches:
- '**'
jobs:
run-linters:
name: ESLint
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v2
- uses: actions/setup-node@v1
with:
node-version: 12.x
- name: Install Packages
run: yarn install
- name: Run Jest
run: yarn test
- name: Run ESLint on changed files
uses: tj-actions/eslint-changed-files@v11
with:
config_path: ".eslintrc.json"
extensions: "js,jsx"
extra_args: "--quiet"
What OS are you seeing the problem on?
all
Expected behavior?
We are expecting the linter to run without warnings.
Relevant log output
14s
##[debug]Evaluating condition for step: 'Run ESLint on changed files'
##[debug]Evaluating: success()
##[debug]Evaluating success:
##[debug]=> true
##[debug]Result: true
##[debug]Starting: Run ESLint on changed files
##[debug]Register post job cleanup for action: tj-actions/eslint-changed-files@v[1](https://github.com/fible/fible-app/runs/6603050516?check_suite_focus=true#step:6:1)1
##[debug]Loading inputs
##[debug]Evaluating: github.token
##[debug]Evaluating Index:
##[debug]..Evaluating github:
##[debug]..=> Object
##[debug]..Evaluating String:
##[debug]..=> 'token'
##[debug]=> '***'
##[debug]Result: '***'
Warning: Unexpected input(s) 'extensions', valid inputs are ['token', 'all_files', 'config_path', 'ignore_path', 'file_extensions', 'extra_args', 'skip_annotations', 'level', 'reporter', 'filter_mode', 'fail_on_error']
##[debug]Loading env
Run tj-actions/eslint-changed-files@v[11](https://github.com/fible/fible-app/runs/6603050516?check_suite_focus=true#step:6:11)
with:
config_path: .eslintrc.json
extensions: js,jsx
extra_args: --quiet
token: ***
all_files: false
file_extensions: **/*.ts
**/*.tsx
**/*.js
**/*.jsx
Anything else?
No response
Code of Conduct
[X] I agree to follow this project's Code of Conduct
@janzzen-tp I’ve added a Migration Guide to help you resolve the warning. Let me know if you need more information but a high-level description of the warning is the extension input is no longer supported and has now been renamed to file_extensions.
|
2025-04-01T06:40:39.654804
| 2021-11-08T08:49:48
|
1047173901
|
{
"authors": [
"schlichtanders",
"tlienart"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11381",
"repo": "tlienart/Franklin.jl",
"url": "https://github.com/tlienart/Franklin.jl/issues/922"
}
|
gharchive/issue
|
inline svg url not working with Franklin because changed by compression
On Franklin server everything just works, but when deploying to gitlab, my background heropatterns (svg images inlined into background-image: url(...)) just don't show up.
Inspecting the issue it turns out that something in the pipeline deleted the surrounding quotes. Unbelievable...
So on my local system I have
background-image: url("data:image/svg+xml,...");
which somehow ends up on gitlab pages as
background-image: url(data:image/svg+xml,...);
no surrounding quotes any longer.
Any help is highly appreciated, as this seems to be due to the standard Franklin gitlab pipeline.
Turning off minify fixed this issue as well.
See https://github.com/tlienart/Franklin.jl/issues/921#issuecomment-962938930
I would vote for turning off minify by default for everyone.
yes so as per other issue, minification will be turned off for everyone (prerendering as well btw even though it causes much fewer issues)
|
2025-04-01T06:40:39.657092
| 2019-08-20T06:06:52
|
482661818
|
{
"authors": [
"Zeus64",
"tlinkowski"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11382",
"repo": "tlinkowski/basic-annotations",
"url": "https://github.com/tlinkowski/basic-annotations/issues/15"
}
|
gharchive/issue
|
Compilation warning about Kotlin's AnnotationTarget enum
The following warning occurs when compiling with basic-annotations in compileOnly scope:
warning: unknown enum constant AnnotationTarget.ANNOTATION_CLASS
reason: class file for kotlin.annotation.AnnotationTarget not found
This happens e.g. when compiling sample-java-api subproject: https://travis-ci.com/tlinkowski/basic-annotations/builds/123617322#L278-L279
Caused by KT-33141.
I have the same problem
|
2025-04-01T06:40:39.658447
| 2022-12-09T21:41:07
|
1487490763
|
{
"authors": [
"bazfum",
"riblus-bandyr"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11383",
"repo": "tloncorp/landscape-apps",
"url": "https://github.com/tloncorp/landscape-apps/issues/1413"
}
|
gharchive/issue
|
profiles: unexpected behavior in sigil color selection
When attempting to edit the sigil color field in the profile settings, it truncates leading zeroes and then pads out the last character. For example, I previously had my color set to 0000b2. After the upgrade, it had converted it to B22222. Attempting to change it back results in the same thing after a refresh. 1000b2 however works.
can confirm this
This is fixed in Groups and Talk, but the Grid still shows the incorrectly padded color.
|
2025-04-01T06:40:39.675550
| 2024-07-22T18:33:37
|
2423524425
|
{
"authors": [
"BrendonPierson",
"tlux"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11384",
"repo": "tlux/sftp_client",
"url": "https://github.com/tlux/sftp_client/pull/36"
}
|
gharchive/pull-request
|
Added preferred_algorithms and packet_size config options
Adding two more config options that get passed to the underlying erlang ssh connect: preferred_algorithms and packet_size
Thank you for adding that!
@BrendonPierson Could you possibly also update config_test.exs to reflect those changes? Thank you.
Sure thing @tlux! Let me know if this looks ok
|
2025-04-01T06:40:39.679185
| 2018-01-08T08:09:11
|
286672279
|
{
"authors": [
"aipsel",
"jimlongo56",
"nubilfi",
"tmcw"
],
"license": "isc",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11385",
"repo": "tmcw/docbox",
"url": "https://github.com/tmcw/docbox/issues/64"
}
|
gharchive/issue
|
After run build script, docbox css broken
Hi,
I'm new to docbox and i have a little problem. Every time i run npm run buildscript it always ends up with a broken CSS, i'm using the default provided CSS, and when i inspect it, i got this error:
Warning: Prop className did not match. Server: "fixed-top fixed-right space-left16" Client: "fixed-top fixed-right undefined"
Also, I found that revert the index.html file can solve this problem.
This would be a great opportunity for someone to help and fix and contribute a PR! I'd love to review and merge such a PR.
Can confirm that changing package.json
"react": "15.6.2",
"react-dom": "15.6.2",
resolves the build issue for me.
|
2025-04-01T06:40:39.692494
| 2018-08-09T19:32:23
|
349260509
|
{
"authors": [
"SquidDev",
"zardyh"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11386",
"repo": "tmpim/amulet",
"url": "https://github.com/tmpim/amulet/pull/84"
}
|
gharchive/pull-request
|
Remove main
We never did anything interesting with main, and our calling of it was rather broken, so it's probably better to remove it. People can now write this instead:
let () = f x
This is closer to how other MLs handle it, which is always nice. We can also remove some pretty ugly code from the backend.
The main disadvantage here is that it breaks a large number of our codegen tests. We used to use main to inject opaque values. As this now has no special meaning, it'll be stripped and so the test won't emit anything!
For now, I've fixed this by calling main with a bottom : 'a value:
let main f = ...
external val bottom : 'a = "nil"
let () = bottom (main bottom)
This definitely is pretty ugly (and can generate some awful code), but I'm OK with this for now. Hopefully the backend rewrite (which is increasingly reaching the status of Urn's resolver rewrite) will clean up a lot of these.
Looks good to me™
|
2025-04-01T06:40:39.716863
| 2017-07-19T10:58:14
|
244001745
|
{
"authors": [
"fabianre",
"ygeelen-tb"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11387",
"repo": "tobania/VSTS.Extension.SqlReportingServices",
"url": "https://github.com/tobania/VSTS.Extension.SqlReportingServices/issues/15"
}
|
gharchive/issue
|
No Release task
Hi,
I am using the Visual Studio Team Services and added your extension. Unfortunately I am not able to find the task when creating a new build sequence. Did I miss something ?
Thanks,
Fabian
Update:
BTW: Just saw that its clearly writtenin the description, shame on me...
Unfortunately the release task is also not there.
Fabian,
It's a task for Release Definitions, not Build Definitions.
Thanks for your fast reply, that make sense. But should I not see it then here ?
(Sorry for those dumb questions, just startet to take a look at the whole Team Services topic today)
Yes you should see it here if the extension is properly installed (the "Deploy" category):
Hmm unfortunately it does not. Any idea what I can do to resolve it ? I already uninstalled and installed it again without luck.
After uninstall/install again, the tasks now show up. Strange, but luckily sovled.
We noticed some temporary issue on VSTS or Azure on the extension page. I guess that was it..
|
2025-04-01T06:40:39.724206
| 2023-10-04T15:14:18
|
1926457436
|
{
"authors": [
"redbar0n",
"sahidrahman404"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11388",
"repo": "tobias-tengler/create-relay-app",
"url": "https://github.com/tobias-tengler/create-relay-app/issues/18"
}
|
gharchive/issue
|
Support for Vike/vite-plugin-ssr
It would be nice to have integration between create-relay-app and Vike/vite-plugin-ssr since create-relay-app currently works with Vite but not with Vike/vite-plugin-ssr
in relation to this, I think create-relay-app would fit nicely into the Bati.js scaffolder for new projects: https://github.com/batijs/bati
|
2025-04-01T06:40:39.747483
| 2023-03-24T21:42:18
|
1640080533
|
{
"authors": [
"TomWagg",
"tobin-wainer"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11389",
"repo": "tobin-wainer/elk",
"url": "https://github.com/tobin-wainer/elk/issues/62"
}
|
gharchive/issue
|
Try giving a list of potential stars that are leading to a given frequency spike
Use the individual pixel periodograms to find likely locations and then use astroquery to search SIMBAD and find some likely candidates
Need to flip the way the pixels are read into the wcs...
if the argwhere returns [48, 62] what we need to put into the pixel_to_world is (62, 48)
If I do "tess_wcs.pixel_to_world(*np.fliplr(solutions[0])).icrs" I get the correct thing, but if there are multiple solutions and I do:
tess_wcs.pixel_to_world(*np.fliplr(solutions)).icrs
then its wrong again....
Ah gotcha so the (x, y) is just backwards? Perhaps we just need *query_pixels.T[::-1] in that case (which will reverse the x-y axes)
I tried that as well... but that reverses the entire array, not the individual pairs
|
2025-04-01T06:40:39.797907
| 2024-12-02T02:53:34
|
2710267551
|
{
"authors": [
"scala-steward"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11390",
"repo": "tofu-tf/typed-schema",
"url": "https://github.com/tofu-tf/typed-schema/pull/916"
}
|
gharchive/pull-request
|
Update sbt, scripted-plugin to 1.10.6
About this PR
📦 Updates
org.scala-sbt:sbt
org.scala-sbt:scripted-plugin
from 1.6.2 to 1.10.6
📜 GitHub Release Notes - Version Diff
Usage
✅ Please merge!
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Configure Scala Steward for your repository with a .scala-steward.conf file.
Have a fantastic day writing Scala!
⚙ Adjust future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "org.scala-sbt" } ]
Or, add this to slow down future updates of this dependency:
dependencyOverrides = [{
pullRequests = { frequency = "30 days" },
dependency = { groupId = "org.scala-sbt" }
}]
labels: library-update, early-semver-minor, semver-spec-minor, version-scheme:early-semver, commit-count:1
Superseded by #918.
|
2025-04-01T06:40:39.801300
| 2024-05-15T16:22:44
|
2298348263
|
{
"authors": [
"creeder-uturn"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11391",
"repo": "tofuutils/tenv",
"url": "https://github.com/tofuutils/tenv/issues/133"
}
|
gharchive/issue
|
Document what priority/order is used for version selection
Right now, I cannot find clear documentation on which order version detection uses. There is a decent listing of which things can be looked at, but it's not immediately obvious which have priority.
It'd be great if the documentation have a section that clearly indicated that order.
If I've missed it feel free to clown on me 😄
It seems like the following order is used, but I've not tested thoroughly:
Project version file
Version file in Project Root (Possibly one anywhere in the tree above the project in order?)
Default system-wide version defined
Notably missing: the requirements listed in the TF files, which can be installed/used with latest-allowed and min-required but it doesn't seem like this is used automatically as a fallback?
Oh, awesome! I didn't see the version resolution order. That's precisely what I was looking for.
I would propose sending that information into a digestible section with version files, or perhaps it's own section.
Weary to make promises I may not deliver but I will probably take a stab at a PR with what I imagine (and I can try to clarify better, too).
|
2025-04-01T06:40:39.804564
| 2016-01-14T20:36:28
|
126742890
|
{
"authors": [
"IndrekV",
"oliverdunk"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11392",
"repo": "toggl/toggl-button",
"url": "https://github.com/toggl/toggl-button/issues/421"
}
|
gharchive/issue
|
Red border?
Had an idea,
Thought gathering some feedback before actually implementing it might make some sense. How about an option which could be enabled, that would add a red border around every page if you were timing? For some people this may be annoying, but I feel for others that it'd be a good reminder that they were still being tracked.
Open to ideas, and if you think this is terrible, be honest! I can take it.
I think it's too "In your face" to have red border on all the pages. I like the idea of always being aware that time is tracking, but I think the red border is too much. Right now the red color of the extension icon is indicating that timer is running.
|
2025-04-01T06:40:39.805948
| 2016-10-10T08:18:28
|
181958278
|
{
"authors": [
"IndrekV"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11393",
"repo": "toggl/toggldesktop",
"url": "https://github.com/toggl/toggldesktop/issues/2011"
}
|
gharchive/issue
|
Ask users to who are on broken updater version to get latest update
Some users are on Macos version that has broken updater. Send them email to update to latest.
Broken updater version should be 7.2.150
This is probably outdated. Double check if anyone is still on this version and then close this.
No users on this ancient version.
|
2025-04-01T06:40:39.808688
| 2019-05-29T08:30:23
|
449668098
|
{
"authors": [
"IndrekV",
"NghiaTranUIT"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11394",
"repo": "toggl/toggldesktop",
"url": "https://github.com/toggl/toggldesktop/issues/2986"
}
|
gharchive/issue
|
Handle cases where project is unsynced
💻 Environment
Platform: All (library)
📒 Description
In some cases, the project or any other item could be not synced and also not marked as "needs sync".
I saw this with an un-synced project that did not have an id (it was NULL), the project was visible in the autocomplete but selecting it would clear the project selection.
First of selecting a non-synced project should still work
Secondly, the project that is not synced should be marked as to be synced when loaded
I wonder that old implementation doesn't store any project_guid after creating, how it handles if the project remains unsync?
I don't see any complain about the unsync project in old code, it means that it always works, isn't it?
Ref: https://github.com/toggl/toggldesktop/blob/b5064ec91eafd104b1d1c7d6baab59da3fe5db42/src/ui/osx/TogglDesktop/TimeEntryEditViewController.m#L396-L421
|
2025-04-01T06:40:39.812157
| 2023-08-11T20:45:07
|
1847408394
|
{
"authors": [
"TheSoloJake"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11395",
"repo": "toggl/track-extension",
"url": "https://github.com/toggl/track-extension/issues/2235"
}
|
gharchive/issue
|
Todoist - Cannot Enter Project in Task View
OS version: macOS latest
Browser version: Chrome latest
Extension version: 3.0.14
Relevant integration (if any):
Todoist webapp
🐛 Describe the bug
User reports that they cannot type into the project window in Toggl when a Todoist task view window is open. Support team member can replicate as well.
Expected behaviour
Project window should be available for typing.
Other details or context
Slack convo
Intercom convo
Intercom convo listed above also contains screenshot video of bug in action
|
2025-04-01T06:40:39.818343
| 2020-10-06T12:46:49
|
715641446
|
{
"authors": [
"LuanFMelo",
"arbutt76",
"loubyl",
"tdardet",
"togiberlin",
"uniAmir"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11396",
"repo": "togiberlin/ui-ux-designer-roadmap",
"url": "https://github.com/togiberlin/ui-ux-designer-roadmap/issues/11"
}
|
gharchive/issue
|
Update for 2020?
Is the plan still to update for 2020's landscape, or is this repo inactive for the time being?
@tdardet I am currently interviewing with several companies.
My plan is to migrate and rewrite some parts in GitBooks towards the end of this year.
Ah awesome, Good luck!
Tomás Dardet
On Sun, Oct 11, 2020 at 1:22 PM togiberlin<EMAIL_ADDRESS>wrote:
@tdardet https://github.com/tdardet I am currently interviewing with
several companies.
My plan is to migrate and rewrite some parts in GitBooks towards the end
of this year.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/togiberlin/ui-ux-designer-roadmap/issues/11#issuecomment-706737909,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/ABEJ3RLVA44AVUML5XAP3A3SKHSWZANCNFSM4SF7HRKA
.
Hi! any chance that you could add in the future a Product Designer roadmap? Maybe not in 2020 but in any update after? Thanks!
Any update?
Any update?
Waiting for your approval
Waiting for your approval
Some updates ?
|
2025-04-01T06:40:39.820290
| 2015-12-01T17:25:59
|
119767063
|
{
"authors": [
"thomcc",
"toji"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11397",
"repo": "toji/gl-matrix",
"url": "https://github.com/toji/gl-matrix/pull/194"
}
|
gharchive/pull-request
|
Implement set and fromValues for mat2, mat2d, mat3, and mat4
Was surprised to find out that these did not already exist.
Let me know if there are any problems.
LGTM, thanks! And thanks for giving me something to test the new Travis CI config with! :D
|
2025-04-01T06:40:39.900114
| 2023-01-12T16:00:50
|
1530972100
|
{
"authors": [
"seanmonstar",
"semtexzv"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11398",
"repo": "tokio-rs/bytes",
"url": "https://github.com/tokio-rs/bytes/issues/586"
}
|
gharchive/issue
|
Add *_unchecked methods
I'm using bytes in an application, where I'm repeatedly slicing to sub-parts of the bytes.
I do my own error-checking of bounds, and handle that out of the hot path.
The methods of (advance, split_off, split_to) do additional bound checking, completely trashing the register allocation.
It'd be nice to have the option to call *_unchecked methods that avoid the bound checking.
Have you been able to try those methods in your app, with a temporary fork? It'd be good to get some data.
|
2025-04-01T06:40:39.901010
| 2019-05-26T18:45:22
|
448594808
|
{
"authors": [
"MSleepyPanda"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11399",
"repo": "tokio-rs/console",
"url": "https://github.com/tokio-rs/console/issues/10"
}
|
gharchive/issue
|
docs: user guides
Write documentation on how to debug applications with the console, how to use it
#3 Would allow us to generate screenshots
|
2025-04-01T06:40:39.925321
| 2015-06-23T05:33:29
|
90292607
|
{
"authors": [
"furueili",
"tolo"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11400",
"repo": "tolo/HHServices",
"url": "https://github.com/tolo/HHServices/issues/14"
}
|
gharchive/issue
|
port number error
hi~ In serviceDidResolve delegate method,
I saw the service data as follows
HHService[0x742E3580, MAC000E53290B23, _afpovertcp._tcp, local., MAC000E53290B23.local., (
"<IP_ADDRESS>:88"
), 1] - Resolved last address
and I try to print the port number
NSLog(@"service port number :%i",service.lastResolvedPort);
what I got is
service port number :22528
please help thx
Hi,
The lastResolvedPort property is an internal property containing a (temporary) port number expressed in network byte order, which is why you get a strange result when you attempt to print it out.
But I've actually just checked in an update to HHService in which I've added new convenience properties, for instance resolvedInetAddresses, which contains a list of resolved addresses on the form IP:PORT.
|
2025-04-01T06:40:39.952243
| 2022-11-12T01:53:12
|
1446170965
|
{
"authors": [
"tomaae",
"wormuths"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11401",
"repo": "tomaae/homeassistant-truenas",
"url": "https://github.com/tomaae/homeassistant-truenas/issues/54"
}
|
gharchive/issue
|
[Bug] No Error, Just Fails
Describe the issue
How to reproduce the issue
Expected behavior
Screenshots
Software versions
Home Assistant version:
TrueNAS integration version:
TrueNAS version:
Diagnostics data
Traceback/Error logs
Additional context
I cant do anything without any info. Make sure to fill the issue template with as much info as there is.
Sorry about that. There really isn't much information to tell. I follow the instructions perfectly, and I get no error message. It just times out or something. What would be most useful to start troubleshooting this? I'm not sure what makes it fail, but I'd love to get it working.
versions, screenshots, diagnostic data, HA log. just provide everything you can.
I just noticed the information isn't there on the bug report. I put all that in when I created it...
Home Assistant 2022.11.2
TrueNAS Core 13.0-RELEASE
Integration v1.2.3
The integration installs, and I clicked add...
All I get is this...
I've tried all sorts of various combinations of SSL, https:// in the URL, etc... Same result. If I put "https:" in the URL i get a "No response from host" error. My firewall logs don't show anything being blocked, but HASS shows connection errors...
======================================
This error originated from a custom integration.
Logger: custom_components.truenas.truenas_api
Source: custom_components/truenas/truenas_api.py:44
Integration: TrueNAS (documentation, issues)
First occurred: 12:15:53 PM (4 occurrences)
Last logged: 12:44:19 PM
TrueNAS <IP_ADDRESS> unable to fetch data "pool" (503)
TrueNAS https://<IP_ADDRESS> unable to fetch data "pool" (no_response)
=======================================
This error originated from a custom integration.
Logger: custom_components.truenas.config_flow
Source: custom_components/truenas/config_flow.py:76
Integration: TrueNAS (documentation, issues)
First occurred: 12:15:53 PM (4 occurrences)
Last logged: 12:44:19 PM
TrueNAS connection error (503)
TrueNAS connection error (no_response)
=======================================
I'm still tinkering in the firewall, but I haven't found any rule which may be blocking this from connecting...
Thanks,
Steve
I have never tried it with 13.0 release version, there could have been some bug.
503 is service unavailable error. seems like there is something broken on truenas side.
I would recommend to actually update your truenas.
Good day,
That would not surprise me. I have found there is another issue with a networking configuration I have reported to them. It prevents me from updating at the moment. My TrueNAS works fine on 13.0, but if I update beyond that all my jails lose internet connectivity. Maybe I'm just special. LOL
https://ixsystems.atlassian.net/browse/NAS-118026
Something under the hood with networking did change along the way with them. The 503... is there a specific service your integration relies on being enabled on the TrueNAS side to check?
Steven
Do they really lose internet connectivity or just DNS? I had such issue with truenas.
Solved it by using resolv.conf syntax in configuration field:
Cant check the ticket, appearantly I dont have access to view it.
I'm accessing official API, I dont think you can control it in any way (officially).
Yeah, sorry. DNS.
yea, you can try to put it into resolver like I did.
DNS server entries in jail configuration never worked for me, even on 12.0
|
2025-04-01T06:40:39.962346
| 2016-11-02T11:34:31
|
186776984
|
{
"authors": [
"fschutt",
"sfackler",
"tomaka",
"wvdschel"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11402",
"repo": "tomaka/android-rs-glue",
"url": "https://github.com/tomaka/android-rs-glue/pull/116"
}
|
gharchive/pull-request
|
Initial JNI support
Feedback welcome!
Example code using these JNI bindings can be seen in https://github.com/tomaka/glutin/pull/822
Here's a gist @sfackler made demonstrating a high-level API for JNI stuff: https://gist.github.com/sfackler/5c036006f8d58a1d38887ba014759f74
It is licensed under MIT/Apache2 so we can copy some stuff.
As I said in the previous comment, this might be out of scope of this PR but I thought I'd leave it somewhere for reference.
Might also be worth depending on https://crates.io/crates/jni-sys to avoid needing to have a second copy in here. Happy to make any changes necessary on that end.
@tomaka, what is your view on using @sfackler's jni-sys crate?
I'd prefer not to re-implement anything that we can readily reuse, but I'm not sure if you're happy to be adding another dependency.
Before I start addressing other feedback, it would be nice to decide if using jni-sys is on the table, and if we should add higher level wrappers to android-rs-glue, jni-sys, or am intermediate crate that wraps raw jni-sys in a nice, safe, Rust-y API.
I agree that adding a higher level API (like JNIEnv's that automatically detach_thread(), or jobject's that automatically DeleteGlobalRef/DeleteLocalRef when dropped). Most of all, I'd like to get rid of the various explicit method calls based on types.
But I'm also not sure this should be part of android-rs-glue.
An alternative approach would be to remove the JNI API from the glue module, and simply using the four methods provided by this patch in injected-glue from jni-sys, guarded by some #![cfg(target_os = "android")].
In my opinion this is the ideal design:
The internal and the external glues both depend on jni-sys and communicate with each other to share the low-level objects (like you have done in your PR).
The external glue also depends on a new crate (named jni for example) that doesn't exist yet and that provides a high-level wrapper around the JNI.
The high-level wrappers are reexported in the external glue's API, but the low-level objects are not exposed.
I can agree with that. I'll look into making the adaptations over the weekend, probably.
The higher level JNI wrapper will probably take a longer time to complete.
@sfackler Are you interested in creating a repo for the higher level wrapper, or should I?
I probably don't have the time to build out the high level wrapper right now unfortunately, so you might want to get it started.
Just a note that I have started on reworking this, but haven't found the time to finish it yet (because life).
So ... what is the current state? The jni_sys crate is pretty mature now and it's been two years ...
Why can't the ffi module simply be exposed to the library user? This way the user can hook up the jni crate and from there on it's rather easy to interact with the JNI. I know about "API concerns", but .. something is better than nothing.
@tomaka ... why? What's wrong about the jni crate? I mean ... do you want to re-do all of the work they've done just to not depend on them - how would you then do this function, if not using jni?
Sorry, what I mean is that we should simply reexport the content of the jni-sys crate from our own crate, instead of exposing it in the API.
|
2025-04-01T06:40:39.966211
| 2015-12-10T18:22:57
|
121547745
|
{
"authors": [
"ReadmeCritic"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11403",
"repo": "tomaka/android-rs-glue",
"url": "https://github.com/tomaka/android-rs-glue/pull/71"
}
|
gharchive/pull-request
|
Update README URLs based on HTTP redirects
Created with https://github.com/dkhamsing/frankenstein
GitHub Corrected URLs
Was
Now
https://github.com/tomaka/gl-init-rs
https://github.com/tomaka/glutin
HTTPS Corrected URLs
Was
Now
http://crates.io/config.html
https://crates.io/config.html
:tada: :fr:
|
2025-04-01T06:40:40.024432
| 2017-04-30T21:05:48
|
225354371
|
{
"authors": [
"alecsandruchirosca",
"grossmane",
"kiliant",
"tomav"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11405",
"repo": "tomav/docker-mailserver",
"url": "https://github.com/tomav/docker-mailserver/issues/586"
}
|
gharchive/issue
|
Cannot start server
Dear all,
I tried for several hours to get this image working and still have the same error:
docker log
docker-mailserver
Initializing setup
Checking configuration
Configuring mail server
No DKIM key provided. Check the documentation to find how to get your keys.
Starting Misc
Starting mail server
Error: Please fix your configuration. Exiting...
docker-compose.yml
version: '2'
services:
mail:
build: docker-mailserver
hostname: xxxxxx
domainname: xxxx.xx
container_name: mail
ports:
- "25:25"
- "143:143"
- "587:587"
- "993:993"
volumes:
- ./maildata/:/var/mail
- ./mailstate/:/var/mail-state
- ./letsencrtypt/:/etc/letsencrypt/
- ./config/:/tmp/docker-mailserver/
environment:
- ENABLE_SPAMASSASSIN=1
- ENABLE_CLAMAV=1
- ENABLE_FAIL2BAN=1
- ENABLE_POSTGREY=1
- ONE_DIR=1
- DMS_DEBUG=0
- SSL_TYPE=letsencrypt
cap_add:
- NET_ADMIN
volumes:
maildata:
driver: local
mailstate:
driver: local
The docker-mailserver directory contains the git clone of the project. Using the suggested image ( tvial/docker-mailserver:latest) yelds the same results.
Can you please help me start this image?
BR,
Alecs
Hi @alecsandruchirosca, to have more detail, use DMS_DEBUG=1, restart container and paste the console log.
`#
ENV
HOSTNAME=xxxxxx.xxxx.xx
DMS_DEBUG=1
VIRUSMAILS_DELETE_DELAY=7
ENABLE_CLAMAV=1
PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin
ENABLE_FAIL2BAN=1
ENABLE_SPAMASSASSIN=1
PWD=/
SHLVL=1
HOME=/root
ONE_DIR=1
SSL_TYPE=letsencrypt
DEBIAN_FRONTEND=noninteractive
ENABLE_POSTGREY=1
_=/usr/bin/printenv
docker-mailserver
Initializing setup
Registering check,setup,fix,misc and start-daemons functions
_check_environment_variables() registered
_check_hostname() registered
_setup_default_vars() registered
_setup_dovecot() registered
_setup_dovecot_local_user() registered
_setup_postgrey() registered
_setup_dkim() registered
_setup_ssl() registered
_setup_docker_permit() registered
_setup_mailname() registered
_setup_amavis() registered
_setup_dmarc_hostname() registered
_setup_postfix_hostname() registered
_setup_dovecot_hostname() registered
_setup_postfix_sasl() registered
_setup_postfix_override_configuration() registered
_setup_postfix_sasl_password() registered
_setup_security_stack() registered
_setup_postfix_aliases() registered
_setup_postfix_vhost() registered
_setup_environment() registered
_fix_var_mail_permissions() registered
_fix_var_amavis_permissions() registered
_misc_save_states() registered
_start_daemons_cron() registered
_start_daemons_rsyslog() registered
_start_daemons_dovecot() registered
_start_daemons_opendkim() registered
_start_daemons_opendmarc() registered
_start_daemons_postgrey() registered
_start_daemons_postfix() registered
_start_daemons_fail2ban() registered
_start_daemons_clamav() registered
_start_daemons_amavis() registered
Checking configuration
Check that there are no conflicts with env variables [_check_environment_variables]
Check that hostname/domainname is provided or overidden (no default docker hostname/kubernetes) [_check_hostname]
Domain has been set to xxxx.xx
Hostname has been set to xxxxxx.xxxx.xx
Configuring mail server
Setting up default variables [_setup_default_vars]
Set ENABLE_LDAP=0
Set ENABLE_FETCHMAIL=0
Set OVERRIDE_HOSTNAME=
Set ENABLE_MANAGESIEVE=0
Set DMS_DEBUG=1
Set ENABLE_POP3=0
Set ENABLE_SASLAUTHD=0
Set ENABLE_CLAMAV=1
Set SMTP_ONLY=0
Set ENABLE_FAIL2BAN=1
Set ENABLE_SPAMASSASSIN=1
Set POSTGREY_DELAY=300
Set POSTGREY_TEXT=Delayed by postgrey
Set ENABLE_POSTGREY=1
Set POSTGREY_MAX_AGE=35
Setting up Dovecot
Setting up Dovecot Local User
'config/docker-mailserver/postfix-accounts.cf' is not provided. No mail account created.
Configuring postgrey
Setting up DKIM
No DKIM key provided. Check the documentation to find how to get your keys.
Setting up SSL
Setting up PERMIT_DOCKER Option
Adding container ip in my networks
Setting up Mailname
Creating /etc/mailname
Setting up Amavis
Applying hostname to /etc/amavis/conf.d/05-node_id
Setting up dmarc
Applying hostname to /etc/opendmarc.conf
Applying hostname and domainname to Postfix
Applying hostname to /etc/postfix/main.cf
Applying hostname to Dovecot
Applying hostname to /etc/dovecot/conf.d/15-lda.conf
Setting up Postfix Override configuration
No extra postfix settings loaded because optional '/tmp/docker-mailserver/postfix-main.cf' not provided.
Setting up Postfix SASL Password
Warning: 'SASL_PASSWD' is not provided. /etc/postfix/sasl_passwd not created.
Setting up Security Stack
Enabling and configuring spamassassin
Enabling clamav
Fail2ban enabled
Setting up Postfix Aliases
Warning 'config/postfix-virtual.cf' is not provided. No mail alias/forward created.
Setting up Postfix vhost
Setting up /etc/environment
Checking /var/mail permissions
Permissions in /var/mail look OK
Checking $amavis_state_dir permissions
Permissions in /var/mail-state/lib-amavis look OK
Starting Misc
Consolidating all state onto /var/mail-state
Destination /var/mail-state/spool-postfix exists, linking /var/spool/postfix to it
Destination /var/mail-state/lib-postfix exists, linking /var/lib/postfix to it
Destination /var/mail-state/lib-amavis exists, linking /var/lib/amavis to it
Destination /var/mail-state/lib-clamav exists, linking /var/lib/clamav to it
Destination /var/mail-state/lib-spamassasin exists, linking /var/lib/spamassasin to it
Destination /var/mail-state/lib-fail2ban exists, linking /var/lib/fail2ban to it
Destination /var/mail-state/lib-postgrey exists, linking /var/lib/postgrey to it
Starting mail server
Starting cron [ OK ]
Starting rsyslog [ OK ]
Starting dovecot services [ OK ]
Starting opendkimfalse
[ FAILED ]
Error: Please fix your configuration. Exiting...
`
With DMS_DEBUG I got the above log (docker log mail ).
BR,
Alecs
Hi, i have exact the same failure.
if i downgrade to docker-mailserver 2.2, then opendkim will be started, but then the error comes during start of amavis.
Output is then:
...
Starting amavisfalse
[ FAILED ] Error: Please fix your configuration. Exiting...
I just upgraded a server to 2.3. No issue with DKIM so far. Only Postgrey was not running.
3 other error to solve:
May 6 12:36:14 mail postgrey[3623]: FATAL: ERROR: can't open lock file: /var/lib/postgrey/postgrey.lock#012
May 6 12:36:15 mail postfix[4414]: Postfix is running with backwards-compatible default settings
May 6 12:36:15 mail postfix[4414]: See http://www.postfix.org/COMPATIBILITY_README.html for details
May 6 12:36:15 mail postfix[4414]: To disable backwards compatibility use "postconf compatibility_level=2" and "postfix reload"
May 6 12:36:15 mail postfix/master[4451]: fatal: open lock file /var/lib/postfix/master.lock: cannot create file exclusively: Permission denied
Tried manually (and had to disable ONE_DIRbut after that it worked well).
No idea yet yo a long term patch. Ideas welcome.
Maybe this also relates to #545.
Although it works for me just well, I also receive backwards-compatibility warnings in my logs.
Will have a look at this later.
Duplicates https://github.com/tomav/docker-mailserver/issues/545
Move this discussion there.
@alecsandruchirosca you had this error because no user account has been provided:
'config/docker-mailserver/postfix-accounts.cf' is not provided. No mail account created. is debug log.
I reproduced this issue today, I'll add a fix.
Here is the PR https://github.com/tomav/docker-mailserver/pull/612
:latest image available.
#491 & #621
|
2025-04-01T06:40:40.028873
| 2017-03-19T19:47:29
|
215289233
|
{
"authors": [
"alinmear",
"tomav"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11406",
"repo": "tomav/docker-mailserver",
"url": "https://github.com/tomav/docker-mailserver/pull/556"
}
|
gharchive/pull-request
|
Fix #526: fatal: no SASL authentication mechanisms
When using the container with SMTP_ONLY = 1, the postfix process within the container fails
on ehlo because there is no valid sasl authentication mechanism available. This happens because sasl has been enabled within postfix/main.cf per default but sasl is not configured.
To fix this _setup_postfix_sasl does not depend anymore on ENABLE_SASLAUTHD and will check in it's logic, whether to enable or disable sasl explicit within postfix/main.cf.
I think you should add an integration test for that.
Added integration test
Just to be accurate on this issue, i just coverd the testing for this particular problem, checking whether sending an email is working or not. But i think we should cover more test cases for the mailserver state SMTP_ONLY use-cases.
Here some examples:
Check Mail send is working
Check SASL authentication is working
Check Relay is working
Check Relay with sasl authentication is working
Maybe we could create a section for this.
|
2025-04-01T06:40:40.033420
| 2014-07-15T18:01:24
|
37909360
|
{
"authors": [
"joeljeske",
"markitosgv",
"nmccready",
"rasmi",
"tombatossals"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11407",
"repo": "tombatossals/angular-leaflet-directive",
"url": "https://github.com/tombatossals/angular-leaflet-directive/issues/421"
}
|
gharchive/issue
|
Listen markercluster click event
Its possible to acces clusterclick, without overlays??
My initial map:
//map initial
angular.extend($scope, {
center: {
autoDiscover: true
},
defaults: {
scrollWheelZoom: false
},
layers: {
baselayers: {
osm: {
name: 'OpenStreetMap',
url: 'http://{s}.tile.openstreetmap.org/{z}/{x}/{y}.png',
type: 'xyz'
}
}
}
});
And then i call my own service to fecth markers data:
//on load map
$scope.$on('leafletDirectiveMap.load', function(event){
api.getMarkersByProximity(function(data){
$scope.markers.push(data);
});
});
Now i have a map with markers and clusters.
With this event i can listen when a marker is clicked, but not a cluster:
$scope.$on("leafletDirectiveMarkersClick", function(event, args){
console.log($scope.markers[args]);
});
Its possible to do?? thanks
I don´t want when cluster marker is clicked to expand child markers... i want to get child markers data to do another thing.
On official Leaflet.markercluster docs says:
markers.on('clusterclick', function (a) {
console.log('cluster ' + a.layer.getAllChildMarkers().length);
});
Is any event directive to listen?
@markitosgv, have you found a way to listen for cluster events?
Has anyone found any ways to listen besides dipping down into pure Leaflet (like @joeljeske did in #648)?
This issue was moved to angular-ui/ui-leaflet#31
I'm going to rework&redesign angular-leaflet-directive to be compatible with Leaflet v1.0. It will mantain almost all its functionality, and will be compatible with the current features of the directive, but I must start from a fresh point, so I'm going to close this issue. If you think it must be worked with the new version, please reopen it.
|
2025-04-01T06:40:40.035038
| 2016-06-16T07:32:39
|
160595720
|
{
"authors": [
"khoerling"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11408",
"repo": "tomchentw/react-google-maps",
"url": "https://github.com/tomchentw/react-google-maps/issues/285"
}
|
gharchive/issue
|
React 15.1.x Support
Is this on the horizon?
On second look, it's working excellently on the latest React! (closing)
|
2025-04-01T06:40:40.070608
| 2022-07-09T20:03:05
|
1299757087
|
{
"authors": [
"awvwgk",
"gnikit"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11409",
"repo": "toml-f/toml-f",
"url": "https://github.com/toml-f/toml-f/issues/90"
}
|
gharchive/issue
|
Support creating machine readable diagnostics
Raised by Giannis @gnikit in https://fortran-lang.discourse.group/t/3949/2
Adding an option or slightly restructuring the diagnostic message to be easily parseable via regex. The message structure is already very good, but fetching the error message via regex would be hard. A good example of that is gfortran-11>= with the flag -fdiagnostic-plain-output. Having an such option would allow for code editors to parse the output of the linter to their Diagnostics console in VS Code PROBLEMS tab.
The current report interface turns an index of a token to a label object
https://github.com/toml-f/toml-f/blob/225174f7a858186507e7bd08c4252918dca9dee0/src/tomlf/de/context.f90#L87-L89
And creates a diagnostic object from it
https://github.com/toml-f/toml-f/blob/225174f7a858186507e7bd08c4252918dca9dee0/src/tomlf/de/context.f90#L92-L96
For the human facing output this is than turned into a string at
https://github.com/toml-f/toml-f/blob/225174f7a858186507e7bd08c4252918dca9dee0/src/tomlf/de/context.f90#L101
Note that the actual source code is only needed when creating the report string, the diagnostic object itself contains only position information from the label objects as well as the messages to display.
To support this without much effort on the user side, we could add a state in the context objects which describes whether the report should be optimized for humans or machines, default to human-friendly output. A tool which wants integrate its TOML Fortran usage for error reporting with the the VS Modern Fortran Extension can provide an option or environment variable to toggle this switch in the context object and make the error output automatically accessible for the Diagnostic console in VS Code.
The actual preferred format for the VS Modern Fortran Extension has to be defined first.
I think I might have been to quick to the current diagnostic message format is okay. I just whipped up a REGEX that should work with VS Code's problemmatcher https://regex101.com/r/q6duAL/4 and in any other tool that uses ECMAScript (Javascript) flavour for regex
The online limitation that I can think of with the current output format, is that the diagnostics are multiline and not all code editors/ linter plugins are able to capture multiline regexs. FYI VS Code is fine with that: https://code.visualstudio.com/Docs/editor/tasks#_defining-a-multiline-problem-matcher
I just whipped up a REGEX that should work with VS Code's problemmatcher regex101.com/r/q6duAL/4 and in any other tool that uses ECMAScript (Javascript) flavour for regex
The indent of the filename and line and column specifier is dependent on the line number, you have to match for multiple spaces there (https://regex101.com/r/9xImjz/2).
If I have understood correctly, this will eventually be present in/part of fpm. Let me know when that happens to add the problemmatcher in the fortran-lang vscode extension.
I think this might be a good opportunity to try the new GitHub projects that span the entire organisation as opposed to the individual repos.
Seems to be not an issue which is worth addressing at the moment.
|
2025-04-01T06:40:40.084471
| 2023-12-06T16:08:50
|
2028903537
|
{
"authors": [
"akashKarmakar02",
"tommyettinger"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11410",
"repo": "tommyettinger/gdx-liftoff",
"url": "https://github.com/tommyettinger/gdx-liftoff/issues/145"
}
|
gharchive/issue
|
Issue in project opening
Class org.jetbrains.kotlin.idea.gradleTooling.KotlinGradleModel must not be requested from main classloader of org.jetbrains.kotlin plugin [Plugin: org.jetbrains.android]
when i open a new project desktop only with java version 21 and gradle version 8.5
It is fixed it is happend due to old kotlin plugin
Interesting, I'll try to update it in Liftoff if possible. I had an issue when launching Gradle with Java 21 and trying to build an Android APK from IDEA, but it wasn't the same bug you reported. The bug I hit was resolved by changing the setting in IDEA to run Gradle with Java 17. In case anyone finds this by searching, the error was:
> Task :android:compileReleaseJavaWithJavac FAILED
Execution failed for task ':android:compileReleaseJavaWithJavac'.
> Could not resolve all files for configuration ':android:androidJdkImage'.
> Failed to transform core-for-system-modules.jar to match attributes {artifactType=_internal_android_jdk_image, org.gradle.libraryelements=jar, org.gradle.usage=java-runtime}.
> Execution failed for JdkImageTransform: C:\Users\tommy\AppData\Local\Android\Sdk\platforms\android-34\core-for-system-modules.jar.
> Error while executing process C:\d\jvm\jdk21-hotspot\bin\jlink.exe with arguments {--module-path C:\Users\tommy\.gradle\caches\transforms-3\553ad13d8645437e7a391ae3ab1b8dbb\transformed\output\temp\jmod --add-modules java.base --output C:\Users\tommy\.gradle\caches\transforms-3\553ad13d8645437e7a391ae3ab1b8dbb\transformed\output\jdkImage --disable-plugin system-modules}
I've never seen this type of bug before, and I didn't know Android even used jlink...
Try ./gradlew clean
After that try building again It sometimes fixed the problem
|
2025-04-01T06:40:40.098982
| 2020-03-06T20:01:22
|
577146503
|
{
"authors": [
"ArnWac",
"marcoaaguiar",
"thismatters",
"tomturner"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11411",
"repo": "tomturner/django-tenants",
"url": "https://github.com/tomturner/django-tenants/issues/349"
}
|
gharchive/issue
|
admin displaying apps inconsistently
I'm seeing some unexpected results in the django admin. Specifically I'm seeing what are strictly tenant apps showing up in the global schema, and I'm seeing what is strictly a shared app not showing up in the global schema admin ("Not available for global schema").
I think this might be because I am using the verbose notation in my SHARED_APPS and TENANT_APPS, e.g. 'polls.apps.PollsConfig', as proposed in the django tutorial. Whenever I switch to the short notation (e.g. 'polls') then the tenant apps only show up in the tenant admin. This is a pretty acceptable workaround, but I wanted you to know that there is some dissonance there.
The primary issue I'm facing is that there doesn't seem to be a way to obscure a shared app from tenants. Specifically, I would like to not show my tenant model to all of my tenants. Have I overlooked a method for making a table in the shared schema private (or for global admin purposes only)?
We also had inconsistencies in the django admin in our project.
I haven't had much time to dig into it to try to find what is happening.
In particular admin.site.register/unregister have been erratic.
The primary issue I'm facing is that there doesn't seem to be a way to obscure a shared app from tenants. Specifically, I would like to not show my tenant model to all of my tenants.
I worked around by implementing a seperate admin site for 'public'
If found out that the issue that we were having was due to having the INSTALLED_APPS like in this line:
https://github.com/tomturner/django-tenants/blob/5128d9e9d2409f3cd0b09c0bd574170ff657725d/examples/tenant_tutorial/tenant_tutorial/settings.py#L175
which was causing to the order of loaded apps to be randomize (set in python does not keep order).
We had an package that was registering in its admin.py, and we were unregistering in one of our apps. If the our app loaded first it would be registered, if the package loaded first, it would be unregistered.
I hope this helps.
I had the same issue use INSTALLED_APPS = list(SHARED_APPS) + [app for app in TENANT_APPS if app not in SHARED_APPS]
In our project, we are using django-configurations which have configurations classes instead a configuration file (one for production, one for development, etc).
Because the scope for list comprehension behave weirdly in a class body the code snippet that you posted won't work. The workaround was to use
INSTALLED_APPS = list(set(SHARED_APPS) - set(TENANT_APPS))+ TENANT_APPS # keep the order of the TENANT_APPS
Since the order of TENANT_APPS that mattered. I failed to find a more elegant solution, but it works
|
2025-04-01T06:40:40.106449
| 2015-09-03T07:28:14
|
104637321
|
{
"authors": [
"saramroo",
"tomvanzummeren"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11412",
"repo": "tomvanzummeren/TZStackView",
"url": "https://github.com/tomvanzummeren/TZStackView/issues/32"
}
|
gharchive/issue
|
Inconsistent podspec version in tags
Podspec version in 1.0.5-swift2 and 1.0.5 tags is still using 1.0.4.
I restructured the whole repository, deleted the separate swift-2 branch, made master compatible with Swift 2.0, so this is fixed now!
|
2025-04-01T06:40:40.114895
| 2021-11-12T19:25:42
|
1052346622
|
{
"authors": [
"fredski-github",
"toniblyx"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11413",
"repo": "toniblyx/prowler",
"url": "https://github.com/toniblyx/prowler/issues/925"
}
|
gharchive/issue
|
sensible versus sensitive in risk messages
need to replace sensible for sensitive in risk associated with finding message.
Publicly accessible load balancers could expose sensible data to bad actors.
Vaults accessible to everyone could expose sensible data to bad actors
Thanks @fredski-github, just fixed it in branch 2.6.1 and will be in master as soon as I fix the other bug
|
2025-04-01T06:40:40.162034
| 2022-04-03T02:28:38
|
1190795746
|
{
"authors": [
"tonyketcham"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11414",
"repo": "tonyketcham/p5-svelte",
"url": "https://github.com/tonyketcham/p5-svelte/issues/125"
}
|
gharchive/issue
|
Add docs example with p5.SVG
Pen plotters + p5 + Svelte would be a great example that some may find inspiring. That was one of the founding pieces of inspiration for starting this project and I was disappointed that p5.SVG didn't support ESM/instance mode so I couldn't get an SVG runtime for p5 working w/ p5-svelte.
In a recent release, p5.SVG added support for those pieces 🥳
Turns out that the instance mode does the thing where it's trying to mod p5 via accessing the window which Svelte no likey. This may not work, or may need a custom shim within p5-svelte in order to support
|
2025-04-01T06:40:40.168302
| 2020-10-16T14:04:00
|
723256846
|
{
"authors": [
"kaizoku-oh",
"tonyp7"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11415",
"repo": "tonyp7/esp32-wifi-manager",
"url": "https://github.com/tonyp7/esp32-wifi-manager/issues/112"
}
|
gharchive/issue
|
Can't build example using esp-idf v4.1
Prerequisites
[x] I have written a descriptive issue title
[x] I have verified that I am using the latest version of esp32-wifi-manager
[x] I have searched open and closed issues to ensure it has not already been reported
Description
I'm facing the same issue similar #103 when working with ESP-IDF 4.1 on platformio.
error: unknown type name 'esp_event_handler_instance_t'; did you mean 'esp_event_handler_t'?
Is there anyway to get around this build error while still using the 4.1 version?
Steps to Reproduce
install platformio extention on vscode
create a new project for esp32 using the esp-idf framework
create a components folder under the root of the project and copy the wifi-manager into it (tried master and other tags)
build
System Configuration
Debian 10 x64
vscode
platformio
Hello,
This is normal, this code only compile with 4.2 and up as per the requirements:
https://github.com/tonyp7/esp32-wifi-manager#requirements
Are you considering to support a stable release in the future? because v4.2 is only pre-release beta now contrary to v4.1 which will be supported until February 2023.
Definitely not. 4.2 should hit RC stage fairly soon and the master tree is already in 4.3. There’s no point holding back the codebase at this stage.
|
2025-04-01T06:40:40.172611
| 2019-08-26T20:53:09
|
485446518
|
{
"authors": [
"rmoritz",
"tonysavon"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11416",
"repo": "tonysavon/Chopper-Command-C64",
"url": "https://github.com/tonysavon/Chopper-Command-C64/issues/1"
}
|
gharchive/issue
|
How to build?
Which assembler must be used to build the sources?
Hi
Please use Kickassembler
http://theweb.dk/KickAssembler/Main.html#frontpage
Just assemble game.asm and that should give you game.prg, which you can run
straight away.
It can also be crunched with exomizer after that, which is what the
included ChopperCommand.prg file is
On Mon, 26 Aug 2019, 22:53 Ralph Möritz<EMAIL_ADDRESS>wrote:
Which assembler must be used to build the sources?
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub
https://github.com/tonysavon/Chopper-Command-C64/issues/1?email_source=notifications&email_token=AMSYTCQYFLCIAY4IYXOW2T3QGQ7DNA5CNFSM4IPVWFEKYY3PNVWWK3TUL52HS4DFUVEXG43VMWVGG33NNVSW45C7NFSM4HHPKN3A,
or mute the thread
https://github.com/notifications/unsubscribe-auth/AMSYTCUNNKSN36Z6QA6V7LLQGQ7DNANCNFSM4IPVWFEA
.
Thank you!
|
2025-04-01T06:40:40.193224
| 2024-11-12T06:24:41
|
2651169326
|
{
"authors": [
"Roberto-Gentili",
"lukehutch",
"michael-simons"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11417",
"repo": "toolfactory/narcissus",
"url": "https://github.com/toolfactory/narcissus/issues/17"
}
|
gharchive/issue
|
MacOS build broken due to encapsulation
@Roberto-Gentili Any idea why this build is failing on macOS?:
Error: ROR] Failed to execute goal org.sonatype.plugins:nexus-staging-maven-plugin:1.6.8:deploy (injected-nexus-deploy) on project narcissus: Execution injected-nexus-deploy of goal org.sonatype.plugins:nexus-staging-maven-plugin:1.6.8:deploy failed: An API incompatibility was encountered while executing org.sonatype.plugins:nexus-staging-maven-plugin:1.6.8:deploy: java.lang.ExceptionInInitializerError: null
Error: ROR] -----------------------------------------------------
Error: ROR] realm = extension>org.sonatype.plugins:nexus-staging-maven-plugin:1.6.8
Error: ROR] strategy = org.codehaus.plexus.classworlds.strategy.SelfFirstStrategy
[...]
Error: ROR] -----------------------------------------------------: Unable to make field private final java.util.Comparator java.util.TreeMap.comparator accessible: module java.base does not "opens java.util" to unnamed module @3681037
Building with JDK 7 fails with the error below: there appears to be an incompatible Maven plugin with this JDK. ave you updated any Maven plugins?
constituent[11]: file:/C:/Users/runneradmin/.m2/wrapper/dists/apache-maven-3.8.8-bin/67c30f74/apache-maven-3.8.8/lib/maven-compat-3.8.8.jar
at org.apache.maven.wrapper.BootstrapMainStarter.start(BootstrapMainStarter.java:53)
at org.apache.maven.wrapper.WrapperExecutor.execute(WrapperExecutor.java:152)
at org.apache.maven.wrapper.MavenWrapperMain.main(MavenWrapperMain.java:76)
Caused by: java.lang.UnsupportedClassVersionError: com/fasterxml/jackson/databind/ObjectMapper: Unsupported major.minor version 52.0
@Roberto-Gentili Any idea why this build is failing on macOS?:
Error: ROR] Failed to execute goal org.sonatype.plugins:nexus-staging-maven-plugin:1.6.8:deploy (injected-nexus-deploy) on project narcissus: Execution injected-nexus-deploy of goal org.sonatype.plugins:nexus-staging-maven-plugin:1.6.8:deploy failed: An API incompatibility was encountered while executing org.sonatype.plugins:nexus-staging-maven-plugin:1.6.8:deploy: java.lang.ExceptionInInitializerError: null
Error: ROR] -----------------------------------------------------
Error: ROR] realm = extension>org.sonatype.plugins:nexus-staging-maven-plugin:1.6.8
Error: ROR] strategy = org.codehaus.plexus.classworlds.strategy.SelfFirstStrategy
[...]
Error: ROR] -----------------------------------------------------: Unable to make field private final java.util.Comparator java.util.TreeMap.comparator accessible: module java.base does not "opens java.util" to unnamed module @3681037
It was already broken, so I tried updating all the Maven plugins to see if that would fix it...
It looks like the Maven wrapper version needs to be updated, or something?
It was already broken, so I tried updating all the Maven plugins to see if that would fix it...
It looks like the Maven wrapper version needs to be updated, or something?
Unfortunately I don't know: I abandoned the build test for JDK 7 due to too many problems even if I do the release build with the version parameter set for compatibility with JDK 7
It was already broken, so I tried updating all the Maven plugins to see if that would fix it...
It looks like the Maven wrapper version needs to be updated, or something?
Unfortunately I don't know: I abandoned the build test for JDK 7 due to too many problems even if I do the release build with the version parameter set for compatibility with JDK 7
@Roberto-Gentili @michael-simons OK, the build seems to be fixed. Thanks for your help! Pushing out 1.0.10.
It's interesting that this failed on JDK 7 :) I explicitly set the Maven Wrapper and Maven versions to work with 7. Seems to be an issue with nexus-staging-maven-plugin.
FWIW There's a way to define tool chains in Maven, so that you don't depend on the JDK that is being used to run Maven, but I never used them.
https://maven.apache.org/guides/mini/guide-using-toolchains.html
@Roberto-Gentili @michael-simons OK, the build seems to be fixed. Thanks for your help! Pushing out 1.0.10.
Great! I will release the new version of JVM-Driver as soon as it is available
Maven now requires a minimum JDK of 7, it seems (or maybe it is Maven-wrapper). I just removed the testing for JDK 7, since I don't have time to mess with this much more!
|
2025-04-01T06:40:40.204574
| 2021-02-26T13:45:55
|
817369501
|
{
"authors": [
"Meemaw",
"coveralls"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11418",
"repo": "toomuchdesign/next-page-tester",
"url": "https://github.com/toomuchdesign/next-page-tester/pull/173"
}
|
gharchive/pull-request
|
Fix client singleton router ii
What kind of change does this PR introduce?
Replaces: https://github.com/toomuchdesign/next-page-tester/pull/161
Fixes: https://github.com/toomuchdesign/next-page-tester/issues/160
What is the current behaviour?
You can also link to an open issue here.
What is the new behaviour?
...
Does this PR introduce a breaking change?
What changes might users need to make in their application due to this PR?
Other information:
Please check if the PR fulfills these requirements:
[x] Tests for the changes have been added
[x] Docs have been added / updated
Coverage remained the same at 100.0% when pulling 187c984537ad68b359ee1f4cad8c5021f0c4e6b9 on fix-client-singleton-router-II into dad5cbb7ca4175290ad32820290575f3318fc263 on master.
|
2025-04-01T06:40:40.354541
| 2016-11-27T19:01:21
|
191878100
|
{
"authors": [
"dbashford",
"dustinfarris",
"justinpark",
"toranb"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11419",
"repo": "toranb/ember-redux",
"url": "https://github.com/toranb/ember-redux/pull/51"
}
|
gharchive/pull-request
|
[PERFORMANCE]: avoid multiple redux.getState() calls in connect
I've simplified the callback workload to reduce the number of times we invoke redux.getState() in handleChange from 2x to 1x and also I've eliminated an unnecessary function (used previously in the shallowEqual days of connect). My hope is that we can achieve the same goal #41 set out to solve but without the tradeoff of holding state local in the component (and further -helping us avoid Ember.set / setProperties).
After the over notify problem was solved, the remaining "Problem Statement" back in #30 was that we call redux.getState() multiple times during re-render work and that could be expensive. This PR should solve a good chunk of that and I'm asking @dbashford to confirm we didn't break his app. In addition, I'm reaching out to @justinpark to confirm he won't run into a performance issue of any kind taking this new connect implementation.
Would you both have time this week to try out this updated connect function and report back?
@dustinfarris awesome! I'm hopeful it helps move us closer to the silver bullet while avoiding whatever it was that @dbashford ran into w/ the setProperties variation from last week. I'll wait until I hear from both of them that this is regression free/ solid before we get a v1.9.2 out the door
the remaining "Problem Statement" back in #30 was that we call redux.getState() multiple times during re-render work and that could be expensive.
I wonder how expensive redux.getState() really is? I would guess under the hood the browser is just kicking around a pointer. I'd be more interested in knowing how expensive the stateToProps function is for a non-trivial component.
in the computed prop:
defineProperty(component, name, computed(function() {
return finalMapStateToComputed(redux.getState())[name];
I'd bet the finalMapStateToComputed which gets run on every single property is much more expensive than the redux.getState() part.
Anyway, probably discussion best saved for another issue/PR.
Looking good here!
All good here too!
|
2025-04-01T06:40:40.365640
| 2023-03-21T17:59:02
|
1634450589
|
{
"authors": [
"sansoy",
"torayeff"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11420",
"repo": "torayeff/fanucpy",
"url": "https://github.com/torayeff/fanucpy/issues/18"
}
|
gharchive/issue
|
INTP-311 (MAPPDK_SERVER,204) Uninitialized data is used
I got fanucpy working well on my first CRX but on my second CRX I get this error message. I've searched the threads and this message pops up for a variety of reasons with other socket based programs. Was wondering if you had any troubleshooting tips.
Hi. Have you copied the latest fanuc driver files from https://github.com/torayeff/fanucpy/tree/main/src/fanuc-driver ?
Basically, you need to install the latest version of fanucpy (0.1.11) with the latest version of fanuc driver files.
i did this morning hoping to resolve the issue. Some threads suggest it
might be a memory issue.
Will try again.
Sabri
BTW great job with integrating ChatGPT!
On Tue, Mar 21, 2023 at 12:06 PM Agajan Torayev @.***>
wrote:
Hi. Have you copied the latest fanuc driver files from
https://github.com/torayeff/fanucpy/tree/main/src/fanuc-driver ?
Basically, you need to install the latest version of fanucpy (0.1.11) with
the latest version of fanuc driver files.
—
Reply to this email directly, view it on GitHub
https://github.com/torayeff/fanucpy/issues/18#issuecomment-1478364315,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AACV24LVMTW2QB4HU5K7IG3W5HUZXANCNFSM6AAAAAAWCZM5QM
.
You are receiving this because you authored the thread.Message ID:
@.***>
@sansoy Let me know how it goes. I will have a closer look at the issue on Thursday. Some far there were not any problems while testing using Roboguide.
I can confirm that v0.1.11 works with the latest .pc and .ls files. Please copy the new files from https://github.com/torayeff/fanucpy/tree/main/src/fanuc-driver
Thanks Agajan! I was able to get it to run my CRX!
Great work!
Sabri
On Thu, Apr 6, 2023 at 7:08 AM Agajan Torayev @.***>
wrote:
Closed #18 https://github.com/torayeff/fanucpy/issues/18 as completed.
—
Reply to this email directly, view it on GitHub
https://github.com/torayeff/fanucpy/issues/18#event-8945813893, or
unsubscribe
https://github.com/notifications/unsubscribe-auth/AACV24KURBH5IQNMKUKDQ5DW7253LANCNFSM6AAAAAAWCZM5QM
.
You are receiving this because you were mentioned.Message ID:
@.***>
|
2025-04-01T06:40:40.367833
| 2016-09-27T05:58:11
|
179403779
|
{
"authors": [
"ArturoDeza",
"ilija139",
"karandwivedi42"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11421",
"repo": "torch/image",
"url": "https://github.com/torch/image/issues/193"
}
|
gharchive/issue
|
lena loading grace bug
When using image.display(image.lena()), it shows grace hopper. As a suggestion, possible creating a new image.grace() function would be better.
I was also curious about this. Is it a joke? :smile:
You can see https://github.com/torch/nn/issues/854.
"we've removed lena. we cant change the name because of breaking API"
|
2025-04-01T06:40:40.377801
| 2024-10-16T06:45:05
|
2590820195
|
{
"authors": [
"songproducer",
"zerolab"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11422",
"repo": "torchbox/wagtail-news-template",
"url": "https://github.com/torchbox/wagtail-news-template/issues/14"
}
|
gharchive/issue
|
Navigation links not working on mobile
https://github.com/torchbox/wagtail-news-template/blob/main/templates/navigation/header.html#L124
a href was blank
should be:
href="{{ link.value.get_url }}"
@songproducer thank you for filing these issues. It is very useful. Feel free to submit pull requests with fixes to any issues you find and have time for
@zerolab Sorry I'm git illiterate, I gave it a shot but not sure if this is right.
https://github.com/songproducer/wagtail-news-template/pull/2
Hey @songproducer, no worries at all.
Perhaps https://docs.github.com/en/pull-requests/collaborating-with-pull-requests/proposing-changes-to-your-work-with-pull-requests/creating-a-pull-request-from-a-fork can help
Thanks! I was trying to make my life more difficult by trying to use command line. Pull request submitted!
|
2025-04-01T06:40:40.451666
| 2023-09-15T18:45:36
|
1898914300
|
{
"authors": [
"LegNeato",
"hawkw"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11423",
"repo": "tosc-rs/mnemos",
"url": "https://github.com/tosc-rs/mnemos/pull/275"
}
|
gharchive/pull-request
|
Add that mn requires bindeps
Without this installing errors with:
error: failed to parse manifest at `mnemos/tools/manganese/Cargo.toml`
Caused by:
`artifact = …` requires `-Z bindeps` (cargo-binutils)
Ah, I had not really considered the use case of installing mn using cargo install. Instead, I had assumed that it would always be run using the cargo mn alias defined in the workspace's .cargo/config.toml, which will always work fine because that file also explicitly enables the bindeps unstable feature.
I'm not sure whether or not we should include this in the README. On one hand, it does tell users how to avoid problems when installing mn using cargo install. On the other hand, I'm not actually sure if installing mn that way is a good idea. The advantage of only ever running it using cargo run is that it is potentially rebuilt whenever the Cargo.lock entries for its dependencies changes. This means that if you run mn using the alias that calls cargo run, you get whatever versions of the bindeps are currently in the lockfile. If those haven't changed, mn and its dependencies are not recompiled every time its run, but if the lockfile has changed, the cargo run command will build new versions of the bindeps. This ensures that everyone gets compatible versions of those dependencies whenever they use cargo mn.
On the other hand, if mn is cargo installed, it's built once, and every subsequent time you run cargo mn, you get the same thing, regardless of the mnemos repo's Cargo.lock. This is fine with regards to the actual mn binary, since it's very simple and won't change often. However, it also means that the versions of the bindeps that are downloaded and compiled are the versions you get when mn is cargo installed, and if the lockfile updates those dependencies, you still get the old ones. This means that if we later update the lockfile to depend on new versions of those tools, any contributors who have cargo installed mn will see their versions of the build dependencies begin to drift from what other contributors are using, potentially introducing compatibility issues with the various scripts run using mn.
Because of the version drift issue, I think it might be better to just have the README suggest that mn should not be cargo installed, and that it should instead always be used through cargo run --package manganese , or through the cargo mn alias defined in the workspace, which expands to that. What do you think?
Ok. I didn't actually know how to run it and assumed it was a normal cargo tool. I didn't know you could define aliases in the workspace, and I work with a lot of different rust projects so this probably will be a tripping point for folks. Thanks for the quick response!
|
2025-04-01T06:40:40.464850
| 2016-02-18T21:24:04
|
134700964
|
{
"authors": [
"GreenGremlin"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11425",
"repo": "toumorokoshi/sprinter",
"url": "https://github.com/toumorokoshi/sprinter/pull/95"
}
|
gharchive/pull-request
|
Adding support for boolean input types
Includes custom prompts for y/n, yes/no, true/false, and t/f. With uppercased default values and optional in/out types.
In/out types are configured with a '/'. "In" type controls the prompting, "out" type controls the string conversion. This allows prompting for "y/n" and outputting as "true/false", for one example.
Maybe in the future, instead of output types we could have some sort of casting use. i.e. %((bool):config:my_var). I'm not sure what the best format would be, but this would lend for greater flexibility.
|
2025-04-01T06:40:40.468784
| 2018-02-01T06:23:08
|
293416889
|
{
"authors": [
"justlikethisdesign",
"tovic"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11427",
"repo": "tovic/rich-text-editor",
"url": "https://github.com/tovic/rich-text-editor/issues/8"
}
|
gharchive/issue
|
Paste on Mobile removes pasted content immediately
When tested on Apple and Android mobile devices (both on chrome so far) the paste function is not working. As far I can tell the function is being triggered (I extended the paste delay) but as soon as the delay is up it seems the content is not found in "view" and is removed.
https://github.com/tovic/rich-text-editor/commit/a330d78ed16cbfd8dded8a55aaf9d6ed2d39bde3
|
2025-04-01T06:40:40.479768
| 2017-11-09T18:21:04
|
272669826
|
{
"authors": [
"meganrm",
"nathanmwilliams",
"pdvorchik"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11428",
"repo": "townhallproject/townHallProject",
"url": "https://github.com/townhallproject/townHallProject/pull/216"
}
|
gharchive/pull-request
|
Fix/style changes to cards
Changes to cards
Added the last updated in a footer, which is nice to see for events that are coming up soon, but my question on this is we have some events that haven't been updated in awhile, does it make things look stale?
Formatted the address so that it breaks at the right point
Changed the box around the address to a line
moved links to the lower right hand corner.
Added the last updated in a footer, which is nice to see for events that are coming up soon, but my question on this is we have some events that haven't been updated in awhile, does it make things look stale?
I assume these are ongoing events (regular office hours)? Maybe we lose the "last updated" for those types of events?
Other changes look solid. A couple questions:
-Do we want the Notes to be above the address? In the examples you used it works but occasionally the Notes can be very long (such as a long list of additional speakers or co-host organizations).
-The bar color is still connected to Party and the little icon/color is connected to the event type on our map? I'm wondering if there is a cleaner way to convey these (rather than blue on blue, with each blue signifying something different).
yeah we can switch notes and address.
The headers are only connected to the type of event. Blue is for in person, purple for staff etc
So the entire header color (including icon) are representing the same thing with two different color shades? We may have already discussed this but I wonder if it's a bit confusing (assuming that one is representing party).
yeah that's right. There's nothing that indicates party by color anymore.
Love the links on the bottom right!
May I make a suggestion about the header? Right now the MOC's name seems understated and the event type icon isn't very grokable. What if we made them columns instead? Something like:
yeah! do you have this styled already or did you do in with screen grabs? if you have the style can you just push it to this branch?
Screen grabs I'm afraid
|
2025-04-01T06:40:40.490466
| 2018-07-14T09:26:48
|
341223329
|
{
"authors": [
"johnmarinelli",
"toxicFork"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11429",
"repo": "toxicFork/react-three-renderer-fiber",
"url": "https://github.com/toxicFork/react-three-renderer-fiber/pull/55"
}
|
gharchive/pull-request
|
Add step to readme
Hey! It's me again. So I had to setup this project on a new machine and ran into problems when I followed the README steps. Specifically, the
<script src="node_modules/react/umd/react.development.js"></script>
<script src="node_modules/react-dom/umd/react-dom.development.js"></script>
lines in examples/index.html were breaking because it couldn't find the files. This is because yarn doesn't install the dependencies for examples when you run yarn install from the root. It's a small detail but cost me ~45 minutes of trying to find a bug in the webpack config or fiddling with the path in the script src.
Hi John, thank you for the PR, it is much appreciated!
I have been meaning to get back to this project and see it to completion, your interest and help is a great motivator.
Apologies for the 45 minutes of lost time.
|
2025-04-01T06:40:40.520317
| 2019-12-10T08:29:31
|
535571658
|
{
"authors": [
"JoshuaDRose",
"tim77"
],
"license": "Vim",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11430",
"repo": "tpope/vim-pathogen",
"url": "https://github.com/tpope/vim-pathogen/issues/209"
}
|
gharchive/issue
|
License file
Would be nice to provide license file. This helps with distro adoption. Simple license file would be enough.
true
|
2025-04-01T06:40:40.521808
| 2021-03-24T22:46:09
|
840266952
|
{
"authors": [
"zachrbrown"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11431",
"repo": "tqtezos/minter-sdk",
"url": "https://github.com/tqtezos/minter-sdk/issues/70"
}
|
gharchive/issue
|
Use canonical Michelson parser implementation to generate JSON representations
#58 added a script which generates a JSON representation of a Michelson contract to facilitate idiomatic importing of contracts into a JavaScript project. This solution outsources Michelson parsing to the @taquito/michel-codec package. Unfortunately, this package often goes out of date with respect to the protocol, encountering parse errors on valid contracts. To mitigate this, we should use a canonical implementation like the tezos-client / RPC.
Closed via #75
|
2025-04-01T06:40:40.525480
| 2022-01-25T11:21:55
|
1113764993
|
{
"authors": [
"devudr",
"tr3v3r"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11432",
"repo": "tr3v3r/react-native-esc-pos-printer",
"url": "https://github.com/tr3v3r/react-native-esc-pos-printer/issues/45"
}
|
gharchive/issue
|
getting error while installing
error logs
/Users/../Documents/work/ReactNative/work/../node_modules/react-native-esc-pos-printer: Command failed.
Exit code: 1
Command: node ./scripts/movelib.js
Arguments:
Directory: /Users/../Documents/work/ReactNative/work/../node_modules/react-native-esc-pos-printer
Output:
internal/modules/cjs/loader.js:905
throw err;
^
Error: Cannot find module '/Users/../Documents/work/ReactNative/work/../node_modules/react-native-esc-pos-printer/scripts/movelib.js'
at Function.Module._resolveFilename (internal/modules/cjs/loader.js:902:15)
at Function.Module._load (internal/modules/cjs/loader.js:746:27)
at Function.executeUserEntryPoint [as runMain] (internal/modules/run_main.js:76:12)
at internal/main/run_main_module.js:17:47 {
Fixed in v1.7.1
hi @tr3v3r what is the major difference in v1.5.0 between v1.7.1
@devudr the only difference is that SDK for iOS is moved to a separate repo.
|
2025-04-01T06:40:40.536213
| 2023-06-05T06:35:11
|
1741080150
|
{
"authors": [
"scala-steward"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11433",
"repo": "trace4cats/trace4cats-dynamic-extras",
"url": "https://github.com/trace4cats/trace4cats-dynamic-extras/pull/160"
}
|
gharchive/pull-request
|
Update sbt-github-actions to 0.15.0
About this PR
📦 Updates com.codecommit:sbt-github-actions from 0.14.2 to 0.15.0
Usage
✅ Please merge!
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Configure Scala Steward for your repository with a .scala-steward.conf file.
Have a fantastic day writing Scala!
🔍 Files still referring to the old version number
The following files still refer to the old version number (0.14.2).
You might want to review and update them manually.
project/Dependencies.scala
⚙ Adjust future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "com.codecommit", artifactId = "sbt-github-actions" } ]
Or, add this to slow down future updates of this dependency:
dependencyOverrides = [{
pullRequests = { frequency = "30 days" },
dependency = { groupId = "com.codecommit", artifactId = "sbt-github-actions" }
}]
labels: sbt-plugin-update, early-semver-major, semver-spec-minor, old-version-remains, commit-count:n:2
Superseded by #195.
|
2025-04-01T06:40:40.541104
| 2023-11-15T12:09:51
|
1994656026
|
{
"authors": [
"alexdesi"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11434",
"repo": "trade-tariff/trade-tariff-frontend",
"url": "https://github.com/trade-tariff/trade-tariff-frontend/pull/1719"
}
|
gharchive/pull-request
|
Update bundler
What?
Update bundler to remove warning "`DidYouMean::SPELL_CHECKERS.merge!(error_name => spell_checker)' has been deprecated.", which is present many times when running spec tests.
Why?
To stop using the deprecated method.
|
2025-04-01T06:40:40.546866
| 2019-03-01T18:20:44
|
416223254
|
{
"authors": [
"chadicus",
"coveralls",
"jncarver"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11435",
"repo": "traderinteractive/filter-php",
"url": "https://github.com/traderinteractive/filter-php/pull/84"
}
|
gharchive/pull-request
|
Add Filterer responseType option
What does this PR do?
This pull request adds a new FilterResponse class and adds a responseType option to the Filterer::filter function to allow an instance of this new class to be returned.
The default return is still an array, which means it is not backwards-breaking. Additionally, the FilterResponse object can be converted to the traditional array result with its toArray() function.
Example use:
$response = Filterer::filter($spec, $value, ['responseType' => Filterer::RESPONSE_TYPE_FILTER]);
if ($response->success === false) {
throw new FilterException($response->errorMessage);
}
return $response->filteredValue;
Checklist
[X] Pull request contains a clear definition of changes
[X] Tests (either unit, integration, or acceptance) written and passing
[X] Relevant documentation produced and/or updated
Coverage decreased (-0.4%) to 97.26% when pulling 859e1caecc2cbd91906dc80584b516dfe77b30ae on jncarver:filter-response into cc04632ec167679e7b0309b50242dab53ab9879a on traderinteractive:master.
@jncarver What advantage is gained by adding the response object?
@chadicus I find it easier to use an object with explicit properties than an array of results that are in a specific order. It's also more compact, which is helpful with line length restrictions.
The biggest difference, however, is that the original array of errors are placed into the response object, so you can iterate through them if you need to. In the traditional array response, they are combined into a newline-delimited string which can be a pain if you're trying to do anything with the errors, like provide an error count.
|
2025-04-01T06:40:40.556720
| 2021-12-06T16:47:35
|
1072366978
|
{
"authors": [
"seedy",
"traefiker"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11436",
"repo": "traefik/faency",
"url": "https://github.com/traefik/faency/pull/221"
}
|
gharchive/pull-request
|
feat: badge cleanup
Description
Cleaning up badge component, NB: this applies no design system changes.
This will be done, hopefully in a future step.
added missing compoundVariants programatically
added all color variants in story programatically
default color variant (gray) is duplicated to make sure it's clear to the user
cleaned up theme file
Package changes
updated rollup-plugin-typescript2 from ^0.29.0 to ^0.30.0 due to a build issue
CI: https://github.com/traefik/faency/runs/4433331267?check_suite_focus=true
https://github.com/alexjoverm/typescript-library-starter/issues/310#issuecomment-811451090
:tada: This PR is included in version 1.0.0 :tada:
The release is available on:
npm package (@latest dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
|
2025-04-01T06:40:40.604031
| 2023-04-18T00:15:33
|
1672147809
|
{
"authors": [
"1-Samuel",
"mpl"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11437",
"repo": "traefik/traefik",
"url": "https://github.com/traefik/traefik/issues/9848"
}
|
gharchive/issue
|
wrong api response for healthcheck interval and timeout values
Welcome!
[X] Yes, I've searched similar issues on GitHub and didn't find any.
[X] Yes, I've searched similar issues on the Traefik community forum and didn't find any.
What did you do?
I configured healthcheck interval or timeout to "wrong" values, e.g. -10s for interval
What did you see instead?
In the api response (and webui) the values for interval and timeout are the same as in the config (in this case -10s), but in reality, they are some sort of fallback values (in this case 30s for interval)
What version of Traefik are you using?
Version: 2.9.10
Codename: banon
Go version: go1.20.3
Built: 2023-04-06T16:15:08Z
OS/Arch: linux/amd64
What is your environment & configuration?
services:
traefik:
image: traefik
command:
- --api.insecure=true
- --providers.docker=true
- --log.level=DEBUG
ports:
- "8080:8080"
volumes:
- /run/user/1000/docker.sock:/var/run/docker.sock
whoami:
image: traefik/whoami
labels:
- "traefik.http.services.myservice.loadbalancer.healthcheck.path=/health"
- "traefik.http.services.myservice.loadbalancer.healthcheck.interval=-10s"
If applicable, please paste the log output in DEBUG level
time="2023-04-18T00:10:28Z" level=info msg="Configuration loaded from flags."
time="2023-04-18T00:10:28Z" level=info msg="Traefik version 2.9.10 built on 2023-04-06T16:15:08Z"
time="2023-04-18T00:10:28Z" level=debug msg="Static configuration loaded {\"global\":{\"checkNewVersion\":true},\"serversTransport\":{\"maxIdleConnsPerHost\":200},\"entryPoints\":{\"http\":{\"address\":\":80\",\"transport\":{\"lifeCycle\":{\"graceTimeOut\":\"10s\"},\"respondingTimeouts\":{\"idleTimeout\":\"3m0s\"}},\"forwardedHeaders\":{},\"http\":{},\"http2\":{\"maxConcurrentStreams\":250},\"udp\":{\"timeout\":\"3s\"}},\"traefik\":{\"address\":\":8080\",\"transport\":{\"lifeCycle\":{\"graceTimeOut\":\"10s\"},\"respondingTimeouts\":{\"idleTimeout\":\"3m0s\"}},\"forwardedHeaders\":{},\"http\":{},\"http2\":{\"maxConcurrentStreams\":250},\"udp\":{\"timeout\":\"3s\"}}},\"providers\":{\"providersThrottleDuration\":\"2s\",\"docker\":{\"watch\":true,\"endpoint\":\"unix:///var/run/docker.sock\",\"defaultRule\":\"Host(`{{ normalize .Name }}`)\",\"exposedByDefault\":true,\"swarmModeRefreshSeconds\":\"15s\"}},\"api\":{\"insecure\":true,\"dashboard\":true},\"log\":{\"level\":\"DEBUG\",\"format\":\"common\"}}"
time="2023-04-18T00:10:28Z" level=info msg="\nStats collection is disabled.\nHelp us improve Traefik by turning this feature on :)\nMore details on: https://doc.traefik.io/traefik/contributing/data-collection/\n"
time="2023-04-18T00:10:28Z" level=info msg="Starting provider aggregator aggregator.ProviderAggregator"
time="2023-04-18T00:10:28Z" level=debug msg="Starting TCP Server" entryPointName=traefik
time="2023-04-18T00:10:28Z" level=debug msg="Starting TCP Server" entryPointName=http
time="2023-04-18T00:10:28Z" level=info msg="Starting provider *traefik.Provider"
time="2023-04-18T00:10:28Z" level=debug msg="*traefik.Provider provider configuration: {}"
time="2023-04-18T00:10:28Z" level=info msg="Starting provider *docker.Provider"
time="2023-04-18T00:10:28Z" level=debug msg="*docker.Provider provider configuration: {\"watch\":true,\"endpoint\":\"unix:///var/run/docker.sock\",\"defaultRule\":\"Host(`{{ normalize .Name }}`)\",\"exposedByDefault\":true,\"swarmModeRefreshSeconds\":\"15s\"}"
time="2023-04-18T00:10:28Z" level=info msg="Starting provider *acme.ChallengeTLSALPN"
time="2023-04-18T00:10:28Z" level=debug msg="*acme.ChallengeTLSALPN provider configuration: {}"
time="2023-04-18T00:10:28Z" level=debug msg="Configuration received: {\"http\":{\"routers\":{\"api\":{\"entryPoints\":[\"traefik\"],\"service\":\"api@internal\",\"rule\":\"PathPrefix(`/api`)\",\"priority\":2147483646},\"dashboard\":{\"entryPoints\":[\"traefik\"],\"middlewares\":[\"dashboard_redirect@internal\",\"dashboard_stripprefix@internal\"],\"service\":\"dashboard@internal\",\"rule\":\"PathPrefix(`/`)\",\"priority\":2147483645}},\"services\":{\"api\":{},\"dashboard\":{},\"noop\":{}},\"middlewares\":{\"dashboard_redirect\":{\"redirectRegex\":{\"regex\":\"^(http:\\\\/\\\\/(\\\\[[\\\\w:.]+\\\\]|[\\\\w\\\\._-]+)(:\\\\d+)?)\\\\/$\",\"replacement\":\"${1}/dashboard/\",\"permanent\":true}},\"dashboard_stripprefix\":{\"stripPrefix\":{\"prefixes\":[\"/dashboard/\",\"/dashboard\"]}}},\"serversTransports\":{\"default\":{\"maxIdleConnsPerHost\":200}}},\"tcp\":{},\"udp\":{},\"tls\":{}}" providerName=internal
time="2023-04-18T00:10:28Z" level=debug msg="Provider connection established with docker 23.0.3 (API 1.42)" providerName=docker
time="2023-04-18T00:10:28Z" level=debug msg="Configuration received: {\"http\":{\"routers\":{\"traefik-traefik\":{\"service\":\"traefik-traefik\",\"rule\":\"Host(`traefik-traefik`)\"},\"whoami-traefik\":{\"service\":\"myservice\",\"rule\":\"Host(`whoami-traefik`)\"}},\"services\":{\"myservice\":{\"loadBalancer\":{\"servers\":[{\"url\":\"http://<IP_ADDRESS>:80\"}],\"healthCheck\":{\"path\":\"/health\",\"interval\":\"-10s\",\"followRedirects\":true},\"passHostHeader\":true}},\"traefik-traefik\":{\"loadBalancer\":{\"servers\":[{\"url\":\"http://<IP_ADDRESS>:80\"}],\"passHostHeader\":true}}}},\"tcp\":{},\"udp\":{}}" providerName=docker
time="2023-04-18T00:10:28Z" level=debug msg="No default certificate, fallback to the internal generated certificate" tlsStoreName=default
time="2023-04-18T00:10:28Z" level=debug msg="Added outgoing tracing middleware dashboard@internal" middlewareType=TracingForwarder entryPointName=traefik routerName=dashboard@internal middlewareName=tracing
time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" entryPointName=traefik middlewareName=dashboard_stripprefix@internal middlewareType=StripPrefix routerName=dashboard@internal
time="2023-04-18T00:10:28Z" level=debug msg="Adding tracing to middleware" entryPointName=traefik middlewareName=dashboard_stripprefix@internal routerName=dashboard@internal
time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" entryPointName=traefik routerName=dashboard@internal middlewareName=dashboard_redirect@internal middlewareType=RedirectRegex
time="2023-04-18T00:10:28Z" level=debug msg="Setting up redirection from ^(http:\\/\\/(\\[[\\w:.]+\\]|[\\w\\._-]+)(:\\d+)?)\\/$ to ${1}/dashboard/" entryPointName=traefik routerName=dashboard@internal middlewareName=dashboard_redirect@internal middlewareType=RedirectRegex
time="2023-04-18T00:10:28Z" level=debug msg="Adding tracing to middleware" entryPointName=traefik middlewareName=dashboard_redirect@internal routerName=dashboard@internal
time="2023-04-18T00:10:28Z" level=debug msg="Added outgoing tracing middleware api@internal" entryPointName=traefik routerName=api@internal middlewareName=tracing middlewareType=TracingForwarder
time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" middlewareName=traefik-internal-recovery entryPointName=traefik middlewareType=Recovery
time="2023-04-18T00:10:28Z" level=debug msg="No entryPoint defined for this router, using the default one(s) instead: [http]" routerName=whoami-traefik
time="2023-04-18T00:10:28Z" level=debug msg="No entryPoint defined for this router, using the default one(s) instead: [http]" routerName=traefik-traefik
time="2023-04-18T00:10:28Z" level=debug msg="No default certificate, fallback to the internal generated certificate" tlsStoreName=default
time="2023-04-18T00:10:28Z" level=debug msg="Added outgoing tracing middleware api@internal" entryPointName=traefik routerName=api@internal middlewareName=tracing middlewareType=TracingForwarder
time="2023-04-18T00:10:28Z" level=debug msg="Added outgoing tracing middleware dashboard@internal" middlewareName=tracing middlewareType=TracingForwarder entryPointName=traefik routerName=dashboard@internal
time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" middlewareName=dashboard_stripprefix@internal middlewareType=StripPrefix entryPointName=traefik routerName=dashboard@internal
time="2023-04-18T00:10:28Z" level=debug msg="Adding tracing to middleware" entryPointName=traefik routerName=dashboard@internal middlewareName=dashboard_stripprefix@internal
time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" entryPointName=traefik routerName=dashboard@internal middlewareType=RedirectRegex middlewareName=dashboard_redirect@internal
time="2023-04-18T00:10:28Z" level=debug msg="Setting up redirection from ^(http:\\/\\/(\\[[\\w:.]+\\]|[\\w\\._-]+)(:\\d+)?)\\/$ to ${1}/dashboard/" entryPointName=traefik routerName=dashboard@internal middlewareType=RedirectRegex middlewareName=dashboard_redirect@internal
time="2023-04-18T00:10:28Z" level=debug msg="Adding tracing to middleware" entryPointName=traefik routerName=dashboard@internal middlewareName=dashboard_redirect@internal
time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" middlewareType=Recovery entryPointName=traefik middlewareName=traefik-internal-recovery
time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" entryPointName=http routerName=whoami-traefik@docker serviceName=myservice middlewareName=pipelining middlewareType=Pipelining
time="2023-04-18T00:10:28Z" level=debug msg="Creating load-balancer" entryPointName=http routerName=whoami-traefik@docker serviceName=myservice
time="2023-04-18T00:10:28Z" level=debug msg="Creating server 0 http://<IP_ADDRESS>:80" serverName=0 entryPointName=http routerName=whoami-traefik@docker serviceName=myservice
time="2023-04-18T00:10:28Z" level=debug msg="child http://<IP_ADDRESS>:80 now UP"
time="2023-04-18T00:10:28Z" level=debug msg="Propagating new UP status"
time="2023-04-18T00:10:28Z" level=debug msg="Added outgoing tracing middleware myservice" entryPointName=http middlewareName=tracing middlewareType=TracingForwarder routerName=whoami-traefik@docker
time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" entryPointName=http routerName=traefik-traefik@docker middlewareType=Pipelining middlewareName=pipelining serviceName=traefik-traefik
time="2023-04-18T00:10:28Z" level=debug msg="Creating load-balancer" routerName=traefik-traefik@docker serviceName=traefik-traefik entryPointName=http
time="2023-04-18T00:10:28Z" level=debug msg="Creating server 0 http://<IP_ADDRESS>:80" routerName=traefik-traefik@docker serverName=0 serviceName=traefik-traefik entryPointName=http
time="2023-04-18T00:10:28Z" level=debug msg="child http://<IP_ADDRESS>:80 now UP"
time="2023-04-18T00:10:28Z" level=debug msg="Propagating new UP status"
time="2023-04-18T00:10:28Z" level=debug msg="Added outgoing tracing middleware traefik-traefik" middlewareType=TracingForwarder entryPointName=http routerName=traefik-traefik@docker middlewareName=tracing
time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" middlewareName=traefik-internal-recovery middlewareType=Recovery entryPointName=http
time="2023-04-18T00:10:28Z" level=error msg="Health check interval smaller than zero for service 'myservice@docker'" serviceName=myservice@docker
time="2023-04-18T00:10:28Z" level=debug msg="Setting up healthcheck for service myservice@docker with [Hostname: Headers: map[] Path: /health Method: Port: 0 Interval: 30s Timeout: 5s FollowRedirects: true]" serviceName=myservice@docker
time="2023-04-18T00:10:28Z" level=debug msg="Initial health check for backend: \"myservice@docker\""
last three lines are relevant here
Hello @1-Samuel ,
We've looked into it, and we agree that the lack of consistency between the effective configuration of traefik and the "raw data" returned by the API is annoying and disturbing (although not very concerning in itself in our opinion).
So here is our plan:
for v3, we actually would like to take another route: in such a case of wonky input from the user (e.g. -10s) we're of the opinion that we should log a clear error about it, and simply fail. And not try to cover up the mistake with a "saner" default. We've actually already taken such steps for similar situations in v3, and we think we should keep that trend going.
for v2, we cannot do the same as above, because it would be a breaking change. We may want to go in the direction you suggested (i.e. fix the raw data so that it fits with the effective configuration), but we're not convinced yet that the exact fix you proposed is the way we want to go. And tbh, we're not even sure that changing anything at all is worth the effort for v2.
So we would like to think about it some more, and also gauge the interest of the community on that matter.
WDYT?
Thanks.
Hello,
thank you for looking into the issue. Sounds like a good plan to me. :+1:
I only stumbled upon the issue while looking into #6550 and I couldn't think of a good way to write a test for retry intervals, so I thought I give the bug fix a try, just to practice a bit. :) I don't think many people with a real use case will run into this issue anyway.
Thank you :)
Hello @mpl ,
yeah I will have a look at implementing the v3 changes (fail if implausible values are configured). But I think making a new PR would make more sense, since the changes would be different all together. I'll close it :)
|
2025-04-01T06:40:40.799491
| 2019-02-01T22:54:57
|
405905377
|
{
"authors": [
"artemdinaburg",
"jberryman",
"pgoodman"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11438",
"repo": "trailofbits/mcsema",
"url": "https://github.com/trailofbits/mcsema/issues/515"
}
|
gharchive/issue
|
What is the status of Binary Ninja support? DynInst support?
These are some questions and a documentation bug report (insofar as the README is a bit confusing or doesn't answer these questions).
The top of the README states:
Control flow recovery is performed using the mcsema-disass tool, which relies on IDA Pro, Binary Ninja, or DynInst to disassemble a binary file and produce a control flow graph.
and later...
Currently IDA and BinaryNinja are supported for control-flow recovery...
but not dyninst...? (I see there are several abandoned or pending PRs).
Then later...
Why do I need IDA Pro to use McSema ... we have a branch with experimental Binary Ninja support ...
It looks like the referenced branch is deleted (presumably merged?). I also see a ninja branch referenced in a PR, and another one hanging out. What is the status of Binary Ninja support?
I might shell out $150 for binary ninja if I was sure it would work (but maybe McSema requires the $600 version for its API?).
Thanks!
Also getting slightly off topic, I tried out the freeware version of IDA on the off chance it worked and no one has noticed. It doesn't and I don't get any helpful errors (is idal64 the headless version that McSema needs? if so I guess they don't ship that with the freeware version), but I wonder if the free version could be made to work in some other way? Or maybe y'all could try to convince them to release the functionality McSema needs in the free version (still for non-commercial use)? Seems mutually beneficial, but what do I know
@jberryman I have not tested IDA Pro Free Version, but if it includes IDAPython it should work. Use of idal64 is not required; you can use idaq64, but you have to do it from an X session since it has to start a GUI. I do this when debugging the CFG recovery scripts.
Thanks. Hm, doesn't seem to. The IDA free tree looks like:
/home/me/.local/idafree-7.0/
├── appico64.png
├── assistant
├── cfg
│ ├── ida.cfg
│ └── idagui.cfg
├── dbgsrv
├── ida64
├── ida64.int
├── ida.hlp
├── idc
│ └── idc.idc
├── ids
│ ├── idsnames
│ └── win7.zip
├── libclpx.so
├── libdwarfx.so
├── libida64.so
├── libQt5CLucene.so.5
├── libQt5Core.so.5
├── libQt5DBus.so.5
├── libQt5Gui.so.5
├── libQt5Help.so.5
├── libQt5Network.so.5
├── libQt5PrintSupport.so.5
├── libQt5Sql.so.5
├── libQt5Widgets.so.5
├── libQt5XcbQpa.so.5
├── libX11-xcb.so.1
├── license.txt
├── loaders
│ ├── elf64.so
│ ├── macho64.so
│ └── pe64.so
├── plugins
│ ├── bdescr64.so
│ ├── dbg64.so
│ ├── dwarf64.so
│ ├── objc64.so
│ ├── platforms
│ │ ├── libqlinuxfb.so
│ │ ├── libqminimal.so
│ │ ├── libqoffscreen.so
│ │ └── libqxcb.so
│ ├── platformthemes
│ │ └── libqgtk2.so
│ ├── plugins.cfg
│ └── tds64.so
├── procs
│ └── pc64.so
├── qidahelpcollection.qhc
├── qidahelp.qch
├── qt.conf
├── qwingraph
├── sig
│ ├── list
│ └── pc
│ ├── bcb5rt.sig
│ ├── iclapp64.sig
│ ├── ms64wdk.sig
│ ├── msmfc64d.sig
│ ├── msmfc64.sig
│ ├── msmfc64u.sig
│ ├── mssdk64.sig
│ ├── pe64.sig
│ ├── vc32rtf.sig
│ ├── vc64_14.sig
│ ├── vc64atl.sig
│ ├── vc64extra.sig
│ ├── vc64mfc.sig
│ ├── vc64rtf.sig
│ ├── vc64seh.sig
│ ├── vc64ucrt.sig
│ ├── vcseh.sig
│ └── win64unx.sig
├── til
│ ├── gnuunx64.til
│ ├── macosx64.til
│ └── pc
│ ├── bcb5win.til
│ ├── gnulnx_x64.til
│ ├── mssdk.til
│ ├── ntddk64.til
│ ├── vc10_64.til
│ ├── vc6win.til
│ └── vc8amd64.til
├── uninstall
├── uninstall.dat
└── Uninstall IDA Free.desktop
Sadly it looks like the free version is missing IDAPython so the script will not work.
Closing for now.
|
2025-04-01T06:40:40.812301
| 2022-02-01T13:48:06
|
1120698177
|
{
"authors": [
"di",
"tetsuo-cpp",
"woodruffw"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11439",
"repo": "trailofbits/pip-audit",
"url": "https://github.com/trailofbits/pip-audit/pull/229"
}
|
gharchive/pull-request
|
requirement, pypi: Add a --require-hashes flag
Closes #172
This needs cleanup + testing but I wanted to engage early on this one since the intended behaviour isn't straightforward.
@di You mentioned that the --require-hashes flag should skip dependency resolution. Did you mean that we shouldn't resolve dependencies for each requirement and instead, only audit what is explicitly listed in the requirements file?
I was also wondering how hashes should work when the --require-hashes flag isn't supplied. My thinking is that when the --require-hashes flag is supplied, every requirement in the file needs to have a hash. But if it isn't supplied, we should still check any hashes that we find against PyPI.
You mentioned that the --require-hashes flag should skip dependency resolution. Did you mean that we shouldn't resolve transitive dependencies for each requirement and instead, only audit what is explicitly listed in the requirements file?
I was also wondering how hashes should work when the --require-hashes flag isn't supplied. My thinking is that when the --require-hashes flag is supplied, every requirement in the file needs to have a hash AND we check them against PyPI. But if the flag isn't supplied, we should still check any hashes that we find against PyPI and fail if it doesn't match.
This is an indirect answer to both: we can skip dependency resolution entirely, since pip stipulates the following:
If any dependency has a content hash, then all dependencies have content hashes
(Transitively) All dependencies must be explicitly listed in the requirements file if hashing is enabled
In other words: our requirements parser in pip-api should preserve those properties: we should raise an error if we're given a requirements file with dependencies that are missing hashes, and we should not do dependency resolution because only the present hashed dependencies should be considered.
(Looking at things more, I'm not actually sure that we need the --require-hashes flag -- pip stipulates that passing it is identical to any requirement having hashes attached to it. But I guess we should preserve it as a no-op, since that's what pip seems to do?
In other words: our requirements parser in pip-api should preserve those properties: we should raise an error if we're given a requirements file with dependencies that are missing hashes, and we should not do dependency resolution because only the present hashed dependencies should be considered.
Ah! Right, I misunderstood how that works. Thanks for explaining.
(Looking at things more, I'm not actually sure that we need the --require-hashes flag -- pip stipulates that passing it is identical to any requirement having hashes attached to it. But I guess we should preserve it as a no-op, since that's what pip seems to do?
I think it's still important because your file may not have ANY hashes in it. If you pass in --require-hashes, it will complain whereas it otherwise would succeed.
I think it's still important because your file may not have ANY hashes in it. If you pass in --require-hashes, it will complain whereas it otherwise would succeed.
Ah! You're absolutely right. Yes, we need it for that case.
@di You mentioned that the --require-hashes flag should skip dependency resolution. Did you mean that we shouldn't resolve transitive dependencies for each requirement and instead, only audit what is explicitly listed in the requirements file?
Thinking about this more, I think we still need to do dependency resolution, and ensure that we have hashes for all subdependencies, because this is what pip does as well:
$ cat req.txt
sampleproject==2.0.0 \
--hash=sha256:2b0c55537193b792098977fdb62f0acbaeb2c3cfc56d0e24ccab775201462e04 \
--hash=sha256:d99de34ffae5515db43916ec47380d3c603e9dead526f96581b48c070cc816d3
$ python -m pip install -r req.txt
Collecting sampleproject==2.0.0 (from -r req.txt (line 1))
Using cached https://files.pythonhosted.org/packages/b8/f7/dd9223b39f683690c30f759c876df0944815e47b588cb517e4b9e652bcf7/sampleproject-2.0.0-py3-none-any.whl
Collecting peppercorn (from sampleproject==2.0.0->-r req.txt (line 1))
ERROR: In --require-hashes mode, all requirements must have their versions pinned with ==. These do not:
peppercorn from https://files.pythonhosted.org/packages/14/84/d8d9c3f17bda2b6f49406982546d6f6bc0fa188a43d4e3ba9169a457ee04/peppercorn-0.6-py3-none-any.whl#sha256=46125cad688a9cf3b08e463bcb797891ee73ece93602a8ea6f14e40d1042d454 (from sampleproject==2.0.0->-r req.txt (line 1))
I think it's still important because your file may not have ANY hashes in it. If you pass in --require-hashes, it will complain whereas it otherwise would succeed.
Agreed!
Thinking about this more, I think we still need to do dependency resolution, and ensure that we have hashes for all subdependencies, because this is what pip does as well:
Can we leave that validation to pip and not attempt to do that validation ourselves? I imagine that's going to introduce a significant chunk of code reinventing what pip does which will have to be removed during integration, since pip already contains this logic.
Can we leave that validation to pip and not attempt to do it ourselves? I imagine that's going to introduce a significant chunk of code reinventing what pip does which will have to be removed during integration, since pip already contains this logic.
I'm inclined to agree with this, although it does deviate our handling of --require-hashes from that of pip (our --require-hashes is more like --no-deps --require-hashes). My reasoning is that pip-audit isn't in the business of telling the user whether pip install ... would succeed, and that the overwhelming majority of hashed requirements files should be valid already (since, if they're like Warehouse's, they're checked in after successful local and CI usage).
Thoughts @di?
I suppose the pip install with an incompletely hashed requirements file, or a hashed requirements file that's missing dependencies, would fail to install anyways, so the user would be protected regardless.
|
2025-04-01T06:40:40.829825
| 2022-12-30T17:36:54
|
1514673788
|
{
"authors": [
"michaelfarrell76"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11440",
"repo": "transcend-io/consent-manager-ui",
"url": "https://github.com/transcend-io/consent-manager-ui/pull/100"
}
|
gharchive/pull-request
|
Downgrades react-intl to fix importStar error
This fixes an issue with websites that have an old version of requirejs. react intl v5 > relies on import *, which the older version of requirejs does not support
see https://github.com/transcend-io/consent-manager-ui/pull/101
|
2025-04-01T06:40:40.837019
| 2018-10-17T04:33:43
|
370895267
|
{
"authors": [
"stangah",
"tinyels"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11441",
"repo": "transcom/mymove",
"url": "https://github.com/transcom/mymove/pull/1164"
}
|
gharchive/pull-request
|
WIP: Adds UI for confirming pre-approval request deletion [delivers #161129636]
Description
Delete UI, some spacing issues to work out. WIP.
Breaks out the PreApprovalRequest row into its own component that owns the state of whether the delete UI is shown
Code Review Verification Steps
[ ] End to end tests pass (make e2e_test).
[ ] Code follows the guidelines for Logging
[ ] The requirements listed in
Querying the Database Safely
have been satisfied.
Any new migrations/schema changes:
[ ] Follow our guidelines for zero-downtime deploys (see Zero-Downtime Deploys)
[ ] Have been communicated to #dp3-engineering
[ ] There are no aXe warnings for UI.
[ ] This works in IE.
Any new client dependencies (Google Analytics, hosted libraries, CDNs, etc) have been:
[ ] Communicated to @willowbl00
[ ] Added to the list of network dependencies
[ ] Request review from a member of a different team.
[ ] Have the Pivotal acceptance criteria been met for this change?
References
Pivotal story for this change
Screenshots
This is a great start, but I'm not sure how you plan on signaling that one of the deletion confirms is open to prevent this:
|
2025-04-01T06:40:40.849186
| 2024-09-18T20:01:40
|
2534607937
|
{
"authors": [
"CoryKleinjanCACI",
"robot-mymove"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11442",
"repo": "transcom/mymove",
"url": "https://github.com/transcom/mymove/pull/13739"
}
|
gharchive/pull-request
|
B 20441 int
Agility ticket
Summary
Updating migrations to add WHERE NOT EXIST to protect against any issues from running the migration more than once
Verification Steps for the Author
These are to be checked by the author.
[ ] Tested in the Experimental environment (for changes to containers, app startup, or connection to data stores)
[ ] Have the Agility acceptance criteria been met for this change?
Verification Steps for Reviewers
These are to be checked by a reviewer.
[ ] Has the branch been pulled in and checked out?
[ ] Have the BL acceptance criteria been met for this change?
[ ] Was the CircleCI build successful?
[ ] Has the code been reviewed from a standards and best practices point of view?
Setup to Run the Code
Instructions for starting storybook
Instructions for starting the MilMove application
Instructions for running tests
How to test
Access the
Login as a
Frontend
[ ] There are no aXe warnings for UI.
[ ] This works in Supported Browsers and their phone views (Chrome, Firefox, Edge).
[ ] There are no new console errors in the browser devtools.
[ ] There are no new console errors in the test output.
[ ] If this PR adds a new component to Storybook, it ensures the component is fully responsive, OR if it is intentionally not, a wrapping div using the officeApp class or custom min-width styling is used to hide any states the would not be visible to the user.
[ ] This change meets the standards for Section 508 compliance.
Backend
[ ] Code follows the guidelines for Logging.
[ ] The requirements listed in Querying the Database Safely have been satisfied.
Database
Any new migrations/schema changes:
[ ] Follows our guidelines for Zero-Downtime Deploys.
[ ] Have been communicated to #g-database.
[ ] Secure migrations have been tested following the instructions in our docs.
Screenshots
Warnings
:warning:
Files located in legacy directories (src/shared or src/scenes) have
been edited. Are you sure you don’t want to also relocate them to the new file structure?
View the frontend file org ADR for more information
Generated by :no_entry_sign: dangerJS against 912398a13be2e07babfe0d644614ac7640536f9b
|
2025-04-01T06:40:40.861864
| 2021-06-30T02:09:58
|
933241127
|
{
"authors": [
"hannaliebl",
"monfresh",
"robot-mymove"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11443",
"repo": "transcom/mymove",
"url": "https://github.com/transcom/mymove/pull/6901"
}
|
gharchive/pull-request
|
[MB-8204] Refactor TXO MTO fetch to not query for MoveTaskOrder
Description
Explain a little about the changes at a high level.
Reviewer Notes
Is there anything you would like reviewers to give additional scrutiny?
Setup
Add any steps or code to run in this section to help others prepare to run your code:
echo "Code goes here"
Code Review Verification Steps
[ ] If the change is risky, it has been tested in experimental before merging.
[ ] Code follows the guidelines for Logging
[ ] The requirements listed in Querying the Database Safely have been satisfied.
Any new migrations/schema changes:
[ ] Follow our guidelines for zero-downtime deploys (see Zero-Downtime Deploys)
[ ] Have been communicated to #g-database
[ ] Secure migrations have been tested following the instructions in our docs
[ ] There are no aXe warnings for UI.
[ ] This works in Supported Browsers and their phone views (Chrome, Firefox, IE, Edge).
[ ] Tested in the Experimental environment (for changes to containers, app startup, or connection to data stores)
[ ] User facing changes have been reviewed by design.
[ ] Request review from a member of a different team.
[ ] Have the Jira acceptance criteria been met for this change?
References
Jira story for this change
this article explains more about the approach used.
Screenshots
If this PR makes visible UI changes, an image of the finished UI can help reviewers and casual
observers understand the context of the changes. A before image is optional and
can be included at the submitter's discretion.
Consider using an animated image to show an entire workflow instead of using multiple images. You may want to use GIPHY CAPTURE for this! 📸
Please frame screenshots to show enough useful context but also highlight the affected regions.
Warnings
:warning:
Files located in legacy directories (src/shared or src/scenes) have
been edited. Are you sure you don’t want to also relocate them to the new file structure?
View the frontend file org ADR for more information
Messages
:book:
:link: MB-8204
Generated by :no_entry_sign: dangerJS against 94d60bf5a36c4b8cb0ee8d8665c616b86dd29ecd
This looks good so far. I love removing code. Here are a few more things we need to remove:
https://github.com/transcom/mymove/blob/master/pkg/services/move_task_order.go#L37
https://github.com/transcom/mymove/blob/master/pkg/services/move_task_order.go#L58
https://github.com/transcom/mymove/blob/master/pkg/services/move_task_order/move_task_order_fetcher.go#L23-L40
And then we'll need to regenerate the mocks:
make mocks_generate
|
2025-04-01T06:40:40.865841
| 2021-07-26T21:20:15
|
953298375
|
{
"authors": [
"pearl-truss",
"robot-mymove"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11444",
"repo": "transcom/mymove",
"url": "https://github.com/transcom/mymove/pull/7071"
}
|
gharchive/pull-request
|
[MB-8827] Bump storybook-related dependencies to 6.3.6
Description
This PR consolidates the following dependabot PRs related to upgrading Storybook to version 6.3.6:
#7068
#7066
#7063
#7061
#7060
Setup
Just need to ensure the tests and build pass with no issues.
Messages
:book:
:link: MB-8827
Generated by :no_entry_sign: dangerJS against 822e9a2a4009fb08faf5fa03ce3c70fc897a2d6b
Can a front-end person confirm the happo diffs are okay? Especially the document viewer one (the diff shows it as loading instead of present)
cc: @YanZ777 @jenniferlynparsons
|
2025-04-01T06:40:40.874671
| 2022-01-12T17:20:42
|
1100598229
|
{
"authors": [
"reggieriser",
"robot-mymove"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11445",
"repo": "transcom/mymove",
"url": "https://github.com/transcom/mymove/pull/8021"
}
|
gharchive/pull-request
|
[MB-11012] Fixes bug where NTS shipment turns into HHG when editing as a SC
Jira ticket for this change
Summary
We ran across a bug in the services counselor UI where editing an NTS shipment turns it into an HHG after submission. This PR addresses that issue (it looks like we had just hardcoded the HHG as an input to the component).
I also noticed that we were using some old naming for ntsSac and ntsTac in some places. That was causing those values (if present on the order record) to not show up on the edit form, for instance. Those are fixed as well in this PR.
Note to reviewers: Are there any particular tests I need to add/edit for this?
Setup to Run Your Code
💻 You will need to use three separate terminals to test this locally.
Terminal 1
Start the Storybook locally.
make storybook
Terminal 2
Start the UI locally.
make client_run
Terminal 3
Start the Go server locally.
make server_run
Additional steps
As a services counselor, try editing an NTS shipment (or NTS-R) and change something like the counselor remarks. Hit save and verify that the shipment remains the same type it was previously.
Also, you can add NTS TAC/SAC values to the associated order and verify that they show up in the edit interface.
Verification Steps for Author
These are to be checked by the author.
[ ] Tested in the Experimental environment (for changes to containers, app startup, or connection to data stores)
[ ] Request review from a member of a different team.
[ ] Have the Jira acceptance criteria been met for this change?
Verification Steps for Reviewers
These are to be checked by a reviewer.
Frontend
[ ] User facing changes have been reviewed by design.
[ ] There are no aXe warnings for UI.
[ ] This works in Supported Browsers and their phone views (Chrome, Firefox, IE, Edge).
[ ] There are no new console errors in the browser devtools
[ ] There are no new console errors in the test output
Warnings
:warning:
This PR does not include changes to unit tests, even though it affects app code.
Messages
:book:
:link: MB-11012
Generated by :no_entry_sign: dangerJS against aea9c431eafe9cc2bc9c8d4023bc31b6c8eca337
|
2025-04-01T06:40:40.881581
| 2022-11-17T22:15:13
|
1454075565
|
{
"authors": [
"reggieriser",
"robot-mymove"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11446",
"repo": "transcom/mymove",
"url": "https://github.com/transcom/mymove/pull/9612"
}
|
gharchive/pull-request
|
[MB-14464] Updated milmove-cypress hash to latest one, now with Cypress 8.5.0
Jira ticket for this change
Summary
First, reference this circleci-docker PR. This MilMove PR is just trying to get MilMove on the latest image with a downgraded Cypress (8.5.0). We're currently on the same version, but we're having to use a circleci-docker hash from back in April because cypress was updated after that to 9.x and 10.x in circleci-docker independently of MilMove. As a result, updating other dependencies becomes more difficult. So this just tries to get everything on the latest main branches back in sync.
Setup to Run Your Code
💻 You will need to use three separate terminals to test this locally.
Terminal 1
Start the Storybook locally.
make storybook
Terminal 2
Start the UI locally.
make client_run
Terminal 3
Start the Go server locally.
make server_run
Additional steps
Just make sure all E2E tests pass.
Verification Steps for Author
These are to be checked by the author.
[ ] Tested in the Experimental environment (for changes to containers, app startup, or connection to data stores)
[ ] Request review from a member of a different team.
[ ] Have the Jira acceptance criteria been met for this change?
Messages
:book:
:link: MB-14464
Generated by :no_entry_sign: dangerJS against ffdd48b6377fec23f2641a781a88ee0bcf573cd9
|
2025-04-01T06:40:40.895541
| 2024-07-26T18:06:56
|
2432740832
|
{
"authors": [
"gauravmm",
"onnimonni",
"transitive-bullshit"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11447",
"repo": "transitive-bullshit/nextjs-notion-starter-kit",
"url": "https://github.com/transitive-bullshit/nextjs-notion-starter-kit/pull/629"
}
|
gharchive/pull-request
|
Modernization
The original project is great!
It relies on a lot of outdated dependencies, though. I've gone through the codebase and bumped dependencies, gotten rid of old code, and implemented a unified way to set the icons on the right bar and the bottom bar.
The key dependency react-notion-x is not altered here. I'm also working on that, but have hit some sort of a roadblock with some dependency.
Thanks @gauravmm for this! While we are waiting for @transitive-bullshit to decide how to continue with this starter I created my own fork with updated react-notion-x dependency.
See more here:
https://github.com/onnimonni/nextjs-notion-starter-kit/pull/2
You can see the forked react-notion-x here: https://github.com/onnimonni/react-notion-x/tree/publish-forked-versions
It works perfectly but few tests are still failing which are written with ts and try to import the internal packages eg `import "notion-utils"' are not working and fail with:
Uncaught exception in src/notion-api-universal.test.ts
Error [ERR_MODULE_NOT_FOUND]: Cannot find package '@onnimonni/notion-utils' imported from /Users/onnimonni/Projects/react-notion-x/packages/notion-client/src/notion-api.ts
I'm happy to add you as contributor to both of these forks if you want to help out 👍.
Fixed in https://github.com/transitive-bullshit/nextjs-notion-starter-kit/pull/641
|
2025-04-01T06:40:40.901646
| 2024-04-23T06:07:29
|
2258024456
|
{
"authors": [
"alexgiano",
"chfritz"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11448",
"repo": "transitiverobotics/transitive",
"url": "https://github.com/transitiverobotics/transitive/issues/38"
}
|
gharchive/issue
|
Services support
Hello, I would like to inquire if the Transient Robotics platform supports sending and receiving ROS services. Currently, I am developing a web interface that requires some services to control certain actions on the robot, and I would like to evaluate your solution for a series of additional features that you offer. I have read that managing services may be a limitation of the MQTT communication protocol. I'm not sure if this is the right place to ask this question; if it isn't, I can move it to where it's more appropriate. Thank you very much.
We don't currently have an existing capability to do this, but it has been requested by others, too, so we'll probably create one soon. Since you are asking this question here, on the github repo of the open-source Transitive framework, I'll also point out that you could write that capability yourself, if that's what you want. To get started with that, you'd follow these steps and then look at the starter code that is generated from it: https://transitiverobotics.com/docs/develop/creating_capabilities
To implement the service calls, I would use mqttSync.mqtt directly, i.e., not use mqttSync (which is for synchronizing stateful data), but the mqtt client itself. You can follow this example from the robot-agent itself, used to trigger agent actions, in this case restarting the agent, from the web:
https://github.com/transitiverobotics/transitive/blob/bf7aadaef4c6508b5195ee9f33079ac4e18370f6/cloud/app/web_components/robot-agent-device.jsx#L294-L298
MQTT does not have any built-in support for RPCs, so you'll need to think about the return value yourself. In the agent we do this by writing it into the synchronized data, where the web client will receive it, e.g., for receiving the package log upon request. This pattern is not ideal though and still needs some improvement.
But again, if this is more involved than you'd like, then you can also wait until we create a capability for it that you could then just install on your robots.
Thanks for your kind reply. I took a look at the solution proposed by you by searching for additional material online like async-mqtt or mqtt-json-rpc but, given my limited knowledge of the mqtt protocol, i would most likely not be able to obtain a good result. Currently I will therefore remain on the websocket protocol but keeping an eye on the developments and implementations of your platform. Thanks so much again
Makes sense. Just to clarify though: MQTTSync is not async-mqtt. The former is a protocol we've developed an it's documented here and here.
|
2025-04-01T06:40:40.902977
| 2024-06-09T23:41:52
|
2342555849
|
{
"authors": [
"devinmatte"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11449",
"repo": "transitmatters/mbta-performance",
"url": "https://github.com/transitmatters/mbta-performance/pull/27"
}
|
gharchive/pull-request
|
Support LAMP based monthly csvs
There's a new dataset for monthly dumps based on LAMP with slightly different formatting
Use that new file and handle its formatting
Already using this for data already in the production dashboard, going to merge to avoid someone parsing data with the old version
|
2025-04-01T06:40:40.931501
| 2020-04-24T08:40:37
|
606144836
|
{
"authors": [
"BettinaTrunk",
"agricolab"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11450",
"repo": "translationalneurosurgery/tool-offspect",
"url": "https://github.com/translationalneurosurgery/tool-offspect/issues/26"
}
|
gharchive/issue
|
Feature: Setting constant onset shift for one cachefile
Is your feature request related to a problem? Please describe.
Currently, it is possibly to shift the onset of MEPs in the GUI. However, for one cachefile it requires one to manually shift the onset for every single MEP.
Describe the solution you'd like
It would be great to have the possibility to set an onset shift and have it constant for the whole data of one Cachefile.
I believe using the python API is the best approach for such batch-manipulation. I wrote a couple of examples and added them to the documentation. The one for your use-case is at https://offline-inspect.readthedocs.io/en/latest/cachefile.html#batch-manipulate and repeated here for ease of use:
from offspect.api import CacheFile, decode, encode
cf = CacheFile("merged.hdf5")
for ix, (data, attrs) in enumerate(cf):
key = "onset_shift"
old = decode(attrs[key])
print(f"Trace {ix} {key}:", old, end=" ")
new = old + 5
attrs["onset_shift"] = encode(new)
cf.set_trace_attrs(ix, attrs)
test = decode(cf.get_trace_attrs(ix)["onset_shift"])
print("to", test)
In this example, we shift all onsets by 5 samples to the right.
Closing this, as we agreed a API approach is most suitable. Find updated documentations with examples for batch manipulation https://offline-inspect.readthedocs.io/en/latest/cachefile.html
|
2025-04-01T06:40:41.011711
| 2017-10-25T17:29:57
|
268483417
|
{
"authors": [
"jacktuck",
"ryanstinson",
"travisjeffery"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11460",
"repo": "travisjeffery/jocko",
"url": "https://github.com/travisjeffery/jocko/issues/69"
}
|
gharchive/issue
|
Code Coverage Plugin
@travisjeffery I would like to enable https://codecov.io/ on the repo. I don't think I can perform all the required steps to set it up but I can assist and add the required test setup. What do you think?
Good idea, I've set it up. Let's get the coverage up before we add the badge to the README though :D
@travisjeffery I messaged on gitter before looking at these issues - i will try and help with coverage too :)
|
2025-04-01T06:40:41.018532
| 2024-10-07T18:46:09
|
2571194352
|
{
"authors": [
"AdamRJensen",
"trchudley"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:11461",
"repo": "trchudley/pdemtools",
"url": "https://github.com/trchudley/pdemtools/issues/31"
}
|
gharchive/issue
|
x and y-units in mosaic and terrain example
In this example https://pdemtools.readthedocs.io/en/latest/examples/mosaic_and_terrain.html there are some very nice figures.
However, the x- and y- units are unknown. Would it be possible to show these as latitude/longitudes or something the reader can understand?
Thanks Adam. The merging process I use to combined the mosaics from different tiles loses the CF-compliant coordinate attributes that the PDT data normally comes with. New addition of code below enforces the CF-compliant attributes and makes the mosaic data match the rest of the data. This is also reflected in the notebooks.
dem['x'].attrs['axis'] = 'X'
dem['x'].attrs['long_name'] = 'x coordinate of projection'
dem['x'].attrs['standard_name'] = 'projection_x_coordinate'
dem['x'].attrs['units'] = 'metre'
dem['y'].attrs['axis'] = 'Y'
dem['y'].attrs['long_name'] = 'y coordinate of projection'
dem['y'].attrs['standard_name'] = 'projection_y_coordinate'
dem['y'].attrs['units'] = 'metre'
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.