added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T06:40:37.992073
2024-06-10T14:03:44
2344015458
{ "authors": [ "iwpnd", "tidwall", "undeadcat" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11346", "repo": "tidwall/tile38", "url": "https://github.com/tidwall/tile38/pull/745" }
gharchive/pull-request
Feature/set circle support Please do not open a pull request without first filing an issue and/or discussing the feature directly with the project maintainer. Please ensure you adhere to every item in this list [ ] This PR was pre-approved by the project maintainer Raised discussion on Slack: https://tile38.slack.com/archives/C0M6QN8VC/p1718027811980879 [x] I have self-reviewed the code [x] I have added all necessary tests [ ] Fix tests, test for concentric circles contains fails. Depends on: https://github.com/tidwall/geojson/pull/27 Describe your changes Adds support for SET ... CIRCLE. <IP_ADDRESS>:9851> SET cities sf POINT 37.769092 -122.451505 OK <IP_ADDRESS>:9851> SET cities 10_km_from_sf CIRCLE 37.769092 -122.451505 10000 OK <IP_ADDRESS>:9851> SET cities 100_km_from_sf CIRCLE 37.769092 -122.451505 100000 OK <IP_ADDRESS>:9851> GET cities 10_km_from_sf "{\"type\":\"Feature\",\"geometry\":{\"type\":\"Point\",\"coordinates\":[-122.451505,37.769092]},\"properties\":{\"type\":\"Circle\",\"radius\":10000,\"radius_units\":\"m\"}}" <IP_ADDRESS>:9851> INTERSECTS cities POINT 37.769092 -122.451505 1) (integer) 0 2) 1) 1) "100_km_from_sf" 2) "{\"type\":\"Feature\",\"geometry\":{\"type\":\"Point\",\"coordinates\":[-122.451505,37.769092]},\"properties\":{\"type\":\"Circle\",\"radius\":100000,\"radius_units\":\"m\"}}" 2) 1) "10_km_from_sf" 2) "{\"type\":\"Feature\",\"geometry\":{\"type\":\"Point\",\"coordinates\":[-122.451505,37.769092]},\"properties\":{\"type\":\"Circle\",\"radius\":10000,\"radius_units\":\"m\"}}" 3) 1) "sf" 2) "{\"type\":\"Point\",\"coordinates\":[-122.451505,37.769092]}" <IP_ADDRESS>:9851> INTERSECTS cities IDS POINT 37.769092 -122.451505 1) (integer) 0 2) 1) "100_km_from_sf" 2) "10_km_from_sf" 3) "sf" <IP_ADDRESS>:9851> INTERSECTS cities IDS POINT 37.554478 -122.020158 1) (integer) 0 2) 1) "100_km_from_sf" This is based on PR https://github.com/tidwall/tile38/pull/649, but with some differences: https://github.com/tidwall/tile38/pull/649 no longer compiles this uses tidwall/geojsons Circle type, while the previous PR stored the polygon representation of the circle. Issue number and link Pull request require a prior issue with discussion. Include the issue number of link here. Duplicate of #649 The problem with a circle type as a geometry stored in a collection is that there is no GeoJSON compatible way to represent it without losing precision on output. I'm not opposed with the idea of a SET ... CIRCLE ... command but I would like to somehow solve the compatibility issue.
2025-04-01T06:40:38.061335
2023-06-13T08:42:39
1754356887
{ "authors": [ "aohsato", "manato" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11348", "repo": "tier4/edge-auto-jetson", "url": "https://github.com/tier4/edge-auto-jetson/pull/26" }
gharchive/pull-request
feat(andible): update environmental setup Related Links https://github.com/tier4/edge-auto/pull/4 Description Update ansible roles to set up edge.auto_jetson environment. Review Procedure Please run setup-dev-env.sh to set up the environment. Remarks This PR contains the netplan role, which fixes the IP address for eth1. If you are using this network interface for ssh connection, running ansilbe-playbook may cause ssh disconnection. Pre-Review Checklist for the PR Author PR Author should check the checkboxes below when creating the PR. [x] Assign PR to reviewer Checklist for the PR Reviewer Reviewers should check the checkboxes below before approval. [ ] Commits are properly organized and messages are according to the guideline [ ] (Optional) Unit tests have been written for new behavior [ ] PR title describes the changes Post-Review Checklist for the PR Author PR Author should check the checkboxes below before merging. [ ] All open points are addressed and tracked via issues or tickets CI Checks Build and test for PR: Required to pass before the merge. @manato How about PR status? @aohsato As well as other PR, I'll open this PR because I guess it's ready. Please note that I set the default camera configuration to be: FSYNC trigger mode fixed exposure time (11ms) in this PR. Please note that I set the default camera configuration to be... @manato Could you add this explain to our tutorials? https://github.com/tier4/edge-auto/pull/3
2025-04-01T06:40:38.088692
2022-06-19T04:14:39
1275948277
{ "authors": [ "caseydavenport", "tmjd", "vikas027" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11349", "repo": "tigera/operator", "url": "https://github.com/tigera/operator/issues/2026" }
gharchive/issue
Question - How do I install Calico add-on in an EKS cluster? I am looking to install Calico add-on in an EKS cluster along with AWS VPC CNI as suggested in the official calico documentation which points to the AWS official documentation, which further states that the recommended way of installing Calico on EKS is by using this operator. I am not sure how to do the same with the operator. What I have tried :) look in the certification courses but that too has the old way of installing it Alternatives Helm Installation helm install calico projectcalico/tigera-operator --version v3.23.1 Yeah it isn't clear how to do this. I think you should be able to follow the Calico quickstart directions and just ensure you properly configure the custom-resources.yaml for the AWS CNI plugin (after you install the AWS CNI plugin). The Amazon EKS documentation has steps for installing Calico listed here: https://docs.aws.amazon.com/eks/latest/userguide/calico.html#calico-install Looks like it could use an update to the latest version, though. The Amazon EKS documentation has steps for installing Calico listed here: https://docs.aws.amazon.com/eks/latest/userguide/calico.html#calico-install Looks like it could use an update to the latest version, though. Yes, this is the link I have in my query as well. It does not state how to install using the operator. @vikas027 I see that the doc is not very clear, but the instructions listed under the "Helm" tab on that page are the correct instructions to install the tigera-operator (using helm), very similar to the instructions we have here: https://projectcalico.docs.tigera.io/getting-started/kubernetes/helm We should clean up the language on the Amazon doc so that it's less confusing, but the content there is correct.
2025-04-01T06:40:38.106387
2023-02-28T19:37:44
1603692095
{ "authors": [ "codecov-commenter", "efirs" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11350", "repo": "tigrisdata/tigris", "url": "https://github.com/tigrisdata/tigris/pull/882" }
gharchive/pull-request
Enable quota metrics by default Describe your changes How best to test these changes Issue ticket number and link Codecov Report Patch coverage: 100.00% and project coverage change: -27.71 :warning: Comparison is base (c0eb3c0) 57.56% compared to head (b071362) 29.85%. :exclamation: Current head b071362 differs from pull request most recent head 93343b8. Consider uploading reports for the commit 93343b8 to get more accurate results :mega: This organization is not using Codecov’s GitHub App Integration. We recommend you install it so Codecov can continue to function properly for your repositories. Learn more Additional details and impacted files @@ Coverage Diff @@ ## main #882 +/- ## =========================================== - Coverage 57.56% 29.85% -27.71% =========================================== Files 175 175 Lines 19569 19568 -1 =========================================== - Hits 11265 5843 -5422 - Misses 7349 13042 +5693 + Partials 955 683 -272 Impacted Files Coverage Δ server/metrics/metrics.go 83.69% <100.00%> (-7.71%) :arrow_down: server/search/factory.go 0.00% <0.00%> (-100.00%) :arrow_down: lib/container/priority_queue.go 0.00% <0.00%> (-100.00%) :arrow_down: server/quota/namespace.go 0.00% <0.00%> (-98.37%) :arrow_down: schema/lang/go.go 0.00% <0.00%> (-90.63%) :arrow_down: schema/lang/java.go 0.00% <0.00%> (-90.63%) :arrow_down: schema/lang/typescript.go 0.00% <0.00%> (-90.63%) :arrow_down: server/search/sorted_hits.go 0.00% <0.00%> (-89.10%) :arrow_down: server/search/facets.go 0.00% <0.00%> (-88.60%) :arrow_down: schema/inference.go 0.00% <0.00%> (-85.06%) :arrow_down: ... and 111 more Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here. :umbrella: View full report at Codecov. :loudspeaker: Do you have feedback about the report comment? Let us know in this issue.
2025-04-01T06:40:38.108712
2023-03-06T13:44:15
1611461086
{ "authors": [ "garrensmith", "pboros" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11351", "repo": "tigrisdata/tigris", "url": "https://github.com/tigrisdata/tigris/pull/902" }
gharchive/pull-request
fix: only log unexpected errors in match Describe your changes We logging an error that is acceptable. There will be cases when the field we looking for does not exist in the document or in the json timestamp object when we do a query. In those cases rather check first the field does not exist and return false. Only if it is an unexpected error do we log it. How best to test these changes Test should pass Issue ticket number and link I verified that this indeed fixes the logging issue.
2025-04-01T06:40:38.118179
2021-10-31T07:56:38
1040402347
{ "authors": [ "apatait", "tikhop" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11352", "repo": "tikhop/TPInAppReceipt", "url": "https://github.com/tikhop/TPInAppReceipt/issues/94" }
gharchive/issue
Objective C example code Hi This is not an issue per se, but I could not find a way to communicate this and I thought it'll also help others in similar boat; hence posting here. Do you have an example of using TPInAppReceipt library in an objective C project? Sample code on how to instantiate the class and call various methods would be great. Or if you have Objective C class documentation, that'd be great too. I am not familiar as much with Swift, and my project is in Objective C hence the request. Thank you in advance! Hi @apatait, Unfortunately, I don't have an example, but you can use it in objc project. Here you can find how to install it: https://github.com/tikhop/TPInAppReceipt/blob/master/Documentation/UseInObjCProject.md Here is a similar question: https://github.com/tikhop/TPInAppReceipt/issues/84 Some old example that probably doesn't work anymore. https://gist.github.com/tikhop/7d1215bbcea7fe8a3be3ce5ef25e15ee Anyway, it must be pretty easy to use the library in objective c project just by checking swift examples from readme file
2025-04-01T06:40:38.122164
2023-06-20T04:36:35
1764620095
{ "authors": [ "cfzjywxk", "zyguan" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11353", "repo": "tikv/client-go", "url": "https://github.com/tikv/client-go/pull/847" }
gharchive/pull-request
lock_resolver: let resolve lock timeout properly fix https://github.com/pingcap/tidb/issues/44822 , after this PR tidb can pass the test. I think it's better to merge PRs affecting critical path with 2-3 approvals.
2025-04-01T06:40:38.138974
2023-09-21T06:48:16
1906235826
{ "authors": [ "Yui-Song" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11354", "repo": "tikv/tikv", "url": "https://github.com/tikv/tikv/issues/15653" }
gharchive/issue
TableFullScan gets slower resulting in a 5% performance regression in TPCDS Bug Report What version of TiKV are you using? good commit: https://github.com/tikv/tikv/pull/15523 bad commit: https://github.com/tikv/tikv/pull/15522 What operating system and CPU are you using? Steps to reproduce deploy a tidb cluster with 1 tidb + 3 tikv with the good and the bad commit run TPCDS 50g What did you expect? no performance regression What did happened? Q88 elapsed time: 34s -> 38.9s Q28 elapsed time: 24.8s->28.3s exection_plan.zip /severity critical It was caused by https://github.com/tikv/tikv/pull/15584 /remove may-affects-5.3 /remove may-affects-5.4 /remove may-affects-6.1 /remove may-affects-6.5 /remove may-affects-7.1 /unlabel may-affects-5.3 /remove-label may-affects-5.3 /remove-label may-affects-5.4 /remove-label may-affects-6.1 /remove-label may-affects-6.5 /remove-label may-affects-7.1
2025-04-01T06:40:38.147331
2021-12-23T07:52:57
1087465884
{ "authors": [ "windtalker", "youjiali1995" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11355", "repo": "tikv/tikv", "url": "https://github.com/tikv/tikv/pull/11708" }
gharchive/pull-request
coprocessor: not return rows when there is no input for simple aggregation What is changed and how it works? Close pingcap/tidb#30923 What's Changed: For SimpleAggregation does not return row if there is no input rows Related changes PR to update pingcap/docs/pingcap/docs-cn: PR to update pingcap/tidb-ansible: Need to cherry-pick to the release branch Check List Tests Unit test Manual test (add detailed scripts or steps below) Side effects Performance regression Consumes more CPU Consumes more MEM Breaking backward compatibility Release note Fix wrong `any_value` result when there are regions returning empty result /label needs-cherry-pick-5.3 /label needs-cherry-pick-5.2 /label needs-cherry-pick-5.1 /label needs-cherry-pick-5.0 /label needs-cherry-pick-4.0 /rebuild pingcap/tidb#30923 TiDB v3.1.2/v3.0.20/v2.1.19, all of them have the same issue. BTW seems that TiKV is keeping this behavior since v2.0.0, so this might be a bug exists for a lot of versions. Better to check v3.0 and v2.1 as well. Tested in TiDB v3.1.2/v3.0.20/v2.1.19, all of them have the same issue. /label needs-cherry-pick-3.1 /label needs-cherry-pick-3.0 /label needs-cherry-pick-2.1 /merge @breeswish can you help merge it? /merge /merge
2025-04-01T06:40:38.153205
2024-09-06T07:07:45
2509688354
{ "authors": [ "overvenus", "wuhuizuo" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11356", "repo": "tikv/tikv", "url": "https://github.com/tikv/tikv/pull/17490" }
gharchive/pull-request
ci: update owners aliases What is changed and how it works? Issue Number: Close #17004 What's Changed: Related changes [ ] PR to update pingcap/docs/pingcap/docs-cn: [ ] Need to cherry-pick to the release branch Check List Tests [ ] Unit test [ ] Integration test [ ] Manual test (add detailed scripts or steps below) [x] No code Side effects [ ] Performance regression: Consumes more CPU [ ] Performance regression: Consumes more Memory [ ] Breaking backward compatibility Release note None /cc @yudongusa @benmeadowcroft @zhangjinpeng87 /cc @overvenus @LykxSassinator /ok-to-test /retest /retest flaky unit test /cc cfzjywxk
2025-04-01T06:40:38.160772
2020-06-03T10:11:27
629869531
{ "authors": [ "youjiali1995" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11357", "repo": "tikv/tikv", "url": "https://github.com/tikv/tikv/pull/8009" }
gharchive/pull-request
test: stabilize lock_manager tests Signed-off-by: youjiali1995<EMAIL_ADDRESS> What problem does this PR solve? Issue Number: close #xxx Problem Summary: What is changed and how it works? Proposal: xxx What's Changed: Related changes PR to update pingcap/docs/pingcap/docs-cn: PR to update pingcap/tidb-ansible: Need to cherry-pick to the release branch Check List Tests Unit test Integration test Manual test (add detailed scripts or steps below) No code Side effects Performance regression Consumes more CPU Consumes more MEM Breaking backward compatibility Release note No release note /test /test /run-all-tests /rebuild /test /test /test /test /test /test /test /test
2025-04-01T06:40:38.170083
2016-07-23T22:53:10
167207515
{ "authors": [ "das-keyboard", "fenryxo", "lonlazer" ], "license": "bsd-2-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11358", "repo": "tiliado/nuvola-app-amazon-cloud-player", "url": "https://github.com/tiliado/nuvola-app-amazon-cloud-player/issues/17" }
gharchive/issue
Nuvola does not recognise the current song Nuvola does not recognise the current song. It shows in the sidebar only (null). So neither the lyrics plugin nor the desktop notifications are working. I installed the Version 5.1 of the Plugin and Version 3.1.0 of Nuvola via the Arch User Repository. Hello @lonlazer. Thanks for taking the time to report this issue. I can confirm it. @SteVwonder, could you take care of it? Thanks! I wonder if there is still someone taking care of updating this service/app. Can someone give us some information about it? @fenryxo @SteVwonder Hello @das-keyboard. This script is currently unmaintained as the maintainer @SteVwonder seems to have lost interest in it. I might take over it next month if there is any spare time. I've adopted this abandoned script and released version 5.2 with fixes. -- Jiri Janousek, Nuvola Player developer • My work on Nuvola Player and user support are funded by Nuvola Patrons. :heart: Make a donation to keep the project going. Thank you.
2025-04-01T06:40:38.202503
2024-03-03T07:51:40
2165235175
{ "authors": [ "zaneduffield" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11359", "repo": "tim-harding/soa-rs", "url": "https://github.com/tim-harding/soa-rs/issues/6" }
gharchive/issue
Compile error from soa-rs-derive: can't leak private type Using version 0.3.1 of soa-rs I cannot compile anything because a private type is being leaked from soa-rs-derive. For me, to reproduce the issue it's as simple as cargo init test_soa_rs; cd test_soa_rs; cargo add soa-rs; cargo check; I get the following error message Compiling soa-rs-derive v0.3.1 error[E0446]: private type `SoaDerive` in public interface --> C:\Users\zane\.cargo\registry\src\index.crates.io-6f17d22bba15001f\soa-rs-derive-0.3.1\src\fields.rs:9:1 | 9 | / pub fn fields_struct( 10 | | ident: Ident, 11 | | vis: Visibility, 12 | | fields: Punctuated<Field, Comma>, 13 | | kind: FieldKind, 14 | | soa_derive: SoaDerive, 15 | | ) -> Result<TokenStream, syn::Error> { | |____________________________________^ can't leak private type | ::: C:\Users\zane\.cargo\registry\src\index.crates.io-6f17d22bba15001f\soa-rs-derive-0.3.1\src\lib.rs:75:1 | 75 | struct SoaDerive { | ---------------- `SoaDerive` declared as private For more information about this error, try `rustc --explain E0446`. error: could not compile `soa-rs-derive` (lib) due to previous error I upgraded to Rust 1.76 from 1.74 and I no longer have the issue.
2025-04-01T06:40:38.208627
2021-11-18T21:05:58
1057799645
{ "authors": [ "BjoernPetersen", "torbencarstens" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11360", "repo": "tim-hat-die-hand-an-der-maus/preliminary-frontend", "url": "https://github.com/tim-hat-die-hand-an-der-maus/preliminary-frontend/issues/12" }
gharchive/issue
Remove size from image url movie.imdb.coverUrl.rsplit("@", 1) lazy loading Bump Could you prioritize this? This is blocking important processes! (My sense of aesthetics) Real talk: shouldn't this happen in the imdb-resolver? I actually want to do this in the frontend since the load-times will go down. The removal of the actual size can happen in imdb-resolver. I don't understand, what do you want to do where? (Keep in mind that I refuse to have a look at the frontend code) imdb-resolver removes the size from the cover url, but then we'd load the full size image in the frontend which isn't necessary since we'll never display a full size cover. That's why I want to add the actual dimensions (which are set in the frontend anyways) so we have fewer resources to load -> better loading times/less bandwidth waste/no resizing necessary. So does imdb-resolver already remove the size from the URL, or do we need to create an issue for that? https://github.com/tim-hat-die-hand-an-der-maus/imdb-resolver/pull/6 I actually need the image size ratio or original width and height. Maybe we should convert coverUrl to cover: { "url": "<String>", "width": <int>, "height": <int> } or replace width/height with ratio: <float> what do you think? I think ratio is probably the best solution additionally: _V1_SY150_CR0,0,101,150_ no idea what CR does, doesn't seem to affect the output (visually at least) SY150 is the height of the image, S can apparently be replaced with U, I couldn't find any differences in those two. Other characters seem to just ignore the size and show the given width/height based from the top left. If Y is replaced by X it takes the width for stretching.
2025-04-01T06:40:38.214915
2023-08-28T18:47:29
1870280901
{ "authors": [ "SlothUS", "fcom000", "tim-hub" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11361", "repo": "tim-hub/obsidian-bible-reference", "url": "https://github.com/tim-hub/obsidian-bible-reference/issues/117" }
gharchive/issue
not working on windows the plugin simply does not work, it is active so I can run the command "Bible reference: verse lookup", but when I write inside the verse there is always no answer, nothing. Also if I write --John1:1 etc I never had any output. Tried also without firewall no difference. Is there some extra plugin I need to run it? I tried it also in android phone, same problem, nothing happens... If on windows, can I try open developer console to see what logs is there? this is the toggle for enable the developer console. If you can get some logs, that will be very helpful. I am very curious why this could happen as well. Also getting the same issue. Here's what my console produces - Also getting the same issue. Here's what my console produces - Hi @SlothUS Thanks for this. @fcom000 @SlothUS Can you try to switch the Bible Version provider by Bolls life? in the settings. This is a CORS issue, which is a server side issue, I know this will happen one day, just don't know it happens so fast. CORS issue, which it blocks the 3rd part API consume to consume its API. This is why it failed. To fix this, need to host a server by my self. For now I will suggest you to use bolls life versions. this should be fixed in 1.9.0 by switching the default version, please re-open this if not. thanks very much, it works with the Bolls life. I was wondering if it is easy to host a Bible Api offline in my computer so even without internet the plugin works (and there will be no update issues?
2025-04-01T06:40:38.233379
2019-07-23T18:43:24
471865221
{ "authors": [ "gustavolanzas" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11362", "repo": "timber/timber", "url": "https://github.com/timber/timber/issues/2048" }
gharchive/issue
wordpress filtered content not available in post object? I'm not getting the post `content` as expected anymore. on a single page, i'm using this in the page.php: `$context = Timber::get_context(); $context['post']= new Timber\Post();' All my templates that use {{ post.content }} stopped working. When I dump the post object, I don't see the 'content' key as being set in the object. Expected behavior {{ post.content }} should output wordpress filtered content (with wpautop, shortcodes, and oembed filters applied. Actual behavior Timber\Post Object ( [ImageClass] => Timber\Image [PostClass] => Timber\Post [TermClass] => Timber\Term [object_type] => post [custom] => Array ( [_edit_lock] =><PHONE_NUMBER>:79 [_edit_last] => 79 [_wp_page_template] => default ) [_content:protected] => [_permalink:protected] => [_next:protected] => Array ( ) [_prev:protected] => Array ( ) [_css_class:protected] => [id] => 15061 [ID] => 15061 [post_author] => 79 [post_content] => this is the content [post_date] => 2019-07-23 09:57:18 [post_excerpt] => [post_parent] => 0 [post_status] => publish [post_title] => test page [post_type] => page [slug] => test-page [__type:protected] => [_edit_lock] =><PHONE_NUMBER>:79 [_edit_last] => 79 [_wp_page_template] => default [post_date_gmt] => 2019-07-23 16:57:18 [comment_status] => closed [ping_status] => closed [post_password] => [post_name] => test-page [to_ping] => [pinged] => [post_modified] => 2019-07-23 09:57:18 [post_modified_gmt] => 2019-07-23 16:57:18 [post_content_filtered] => [guid] => http://localhost:8080/library/?page_id=15061 [menu_order] => 0 [post_mime_type] => [comment_count] => 0 [filter] => raw [status] => publish ) Steps to reproduce behavior page.php: $context = Timber::get_context(); $context['post']= new Timber\Post(); Timber::render( array( 'page.twig' ), $context ); page.twig: {{ post.content }} or {{ post | print_r }} What version of WordPress, PHP and Timber are you using? PHP 7.1.3, Wordpress 5.2.2, Timber 1.9.2 How did you install Timber? (for example, from GitHub, Composer/Packagist, WP.org?) Latest from Wordpress plugins Sorry - discovered some changes to a custom plugin that were breaking this. Not a timber issue.
2025-04-01T06:40:38.267557
2021-02-07T10:55:30
802924213
{ "authors": [ "denouche", "timdorr" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11363", "repo": "timdorr/tesla-api", "url": "https://github.com/timdorr/tesla-api/issues/299" }
gharchive/issue
auth.tesla.cn ? Hello One of my users reported me that he cannot login anymore since I updated my app on the new SSO login. He also told me that on his side the login page is: https://auth.tesla.cn/oauth2/v1/authorize?client_id=teslaweb-cn&response_type=code&scope=openid email profile&redirect_uri=https%3A//www.tesla.cn/openid-connect/generic&state=Ht7mUM_AjhcOj5gNaU_xHLUqy-v9TljYQOtnLUghodY&locale=zh-CN When I asked him to login on Tesla.com he told me that this does not works. On my side, I cannot login to auth.tesla.cn. So there would be multiple SSO servers not sharing their users databases? Have you encountered this case? Last thing I tested on my side, when I try to load: https://auth.tesla.cn/oauth2/v3/authorize?client_id=ownerapi&redirect_uri=https%3A%2F%2Fauth.tesla.cn%2Fvoid%2Fcallback&response_type=code&scope=openid email offline_access&state=xxx I have an "invalid-redirect-uri" error. With an account created on tesla.cn I succeed to login using this URL: https://auth.tesla.cn/oauth2/v3/authorize?client_id=ownerapi&redirect_uri=https%3A%2F%2Fauth.tesla.com%2Fvoid%2Fcallback&response_type=code&scope=openid email offline_access&state=xxxxxx So the SSO server seems to be on auth.tesla.cn and the redirect_uri seems to be common between all SSO servers (would be logical because the Android app is the same everywhere) Last thing, the /authorize page seems to be different on auth.tesla.cn and auth.tesla.com , there is one more step on auth.tesla.cn (the password field is hidden when you load the page and you have to click on "Next" to see it). The first POST to /authorize is sent with the following body: _csrf: azertyuiopzertyuiop _phase: identity transaction_id: azerty cancel: identity: user@email It answer with a 200 and load the /authorize page again, with the password field this time. Then I post the form with the password and the body contains: _csrf: qsdfghjklqsdfghjkl _phase: authenticate _process: 1 transaction_id: azerty cancel: identity: user@email credential: userPassword And this time I have the 302 response with the code in the location header. So it seems there is an additional "identity" phase on the CN login page! Ok so by adding the extra POST /authorize with the "identity" phase I succeed to obtain an SSO token. But... After getting it I cannot obtain an owner API token using the step 4. The answer is: Response{protocol=h2, code=404, message=, url=https://owner-api.teslamotors.com/oauth/token} {"response":null,"error":"https://mobilesvc.teslamotors.com:443/accounts/GetUidByFederationId => {:code=>404, :message=>\"Could not get account by federation id\"}","error_description":""} login_hint query parameter on first GET /authorize seems to be the way to get the good environment, see here: https://github.com/adriankumpf/teslamate/issues/1297 I'm guessing they distribute a different app build in China, because there are references to the tesla.cn domain in there, but I don't see any code path to actually have them set as the SSO URL or other related variables. So, there may not be that much to glean from the build in the global app stores. Given I'm not a citizen of China, I don't really have much to offer here, unfortunately. I think the last error I have when trying to exchange my bearer an owner API token is because my account in CN is not associated to any vehicle. So I sent an alpha version of my app with the login process using the login_hint query parameter, following the redirects on step 1 GET and reuse the Host value of the last redirect. I will keep you informed. Ok so I got a feedback from the CN user I sent the alpha version of my app, and this is working for him now! So I guess the error I got Could not get account by federation id with the CN account I created for test purpose was because I have no vehicle associated with this account. To summarize, in order to be compliant with all Tesla SSO servers, the GET at step 1 should include a login_hint query parameter with the email as value, and then 2 cases: you get a 200 response you get a 303 response redirecting you to another SSO server (for example auth.tesla.cn) And the rest of the process (step 2 and step 3) should be executed on this SSO server. So in my case I just followed the redirect if any, then get the request Host, and use this new value as Tesla SSO Environment. A big thanks to @cliffchen and @adriankumpf for finding a way to handle this! Documented in #309. Feel free to submit a PR if you want to add anything else. Thanks! And thanks for this documentation and this repository!!
2025-04-01T06:40:38.270014
2022-05-05T03:30:17
1226163816
{ "authors": [ "insaindesign", "timdorr" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11364", "repo": "timdorr/tesla-api", "url": "https://github.com/timdorr/tesla-api/issues/572" }
gharchive/issue
Powerwall and energy endpoints Between this (timdorr/tesla-api) and joeblau/teslaapi.io, there is pretty good coverage of the API. timdorr/tesla-api has a lot more updated information on the auth endpoints, but joeblau/teslaapi.io has much more on the powerwall and energy sites. Can we make these one source of truth? We've got a start with #580!
2025-04-01T06:40:38.296638
2018-01-24T12:22:52
291189262
{ "authors": [ "vpp" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11365", "repo": "timeoff-management/application", "url": "https://github.com/timeoff-management/application/issues/232" }
gharchive/issue
Black out period on department basis Would be VERY helpful if we could block out a section of the calendar in a different colour and put a message on that months team view or a legend of the colours on it so that it says that people should not book leave on those dates - we have a period in the year where we don't allow leave between certain dates (although we need the option to still book it if its approved by directors) - just a colour indication I think would be good. So I would set a date period to be in red, showing that employees shouldn't book leave on those dates. +1 as in #283
2025-04-01T06:40:38.298876
2024-02-02T06:48:13
2114217803
{ "authors": [ "jovezhong", "zliang-min" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11366", "repo": "timeplus-io/proton", "url": "https://github.com/timeplus-io/proton/pull/550" }
gharchive/pull-request
enhancement: input_format_skip_unknown_fields default to true for external streams PR checklist: Did you run ClangFormat ? Did you separate headers to a different section in existing community code base ? Did you surround proton: starts/ends for new code in existing community code base ? Please write user-readable short description of the changes: Closes #549 . (Jove Github Bot) added it to the current sprint. (Jove Github Bot) moved this ticket out of the GitHub project(up to 1200 tickets for one project).
2025-04-01T06:40:38.300927
2020-09-01T05:47:22
689833655
{ "authors": [ "drpebcak", "feikesteenbergen" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11367", "repo": "timescale/timescaledb-kubernetes", "url": "https://github.com/timescale/timescaledb-kubernetes/pull/216" }
gharchive/pull-request
use named port for pgbouncer The services(s) call for targetPort: pgbouncer but the sts does not currently define that name. This leads to the endpoint not getting computed, and it causes issues with service loadbalancers. Thanks for the PR, this is part of https://github.com/timescale/timescaledb-kubernetes/pull/218 now (including this commit). We cannot merge this PR directly, as this commit is now a bugfix on top of 0.7.0, whereas master has diverged a lot already.
2025-04-01T06:40:38.373646
2015-08-09T21:32:34
99930854
{ "authors": [ "scriptingosx", "timsutton" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11368", "repo": "timsutton/mcxToProfile", "url": "https://github.com/timsutton/mcxToProfile/pull/11" }
gharchive/pull-request
added '--defaults' option that gets data with the defaults command This adds a --defaults option as an alternative to --plist and --dsobject. Basically this will run the defaults read command and create the mobileconfig from that output. That way you do not have to worry where the actual plist is stored and wether cfprefsd has already updated. There is an extra option --currentHost which will enable the -currentHost flag when running the defaults command. Example: ./mcxToProfile.py --defaults com.apple.screensaver --identifier com.example.screensaver --currentHost --manage Once I like the idea! However, since we have Python and Foundation at our fingertips, we could do this without shelling out to Python and use CFPreferences natively. I'm also not sure if the output from defaults properly distinguishes between an integer 0 and a Boolean false. I've never tried to throw the defaults output at plistlib directly as you did here. I don't know if there's a convenient way to do the equivalent of defaults read with a single API call but there is a method to list all keys, and then each value can be retrieved. Then you'd still get things in a nice dict structure that should map to a plist. You're right, it wasn't that hard at all. I have also added flags (or a placeholder) to access the 'anyUser' and 'anyApplication' (Global) domain. This is how you get at the different files: Application User Host File Location app_id current current ~/Library/Preferences/ByHost/app_id.host_id.plist app_id current any ~/Library/Preferences/app_id.plist app_id any current /Library/Preferences/app_id.plist app_id any any ??? any current current ~/Library/Preferences/ByHost/.GlobalPrefences.host_id.plist any current any ~/Library/Preferences/.GlobalPrefences.plist any any current /Library/Preferences/.GlobalPreferences.plist any any any ??? I wish I knew what the ??? are. I thought it might read forced profiles if present, but that does not seem to work. Setting those parameter always seems to return empty dicts. Maybe with MCX? If there really is no (current) application for this combinations, it might make sense to assume --currentHost when --anyUser is set. Examples: ./mcxToProfile.py --defaults com.apple.screensaver --identifier ScreenSaver --currentHost reads ~/Library/Preferences/ByHost/com.apple.screensaver.host_id.plist ./mcxToProfile.py --defaults com.apple.screensaver --identifier ScreenSaver reads ~/Library/Preferences/com.apple.screensaver.plist ./mcxToProfile.py --defaults com.apple.loginwindow --identifier loginwindow --anyUser --currentHost reads /Library/Preferences/com.apple.loginwindow.plist ./mcxToProfile.py --defaults ManagedInstalls --identifier MunkiSettings --currentHost --anyUser reads `/Library/Preferences/ManagedInstalls.plist ./mcxToProfile.py --defaults NSGlobalDomain --identifier GlobalPrefs --currentHost reads ~/Library/Preferences/ByHost/.GlobalPrefences.host_id.plist ./mcxToProfile.py --defaults NSGlobalDomain --identifier GlobalPrefs reads ~/Library/Preferences/.GlobalPrefences.plist ./mcxToProfile.py --defaults NSGlobalDomain --identifier GlobalPrefs --anyUser --currentHost reads /Library/Preferences/.GlobalPreferences.plist I think the any-host / any-user combination may have been used in the old days of network-mounted storage and user data. I think it's reasonable enough to make these options for --defaults work similarly to how the defaults command works. Thanks for adding this! Oh, and this was just merged: https://github.com/timsutton/mcxToProfile/commit/cd1314f31cec9fe8b874b6727c0e998097269153 Thinking again about the behaviour when --currentHost and --anyUser are both set, it doesn't generate the kind of "scope" we want. Say I want to capture the settings that are effectively stored in /L/P/com.apple.loginwindow: I need to type this: ./mcxToProfile.py --defaults com.apple.loginwindow --currentHost --anyUser --identifier Loginwindow But this gives me something like: <key>PayloadContent</key> <dict> <key>com.apple.loginwindow.ByHost</key> <dict> <key>Forced</key> <array> <dict> <key>mcx_preference_settings</key> <dict> <key>Hide500Users</key> <true/> </dict> </dict> </array> </dict> </dict> Notice the ".ByHost" at the end of the domain - except that this is not actually a ByHost pref. I'm starting to wonder if we should instead try and make it behave analogous to defaults, and support the --currentHost flag only. To read from the any-user domain, we could simply take the /L/P/com.whatever.domain syntax that defaults uses, and not put the .ByHost for the domain. I kind of like the extra behavior, we could also fix the logic to add the "ByHost" to the preference domain: https://github.com/timsutton/mcxToProfile/pull/12 Thanks, I just wonder what the extra behavior of --anyUser gains us if it only gives expected results when combined with --currentHost. But in practice, it's going to be less common that you want to pull preferences from the any-user domain anyway. We can just document this in an example.
2025-04-01T06:40:38.382976
2021-06-18T13:04:03
924884557
{ "authors": [ "timtorChen" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11369", "repo": "timtorChen/pi-cluster", "url": "https://github.com/timtorChen/pi-cluster/pull/154" }
gharchive/pull-request
(bot) Update ghcr.io/fluxcd/helm-controller Docker tag to v0.11.1 This PR contains the following updates: Package Update Change ghcr.io/fluxcd/helm-controller minor v0.10.1 -> v0.11.1 Release Notes fluxcd/helm-controller v0.11.1 Compare Source CHANGELOG v0.11.0 Compare Source CHANGELOG Configuration 📅 Schedule: At any time (no schedule defined). 🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied. ♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 🔕 Ignore: Close this PR and you won't be reminded about this update again. [ ] If you want to rebase/retry this PR, check this box. This PR has been generated by Renovate Bot. Renovate Ignore Notification As this PR has been closed unmerged, Renovate will now ignore this update (v0.11.1). You will still receive a PR once a newer version is released, so if you wish to permanently ignore this dependency, please add it to the ignoreDeps array of your renovate config. If this PR was closed by mistake or you changed your mind, you can simply rename this PR and you will soon get a fresh replacement PR opened.
2025-04-01T06:40:38.412269
2021-02-12T07:31:42
807018179
{ "authors": [ "atifsyedali", "r00dY", "spbyrne" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11370", "repo": "tinacms/tinacms", "url": "https://github.com/tinacms/tinacms/issues/1743" }
gharchive/issue
StyledFocusRing has width:100%;height:100% which can break CSS layout Hey guys, Awesome job with Tina, couldn't love your project more!!! Just one issue I noticed. react-tina-inline adds StyledFocusRing to the DOM. It's all fine but by making it width: 100%; height: 100% it can break CSS layout. I'm talking about this line of code: https://github.com/tinacms/tinacms/blob/a22063fdf2e10f873df9bd401f5d224c73ef6729/packages/react-tinacms-inline/src/styles/focus-ring.tsx#L85-L86 Imagine you have an InlineGroup or InlineBlock with content that has some "intrinsic natural width" which is not 100%. Like a button with width: 200px. Imagine that the parent of this element is display: flex. Flex parent makes a use of natural width of its children during layout (uses width as flex-basis if the latter is missing). However, if you add width: 100% between parent and child (StyledFocusRing) then layout breaks, because flex parent thinks that its child is 100% width. Basically any kind of layout that makes a use of content size of its children will break if we add Tina container in between. In my opinion StyledFocusRing shouldn't have neither width nor height. Its size should be determined by the content and it should be as non-intrusive as possible. (can't wait for display: contents to have better support btw). I know this change could break backward compatibility in some projects. Maybe we could add a flag? I'd be happy to do a PR, but curious about rationale behind the current behaviour. Just with some quick testing it seems that this was done for components that use relative or absolute positioning within their parent, which is probably the minority compared to intrinsic sizing. I think possibly just accepting a class string and passing it to the focus ring, or only setting width/height when some prop is set could solve the issue for now. This component is used by the inline group and inline block components, and they already accept focus ring options which could accommodate this change. Currently for us, certain types of elements that expect direct children to have certain CSS constraints ends up with incorrect layouts while editing. For example, a FlexGroup that has FlexItems requires FlexItems to have flex properties. It would be great if we can just pass a className to be appended to the StyledFocusRing elements. This way, if I have an element like FlexGroup and another set called FlexItems, then I can give classNames to StyledFocusRing elements that would make them behave like FlexItems. In the long term, it would be great to not even have feedback elements injected. For example, notion.so adds separate feedback elements that are positioned absolutely for mouse move and dnd effects. Also, thanks for an amazing project!
2025-04-01T06:40:39.479052
2022-07-22T12:28:42
1314946708
{ "authors": [ "codecov-commenter", "martins0n" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11371", "repo": "tinkoff-ai/etna", "url": "https://github.com/tinkoff-ai/etna/pull/813" }
gharchive/pull-request
[BUG] Raise errors in models.nn if they can't make in-sample and some cases out-sample predictions Before submitting (must do checklist) [x] Did you read the contribution guide? [x] Did you update the docs? We use Numpy format for all the methods and classes. [x] Did you write any new necessary tests? [x] Did you update the CHANGELOG? Proposed Changes Closing issues closes #787 Codecov Report Merging #813 (2cc0ddd) into master (41fcce4) will decrease coverage by 34.59%. The diff coverage is 70.00%. @@ Coverage Diff @@ ## master #813 +/- ## =========================================== - Coverage 84.05% 49.46% -34.60% =========================================== Files 126 126 Lines 7220 7236 +16 =========================================== - Hits 6069 3579 -2490 - Misses 1151 3657 +2506 Impacted Files Coverage Δ etna/models/nn/deepar.py 82.95% <70.00%> (-17.05%) :arrow_down: etna/models/nn/tft.py 71.69% <70.00%> (-28.31%) :arrow_down: etna/commands/__init__.py 0.00% <0.00%> (-100.00%) :arrow_down: etna/commands/backtest_command.py 0.00% <0.00%> (-97.06%) :arrow_down: etna/commands/forecast_command.py 0.00% <0.00%> (-94.88%) :arrow_down: etna/commands/__main__.py 0.00% <0.00%> (-87.50%) :arrow_down: etna/models/utils.py 14.28% <0.00%> (-85.72%) :arrow_down: etna/commands/resolvers.py 0.00% <0.00%> (-80.00%) :arrow_down: etna/analysis/outliers/density_outliers.py 22.44% <0.00%> (-75.52%) :arrow_down: etna/datasets/datasets_generation.py 27.02% <0.00%> (-72.98%) :arrow_down: ... and 77 more :mega: Codecov can now indicate which changes are the most critical in Pull Requests. Learn more
2025-04-01T06:40:39.481597
2018-11-12T15:54:25
379839322
{ "authors": [ "Scottmar93", "tinosulzer" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11372", "repo": "tinosulzer/PyBaMM", "url": "https://github.com/tinosulzer/PyBaMM/issues/9" }
gharchive/issue
Add mesh and operators for spherical domain Summary Add (dimensionless?) spherical domain to mesh.py, independently of the existing Cartesian domain Add grad and div operators (Finite Volumes) for the spherical domain Mesh shall be unchanged. Will use if domain is "negative particle" or "positive particle" to determine when spherical operators should be used. Closed by #142
2025-04-01T06:40:39.517500
2024-02-22T11:11:57
2148816076
{ "authors": [ "geohot" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11373", "repo": "tinygrad/tinygrad", "url": "https://github.com/tinygrad/tinygrad/pull/3475" }
gharchive/pull-request
move all reduces to the end in lazy This is just an experiment. I don't like how the reduce LazyOp doesn't specify the actual reduce operation. Neither new_shape nor axis list expresses it fully, but rather depends on what shape the input buffer has. One idea is to have the arg just be the number of elements that are being reduced and move them to the end, but I'm not sure this is expressive enough for multireduce. Oh I think this actually might work! The final permute isn't really a permute, it's a reshape of only 1s. This is also already happening in the Linearizer
2025-04-01T06:40:39.530446
2021-04-03T10:48:45
849632155
{ "authors": [ "CLAassistant", "Mickls", "leonwanghui" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11374", "repo": "tinyms-ai/tinyms", "url": "https://github.com/tinyms-ai/tinyms/pull/66" }
gharchive/pull-request
fix: start_server不支持windows启动flask的问题 What type of PR is this? enhancement What does this PR do / why do we need it: The start_server method in tinyms does not support running in the windows environment. It is a huge blow to the friends who use the windows environment to learn tinyms. We have also seen that tinyms stated in the introduction that it supports the operation of the windows system, but it is a very uncomfortable thing that it cannot be used as one of its features. Which issue(s) this PR fixes: Fixes #65 Special notes for your reviewers: Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.Dragonfruit_JC seems not to be a GitHub user. You need a GitHub account to be able to sign the CLA. If you have already a GitHub account, please add the email address used for this commit to your account.You have signed the CLA already but the status is still pending? Let us recheck it. 在FlaskServer类中,linux_run_server方法中,有一行循环监听信号的代码,我给了一些注释,这个地方是我存疑的地方,希望你们看完谨慎合并 @lyd911 Please take a look, thanks!
2025-04-01T06:40:39.540473
2021-02-27T22:57:54
818050014
{ "authors": [ "drewroberts" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11375", "repo": "tipoff/laravel-google-api", "url": "https://github.com/tipoff/laravel-google-api/issues/8" }
gharchive/issue
GMB Accounts Multiple Accounts need to be present to implement data pulls through the Google My Business API (#7) for Locations (https://github.com/tipoff/locations/issues/42). This is also mentioned in this issue: https://github.com/tipoff/locations/issues/50 https://developers.google.com/my-business/reference/rest/v4/accounts https://support.google.com/business/answer/6085339?ref_topic=6085325
2025-04-01T06:40:39.568938
2024-02-06T21:31:26
2121699475
{ "authors": [ "KeisukeYamashita", "jackton1" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11376", "repo": "tj-actions/changed-files", "url": "https://github.com/tj-actions/changed-files/pull/1918" }
gharchive/pull-request
fix: broken link to matrix example workflow Why Because the link was broken, I got 404. It should be linked to this file → https://github.com/tj-actions/changed-files/blob/main/.github/workflows/matrix-example.yml @all-contributors please add @KeisukeYamashita for docs
2025-04-01T06:40:39.575426
2022-05-26T02:32:15
1249014101
{ "authors": [ "jackton1", "janzzen-tp" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11377", "repo": "tj-actions/eslint-changed-files", "url": "https://github.com/tj-actions/eslint-changed-files/issues/632" }
gharchive/issue
[BUG] - Warning: Unexpected input(s) 'extensions', Is there an existing issue for this? [X] I have searched the existing issues Does this issue exist in the latest version? [X] I'm using the latest release Describe the bug? To eslint-changed-files maintainers, First off, i would like to extend gratitude to the maintainers for creating this amazing workflow to automate the linting process. This workflow has become a great help to our team to ensure code readability and code quality. However, we found that there were some breaking changes from @v10 so we updated the package to @v11 after updating the latest v11, we have started to receive a warning that is blocking the github actions in the workflow please see the following snippet: 14s ##[debug]Evaluating condition for step: 'Run ESLint on changed files' ##[debug]Evaluating: success() ##[debug]Evaluating success: ##[debug]=> true ##[debug]Result: true ##[debug]Starting: Run ESLint on changed files ##[debug]Register post job cleanup for action: tj-actions/eslint-changed-files@v[1](https://github.com/fible/fible-app/runs/6603050516?check_suite_focus=true#step:6:1)1 ##[debug]Loading inputs ##[debug]Evaluating: github.token ##[debug]Evaluating Index: ##[debug]..Evaluating github: ##[debug]..=> Object ##[debug]..Evaluating String: ##[debug]..=> 'token' ##[debug]=> '***' ##[debug]Result: '***' Warning: Unexpected input(s) 'extensions', valid inputs are ['token', 'all_files', 'config_path', 'ignore_path', 'file_extensions', 'extra_args', 'skip_annotations', 'level', 'reporter', 'filter_mode', 'fail_on_error'] ##[debug]Loading env Run tj-actions/eslint-changed-files@v[11](https://github.com/fible/fible-app/runs/6603050516?check_suite_focus=true#step:6:11) with: config_path: .eslintrc.json extensions: js,jsx extra_args: --quiet token: *** all_files: false file_extensions: **/*.ts **/*.tsx **/*.js **/*.jsx With that said, would there be something missing the configuration that's causing this issue? Thanks for accommodating me. To Reproduce This is how the github action is written name: Run Linter on: pull_request: branches: - '**' jobs: run-linters: name: ESLint runs-on: ubuntu-latest steps: - uses: actions/checkout@v2 - uses: actions/setup-node@v1 with: node-version: 12.x - name: Install Packages run: yarn install - name: Run Jest run: yarn test - name: Run ESLint on changed files uses: tj-actions/eslint-changed-files@v11 with: config_path: ".eslintrc.json" extensions: "js,jsx" extra_args: "--quiet" What OS are you seeing the problem on? all Expected behavior? We are expecting the linter to run without warnings. Relevant log output 14s ##[debug]Evaluating condition for step: 'Run ESLint on changed files' ##[debug]Evaluating: success() ##[debug]Evaluating success: ##[debug]=> true ##[debug]Result: true ##[debug]Starting: Run ESLint on changed files ##[debug]Register post job cleanup for action: tj-actions/eslint-changed-files@v[1](https://github.com/fible/fible-app/runs/6603050516?check_suite_focus=true#step:6:1)1 ##[debug]Loading inputs ##[debug]Evaluating: github.token ##[debug]Evaluating Index: ##[debug]..Evaluating github: ##[debug]..=> Object ##[debug]..Evaluating String: ##[debug]..=> 'token' ##[debug]=> '***' ##[debug]Result: '***' Warning: Unexpected input(s) 'extensions', valid inputs are ['token', 'all_files', 'config_path', 'ignore_path', 'file_extensions', 'extra_args', 'skip_annotations', 'level', 'reporter', 'filter_mode', 'fail_on_error'] ##[debug]Loading env Run tj-actions/eslint-changed-files@v[11](https://github.com/fible/fible-app/runs/6603050516?check_suite_focus=true#step:6:11) with: config_path: .eslintrc.json extensions: js,jsx extra_args: --quiet token: *** all_files: false file_extensions: **/*.ts **/*.tsx **/*.js **/*.jsx Anything else? No response Code of Conduct [X] I agree to follow this project's Code of Conduct @janzzen-tp I’ve added a Migration Guide to help you resolve the warning. Let me know if you need more information but a high-level description of the warning is the extension input is no longer supported and has now been renamed to file_extensions.
2025-04-01T06:40:39.654804
2021-11-08T08:49:48
1047173901
{ "authors": [ "schlichtanders", "tlienart" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11381", "repo": "tlienart/Franklin.jl", "url": "https://github.com/tlienart/Franklin.jl/issues/922" }
gharchive/issue
inline svg url not working with Franklin because changed by compression On Franklin server everything just works, but when deploying to gitlab, my background heropatterns (svg images inlined into background-image: url(...)) just don't show up. Inspecting the issue it turns out that something in the pipeline deleted the surrounding quotes. Unbelievable... So on my local system I have background-image: url("data:image/svg+xml,..."); which somehow ends up on gitlab pages as background-image: url(data:image/svg+xml,...); no surrounding quotes any longer. Any help is highly appreciated, as this seems to be due to the standard Franklin gitlab pipeline. Turning off minify fixed this issue as well. See https://github.com/tlienart/Franklin.jl/issues/921#issuecomment-962938930 I would vote for turning off minify by default for everyone. yes so as per other issue, minification will be turned off for everyone (prerendering as well btw even though it causes much fewer issues)
2025-04-01T06:40:39.657092
2019-08-20T06:06:52
482661818
{ "authors": [ "Zeus64", "tlinkowski" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11382", "repo": "tlinkowski/basic-annotations", "url": "https://github.com/tlinkowski/basic-annotations/issues/15" }
gharchive/issue
Compilation warning about Kotlin's AnnotationTarget enum The following warning occurs when compiling with basic-annotations in compileOnly scope: warning: unknown enum constant AnnotationTarget.ANNOTATION_CLASS reason: class file for kotlin.annotation.AnnotationTarget not found This happens e.g. when compiling sample-java-api subproject: https://travis-ci.com/tlinkowski/basic-annotations/builds/123617322#L278-L279 Caused by KT-33141. I have the same problem
2025-04-01T06:40:39.658447
2022-12-09T21:41:07
1487490763
{ "authors": [ "bazfum", "riblus-bandyr" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11383", "repo": "tloncorp/landscape-apps", "url": "https://github.com/tloncorp/landscape-apps/issues/1413" }
gharchive/issue
profiles: unexpected behavior in sigil color selection When attempting to edit the sigil color field in the profile settings, it truncates leading zeroes and then pads out the last character. For example, I previously had my color set to 0000b2. After the upgrade, it had converted it to B22222. Attempting to change it back results in the same thing after a refresh. 1000b2 however works. can confirm this This is fixed in Groups and Talk, but the Grid still shows the incorrectly padded color.
2025-04-01T06:40:39.675550
2024-07-22T18:33:37
2423524425
{ "authors": [ "BrendonPierson", "tlux" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11384", "repo": "tlux/sftp_client", "url": "https://github.com/tlux/sftp_client/pull/36" }
gharchive/pull-request
Added preferred_algorithms and packet_size config options Adding two more config options that get passed to the underlying erlang ssh connect: preferred_algorithms and packet_size Thank you for adding that! @BrendonPierson Could you possibly also update config_test.exs to reflect those changes? Thank you. Sure thing @tlux! Let me know if this looks ok
2025-04-01T06:40:39.679185
2018-01-08T08:09:11
286672279
{ "authors": [ "aipsel", "jimlongo56", "nubilfi", "tmcw" ], "license": "isc", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11385", "repo": "tmcw/docbox", "url": "https://github.com/tmcw/docbox/issues/64" }
gharchive/issue
After run build script, docbox css broken Hi, I'm new to docbox and i have a little problem. Every time i run npm run buildscript it always ends up with a broken CSS, i'm using the default provided CSS, and when i inspect it, i got this error: Warning: Prop className did not match. Server: "fixed-top fixed-right space-left16" Client: "fixed-top fixed-right undefined" Also, I found that revert the index.html file can solve this problem. This would be a great opportunity for someone to help and fix and contribute a PR! I'd love to review and merge such a PR. Can confirm that changing package.json "react": "15.6.2", "react-dom": "15.6.2", resolves the build issue for me.
2025-04-01T06:40:39.692494
2018-08-09T19:32:23
349260509
{ "authors": [ "SquidDev", "zardyh" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11386", "repo": "tmpim/amulet", "url": "https://github.com/tmpim/amulet/pull/84" }
gharchive/pull-request
Remove main We never did anything interesting with main, and our calling of it was rather broken, so it's probably better to remove it. People can now write this instead: let () = f x This is closer to how other MLs handle it, which is always nice. We can also remove some pretty ugly code from the backend. The main disadvantage here is that it breaks a large number of our codegen tests. We used to use main to inject opaque values. As this now has no special meaning, it'll be stripped and so the test won't emit anything! For now, I've fixed this by calling main with a bottom : 'a value: let main f = ... external val bottom : 'a = "nil" let () = bottom (main bottom) This definitely is pretty ugly (and can generate some awful code), but I'm OK with this for now. Hopefully the backend rewrite (which is increasingly reaching the status of Urn's resolver rewrite) will clean up a lot of these. Looks good to me™
2025-04-01T06:40:39.716863
2017-07-19T10:58:14
244001745
{ "authors": [ "fabianre", "ygeelen-tb" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11387", "repo": "tobania/VSTS.Extension.SqlReportingServices", "url": "https://github.com/tobania/VSTS.Extension.SqlReportingServices/issues/15" }
gharchive/issue
No Release task Hi, I am using the Visual Studio Team Services and added your extension. Unfortunately I am not able to find the task when creating a new build sequence. Did I miss something ? Thanks, Fabian Update: BTW: Just saw that its clearly writtenin the description, shame on me... Unfortunately the release task is also not there. Fabian, It's a task for Release Definitions, not Build Definitions. Thanks for your fast reply, that make sense. But should I not see it then here ? (Sorry for those dumb questions, just startet to take a look at the whole Team Services topic today) Yes you should see it here if the extension is properly installed (the "Deploy" category): Hmm unfortunately it does not. Any idea what I can do to resolve it ? I already uninstalled and installed it again without luck. After uninstall/install again, the tasks now show up. Strange, but luckily sovled. We noticed some temporary issue on VSTS or Azure on the extension page. I guess that was it..
2025-04-01T06:40:39.724206
2023-10-04T15:14:18
1926457436
{ "authors": [ "redbar0n", "sahidrahman404" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11388", "repo": "tobias-tengler/create-relay-app", "url": "https://github.com/tobias-tengler/create-relay-app/issues/18" }
gharchive/issue
Support for Vike/vite-plugin-ssr It would be nice to have integration between create-relay-app and Vike/vite-plugin-ssr since create-relay-app currently works with Vite but not with Vike/vite-plugin-ssr in relation to this, I think create-relay-app would fit nicely into the Bati.js scaffolder for new projects: https://github.com/batijs/bati
2025-04-01T06:40:39.747483
2023-03-24T21:42:18
1640080533
{ "authors": [ "TomWagg", "tobin-wainer" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11389", "repo": "tobin-wainer/elk", "url": "https://github.com/tobin-wainer/elk/issues/62" }
gharchive/issue
Try giving a list of potential stars that are leading to a given frequency spike Use the individual pixel periodograms to find likely locations and then use astroquery to search SIMBAD and find some likely candidates Need to flip the way the pixels are read into the wcs... if the argwhere returns [48, 62] what we need to put into the pixel_to_world is (62, 48) If I do "tess_wcs.pixel_to_world(*np.fliplr(solutions[0])).icrs" I get the correct thing, but if there are multiple solutions and I do: tess_wcs.pixel_to_world(*np.fliplr(solutions)).icrs then its wrong again.... Ah gotcha so the (x, y) is just backwards? Perhaps we just need *query_pixels.T[::-1] in that case (which will reverse the x-y axes) I tried that as well... but that reverses the entire array, not the individual pairs
2025-04-01T06:40:39.797907
2024-12-02T02:53:34
2710267551
{ "authors": [ "scala-steward" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11390", "repo": "tofu-tf/typed-schema", "url": "https://github.com/tofu-tf/typed-schema/pull/916" }
gharchive/pull-request
Update sbt, scripted-plugin to 1.10.6 About this PR 📦 Updates org.scala-sbt:sbt org.scala-sbt:scripted-plugin from 1.6.2 to 1.10.6 📜 GitHub Release Notes - Version Diff Usage ✅ Please merge! I'll automatically update this PR to resolve conflicts as long as you don't change it yourself. If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below. Configure Scala Steward for your repository with a .scala-steward.conf file. Have a fantastic day writing Scala! ⚙ Adjust future updates Add this to your .scala-steward.conf file to ignore future updates of this dependency: updates.ignore = [ { groupId = "org.scala-sbt" } ] Or, add this to slow down future updates of this dependency: dependencyOverrides = [{ pullRequests = { frequency = "30 days" }, dependency = { groupId = "org.scala-sbt" } }] labels: library-update, early-semver-minor, semver-spec-minor, version-scheme:early-semver, commit-count:1 Superseded by #918.
2025-04-01T06:40:39.801300
2024-05-15T16:22:44
2298348263
{ "authors": [ "creeder-uturn" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11391", "repo": "tofuutils/tenv", "url": "https://github.com/tofuutils/tenv/issues/133" }
gharchive/issue
Document what priority/order is used for version selection Right now, I cannot find clear documentation on which order version detection uses. There is a decent listing of which things can be looked at, but it's not immediately obvious which have priority. It'd be great if the documentation have a section that clearly indicated that order. If I've missed it feel free to clown on me 😄 It seems like the following order is used, but I've not tested thoroughly: Project version file Version file in Project Root (Possibly one anywhere in the tree above the project in order?) Default system-wide version defined Notably missing: the requirements listed in the TF files, which can be installed/used with latest-allowed and min-required but it doesn't seem like this is used automatically as a fallback? Oh, awesome! I didn't see the version resolution order. That's precisely what I was looking for. I would propose sending that information into a digestible section with version files, or perhaps it's own section. Weary to make promises I may not deliver but I will probably take a stab at a PR with what I imagine (and I can try to clarify better, too).
2025-04-01T06:40:39.804564
2016-01-14T20:36:28
126742890
{ "authors": [ "IndrekV", "oliverdunk" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11392", "repo": "toggl/toggl-button", "url": "https://github.com/toggl/toggl-button/issues/421" }
gharchive/issue
Red border? Had an idea, Thought gathering some feedback before actually implementing it might make some sense. How about an option which could be enabled, that would add a red border around every page if you were timing? For some people this may be annoying, but I feel for others that it'd be a good reminder that they were still being tracked. Open to ideas, and if you think this is terrible, be honest! I can take it. I think it's too "In your face" to have red border on all the pages. I like the idea of always being aware that time is tracking, but I think the red border is too much. Right now the red color of the extension icon is indicating that timer is running.
2025-04-01T06:40:39.805948
2016-10-10T08:18:28
181958278
{ "authors": [ "IndrekV" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11393", "repo": "toggl/toggldesktop", "url": "https://github.com/toggl/toggldesktop/issues/2011" }
gharchive/issue
Ask users to who are on broken updater version to get latest update Some users are on Macos version that has broken updater. Send them email to update to latest. Broken updater version should be 7.2.150 This is probably outdated. Double check if anyone is still on this version and then close this. No users on this ancient version.
2025-04-01T06:40:39.808688
2019-05-29T08:30:23
449668098
{ "authors": [ "IndrekV", "NghiaTranUIT" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11394", "repo": "toggl/toggldesktop", "url": "https://github.com/toggl/toggldesktop/issues/2986" }
gharchive/issue
Handle cases where project is unsynced 💻 Environment Platform: All (library) 📒 Description In some cases, the project or any other item could be not synced and also not marked as "needs sync". I saw this with an un-synced project that did not have an id (it was NULL), the project was visible in the autocomplete but selecting it would clear the project selection. First of selecting a non-synced project should still work Secondly, the project that is not synced should be marked as to be synced when loaded I wonder that old implementation doesn't store any project_guid after creating, how it handles if the project remains unsync? I don't see any complain about the unsync project in old code, it means that it always works, isn't it? Ref: https://github.com/toggl/toggldesktop/blob/b5064ec91eafd104b1d1c7d6baab59da3fe5db42/src/ui/osx/TogglDesktop/TimeEntryEditViewController.m#L396-L421
2025-04-01T06:40:39.812157
2023-08-11T20:45:07
1847408394
{ "authors": [ "TheSoloJake" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11395", "repo": "toggl/track-extension", "url": "https://github.com/toggl/track-extension/issues/2235" }
gharchive/issue
Todoist - Cannot Enter Project in Task View OS version: macOS latest Browser version: Chrome latest Extension version: 3.0.14 Relevant integration (if any): Todoist webapp 🐛 Describe the bug User reports that they cannot type into the project window in Toggl when a Todoist task view window is open. Support team member can replicate as well. Expected behaviour Project window should be available for typing. Other details or context Slack convo Intercom convo Intercom convo listed above also contains screenshot video of bug in action
2025-04-01T06:40:39.818343
2020-10-06T12:46:49
715641446
{ "authors": [ "LuanFMelo", "arbutt76", "loubyl", "tdardet", "togiberlin", "uniAmir" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11396", "repo": "togiberlin/ui-ux-designer-roadmap", "url": "https://github.com/togiberlin/ui-ux-designer-roadmap/issues/11" }
gharchive/issue
Update for 2020? Is the plan still to update for 2020's landscape, or is this repo inactive for the time being? @tdardet I am currently interviewing with several companies. My plan is to migrate and rewrite some parts in GitBooks towards the end of this year. Ah awesome, Good luck! Tomás Dardet On Sun, Oct 11, 2020 at 1:22 PM togiberlin<EMAIL_ADDRESS>wrote: @tdardet https://github.com/tdardet I am currently interviewing with several companies. My plan is to migrate and rewrite some parts in GitBooks towards the end of this year. — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/togiberlin/ui-ux-designer-roadmap/issues/11#issuecomment-706737909, or unsubscribe https://github.com/notifications/unsubscribe-auth/ABEJ3RLVA44AVUML5XAP3A3SKHSWZANCNFSM4SF7HRKA . Hi! any chance that you could add in the future a Product Designer roadmap? Maybe not in 2020 but in any update after? Thanks! Any update? Any update? Waiting for your approval Waiting for your approval Some updates ?
2025-04-01T06:40:39.820290
2015-12-01T17:25:59
119767063
{ "authors": [ "thomcc", "toji" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11397", "repo": "toji/gl-matrix", "url": "https://github.com/toji/gl-matrix/pull/194" }
gharchive/pull-request
Implement set and fromValues for mat2, mat2d, mat3, and mat4 Was surprised to find out that these did not already exist. Let me know if there are any problems. LGTM, thanks! And thanks for giving me something to test the new Travis CI config with! :D
2025-04-01T06:40:39.900114
2023-01-12T16:00:50
1530972100
{ "authors": [ "seanmonstar", "semtexzv" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11398", "repo": "tokio-rs/bytes", "url": "https://github.com/tokio-rs/bytes/issues/586" }
gharchive/issue
Add *_unchecked methods I'm using bytes in an application, where I'm repeatedly slicing to sub-parts of the bytes. I do my own error-checking of bounds, and handle that out of the hot path. The methods of (advance, split_off, split_to) do additional bound checking, completely trashing the register allocation. It'd be nice to have the option to call *_unchecked methods that avoid the bound checking. Have you been able to try those methods in your app, with a temporary fork? It'd be good to get some data.
2025-04-01T06:40:39.901010
2019-05-26T18:45:22
448594808
{ "authors": [ "MSleepyPanda" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11399", "repo": "tokio-rs/console", "url": "https://github.com/tokio-rs/console/issues/10" }
gharchive/issue
docs: user guides Write documentation on how to debug applications with the console, how to use it #3 Would allow us to generate screenshots
2025-04-01T06:40:39.925321
2015-06-23T05:33:29
90292607
{ "authors": [ "furueili", "tolo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11400", "repo": "tolo/HHServices", "url": "https://github.com/tolo/HHServices/issues/14" }
gharchive/issue
port number error hi~ In serviceDidResolve delegate method, I saw the service data as follows HHService[0x742E3580, MAC000E53290B23, _afpovertcp._tcp, local., MAC000E53290B23.local., ( "<IP_ADDRESS>:88" ), 1] - Resolved last address and I try to print the port number NSLog(@"service port number :%i",service.lastResolvedPort); what I got is service port number :22528 please help thx Hi, The lastResolvedPort property is an internal property containing a (temporary) port number expressed in network byte order, which is why you get a strange result when you attempt to print it out. But I've actually just checked in an update to HHService in which I've added new convenience properties, for instance resolvedInetAddresses, which contains a list of resolved addresses on the form IP:PORT.
2025-04-01T06:40:39.952243
2022-11-12T01:53:12
1446170965
{ "authors": [ "tomaae", "wormuths" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11401", "repo": "tomaae/homeassistant-truenas", "url": "https://github.com/tomaae/homeassistant-truenas/issues/54" }
gharchive/issue
[Bug] No Error, Just Fails Describe the issue How to reproduce the issue Expected behavior Screenshots Software versions Home Assistant version: TrueNAS integration version: TrueNAS version: Diagnostics data Traceback/Error logs Additional context I cant do anything without any info. Make sure to fill the issue template with as much info as there is. Sorry about that. There really isn't much information to tell. I follow the instructions perfectly, and I get no error message. It just times out or something. What would be most useful to start troubleshooting this? I'm not sure what makes it fail, but I'd love to get it working. versions, screenshots, diagnostic data, HA log. just provide everything you can. I just noticed the information isn't there on the bug report. I put all that in when I created it... Home Assistant 2022.11.2 TrueNAS Core 13.0-RELEASE Integration v1.2.3 The integration installs, and I clicked add... All I get is this... I've tried all sorts of various combinations of SSL, https:// in the URL, etc... Same result. If I put "https:" in the URL i get a "No response from host" error. My firewall logs don't show anything being blocked, but HASS shows connection errors... ====================================== This error originated from a custom integration. Logger: custom_components.truenas.truenas_api Source: custom_components/truenas/truenas_api.py:44 Integration: TrueNAS (documentation, issues) First occurred: 12:15:53 PM (4 occurrences) Last logged: 12:44:19 PM TrueNAS <IP_ADDRESS> unable to fetch data "pool" (503) TrueNAS https://<IP_ADDRESS> unable to fetch data "pool" (no_response) ======================================= This error originated from a custom integration. Logger: custom_components.truenas.config_flow Source: custom_components/truenas/config_flow.py:76 Integration: TrueNAS (documentation, issues) First occurred: 12:15:53 PM (4 occurrences) Last logged: 12:44:19 PM TrueNAS connection error (503) TrueNAS connection error (no_response) ======================================= I'm still tinkering in the firewall, but I haven't found any rule which may be blocking this from connecting... Thanks, Steve I have never tried it with 13.0 release version, there could have been some bug. 503 is service unavailable error. seems like there is something broken on truenas side. I would recommend to actually update your truenas. Good day, That would not surprise me. I have found there is another issue with a networking configuration I have reported to them. It prevents me from updating at the moment. My TrueNAS works fine on 13.0, but if I update beyond that all my jails lose internet connectivity. Maybe I'm just special. LOL https://ixsystems.atlassian.net/browse/NAS-118026 Something under the hood with networking did change along the way with them. The 503... is there a specific service your integration relies on being enabled on the TrueNAS side to check? Steven Do they really lose internet connectivity or just DNS? I had such issue with truenas. Solved it by using resolv.conf syntax in configuration field: Cant check the ticket, appearantly I dont have access to view it. I'm accessing official API, I dont think you can control it in any way (officially). Yeah, sorry. DNS. yea, you can try to put it into resolver like I did. DNS server entries in jail configuration never worked for me, even on 12.0
2025-04-01T06:40:39.962346
2016-11-02T11:34:31
186776984
{ "authors": [ "fschutt", "sfackler", "tomaka", "wvdschel" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11402", "repo": "tomaka/android-rs-glue", "url": "https://github.com/tomaka/android-rs-glue/pull/116" }
gharchive/pull-request
Initial JNI support Feedback welcome! Example code using these JNI bindings can be seen in https://github.com/tomaka/glutin/pull/822 Here's a gist @sfackler made demonstrating a high-level API for JNI stuff: https://gist.github.com/sfackler/5c036006f8d58a1d38887ba014759f74 It is licensed under MIT/Apache2 so we can copy some stuff. As I said in the previous comment, this might be out of scope of this PR but I thought I'd leave it somewhere for reference. Might also be worth depending on https://crates.io/crates/jni-sys to avoid needing to have a second copy in here. Happy to make any changes necessary on that end. @tomaka, what is your view on using @sfackler's jni-sys crate? I'd prefer not to re-implement anything that we can readily reuse, but I'm not sure if you're happy to be adding another dependency. Before I start addressing other feedback, it would be nice to decide if using jni-sys is on the table, and if we should add higher level wrappers to android-rs-glue, jni-sys, or am intermediate crate that wraps raw jni-sys in a nice, safe, Rust-y API. I agree that adding a higher level API (like JNIEnv's that automatically detach_thread(), or jobject's that automatically DeleteGlobalRef/DeleteLocalRef when dropped). Most of all, I'd like to get rid of the various explicit method calls based on types. But I'm also not sure this should be part of android-rs-glue. An alternative approach would be to remove the JNI API from the glue module, and simply using the four methods provided by this patch in injected-glue from jni-sys, guarded by some #![cfg(target_os = "android")]. In my opinion this is the ideal design: The internal and the external glues both depend on jni-sys and communicate with each other to share the low-level objects (like you have done in your PR). The external glue also depends on a new crate (named jni for example) that doesn't exist yet and that provides a high-level wrapper around the JNI. The high-level wrappers are reexported in the external glue's API, but the low-level objects are not exposed. I can agree with that. I'll look into making the adaptations over the weekend, probably. The higher level JNI wrapper will probably take a longer time to complete. @sfackler Are you interested in creating a repo for the higher level wrapper, or should I? I probably don't have the time to build out the high level wrapper right now unfortunately, so you might want to get it started. Just a note that I have started on reworking this, but haven't found the time to finish it yet (because life). So ... what is the current state? The jni_sys crate is pretty mature now and it's been two years ... Why can't the ffi module simply be exposed to the library user? This way the user can hook up the jni crate and from there on it's rather easy to interact with the JNI. I know about "API concerns", but .. something is better than nothing. @tomaka ... why? What's wrong about the jni crate? I mean ... do you want to re-do all of the work they've done just to not depend on them - how would you then do this function, if not using jni? Sorry, what I mean is that we should simply reexport the content of the jni-sys crate from our own crate, instead of exposing it in the API.
2025-04-01T06:40:39.966211
2015-12-10T18:22:57
121547745
{ "authors": [ "ReadmeCritic" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11403", "repo": "tomaka/android-rs-glue", "url": "https://github.com/tomaka/android-rs-glue/pull/71" }
gharchive/pull-request
Update README URLs based on HTTP redirects Created with https://github.com/dkhamsing/frankenstein GitHub Corrected URLs Was Now https://github.com/tomaka/gl-init-rs https://github.com/tomaka/glutin HTTPS Corrected URLs Was Now http://crates.io/config.html https://crates.io/config.html :tada: :fr:
2025-04-01T06:40:40.024432
2017-04-30T21:05:48
225354371
{ "authors": [ "alecsandruchirosca", "grossmane", "kiliant", "tomav" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11405", "repo": "tomav/docker-mailserver", "url": "https://github.com/tomav/docker-mailserver/issues/586" }
gharchive/issue
Cannot start server Dear all, I tried for several hours to get this image working and still have the same error: docker log docker-mailserver Initializing setup Checking configuration Configuring mail server No DKIM key provided. Check the documentation to find how to get your keys. Starting Misc Starting mail server Error: Please fix your configuration. Exiting... docker-compose.yml version: '2' services: mail: build: docker-mailserver hostname: xxxxxx domainname: xxxx.xx container_name: mail ports: - "25:25" - "143:143" - "587:587" - "993:993" volumes: - ./maildata/:/var/mail - ./mailstate/:/var/mail-state - ./letsencrtypt/:/etc/letsencrypt/ - ./config/:/tmp/docker-mailserver/ environment: - ENABLE_SPAMASSASSIN=1 - ENABLE_CLAMAV=1 - ENABLE_FAIL2BAN=1 - ENABLE_POSTGREY=1 - ONE_DIR=1 - DMS_DEBUG=0 - SSL_TYPE=letsencrypt cap_add: - NET_ADMIN volumes: maildata: driver: local mailstate: driver: local The docker-mailserver directory contains the git clone of the project. Using the suggested image ( tvial/docker-mailserver:latest) yelds the same results. Can you please help me start this image? BR, Alecs Hi @alecsandruchirosca, to have more detail, use DMS_DEBUG=1, restart container and paste the console log. `# ENV HOSTNAME=xxxxxx.xxxx.xx DMS_DEBUG=1 VIRUSMAILS_DELETE_DELAY=7 ENABLE_CLAMAV=1 PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin ENABLE_FAIL2BAN=1 ENABLE_SPAMASSASSIN=1 PWD=/ SHLVL=1 HOME=/root ONE_DIR=1 SSL_TYPE=letsencrypt DEBIAN_FRONTEND=noninteractive ENABLE_POSTGREY=1 _=/usr/bin/printenv docker-mailserver Initializing setup Registering check,setup,fix,misc and start-daemons functions _check_environment_variables() registered _check_hostname() registered _setup_default_vars() registered _setup_dovecot() registered _setup_dovecot_local_user() registered _setup_postgrey() registered _setup_dkim() registered _setup_ssl() registered _setup_docker_permit() registered _setup_mailname() registered _setup_amavis() registered _setup_dmarc_hostname() registered _setup_postfix_hostname() registered _setup_dovecot_hostname() registered _setup_postfix_sasl() registered _setup_postfix_override_configuration() registered _setup_postfix_sasl_password() registered _setup_security_stack() registered _setup_postfix_aliases() registered _setup_postfix_vhost() registered _setup_environment() registered _fix_var_mail_permissions() registered _fix_var_amavis_permissions() registered _misc_save_states() registered _start_daemons_cron() registered _start_daemons_rsyslog() registered _start_daemons_dovecot() registered _start_daemons_opendkim() registered _start_daemons_opendmarc() registered _start_daemons_postgrey() registered _start_daemons_postfix() registered _start_daemons_fail2ban() registered _start_daemons_clamav() registered _start_daemons_amavis() registered Checking configuration Check that there are no conflicts with env variables [_check_environment_variables] Check that hostname/domainname is provided or overidden (no default docker hostname/kubernetes) [_check_hostname] Domain has been set to xxxx.xx Hostname has been set to xxxxxx.xxxx.xx Configuring mail server Setting up default variables [_setup_default_vars] Set ENABLE_LDAP=0 Set ENABLE_FETCHMAIL=0 Set OVERRIDE_HOSTNAME= Set ENABLE_MANAGESIEVE=0 Set DMS_DEBUG=1 Set ENABLE_POP3=0 Set ENABLE_SASLAUTHD=0 Set ENABLE_CLAMAV=1 Set SMTP_ONLY=0 Set ENABLE_FAIL2BAN=1 Set ENABLE_SPAMASSASSIN=1 Set POSTGREY_DELAY=300 Set POSTGREY_TEXT=Delayed by postgrey Set ENABLE_POSTGREY=1 Set POSTGREY_MAX_AGE=35 Setting up Dovecot Setting up Dovecot Local User 'config/docker-mailserver/postfix-accounts.cf' is not provided. No mail account created. Configuring postgrey Setting up DKIM No DKIM key provided. Check the documentation to find how to get your keys. Setting up SSL Setting up PERMIT_DOCKER Option Adding container ip in my networks Setting up Mailname Creating /etc/mailname Setting up Amavis Applying hostname to /etc/amavis/conf.d/05-node_id Setting up dmarc Applying hostname to /etc/opendmarc.conf Applying hostname and domainname to Postfix Applying hostname to /etc/postfix/main.cf Applying hostname to Dovecot Applying hostname to /etc/dovecot/conf.d/15-lda.conf Setting up Postfix Override configuration No extra postfix settings loaded because optional '/tmp/docker-mailserver/postfix-main.cf' not provided. Setting up Postfix SASL Password Warning: 'SASL_PASSWD' is not provided. /etc/postfix/sasl_passwd not created. Setting up Security Stack Enabling and configuring spamassassin Enabling clamav Fail2ban enabled Setting up Postfix Aliases Warning 'config/postfix-virtual.cf' is not provided. No mail alias/forward created. Setting up Postfix vhost Setting up /etc/environment Checking /var/mail permissions Permissions in /var/mail look OK Checking $amavis_state_dir permissions Permissions in /var/mail-state/lib-amavis look OK Starting Misc Consolidating all state onto /var/mail-state Destination /var/mail-state/spool-postfix exists, linking /var/spool/postfix to it Destination /var/mail-state/lib-postfix exists, linking /var/lib/postfix to it Destination /var/mail-state/lib-amavis exists, linking /var/lib/amavis to it Destination /var/mail-state/lib-clamav exists, linking /var/lib/clamav to it Destination /var/mail-state/lib-spamassasin exists, linking /var/lib/spamassasin to it Destination /var/mail-state/lib-fail2ban exists, linking /var/lib/fail2ban to it Destination /var/mail-state/lib-postgrey exists, linking /var/lib/postgrey to it Starting mail server Starting cron [ OK ] Starting rsyslog [ OK ] Starting dovecot services [ OK ] Starting opendkimfalse [ FAILED ] Error: Please fix your configuration. Exiting... ` With DMS_DEBUG I got the above log (docker log mail ). BR, Alecs Hi, i have exact the same failure. if i downgrade to docker-mailserver 2.2, then opendkim will be started, but then the error comes during start of amavis. Output is then: ... Starting amavisfalse [ FAILED ] Error: Please fix your configuration. Exiting... I just upgraded a server to 2.3. No issue with DKIM so far. Only Postgrey was not running. 3 other error to solve: May 6 12:36:14 mail postgrey[3623]: FATAL: ERROR: can't open lock file: /var/lib/postgrey/postgrey.lock#012 May 6 12:36:15 mail postfix[4414]: Postfix is running with backwards-compatible default settings May 6 12:36:15 mail postfix[4414]: See http://www.postfix.org/COMPATIBILITY_README.html for details May 6 12:36:15 mail postfix[4414]: To disable backwards compatibility use "postconf compatibility_level=2" and "postfix reload" May 6 12:36:15 mail postfix/master[4451]: fatal: open lock file /var/lib/postfix/master.lock: cannot create file exclusively: Permission denied Tried manually (and had to disable ONE_DIRbut after that it worked well). No idea yet yo a long term patch. Ideas welcome. Maybe this also relates to #545. Although it works for me just well, I also receive backwards-compatibility warnings in my logs. Will have a look at this later. Duplicates https://github.com/tomav/docker-mailserver/issues/545 Move this discussion there. @alecsandruchirosca you had this error because no user account has been provided: 'config/docker-mailserver/postfix-accounts.cf' is not provided. No mail account created. is debug log. I reproduced this issue today, I'll add a fix. Here is the PR https://github.com/tomav/docker-mailserver/pull/612 :latest image available. #491 & #621
2025-04-01T06:40:40.028873
2017-03-19T19:47:29
215289233
{ "authors": [ "alinmear", "tomav" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11406", "repo": "tomav/docker-mailserver", "url": "https://github.com/tomav/docker-mailserver/pull/556" }
gharchive/pull-request
Fix #526: fatal: no SASL authentication mechanisms When using the container with SMTP_ONLY = 1, the postfix process within the container fails on ehlo because there is no valid sasl authentication mechanism available. This happens because sasl has been enabled within postfix/main.cf per default but sasl is not configured. To fix this _setup_postfix_sasl does not depend anymore on ENABLE_SASLAUTHD and will check in it's logic, whether to enable or disable sasl explicit within postfix/main.cf. I think you should add an integration test for that. Added integration test Just to be accurate on this issue, i just coverd the testing for this particular problem, checking whether sending an email is working or not. But i think we should cover more test cases for the mailserver state SMTP_ONLY use-cases. Here some examples: Check Mail send is working Check SASL authentication is working Check Relay is working Check Relay with sasl authentication is working Maybe we could create a section for this.
2025-04-01T06:40:40.033420
2014-07-15T18:01:24
37909360
{ "authors": [ "joeljeske", "markitosgv", "nmccready", "rasmi", "tombatossals" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11407", "repo": "tombatossals/angular-leaflet-directive", "url": "https://github.com/tombatossals/angular-leaflet-directive/issues/421" }
gharchive/issue
Listen markercluster click event Its possible to acces clusterclick, without overlays?? My initial map: //map initial angular.extend($scope, { center: { autoDiscover: true }, defaults: { scrollWheelZoom: false }, layers: { baselayers: { osm: { name: 'OpenStreetMap', url: 'http://{s}.tile.openstreetmap.org/{z}/{x}/{y}.png', type: 'xyz' } } } }); And then i call my own service to fecth markers data: //on load map $scope.$on('leafletDirectiveMap.load', function(event){ api.getMarkersByProximity(function(data){ $scope.markers.push(data); }); }); Now i have a map with markers and clusters. With this event i can listen when a marker is clicked, but not a cluster: $scope.$on("leafletDirectiveMarkersClick", function(event, args){ console.log($scope.markers[args]); }); Its possible to do?? thanks I don´t want when cluster marker is clicked to expand child markers... i want to get child markers data to do another thing. On official Leaflet.markercluster docs says: markers.on('clusterclick', function (a) { console.log('cluster ' + a.layer.getAllChildMarkers().length); }); Is any event directive to listen? @markitosgv, have you found a way to listen for cluster events? Has anyone found any ways to listen besides dipping down into pure Leaflet (like @joeljeske did in #648)? This issue was moved to angular-ui/ui-leaflet#31 I'm going to rework&redesign angular-leaflet-directive to be compatible with Leaflet v1.0. It will mantain almost all its functionality, and will be compatible with the current features of the directive, but I must start from a fresh point, so I'm going to close this issue. If you think it must be worked with the new version, please reopen it.
2025-04-01T06:40:40.035038
2016-06-16T07:32:39
160595720
{ "authors": [ "khoerling" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11408", "repo": "tomchentw/react-google-maps", "url": "https://github.com/tomchentw/react-google-maps/issues/285" }
gharchive/issue
React 15.1.x Support Is this on the horizon? On second look, it's working excellently on the latest React! (closing)
2025-04-01T06:40:40.070608
2022-07-09T20:03:05
1299757087
{ "authors": [ "awvwgk", "gnikit" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11409", "repo": "toml-f/toml-f", "url": "https://github.com/toml-f/toml-f/issues/90" }
gharchive/issue
Support creating machine readable diagnostics Raised by Giannis @gnikit in https://fortran-lang.discourse.group/t/3949/2 Adding an option or slightly restructuring the diagnostic message to be easily parseable via regex. The message structure is already very good, but fetching the error message via regex would be hard. A good example of that is gfortran-11>= with the flag -fdiagnostic-plain-output. Having an such option would allow for code editors to parse the output of the linter to their Diagnostics console in VS Code PROBLEMS tab. The current report interface turns an index of a token to a label object https://github.com/toml-f/toml-f/blob/225174f7a858186507e7bd08c4252918dca9dee0/src/tomlf/de/context.f90#L87-L89 And creates a diagnostic object from it https://github.com/toml-f/toml-f/blob/225174f7a858186507e7bd08c4252918dca9dee0/src/tomlf/de/context.f90#L92-L96 For the human facing output this is than turned into a string at https://github.com/toml-f/toml-f/blob/225174f7a858186507e7bd08c4252918dca9dee0/src/tomlf/de/context.f90#L101 Note that the actual source code is only needed when creating the report string, the diagnostic object itself contains only position information from the label objects as well as the messages to display. To support this without much effort on the user side, we could add a state in the context objects which describes whether the report should be optimized for humans or machines, default to human-friendly output. A tool which wants integrate its TOML Fortran usage for error reporting with the the VS Modern Fortran Extension can provide an option or environment variable to toggle this switch in the context object and make the error output automatically accessible for the Diagnostic console in VS Code. The actual preferred format for the VS Modern Fortran Extension has to be defined first. I think I might have been to quick to the current diagnostic message format is okay. I just whipped up a REGEX that should work with VS Code's problemmatcher https://regex101.com/r/q6duAL/4 and in any other tool that uses ECMAScript (Javascript) flavour for regex The online limitation that I can think of with the current output format, is that the diagnostics are multiline and not all code editors/ linter plugins are able to capture multiline regexs. FYI VS Code is fine with that: https://code.visualstudio.com/Docs/editor/tasks#_defining-a-multiline-problem-matcher I just whipped up a REGEX that should work with VS Code's problemmatcher regex101.com/r/q6duAL/4 and in any other tool that uses ECMAScript (Javascript) flavour for regex The indent of the filename and line and column specifier is dependent on the line number, you have to match for multiple spaces there (https://regex101.com/r/9xImjz/2). If I have understood correctly, this will eventually be present in/part of fpm. Let me know when that happens to add the problemmatcher in the fortran-lang vscode extension. I think this might be a good opportunity to try the new GitHub projects that span the entire organisation as opposed to the individual repos. Seems to be not an issue which is worth addressing at the moment.
2025-04-01T06:40:40.084471
2023-12-06T16:08:50
2028903537
{ "authors": [ "akashKarmakar02", "tommyettinger" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11410", "repo": "tommyettinger/gdx-liftoff", "url": "https://github.com/tommyettinger/gdx-liftoff/issues/145" }
gharchive/issue
Issue in project opening Class org.jetbrains.kotlin.idea.gradleTooling.KotlinGradleModel must not be requested from main classloader of org.jetbrains.kotlin plugin [Plugin: org.jetbrains.android] when i open a new project desktop only with java version 21 and gradle version 8.5 It is fixed it is happend due to old kotlin plugin Interesting, I'll try to update it in Liftoff if possible. I had an issue when launching Gradle with Java 21 and trying to build an Android APK from IDEA, but it wasn't the same bug you reported. The bug I hit was resolved by changing the setting in IDEA to run Gradle with Java 17. In case anyone finds this by searching, the error was: > Task :android:compileReleaseJavaWithJavac FAILED Execution failed for task ':android:compileReleaseJavaWithJavac'. > Could not resolve all files for configuration ':android:androidJdkImage'. > Failed to transform core-for-system-modules.jar to match attributes {artifactType=_internal_android_jdk_image, org.gradle.libraryelements=jar, org.gradle.usage=java-runtime}. > Execution failed for JdkImageTransform: C:\Users\tommy\AppData\Local\Android\Sdk\platforms\android-34\core-for-system-modules.jar. > Error while executing process C:\d\jvm\jdk21-hotspot\bin\jlink.exe with arguments {--module-path C:\Users\tommy\.gradle\caches\transforms-3\553ad13d8645437e7a391ae3ab1b8dbb\transformed\output\temp\jmod --add-modules java.base --output C:\Users\tommy\.gradle\caches\transforms-3\553ad13d8645437e7a391ae3ab1b8dbb\transformed\output\jdkImage --disable-plugin system-modules} I've never seen this type of bug before, and I didn't know Android even used jlink... Try ./gradlew clean After that try building again It sometimes fixed the problem
2025-04-01T06:40:40.098982
2020-03-06T20:01:22
577146503
{ "authors": [ "ArnWac", "marcoaaguiar", "thismatters", "tomturner" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11411", "repo": "tomturner/django-tenants", "url": "https://github.com/tomturner/django-tenants/issues/349" }
gharchive/issue
admin displaying apps inconsistently I'm seeing some unexpected results in the django admin. Specifically I'm seeing what are strictly tenant apps showing up in the global schema, and I'm seeing what is strictly a shared app not showing up in the global schema admin ("Not available for global schema"). I think this might be because I am using the verbose notation in my SHARED_APPS and TENANT_APPS, e.g. 'polls.apps.PollsConfig', as proposed in the django tutorial. Whenever I switch to the short notation (e.g. 'polls') then the tenant apps only show up in the tenant admin. This is a pretty acceptable workaround, but I wanted you to know that there is some dissonance there. The primary issue I'm facing is that there doesn't seem to be a way to obscure a shared app from tenants. Specifically, I would like to not show my tenant model to all of my tenants. Have I overlooked a method for making a table in the shared schema private (or for global admin purposes only)? We also had inconsistencies in the django admin in our project. I haven't had much time to dig into it to try to find what is happening. In particular admin.site.register/unregister have been erratic. The primary issue I'm facing is that there doesn't seem to be a way to obscure a shared app from tenants. Specifically, I would like to not show my tenant model to all of my tenants. I worked around by implementing a seperate admin site for 'public' If found out that the issue that we were having was due to having the INSTALLED_APPS like in this line: https://github.com/tomturner/django-tenants/blob/5128d9e9d2409f3cd0b09c0bd574170ff657725d/examples/tenant_tutorial/tenant_tutorial/settings.py#L175 which was causing to the order of loaded apps to be randomize (set in python does not keep order). We had an package that was registering in its admin.py, and we were unregistering in one of our apps. If the our app loaded first it would be registered, if the package loaded first, it would be unregistered. I hope this helps. I had the same issue use INSTALLED_APPS = list(SHARED_APPS) + [app for app in TENANT_APPS if app not in SHARED_APPS] In our project, we are using django-configurations which have configurations classes instead a configuration file (one for production, one for development, etc). Because the scope for list comprehension behave weirdly in a class body the code snippet that you posted won't work. The workaround was to use INSTALLED_APPS = list(set(SHARED_APPS) - set(TENANT_APPS))+ TENANT_APPS # keep the order of the TENANT_APPS Since the order of TENANT_APPS that mattered. I failed to find a more elegant solution, but it works
2025-04-01T06:40:40.106449
2015-09-03T07:28:14
104637321
{ "authors": [ "saramroo", "tomvanzummeren" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11412", "repo": "tomvanzummeren/TZStackView", "url": "https://github.com/tomvanzummeren/TZStackView/issues/32" }
gharchive/issue
Inconsistent podspec version in tags Podspec version in 1.0.5-swift2 and 1.0.5 tags is still using 1.0.4. I restructured the whole repository, deleted the separate swift-2 branch, made master compatible with Swift 2.0, so this is fixed now!
2025-04-01T06:40:40.114895
2021-11-12T19:25:42
1052346622
{ "authors": [ "fredski-github", "toniblyx" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11413", "repo": "toniblyx/prowler", "url": "https://github.com/toniblyx/prowler/issues/925" }
gharchive/issue
sensible versus sensitive in risk messages need to replace sensible for sensitive in risk associated with finding message. Publicly accessible load balancers could expose sensible data to bad actors. Vaults accessible to everyone could expose sensible data to bad actors Thanks @fredski-github, just fixed it in branch 2.6.1 and will be in master as soon as I fix the other bug
2025-04-01T06:40:40.162034
2022-04-03T02:28:38
1190795746
{ "authors": [ "tonyketcham" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11414", "repo": "tonyketcham/p5-svelte", "url": "https://github.com/tonyketcham/p5-svelte/issues/125" }
gharchive/issue
Add docs example with p5.SVG Pen plotters + p5 + Svelte would be a great example that some may find inspiring. That was one of the founding pieces of inspiration for starting this project and I was disappointed that p5.SVG didn't support ESM/instance mode so I couldn't get an SVG runtime for p5 working w/ p5-svelte. In a recent release, p5.SVG added support for those pieces 🥳 Turns out that the instance mode does the thing where it's trying to mod p5 via accessing the window which Svelte no likey. This may not work, or may need a custom shim within p5-svelte in order to support
2025-04-01T06:40:40.168302
2020-10-16T14:04:00
723256846
{ "authors": [ "kaizoku-oh", "tonyp7" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11415", "repo": "tonyp7/esp32-wifi-manager", "url": "https://github.com/tonyp7/esp32-wifi-manager/issues/112" }
gharchive/issue
Can't build example using esp-idf v4.1 Prerequisites [x] I have written a descriptive issue title [x] I have verified that I am using the latest version of esp32-wifi-manager [x] I have searched open and closed issues to ensure it has not already been reported Description I'm facing the same issue similar #103 when working with ESP-IDF 4.1 on platformio. error: unknown type name 'esp_event_handler_instance_t'; did you mean 'esp_event_handler_t'? Is there anyway to get around this build error while still using the 4.1 version? Steps to Reproduce install platformio extention on vscode create a new project for esp32 using the esp-idf framework create a components folder under the root of the project and copy the wifi-manager into it (tried master and other tags) build System Configuration Debian 10 x64 vscode platformio Hello, This is normal, this code only compile with 4.2 and up as per the requirements: https://github.com/tonyp7/esp32-wifi-manager#requirements Are you considering to support a stable release in the future? because v4.2 is only pre-release beta now contrary to v4.1 which will be supported until February 2023. Definitely not. 4.2 should hit RC stage fairly soon and the master tree is already in 4.3. There’s no point holding back the codebase at this stage.
2025-04-01T06:40:40.172611
2019-08-26T20:53:09
485446518
{ "authors": [ "rmoritz", "tonysavon" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11416", "repo": "tonysavon/Chopper-Command-C64", "url": "https://github.com/tonysavon/Chopper-Command-C64/issues/1" }
gharchive/issue
How to build? Which assembler must be used to build the sources? Hi Please use Kickassembler http://theweb.dk/KickAssembler/Main.html#frontpage Just assemble game.asm and that should give you game.prg, which you can run straight away. It can also be crunched with exomizer after that, which is what the included ChopperCommand.prg file is On Mon, 26 Aug 2019, 22:53 Ralph Möritz<EMAIL_ADDRESS>wrote: Which assembler must be used to build the sources? — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/tonysavon/Chopper-Command-C64/issues/1?email_source=notifications&email_token=AMSYTCQYFLCIAY4IYXOW2T3QGQ7DNA5CNFSM4IPVWFEKYY3PNVWWK3TUL52HS4DFUVEXG43VMWVGG33NNVSW45C7NFSM4HHPKN3A, or mute the thread https://github.com/notifications/unsubscribe-auth/AMSYTCUNNKSN36Z6QA6V7LLQGQ7DNANCNFSM4IPVWFEA . Thank you!
2025-04-01T06:40:40.193224
2024-11-12T06:24:41
2651169326
{ "authors": [ "Roberto-Gentili", "lukehutch", "michael-simons" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11417", "repo": "toolfactory/narcissus", "url": "https://github.com/toolfactory/narcissus/issues/17" }
gharchive/issue
MacOS build broken due to encapsulation @Roberto-Gentili Any idea why this build is failing on macOS?: Error: ROR] Failed to execute goal org.sonatype.plugins:nexus-staging-maven-plugin:1.6.8:deploy (injected-nexus-deploy) on project narcissus: Execution injected-nexus-deploy of goal org.sonatype.plugins:nexus-staging-maven-plugin:1.6.8:deploy failed: An API incompatibility was encountered while executing org.sonatype.plugins:nexus-staging-maven-plugin:1.6.8:deploy: java.lang.ExceptionInInitializerError: null Error: ROR] ----------------------------------------------------- Error: ROR] realm = extension>org.sonatype.plugins:nexus-staging-maven-plugin:1.6.8 Error: ROR] strategy = org.codehaus.plexus.classworlds.strategy.SelfFirstStrategy [...] Error: ROR] -----------------------------------------------------: Unable to make field private final java.util.Comparator java.util.TreeMap.comparator accessible: module java.base does not "opens java.util" to unnamed module @3681037 Building with JDK 7 fails with the error below: there appears to be an incompatible Maven plugin with this JDK. ave you updated any Maven plugins? constituent[11]: file:/C:/Users/runneradmin/.m2/wrapper/dists/apache-maven-3.8.8-bin/67c30f74/apache-maven-3.8.8/lib/maven-compat-3.8.8.jar at org.apache.maven.wrapper.BootstrapMainStarter.start(BootstrapMainStarter.java:53) at org.apache.maven.wrapper.WrapperExecutor.execute(WrapperExecutor.java:152) at org.apache.maven.wrapper.MavenWrapperMain.main(MavenWrapperMain.java:76) Caused by: java.lang.UnsupportedClassVersionError: com/fasterxml/jackson/databind/ObjectMapper: Unsupported major.minor version 52.0 @Roberto-Gentili Any idea why this build is failing on macOS?: Error: ROR] Failed to execute goal org.sonatype.plugins:nexus-staging-maven-plugin:1.6.8:deploy (injected-nexus-deploy) on project narcissus: Execution injected-nexus-deploy of goal org.sonatype.plugins:nexus-staging-maven-plugin:1.6.8:deploy failed: An API incompatibility was encountered while executing org.sonatype.plugins:nexus-staging-maven-plugin:1.6.8:deploy: java.lang.ExceptionInInitializerError: null Error: ROR] ----------------------------------------------------- Error: ROR] realm = extension>org.sonatype.plugins:nexus-staging-maven-plugin:1.6.8 Error: ROR] strategy = org.codehaus.plexus.classworlds.strategy.SelfFirstStrategy [...] Error: ROR] -----------------------------------------------------: Unable to make field private final java.util.Comparator java.util.TreeMap.comparator accessible: module java.base does not "opens java.util" to unnamed module @3681037 It was already broken, so I tried updating all the Maven plugins to see if that would fix it... It looks like the Maven wrapper version needs to be updated, or something? It was already broken, so I tried updating all the Maven plugins to see if that would fix it... It looks like the Maven wrapper version needs to be updated, or something? Unfortunately I don't know: I abandoned the build test for JDK 7 due to too many problems even if I do the release build with the version parameter set for compatibility with JDK 7 It was already broken, so I tried updating all the Maven plugins to see if that would fix it... It looks like the Maven wrapper version needs to be updated, or something? Unfortunately I don't know: I abandoned the build test for JDK 7 due to too many problems even if I do the release build with the version parameter set for compatibility with JDK 7 @Roberto-Gentili @michael-simons OK, the build seems to be fixed. Thanks for your help! Pushing out 1.0.10. It's interesting that this failed on JDK 7 :) I explicitly set the Maven Wrapper and Maven versions to work with 7. Seems to be an issue with nexus-staging-maven-plugin. FWIW There's a way to define tool chains in Maven, so that you don't depend on the JDK that is being used to run Maven, but I never used them. https://maven.apache.org/guides/mini/guide-using-toolchains.html @Roberto-Gentili @michael-simons OK, the build seems to be fixed. Thanks for your help! Pushing out 1.0.10. Great! I will release the new version of JVM-Driver as soon as it is available Maven now requires a minimum JDK of 7, it seems (or maybe it is Maven-wrapper). I just removed the testing for JDK 7, since I don't have time to mess with this much more!
2025-04-01T06:40:40.204574
2021-02-26T13:45:55
817369501
{ "authors": [ "Meemaw", "coveralls" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11418", "repo": "toomuchdesign/next-page-tester", "url": "https://github.com/toomuchdesign/next-page-tester/pull/173" }
gharchive/pull-request
Fix client singleton router ii What kind of change does this PR introduce? Replaces: https://github.com/toomuchdesign/next-page-tester/pull/161 Fixes: https://github.com/toomuchdesign/next-page-tester/issues/160 What is the current behaviour? You can also link to an open issue here. What is the new behaviour? ... Does this PR introduce a breaking change? What changes might users need to make in their application due to this PR? Other information: Please check if the PR fulfills these requirements: [x] Tests for the changes have been added [x] Docs have been added / updated Coverage remained the same at 100.0% when pulling 187c984537ad68b359ee1f4cad8c5021f0c4e6b9 on fix-client-singleton-router-II into dad5cbb7ca4175290ad32820290575f3318fc263 on master.
2025-04-01T06:40:40.354541
2016-11-27T19:01:21
191878100
{ "authors": [ "dbashford", "dustinfarris", "justinpark", "toranb" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11419", "repo": "toranb/ember-redux", "url": "https://github.com/toranb/ember-redux/pull/51" }
gharchive/pull-request
[PERFORMANCE]: avoid multiple redux.getState() calls in connect I've simplified the callback workload to reduce the number of times we invoke redux.getState() in handleChange from 2x to 1x and also I've eliminated an unnecessary function (used previously in the shallowEqual days of connect). My hope is that we can achieve the same goal #41 set out to solve but without the tradeoff of holding state local in the component (and further -helping us avoid Ember.set / setProperties). After the over notify problem was solved, the remaining "Problem Statement" back in #30 was that we call redux.getState() multiple times during re-render work and that could be expensive. This PR should solve a good chunk of that and I'm asking @dbashford to confirm we didn't break his app. In addition, I'm reaching out to @justinpark to confirm he won't run into a performance issue of any kind taking this new connect implementation. Would you both have time this week to try out this updated connect function and report back? @dustinfarris awesome! I'm hopeful it helps move us closer to the silver bullet while avoiding whatever it was that @dbashford ran into w/ the setProperties variation from last week. I'll wait until I hear from both of them that this is regression free/ solid before we get a v1.9.2 out the door the remaining "Problem Statement" back in #30 was that we call redux.getState() multiple times during re-render work and that could be expensive. I wonder how expensive redux.getState() really is? I would guess under the hood the browser is just kicking around a pointer. I'd be more interested in knowing how expensive the stateToProps function is for a non-trivial component. in the computed prop: defineProperty(component, name, computed(function() { return finalMapStateToComputed(redux.getState())[name]; I'd bet the finalMapStateToComputed which gets run on every single property is much more expensive than the redux.getState() part. Anyway, probably discussion best saved for another issue/PR. Looking good here! All good here too!
2025-04-01T06:40:40.365640
2023-03-21T17:59:02
1634450589
{ "authors": [ "sansoy", "torayeff" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11420", "repo": "torayeff/fanucpy", "url": "https://github.com/torayeff/fanucpy/issues/18" }
gharchive/issue
INTP-311 (MAPPDK_SERVER,204) Uninitialized data is used I got fanucpy working well on my first CRX but on my second CRX I get this error message. I've searched the threads and this message pops up for a variety of reasons with other socket based programs. Was wondering if you had any troubleshooting tips. Hi. Have you copied the latest fanuc driver files from https://github.com/torayeff/fanucpy/tree/main/src/fanuc-driver ? Basically, you need to install the latest version of fanucpy (0.1.11) with the latest version of fanuc driver files. i did this morning hoping to resolve the issue. Some threads suggest it might be a memory issue. Will try again. Sabri BTW great job with integrating ChatGPT! On Tue, Mar 21, 2023 at 12:06 PM Agajan Torayev @.***> wrote: Hi. Have you copied the latest fanuc driver files from https://github.com/torayeff/fanucpy/tree/main/src/fanuc-driver ? Basically, you need to install the latest version of fanucpy (0.1.11) with the latest version of fanuc driver files. — Reply to this email directly, view it on GitHub https://github.com/torayeff/fanucpy/issues/18#issuecomment-1478364315, or unsubscribe https://github.com/notifications/unsubscribe-auth/AACV24LVMTW2QB4HU5K7IG3W5HUZXANCNFSM6AAAAAAWCZM5QM . You are receiving this because you authored the thread.Message ID: @.***> @sansoy Let me know how it goes. I will have a closer look at the issue on Thursday. Some far there were not any problems while testing using Roboguide. I can confirm that v0.1.11 works with the latest .pc and .ls files. Please copy the new files from https://github.com/torayeff/fanucpy/tree/main/src/fanuc-driver Thanks Agajan! I was able to get it to run my CRX! Great work! Sabri On Thu, Apr 6, 2023 at 7:08 AM Agajan Torayev @.***> wrote: Closed #18 https://github.com/torayeff/fanucpy/issues/18 as completed. — Reply to this email directly, view it on GitHub https://github.com/torayeff/fanucpy/issues/18#event-8945813893, or unsubscribe https://github.com/notifications/unsubscribe-auth/AACV24KURBH5IQNMKUKDQ5DW7253LANCNFSM6AAAAAAWCZM5QM . You are receiving this because you were mentioned.Message ID: @.***>
2025-04-01T06:40:40.367833
2016-09-27T05:58:11
179403779
{ "authors": [ "ArturoDeza", "ilija139", "karandwivedi42" ], "license": "bsd-3-clause", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11421", "repo": "torch/image", "url": "https://github.com/torch/image/issues/193" }
gharchive/issue
lena loading grace bug When using image.display(image.lena()), it shows grace hopper. As a suggestion, possible creating a new image.grace() function would be better. I was also curious about this. Is it a joke? :smile: You can see https://github.com/torch/nn/issues/854. "we've removed lena. we cant change the name because of breaking API"
2025-04-01T06:40:40.377801
2024-10-16T06:45:05
2590820195
{ "authors": [ "songproducer", "zerolab" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11422", "repo": "torchbox/wagtail-news-template", "url": "https://github.com/torchbox/wagtail-news-template/issues/14" }
gharchive/issue
Navigation links not working on mobile https://github.com/torchbox/wagtail-news-template/blob/main/templates/navigation/header.html#L124 a href was blank should be: href="{{ link.value.get_url }}" @songproducer thank you for filing these issues. It is very useful. Feel free to submit pull requests with fixes to any issues you find and have time for @zerolab Sorry I'm git illiterate, I gave it a shot but not sure if this is right. https://github.com/songproducer/wagtail-news-template/pull/2 Hey @songproducer, no worries at all. Perhaps https://docs.github.com/en/pull-requests/collaborating-with-pull-requests/proposing-changes-to-your-work-with-pull-requests/creating-a-pull-request-from-a-fork can help Thanks! I was trying to make my life more difficult by trying to use command line. Pull request submitted!
2025-04-01T06:40:40.451666
2023-09-15T18:45:36
1898914300
{ "authors": [ "LegNeato", "hawkw" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11423", "repo": "tosc-rs/mnemos", "url": "https://github.com/tosc-rs/mnemos/pull/275" }
gharchive/pull-request
Add that mn requires bindeps Without this installing errors with: error: failed to parse manifest at `mnemos/tools/manganese/Cargo.toml` Caused by: `artifact = …` requires `-Z bindeps` (cargo-binutils) Ah, I had not really considered the use case of installing mn using cargo install. Instead, I had assumed that it would always be run using the cargo mn alias defined in the workspace's .cargo/config.toml, which will always work fine because that file also explicitly enables the bindeps unstable feature. I'm not sure whether or not we should include this in the README. On one hand, it does tell users how to avoid problems when installing mn using cargo install. On the other hand, I'm not actually sure if installing mn that way is a good idea. The advantage of only ever running it using cargo run is that it is potentially rebuilt whenever the Cargo.lock entries for its dependencies changes. This means that if you run mn using the alias that calls cargo run, you get whatever versions of the bindeps are currently in the lockfile. If those haven't changed, mn and its dependencies are not recompiled every time its run, but if the lockfile has changed, the cargo run command will build new versions of the bindeps. This ensures that everyone gets compatible versions of those dependencies whenever they use cargo mn. On the other hand, if mn is cargo installed, it's built once, and every subsequent time you run cargo mn, you get the same thing, regardless of the mnemos repo's Cargo.lock. This is fine with regards to the actual mn binary, since it's very simple and won't change often. However, it also means that the versions of the bindeps that are downloaded and compiled are the versions you get when mn is cargo installed, and if the lockfile updates those dependencies, you still get the old ones. This means that if we later update the lockfile to depend on new versions of those tools, any contributors who have cargo installed mn will see their versions of the build dependencies begin to drift from what other contributors are using, potentially introducing compatibility issues with the various scripts run using mn. Because of the version drift issue, I think it might be better to just have the README suggest that mn should not be cargo installed, and that it should instead always be used through cargo run --package manganese , or through the cargo mn alias defined in the workspace, which expands to that. What do you think? Ok. I didn't actually know how to run it and assumed it was a normal cargo tool. I didn't know you could define aliases in the workspace, and I work with a lot of different rust projects so this probably will be a tripping point for folks. Thanks for the quick response!
2025-04-01T06:40:40.464850
2016-02-18T21:24:04
134700964
{ "authors": [ "GreenGremlin" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11425", "repo": "toumorokoshi/sprinter", "url": "https://github.com/toumorokoshi/sprinter/pull/95" }
gharchive/pull-request
Adding support for boolean input types Includes custom prompts for y/n, yes/no, true/false, and t/f. With uppercased default values and optional in/out types. In/out types are configured with a '/'. "In" type controls the prompting, "out" type controls the string conversion. This allows prompting for "y/n" and outputting as "true/false", for one example. Maybe in the future, instead of output types we could have some sort of casting use. i.e. %((bool):config:my_var). I'm not sure what the best format would be, but this would lend for greater flexibility.
2025-04-01T06:40:40.468784
2018-02-01T06:23:08
293416889
{ "authors": [ "justlikethisdesign", "tovic" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11427", "repo": "tovic/rich-text-editor", "url": "https://github.com/tovic/rich-text-editor/issues/8" }
gharchive/issue
Paste on Mobile removes pasted content immediately When tested on Apple and Android mobile devices (both on chrome so far) the paste function is not working. As far I can tell the function is being triggered (I extended the paste delay) but as soon as the delay is up it seems the content is not found in "view" and is removed. https://github.com/tovic/rich-text-editor/commit/a330d78ed16cbfd8dded8a55aaf9d6ed2d39bde3
2025-04-01T06:40:40.479768
2017-11-09T18:21:04
272669826
{ "authors": [ "meganrm", "nathanmwilliams", "pdvorchik" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11428", "repo": "townhallproject/townHallProject", "url": "https://github.com/townhallproject/townHallProject/pull/216" }
gharchive/pull-request
Fix/style changes to cards Changes to cards Added the last updated in a footer, which is nice to see for events that are coming up soon, but my question on this is we have some events that haven't been updated in awhile, does it make things look stale? Formatted the address so that it breaks at the right point Changed the box around the address to a line moved links to the lower right hand corner. Added the last updated in a footer, which is nice to see for events that are coming up soon, but my question on this is we have some events that haven't been updated in awhile, does it make things look stale? I assume these are ongoing events (regular office hours)? Maybe we lose the "last updated" for those types of events? Other changes look solid. A couple questions: -Do we want the Notes to be above the address? In the examples you used it works but occasionally the Notes can be very long (such as a long list of additional speakers or co-host organizations). -The bar color is still connected to Party and the little icon/color is connected to the event type on our map? I'm wondering if there is a cleaner way to convey these (rather than blue on blue, with each blue signifying something different). yeah we can switch notes and address. The headers are only connected to the type of event. Blue is for in person, purple for staff etc So the entire header color (including icon) are representing the same thing with two different color shades? We may have already discussed this but I wonder if it's a bit confusing (assuming that one is representing party). yeah that's right. There's nothing that indicates party by color anymore. Love the links on the bottom right! May I make a suggestion about the header? Right now the MOC's name seems understated and the event type icon isn't very grokable. What if we made them columns instead? Something like: yeah! do you have this styled already or did you do in with screen grabs? if you have the style can you just push it to this branch? Screen grabs I'm afraid
2025-04-01T06:40:40.490466
2018-07-14T09:26:48
341223329
{ "authors": [ "johnmarinelli", "toxicFork" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11429", "repo": "toxicFork/react-three-renderer-fiber", "url": "https://github.com/toxicFork/react-three-renderer-fiber/pull/55" }
gharchive/pull-request
Add step to readme Hey! It's me again. So I had to setup this project on a new machine and ran into problems when I followed the README steps. Specifically, the <script src="node_modules/react/umd/react.development.js"></script> <script src="node_modules/react-dom/umd/react-dom.development.js"></script> lines in examples/index.html were breaking because it couldn't find the files. This is because yarn doesn't install the dependencies for examples when you run yarn install from the root. It's a small detail but cost me ~45 minutes of trying to find a bug in the webpack config or fiddling with the path in the script src. Hi John, thank you for the PR, it is much appreciated! I have been meaning to get back to this project and see it to completion, your interest and help is a great motivator. Apologies for the 45 minutes of lost time.
2025-04-01T06:40:40.520317
2019-12-10T08:29:31
535571658
{ "authors": [ "JoshuaDRose", "tim77" ], "license": "Vim", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11430", "repo": "tpope/vim-pathogen", "url": "https://github.com/tpope/vim-pathogen/issues/209" }
gharchive/issue
License file Would be nice to provide license file. This helps with distro adoption. Simple license file would be enough. true
2025-04-01T06:40:40.521808
2021-03-24T22:46:09
840266952
{ "authors": [ "zachrbrown" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11431", "repo": "tqtezos/minter-sdk", "url": "https://github.com/tqtezos/minter-sdk/issues/70" }
gharchive/issue
Use canonical Michelson parser implementation to generate JSON representations #58 added a script which generates a JSON representation of a Michelson contract to facilitate idiomatic importing of contracts into a JavaScript project. This solution outsources Michelson parsing to the @taquito/michel-codec package. Unfortunately, this package often goes out of date with respect to the protocol, encountering parse errors on valid contracts. To mitigate this, we should use a canonical implementation like the tezos-client / RPC. Closed via #75
2025-04-01T06:40:40.525480
2022-01-25T11:21:55
1113764993
{ "authors": [ "devudr", "tr3v3r" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11432", "repo": "tr3v3r/react-native-esc-pos-printer", "url": "https://github.com/tr3v3r/react-native-esc-pos-printer/issues/45" }
gharchive/issue
getting error while installing error logs /Users/../Documents/work/ReactNative/work/../node_modules/react-native-esc-pos-printer: Command failed. Exit code: 1 Command: node ./scripts/movelib.js Arguments: Directory: /Users/../Documents/work/ReactNative/work/../node_modules/react-native-esc-pos-printer Output: internal/modules/cjs/loader.js:905 throw err; ^ Error: Cannot find module '/Users/../Documents/work/ReactNative/work/../node_modules/react-native-esc-pos-printer/scripts/movelib.js' at Function.Module._resolveFilename (internal/modules/cjs/loader.js:902:15) at Function.Module._load (internal/modules/cjs/loader.js:746:27) at Function.executeUserEntryPoint [as runMain] (internal/modules/run_main.js:76:12) at internal/main/run_main_module.js:17:47 { Fixed in v1.7.1 hi @tr3v3r what is the major difference in v1.5.0 between v1.7.1 @devudr the only difference is that SDK for iOS is moved to a separate repo.
2025-04-01T06:40:40.536213
2023-06-05T06:35:11
1741080150
{ "authors": [ "scala-steward" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11433", "repo": "trace4cats/trace4cats-dynamic-extras", "url": "https://github.com/trace4cats/trace4cats-dynamic-extras/pull/160" }
gharchive/pull-request
Update sbt-github-actions to 0.15.0 About this PR 📦 Updates com.codecommit:sbt-github-actions from 0.14.2 to 0.15.0 Usage ✅ Please merge! I'll automatically update this PR to resolve conflicts as long as you don't change it yourself. If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below. Configure Scala Steward for your repository with a .scala-steward.conf file. Have a fantastic day writing Scala! 🔍 Files still referring to the old version number The following files still refer to the old version number (0.14.2). You might want to review and update them manually. project/Dependencies.scala ⚙ Adjust future updates Add this to your .scala-steward.conf file to ignore future updates of this dependency: updates.ignore = [ { groupId = "com.codecommit", artifactId = "sbt-github-actions" } ] Or, add this to slow down future updates of this dependency: dependencyOverrides = [{ pullRequests = { frequency = "30 days" }, dependency = { groupId = "com.codecommit", artifactId = "sbt-github-actions" } }] labels: sbt-plugin-update, early-semver-major, semver-spec-minor, old-version-remains, commit-count:n:2 Superseded by #195.
2025-04-01T06:40:40.541104
2023-11-15T12:09:51
1994656026
{ "authors": [ "alexdesi" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11434", "repo": "trade-tariff/trade-tariff-frontend", "url": "https://github.com/trade-tariff/trade-tariff-frontend/pull/1719" }
gharchive/pull-request
Update bundler What? Update bundler to remove warning "`DidYouMean::SPELL_CHECKERS.merge!(error_name => spell_checker)' has been deprecated.", which is present many times when running spec tests. Why? To stop using the deprecated method.
2025-04-01T06:40:40.546866
2019-03-01T18:20:44
416223254
{ "authors": [ "chadicus", "coveralls", "jncarver" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11435", "repo": "traderinteractive/filter-php", "url": "https://github.com/traderinteractive/filter-php/pull/84" }
gharchive/pull-request
Add Filterer responseType option What does this PR do? This pull request adds a new FilterResponse class and adds a responseType option to the Filterer::filter function to allow an instance of this new class to be returned. The default return is still an array, which means it is not backwards-breaking. Additionally, the FilterResponse object can be converted to the traditional array result with its toArray() function. Example use: $response = Filterer::filter($spec, $value, ['responseType' => Filterer::RESPONSE_TYPE_FILTER]); if ($response->success === false) { throw new FilterException($response->errorMessage); } return $response->filteredValue; Checklist [X] Pull request contains a clear definition of changes [X] Tests (either unit, integration, or acceptance) written and passing [X] Relevant documentation produced and/or updated Coverage decreased (-0.4%) to 97.26% when pulling 859e1caecc2cbd91906dc80584b516dfe77b30ae on jncarver:filter-response into cc04632ec167679e7b0309b50242dab53ab9879a on traderinteractive:master. @jncarver What advantage is gained by adding the response object? @chadicus I find it easier to use an object with explicit properties than an array of results that are in a specific order. It's also more compact, which is helpful with line length restrictions. The biggest difference, however, is that the original array of errors are placed into the response object, so you can iterate through them if you need to. In the traditional array response, they are combined into a newline-delimited string which can be a pain if you're trying to do anything with the errors, like provide an error count.
2025-04-01T06:40:40.556720
2021-12-06T16:47:35
1072366978
{ "authors": [ "seedy", "traefiker" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11436", "repo": "traefik/faency", "url": "https://github.com/traefik/faency/pull/221" }
gharchive/pull-request
feat: badge cleanup Description Cleaning up badge component, NB: this applies no design system changes. This will be done, hopefully in a future step. added missing compoundVariants programatically added all color variants in story programatically default color variant (gray) is duplicated to make sure it's clear to the user cleaned up theme file Package changes updated rollup-plugin-typescript2 from ^0.29.0 to ^0.30.0 due to a build issue CI: https://github.com/traefik/faency/runs/4433331267?check_suite_focus=true https://github.com/alexjoverm/typescript-library-starter/issues/310#issuecomment-811451090 :tada: This PR is included in version 1.0.0 :tada: The release is available on: npm package (@latest dist-tag) GitHub release Your semantic-release bot :package::rocket:
2025-04-01T06:40:40.604031
2023-04-18T00:15:33
1672147809
{ "authors": [ "1-Samuel", "mpl" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11437", "repo": "traefik/traefik", "url": "https://github.com/traefik/traefik/issues/9848" }
gharchive/issue
wrong api response for healthcheck interval and timeout values Welcome! [X] Yes, I've searched similar issues on GitHub and didn't find any. [X] Yes, I've searched similar issues on the Traefik community forum and didn't find any. What did you do? I configured healthcheck interval or timeout to "wrong" values, e.g. -10s for interval What did you see instead? In the api response (and webui) the values for interval and timeout are the same as in the config (in this case -10s), but in reality, they are some sort of fallback values (in this case 30s for interval) What version of Traefik are you using? Version: 2.9.10 Codename: banon Go version: go1.20.3 Built: 2023-04-06T16:15:08Z OS/Arch: linux/amd64 What is your environment & configuration? services: traefik: image: traefik command: - --api.insecure=true - --providers.docker=true - --log.level=DEBUG ports: - "8080:8080" volumes: - /run/user/1000/docker.sock:/var/run/docker.sock whoami: image: traefik/whoami labels: - "traefik.http.services.myservice.loadbalancer.healthcheck.path=/health" - "traefik.http.services.myservice.loadbalancer.healthcheck.interval=-10s" If applicable, please paste the log output in DEBUG level time="2023-04-18T00:10:28Z" level=info msg="Configuration loaded from flags." time="2023-04-18T00:10:28Z" level=info msg="Traefik version 2.9.10 built on 2023-04-06T16:15:08Z" time="2023-04-18T00:10:28Z" level=debug msg="Static configuration loaded {\"global\":{\"checkNewVersion\":true},\"serversTransport\":{\"maxIdleConnsPerHost\":200},\"entryPoints\":{\"http\":{\"address\":\":80\",\"transport\":{\"lifeCycle\":{\"graceTimeOut\":\"10s\"},\"respondingTimeouts\":{\"idleTimeout\":\"3m0s\"}},\"forwardedHeaders\":{},\"http\":{},\"http2\":{\"maxConcurrentStreams\":250},\"udp\":{\"timeout\":\"3s\"}},\"traefik\":{\"address\":\":8080\",\"transport\":{\"lifeCycle\":{\"graceTimeOut\":\"10s\"},\"respondingTimeouts\":{\"idleTimeout\":\"3m0s\"}},\"forwardedHeaders\":{},\"http\":{},\"http2\":{\"maxConcurrentStreams\":250},\"udp\":{\"timeout\":\"3s\"}}},\"providers\":{\"providersThrottleDuration\":\"2s\",\"docker\":{\"watch\":true,\"endpoint\":\"unix:///var/run/docker.sock\",\"defaultRule\":\"Host(`{{ normalize .Name }}`)\",\"exposedByDefault\":true,\"swarmModeRefreshSeconds\":\"15s\"}},\"api\":{\"insecure\":true,\"dashboard\":true},\"log\":{\"level\":\"DEBUG\",\"format\":\"common\"}}" time="2023-04-18T00:10:28Z" level=info msg="\nStats collection is disabled.\nHelp us improve Traefik by turning this feature on :)\nMore details on: https://doc.traefik.io/traefik/contributing/data-collection/\n" time="2023-04-18T00:10:28Z" level=info msg="Starting provider aggregator aggregator.ProviderAggregator" time="2023-04-18T00:10:28Z" level=debug msg="Starting TCP Server" entryPointName=traefik time="2023-04-18T00:10:28Z" level=debug msg="Starting TCP Server" entryPointName=http time="2023-04-18T00:10:28Z" level=info msg="Starting provider *traefik.Provider" time="2023-04-18T00:10:28Z" level=debug msg="*traefik.Provider provider configuration: {}" time="2023-04-18T00:10:28Z" level=info msg="Starting provider *docker.Provider" time="2023-04-18T00:10:28Z" level=debug msg="*docker.Provider provider configuration: {\"watch\":true,\"endpoint\":\"unix:///var/run/docker.sock\",\"defaultRule\":\"Host(`{{ normalize .Name }}`)\",\"exposedByDefault\":true,\"swarmModeRefreshSeconds\":\"15s\"}" time="2023-04-18T00:10:28Z" level=info msg="Starting provider *acme.ChallengeTLSALPN" time="2023-04-18T00:10:28Z" level=debug msg="*acme.ChallengeTLSALPN provider configuration: {}" time="2023-04-18T00:10:28Z" level=debug msg="Configuration received: {\"http\":{\"routers\":{\"api\":{\"entryPoints\":[\"traefik\"],\"service\":\"api@internal\",\"rule\":\"PathPrefix(`/api`)\",\"priority\":2147483646},\"dashboard\":{\"entryPoints\":[\"traefik\"],\"middlewares\":[\"dashboard_redirect@internal\",\"dashboard_stripprefix@internal\"],\"service\":\"dashboard@internal\",\"rule\":\"PathPrefix(`/`)\",\"priority\":2147483645}},\"services\":{\"api\":{},\"dashboard\":{},\"noop\":{}},\"middlewares\":{\"dashboard_redirect\":{\"redirectRegex\":{\"regex\":\"^(http:\\\\/\\\\/(\\\\[[\\\\w:.]+\\\\]|[\\\\w\\\\._-]+)(:\\\\d+)?)\\\\/$\",\"replacement\":\"${1}/dashboard/\",\"permanent\":true}},\"dashboard_stripprefix\":{\"stripPrefix\":{\"prefixes\":[\"/dashboard/\",\"/dashboard\"]}}},\"serversTransports\":{\"default\":{\"maxIdleConnsPerHost\":200}}},\"tcp\":{},\"udp\":{},\"tls\":{}}" providerName=internal time="2023-04-18T00:10:28Z" level=debug msg="Provider connection established with docker 23.0.3 (API 1.42)" providerName=docker time="2023-04-18T00:10:28Z" level=debug msg="Configuration received: {\"http\":{\"routers\":{\"traefik-traefik\":{\"service\":\"traefik-traefik\",\"rule\":\"Host(`traefik-traefik`)\"},\"whoami-traefik\":{\"service\":\"myservice\",\"rule\":\"Host(`whoami-traefik`)\"}},\"services\":{\"myservice\":{\"loadBalancer\":{\"servers\":[{\"url\":\"http://<IP_ADDRESS>:80\"}],\"healthCheck\":{\"path\":\"/health\",\"interval\":\"-10s\",\"followRedirects\":true},\"passHostHeader\":true}},\"traefik-traefik\":{\"loadBalancer\":{\"servers\":[{\"url\":\"http://<IP_ADDRESS>:80\"}],\"passHostHeader\":true}}}},\"tcp\":{},\"udp\":{}}" providerName=docker time="2023-04-18T00:10:28Z" level=debug msg="No default certificate, fallback to the internal generated certificate" tlsStoreName=default time="2023-04-18T00:10:28Z" level=debug msg="Added outgoing tracing middleware dashboard@internal" middlewareType=TracingForwarder entryPointName=traefik routerName=dashboard@internal middlewareName=tracing time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" entryPointName=traefik middlewareName=dashboard_stripprefix@internal middlewareType=StripPrefix routerName=dashboard@internal time="2023-04-18T00:10:28Z" level=debug msg="Adding tracing to middleware" entryPointName=traefik middlewareName=dashboard_stripprefix@internal routerName=dashboard@internal time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" entryPointName=traefik routerName=dashboard@internal middlewareName=dashboard_redirect@internal middlewareType=RedirectRegex time="2023-04-18T00:10:28Z" level=debug msg="Setting up redirection from ^(http:\\/\\/(\\[[\\w:.]+\\]|[\\w\\._-]+)(:\\d+)?)\\/$ to ${1}/dashboard/" entryPointName=traefik routerName=dashboard@internal middlewareName=dashboard_redirect@internal middlewareType=RedirectRegex time="2023-04-18T00:10:28Z" level=debug msg="Adding tracing to middleware" entryPointName=traefik middlewareName=dashboard_redirect@internal routerName=dashboard@internal time="2023-04-18T00:10:28Z" level=debug msg="Added outgoing tracing middleware api@internal" entryPointName=traefik routerName=api@internal middlewareName=tracing middlewareType=TracingForwarder time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" middlewareName=traefik-internal-recovery entryPointName=traefik middlewareType=Recovery time="2023-04-18T00:10:28Z" level=debug msg="No entryPoint defined for this router, using the default one(s) instead: [http]" routerName=whoami-traefik time="2023-04-18T00:10:28Z" level=debug msg="No entryPoint defined for this router, using the default one(s) instead: [http]" routerName=traefik-traefik time="2023-04-18T00:10:28Z" level=debug msg="No default certificate, fallback to the internal generated certificate" tlsStoreName=default time="2023-04-18T00:10:28Z" level=debug msg="Added outgoing tracing middleware api@internal" entryPointName=traefik routerName=api@internal middlewareName=tracing middlewareType=TracingForwarder time="2023-04-18T00:10:28Z" level=debug msg="Added outgoing tracing middleware dashboard@internal" middlewareName=tracing middlewareType=TracingForwarder entryPointName=traefik routerName=dashboard@internal time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" middlewareName=dashboard_stripprefix@internal middlewareType=StripPrefix entryPointName=traefik routerName=dashboard@internal time="2023-04-18T00:10:28Z" level=debug msg="Adding tracing to middleware" entryPointName=traefik routerName=dashboard@internal middlewareName=dashboard_stripprefix@internal time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" entryPointName=traefik routerName=dashboard@internal middlewareType=RedirectRegex middlewareName=dashboard_redirect@internal time="2023-04-18T00:10:28Z" level=debug msg="Setting up redirection from ^(http:\\/\\/(\\[[\\w:.]+\\]|[\\w\\._-]+)(:\\d+)?)\\/$ to ${1}/dashboard/" entryPointName=traefik routerName=dashboard@internal middlewareType=RedirectRegex middlewareName=dashboard_redirect@internal time="2023-04-18T00:10:28Z" level=debug msg="Adding tracing to middleware" entryPointName=traefik routerName=dashboard@internal middlewareName=dashboard_redirect@internal time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" middlewareType=Recovery entryPointName=traefik middlewareName=traefik-internal-recovery time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" entryPointName=http routerName=whoami-traefik@docker serviceName=myservice middlewareName=pipelining middlewareType=Pipelining time="2023-04-18T00:10:28Z" level=debug msg="Creating load-balancer" entryPointName=http routerName=whoami-traefik@docker serviceName=myservice time="2023-04-18T00:10:28Z" level=debug msg="Creating server 0 http://<IP_ADDRESS>:80" serverName=0 entryPointName=http routerName=whoami-traefik@docker serviceName=myservice time="2023-04-18T00:10:28Z" level=debug msg="child http://<IP_ADDRESS>:80 now UP" time="2023-04-18T00:10:28Z" level=debug msg="Propagating new UP status" time="2023-04-18T00:10:28Z" level=debug msg="Added outgoing tracing middleware myservice" entryPointName=http middlewareName=tracing middlewareType=TracingForwarder routerName=whoami-traefik@docker time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" entryPointName=http routerName=traefik-traefik@docker middlewareType=Pipelining middlewareName=pipelining serviceName=traefik-traefik time="2023-04-18T00:10:28Z" level=debug msg="Creating load-balancer" routerName=traefik-traefik@docker serviceName=traefik-traefik entryPointName=http time="2023-04-18T00:10:28Z" level=debug msg="Creating server 0 http://<IP_ADDRESS>:80" routerName=traefik-traefik@docker serverName=0 serviceName=traefik-traefik entryPointName=http time="2023-04-18T00:10:28Z" level=debug msg="child http://<IP_ADDRESS>:80 now UP" time="2023-04-18T00:10:28Z" level=debug msg="Propagating new UP status" time="2023-04-18T00:10:28Z" level=debug msg="Added outgoing tracing middleware traefik-traefik" middlewareType=TracingForwarder entryPointName=http routerName=traefik-traefik@docker middlewareName=tracing time="2023-04-18T00:10:28Z" level=debug msg="Creating middleware" middlewareName=traefik-internal-recovery middlewareType=Recovery entryPointName=http time="2023-04-18T00:10:28Z" level=error msg="Health check interval smaller than zero for service 'myservice@docker'" serviceName=myservice@docker time="2023-04-18T00:10:28Z" level=debug msg="Setting up healthcheck for service myservice@docker with [Hostname: Headers: map[] Path: /health Method: Port: 0 Interval: 30s Timeout: 5s FollowRedirects: true]" serviceName=myservice@docker time="2023-04-18T00:10:28Z" level=debug msg="Initial health check for backend: \"myservice@docker\"" last three lines are relevant here Hello @1-Samuel , We've looked into it, and we agree that the lack of consistency between the effective configuration of traefik and the "raw data" returned by the API is annoying and disturbing (although not very concerning in itself in our opinion). So here is our plan: for v3, we actually would like to take another route: in such a case of wonky input from the user (e.g. -10s) we're of the opinion that we should log a clear error about it, and simply fail. And not try to cover up the mistake with a "saner" default. We've actually already taken such steps for similar situations in v3, and we think we should keep that trend going. for v2, we cannot do the same as above, because it would be a breaking change. We may want to go in the direction you suggested (i.e. fix the raw data so that it fits with the effective configuration), but we're not convinced yet that the exact fix you proposed is the way we want to go. And tbh, we're not even sure that changing anything at all is worth the effort for v2. So we would like to think about it some more, and also gauge the interest of the community on that matter. WDYT? Thanks. Hello, thank you for looking into the issue. Sounds like a good plan to me. :+1: I only stumbled upon the issue while looking into #6550 and I couldn't think of a good way to write a test for retry intervals, so I thought I give the bug fix a try, just to practice a bit. :) I don't think many people with a real use case will run into this issue anyway. Thank you :) Hello @mpl , yeah I will have a look at implementing the v3 changes (fail if implausible values are configured). But I think making a new PR would make more sense, since the changes would be different all together. I'll close it :)
2025-04-01T06:40:40.799491
2019-02-01T22:54:57
405905377
{ "authors": [ "artemdinaburg", "jberryman", "pgoodman" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11438", "repo": "trailofbits/mcsema", "url": "https://github.com/trailofbits/mcsema/issues/515" }
gharchive/issue
What is the status of Binary Ninja support? DynInst support? These are some questions and a documentation bug report (insofar as the README is a bit confusing or doesn't answer these questions). The top of the README states: Control flow recovery is performed using the mcsema-disass tool, which relies on IDA Pro, Binary Ninja, or DynInst to disassemble a binary file and produce a control flow graph. and later... Currently IDA and BinaryNinja are supported for control-flow recovery... but not dyninst...? (I see there are several abandoned or pending PRs). Then later... Why do I need IDA Pro to use McSema ... we have a branch with experimental Binary Ninja support ... It looks like the referenced branch is deleted (presumably merged?). I also see a ninja branch referenced in a PR, and another one hanging out. What is the status of Binary Ninja support? I might shell out $150 for binary ninja if I was sure it would work (but maybe McSema requires the $600 version for its API?). Thanks! Also getting slightly off topic, I tried out the freeware version of IDA on the off chance it worked and no one has noticed. It doesn't and I don't get any helpful errors (is idal64 the headless version that McSema needs? if so I guess they don't ship that with the freeware version), but I wonder if the free version could be made to work in some other way? Or maybe y'all could try to convince them to release the functionality McSema needs in the free version (still for non-commercial use)? Seems mutually beneficial, but what do I know @jberryman I have not tested IDA Pro Free Version, but if it includes IDAPython it should work. Use of idal64 is not required; you can use idaq64, but you have to do it from an X session since it has to start a GUI. I do this when debugging the CFG recovery scripts. Thanks. Hm, doesn't seem to. The IDA free tree looks like: /home/me/.local/idafree-7.0/ ├── appico64.png ├── assistant ├── cfg │   ├── ida.cfg │   └── idagui.cfg ├── dbgsrv ├── ida64 ├── ida64.int ├── ida.hlp ├── idc │   └── idc.idc ├── ids │   ├── idsnames │   └── win7.zip ├── libclpx.so ├── libdwarfx.so ├── libida64.so ├── libQt5CLucene.so.5 ├── libQt5Core.so.5 ├── libQt5DBus.so.5 ├── libQt5Gui.so.5 ├── libQt5Help.so.5 ├── libQt5Network.so.5 ├── libQt5PrintSupport.so.5 ├── libQt5Sql.so.5 ├── libQt5Widgets.so.5 ├── libQt5XcbQpa.so.5 ├── libX11-xcb.so.1 ├── license.txt ├── loaders │   ├── elf64.so │   ├── macho64.so │   └── pe64.so ├── plugins │   ├── bdescr64.so │   ├── dbg64.so │   ├── dwarf64.so │   ├── objc64.so │   ├── platforms │   │   ├── libqlinuxfb.so │   │   ├── libqminimal.so │   │   ├── libqoffscreen.so │   │   └── libqxcb.so │   ├── platformthemes │   │   └── libqgtk2.so │   ├── plugins.cfg │   └── tds64.so ├── procs │   └── pc64.so ├── qidahelpcollection.qhc ├── qidahelp.qch ├── qt.conf ├── qwingraph ├── sig │   ├── list │   └── pc │   ├── bcb5rt.sig │   ├── iclapp64.sig │   ├── ms64wdk.sig │   ├── msmfc64d.sig │   ├── msmfc64.sig │   ├── msmfc64u.sig │   ├── mssdk64.sig │   ├── pe64.sig │   ├── vc32rtf.sig │   ├── vc64_14.sig │   ├── vc64atl.sig │   ├── vc64extra.sig │   ├── vc64mfc.sig │   ├── vc64rtf.sig │   ├── vc64seh.sig │   ├── vc64ucrt.sig │   ├── vcseh.sig │   └── win64unx.sig ├── til │   ├── gnuunx64.til │   ├── macosx64.til │   └── pc │   ├── bcb5win.til │   ├── gnulnx_x64.til │   ├── mssdk.til │   ├── ntddk64.til │   ├── vc10_64.til │   ├── vc6win.til │   └── vc8amd64.til ├── uninstall ├── uninstall.dat └── Uninstall IDA Free.desktop Sadly it looks like the free version is missing IDAPython so the script will not work. Closing for now.
2025-04-01T06:40:40.812301
2022-02-01T13:48:06
1120698177
{ "authors": [ "di", "tetsuo-cpp", "woodruffw" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11439", "repo": "trailofbits/pip-audit", "url": "https://github.com/trailofbits/pip-audit/pull/229" }
gharchive/pull-request
requirement, pypi: Add a --require-hashes flag Closes #172 This needs cleanup + testing but I wanted to engage early on this one since the intended behaviour isn't straightforward. @di You mentioned that the --require-hashes flag should skip dependency resolution. Did you mean that we shouldn't resolve dependencies for each requirement and instead, only audit what is explicitly listed in the requirements file? I was also wondering how hashes should work when the --require-hashes flag isn't supplied. My thinking is that when the --require-hashes flag is supplied, every requirement in the file needs to have a hash. But if it isn't supplied, we should still check any hashes that we find against PyPI. You mentioned that the --require-hashes flag should skip dependency resolution. Did you mean that we shouldn't resolve transitive dependencies for each requirement and instead, only audit what is explicitly listed in the requirements file? I was also wondering how hashes should work when the --require-hashes flag isn't supplied. My thinking is that when the --require-hashes flag is supplied, every requirement in the file needs to have a hash AND we check them against PyPI. But if the flag isn't supplied, we should still check any hashes that we find against PyPI and fail if it doesn't match. This is an indirect answer to both: we can skip dependency resolution entirely, since pip stipulates the following: If any dependency has a content hash, then all dependencies have content hashes (Transitively) All dependencies must be explicitly listed in the requirements file if hashing is enabled In other words: our requirements parser in pip-api should preserve those properties: we should raise an error if we're given a requirements file with dependencies that are missing hashes, and we should not do dependency resolution because only the present hashed dependencies should be considered. (Looking at things more, I'm not actually sure that we need the --require-hashes flag -- pip stipulates that passing it is identical to any requirement having hashes attached to it. But I guess we should preserve it as a no-op, since that's what pip seems to do? In other words: our requirements parser in pip-api should preserve those properties: we should raise an error if we're given a requirements file with dependencies that are missing hashes, and we should not do dependency resolution because only the present hashed dependencies should be considered. Ah! Right, I misunderstood how that works. Thanks for explaining. (Looking at things more, I'm not actually sure that we need the --require-hashes flag -- pip stipulates that passing it is identical to any requirement having hashes attached to it. But I guess we should preserve it as a no-op, since that's what pip seems to do? I think it's still important because your file may not have ANY hashes in it. If you pass in --require-hashes, it will complain whereas it otherwise would succeed. I think it's still important because your file may not have ANY hashes in it. If you pass in --require-hashes, it will complain whereas it otherwise would succeed. Ah! You're absolutely right. Yes, we need it for that case. @di You mentioned that the --require-hashes flag should skip dependency resolution. Did you mean that we shouldn't resolve transitive dependencies for each requirement and instead, only audit what is explicitly listed in the requirements file? Thinking about this more, I think we still need to do dependency resolution, and ensure that we have hashes for all subdependencies, because this is what pip does as well: $ cat req.txt sampleproject==2.0.0 \ --hash=sha256:2b0c55537193b792098977fdb62f0acbaeb2c3cfc56d0e24ccab775201462e04 \ --hash=sha256:d99de34ffae5515db43916ec47380d3c603e9dead526f96581b48c070cc816d3 $ python -m pip install -r req.txt Collecting sampleproject==2.0.0 (from -r req.txt (line 1)) Using cached https://files.pythonhosted.org/packages/b8/f7/dd9223b39f683690c30f759c876df0944815e47b588cb517e4b9e652bcf7/sampleproject-2.0.0-py3-none-any.whl Collecting peppercorn (from sampleproject==2.0.0->-r req.txt (line 1)) ERROR: In --require-hashes mode, all requirements must have their versions pinned with ==. These do not: peppercorn from https://files.pythonhosted.org/packages/14/84/d8d9c3f17bda2b6f49406982546d6f6bc0fa188a43d4e3ba9169a457ee04/peppercorn-0.6-py3-none-any.whl#sha256=46125cad688a9cf3b08e463bcb797891ee73ece93602a8ea6f14e40d1042d454 (from sampleproject==2.0.0->-r req.txt (line 1)) I think it's still important because your file may not have ANY hashes in it. If you pass in --require-hashes, it will complain whereas it otherwise would succeed. Agreed! Thinking about this more, I think we still need to do dependency resolution, and ensure that we have hashes for all subdependencies, because this is what pip does as well: Can we leave that validation to pip and not attempt to do that validation ourselves? I imagine that's going to introduce a significant chunk of code reinventing what pip does which will have to be removed during integration, since pip already contains this logic. Can we leave that validation to pip and not attempt to do it ourselves? I imagine that's going to introduce a significant chunk of code reinventing what pip does which will have to be removed during integration, since pip already contains this logic. I'm inclined to agree with this, although it does deviate our handling of --require-hashes from that of pip (our --require-hashes is more like --no-deps --require-hashes). My reasoning is that pip-audit isn't in the business of telling the user whether pip install ... would succeed, and that the overwhelming majority of hashed requirements files should be valid already (since, if they're like Warehouse's, they're checked in after successful local and CI usage). Thoughts @di? I suppose the pip install with an incompletely hashed requirements file, or a hashed requirements file that's missing dependencies, would fail to install anyways, so the user would be protected regardless.
2025-04-01T06:40:40.829825
2022-12-30T17:36:54
1514673788
{ "authors": [ "michaelfarrell76" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11440", "repo": "transcend-io/consent-manager-ui", "url": "https://github.com/transcend-io/consent-manager-ui/pull/100" }
gharchive/pull-request
Downgrades react-intl to fix importStar error This fixes an issue with websites that have an old version of requirejs. react intl v5 > relies on import *, which the older version of requirejs does not support see https://github.com/transcend-io/consent-manager-ui/pull/101
2025-04-01T06:40:40.837019
2018-10-17T04:33:43
370895267
{ "authors": [ "stangah", "tinyels" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11441", "repo": "transcom/mymove", "url": "https://github.com/transcom/mymove/pull/1164" }
gharchive/pull-request
WIP: Adds UI for confirming pre-approval request deletion [delivers #161129636] Description Delete UI, some spacing issues to work out. WIP. Breaks out the PreApprovalRequest row into its own component that owns the state of whether the delete UI is shown Code Review Verification Steps [ ] End to end tests pass (make e2e_test). [ ] Code follows the guidelines for Logging [ ] The requirements listed in Querying the Database Safely have been satisfied. Any new migrations/schema changes: [ ] Follow our guidelines for zero-downtime deploys (see Zero-Downtime Deploys) [ ] Have been communicated to #dp3-engineering [ ] There are no aXe warnings for UI. [ ] This works in IE. Any new client dependencies (Google Analytics, hosted libraries, CDNs, etc) have been: [ ] Communicated to @willowbl00 [ ] Added to the list of network dependencies [ ] Request review from a member of a different team. [ ] Have the Pivotal acceptance criteria been met for this change? References Pivotal story for this change Screenshots This is a great start, but I'm not sure how you plan on signaling that one of the deletion confirms is open to prevent this:
2025-04-01T06:40:40.849186
2024-09-18T20:01:40
2534607937
{ "authors": [ "CoryKleinjanCACI", "robot-mymove" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11442", "repo": "transcom/mymove", "url": "https://github.com/transcom/mymove/pull/13739" }
gharchive/pull-request
B 20441 int Agility ticket Summary Updating migrations to add WHERE NOT EXIST to protect against any issues from running the migration more than once Verification Steps for the Author These are to be checked by the author. [ ] Tested in the Experimental environment (for changes to containers, app startup, or connection to data stores) [ ] Have the Agility acceptance criteria been met for this change? Verification Steps for Reviewers These are to be checked by a reviewer. [ ] Has the branch been pulled in and checked out? [ ] Have the BL acceptance criteria been met for this change? [ ] Was the CircleCI build successful? [ ] Has the code been reviewed from a standards and best practices point of view? Setup to Run the Code Instructions for starting storybook Instructions for starting the MilMove application Instructions for running tests How to test Access the Login as a Frontend [ ] There are no aXe warnings for UI. [ ] This works in Supported Browsers and their phone views (Chrome, Firefox, Edge). [ ] There are no new console errors in the browser devtools. [ ] There are no new console errors in the test output. [ ] If this PR adds a new component to Storybook, it ensures the component is fully responsive, OR if it is intentionally not, a wrapping div using the officeApp class or custom min-width styling is used to hide any states the would not be visible to the user. [ ] This change meets the standards for Section 508 compliance. Backend [ ] Code follows the guidelines for Logging. [ ] The requirements listed in Querying the Database Safely have been satisfied. Database Any new migrations/schema changes: [ ] Follows our guidelines for Zero-Downtime Deploys. [ ] Have been communicated to #g-database. [ ] Secure migrations have been tested following the instructions in our docs. Screenshots Warnings :warning: Files located in legacy directories (src/shared or src/scenes) have been edited. Are you sure you don’t want to also relocate them to the new file structure? View the frontend file org ADR for more information Generated by :no_entry_sign: dangerJS against 912398a13be2e07babfe0d644614ac7640536f9b
2025-04-01T06:40:40.861864
2021-06-30T02:09:58
933241127
{ "authors": [ "hannaliebl", "monfresh", "robot-mymove" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11443", "repo": "transcom/mymove", "url": "https://github.com/transcom/mymove/pull/6901" }
gharchive/pull-request
[MB-8204] Refactor TXO MTO fetch to not query for MoveTaskOrder Description Explain a little about the changes at a high level. Reviewer Notes Is there anything you would like reviewers to give additional scrutiny? Setup Add any steps or code to run in this section to help others prepare to run your code: echo "Code goes here" Code Review Verification Steps [ ] If the change is risky, it has been tested in experimental before merging. [ ] Code follows the guidelines for Logging [ ] The requirements listed in Querying the Database Safely have been satisfied. Any new migrations/schema changes: [ ] Follow our guidelines for zero-downtime deploys (see Zero-Downtime Deploys) [ ] Have been communicated to #g-database [ ] Secure migrations have been tested following the instructions in our docs [ ] There are no aXe warnings for UI. [ ] This works in Supported Browsers and their phone views (Chrome, Firefox, IE, Edge). [ ] Tested in the Experimental environment (for changes to containers, app startup, or connection to data stores) [ ] User facing changes have been reviewed by design. [ ] Request review from a member of a different team. [ ] Have the Jira acceptance criteria been met for this change? References Jira story for this change this article explains more about the approach used. Screenshots If this PR makes visible UI changes, an image of the finished UI can help reviewers and casual observers understand the context of the changes. A before image is optional and can be included at the submitter's discretion. Consider using an animated image to show an entire workflow instead of using multiple images. You may want to use GIPHY CAPTURE for this! 📸 Please frame screenshots to show enough useful context but also highlight the affected regions. Warnings :warning: Files located in legacy directories (src/shared or src/scenes) have been edited. Are you sure you don’t want to also relocate them to the new file structure? View the frontend file org ADR for more information Messages :book: :link: MB-8204 Generated by :no_entry_sign: dangerJS against 94d60bf5a36c4b8cb0ee8d8665c616b86dd29ecd This looks good so far. I love removing code. Here are a few more things we need to remove: https://github.com/transcom/mymove/blob/master/pkg/services/move_task_order.go#L37 https://github.com/transcom/mymove/blob/master/pkg/services/move_task_order.go#L58 https://github.com/transcom/mymove/blob/master/pkg/services/move_task_order/move_task_order_fetcher.go#L23-L40 And then we'll need to regenerate the mocks: make mocks_generate
2025-04-01T06:40:40.865841
2021-07-26T21:20:15
953298375
{ "authors": [ "pearl-truss", "robot-mymove" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11444", "repo": "transcom/mymove", "url": "https://github.com/transcom/mymove/pull/7071" }
gharchive/pull-request
[MB-8827] Bump storybook-related dependencies to 6.3.6 Description This PR consolidates the following dependabot PRs related to upgrading Storybook to version 6.3.6: #7068 #7066 #7063 #7061 #7060 Setup Just need to ensure the tests and build pass with no issues. Messages :book: :link: MB-8827 Generated by :no_entry_sign: dangerJS against 822e9a2a4009fb08faf5fa03ce3c70fc897a2d6b Can a front-end person confirm the happo diffs are okay? Especially the document viewer one (the diff shows it as loading instead of present) cc: @YanZ777 @jenniferlynparsons
2025-04-01T06:40:40.874671
2022-01-12T17:20:42
1100598229
{ "authors": [ "reggieriser", "robot-mymove" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11445", "repo": "transcom/mymove", "url": "https://github.com/transcom/mymove/pull/8021" }
gharchive/pull-request
[MB-11012] Fixes bug where NTS shipment turns into HHG when editing as a SC Jira ticket for this change Summary We ran across a bug in the services counselor UI where editing an NTS shipment turns it into an HHG after submission. This PR addresses that issue (it looks like we had just hardcoded the HHG as an input to the component). I also noticed that we were using some old naming for ntsSac and ntsTac in some places. That was causing those values (if present on the order record) to not show up on the edit form, for instance. Those are fixed as well in this PR. Note to reviewers: Are there any particular tests I need to add/edit for this? Setup to Run Your Code 💻 You will need to use three separate terminals to test this locally. Terminal 1 Start the Storybook locally. make storybook Terminal 2 Start the UI locally. make client_run Terminal 3 Start the Go server locally. make server_run Additional steps As a services counselor, try editing an NTS shipment (or NTS-R) and change something like the counselor remarks. Hit save and verify that the shipment remains the same type it was previously. Also, you can add NTS TAC/SAC values to the associated order and verify that they show up in the edit interface. Verification Steps for Author These are to be checked by the author. [ ] Tested in the Experimental environment (for changes to containers, app startup, or connection to data stores) [ ] Request review from a member of a different team. [ ] Have the Jira acceptance criteria been met for this change? Verification Steps for Reviewers These are to be checked by a reviewer. Frontend [ ] User facing changes have been reviewed by design. [ ] There are no aXe warnings for UI. [ ] This works in Supported Browsers and their phone views (Chrome, Firefox, IE, Edge). [ ] There are no new console errors in the browser devtools [ ] There are no new console errors in the test output Warnings :warning: This PR does not include changes to unit tests, even though it affects app code. Messages :book: :link: MB-11012 Generated by :no_entry_sign: dangerJS against aea9c431eafe9cc2bc9c8d4023bc31b6c8eca337
2025-04-01T06:40:40.881581
2022-11-17T22:15:13
1454075565
{ "authors": [ "reggieriser", "robot-mymove" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11446", "repo": "transcom/mymove", "url": "https://github.com/transcom/mymove/pull/9612" }
gharchive/pull-request
[MB-14464] Updated milmove-cypress hash to latest one, now with Cypress 8.5.0 Jira ticket for this change Summary First, reference this circleci-docker PR. This MilMove PR is just trying to get MilMove on the latest image with a downgraded Cypress (8.5.0). We're currently on the same version, but we're having to use a circleci-docker hash from back in April because cypress was updated after that to 9.x and 10.x in circleci-docker independently of MilMove. As a result, updating other dependencies becomes more difficult. So this just tries to get everything on the latest main branches back in sync. Setup to Run Your Code 💻 You will need to use three separate terminals to test this locally. Terminal 1 Start the Storybook locally. make storybook Terminal 2 Start the UI locally. make client_run Terminal 3 Start the Go server locally. make server_run Additional steps Just make sure all E2E tests pass. Verification Steps for Author These are to be checked by the author. [ ] Tested in the Experimental environment (for changes to containers, app startup, or connection to data stores) [ ] Request review from a member of a different team. [ ] Have the Jira acceptance criteria been met for this change? Messages :book: :link: MB-14464 Generated by :no_entry_sign: dangerJS against ffdd48b6377fec23f2641a781a88ee0bcf573cd9
2025-04-01T06:40:40.895541
2024-07-26T18:06:56
2432740832
{ "authors": [ "gauravmm", "onnimonni", "transitive-bullshit" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11447", "repo": "transitive-bullshit/nextjs-notion-starter-kit", "url": "https://github.com/transitive-bullshit/nextjs-notion-starter-kit/pull/629" }
gharchive/pull-request
Modernization The original project is great! It relies on a lot of outdated dependencies, though. I've gone through the codebase and bumped dependencies, gotten rid of old code, and implemented a unified way to set the icons on the right bar and the bottom bar. The key dependency react-notion-x is not altered here. I'm also working on that, but have hit some sort of a roadblock with some dependency. Thanks @gauravmm for this! While we are waiting for @transitive-bullshit to decide how to continue with this starter I created my own fork with updated react-notion-x dependency. See more here: https://github.com/onnimonni/nextjs-notion-starter-kit/pull/2 You can see the forked react-notion-x here: https://github.com/onnimonni/react-notion-x/tree/publish-forked-versions It works perfectly but few tests are still failing which are written with ts and try to import the internal packages eg `import "notion-utils"' are not working and fail with: Uncaught exception in src/notion-api-universal.test.ts Error [ERR_MODULE_NOT_FOUND]: Cannot find package '@onnimonni/notion-utils' imported from /Users/onnimonni/Projects/react-notion-x/packages/notion-client/src/notion-api.ts I'm happy to add you as contributor to both of these forks if you want to help out 👍. Fixed in https://github.com/transitive-bullshit/nextjs-notion-starter-kit/pull/641
2025-04-01T06:40:40.901646
2024-04-23T06:07:29
2258024456
{ "authors": [ "alexgiano", "chfritz" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11448", "repo": "transitiverobotics/transitive", "url": "https://github.com/transitiverobotics/transitive/issues/38" }
gharchive/issue
Services support Hello, I would like to inquire if the Transient Robotics platform supports sending and receiving ROS services. Currently, I am developing a web interface that requires some services to control certain actions on the robot, and I would like to evaluate your solution for a series of additional features that you offer. I have read that managing services may be a limitation of the MQTT communication protocol. I'm not sure if this is the right place to ask this question; if it isn't, I can move it to where it's more appropriate. Thank you very much. We don't currently have an existing capability to do this, but it has been requested by others, too, so we'll probably create one soon. Since you are asking this question here, on the github repo of the open-source Transitive framework, I'll also point out that you could write that capability yourself, if that's what you want. To get started with that, you'd follow these steps and then look at the starter code that is generated from it: https://transitiverobotics.com/docs/develop/creating_capabilities To implement the service calls, I would use mqttSync.mqtt directly, i.e., not use mqttSync (which is for synchronizing stateful data), but the mqtt client itself. You can follow this example from the robot-agent itself, used to trigger agent actions, in this case restarting the agent, from the web: https://github.com/transitiverobotics/transitive/blob/bf7aadaef4c6508b5195ee9f33079ac4e18370f6/cloud/app/web_components/robot-agent-device.jsx#L294-L298 MQTT does not have any built-in support for RPCs, so you'll need to think about the return value yourself. In the agent we do this by writing it into the synchronized data, where the web client will receive it, e.g., for receiving the package log upon request. This pattern is not ideal though and still needs some improvement. But again, if this is more involved than you'd like, then you can also wait until we create a capability for it that you could then just install on your robots. Thanks for your kind reply. I took a look at the solution proposed by you by searching for additional material online like async-mqtt or mqtt-json-rpc but, given my limited knowledge of the mqtt protocol, i would most likely not be able to obtain a good result. Currently I will therefore remain on the websocket protocol but keeping an eye on the developments and implementations of your platform. Thanks so much again Makes sense. Just to clarify though: MQTTSync is not async-mqtt. The former is a protocol we've developed an it's documented here and here.
2025-04-01T06:40:40.902977
2024-06-09T23:41:52
2342555849
{ "authors": [ "devinmatte" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11449", "repo": "transitmatters/mbta-performance", "url": "https://github.com/transitmatters/mbta-performance/pull/27" }
gharchive/pull-request
Support LAMP based monthly csvs There's a new dataset for monthly dumps based on LAMP with slightly different formatting Use that new file and handle its formatting Already using this for data already in the production dashboard, going to merge to avoid someone parsing data with the old version
2025-04-01T06:40:40.931501
2020-04-24T08:40:37
606144836
{ "authors": [ "BettinaTrunk", "agricolab" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11450", "repo": "translationalneurosurgery/tool-offspect", "url": "https://github.com/translationalneurosurgery/tool-offspect/issues/26" }
gharchive/issue
Feature: Setting constant onset shift for one cachefile Is your feature request related to a problem? Please describe. Currently, it is possibly to shift the onset of MEPs in the GUI. However, for one cachefile it requires one to manually shift the onset for every single MEP. Describe the solution you'd like It would be great to have the possibility to set an onset shift and have it constant for the whole data of one Cachefile. I believe using the python API is the best approach for such batch-manipulation. I wrote a couple of examples and added them to the documentation. The one for your use-case is at https://offline-inspect.readthedocs.io/en/latest/cachefile.html#batch-manipulate and repeated here for ease of use: from offspect.api import CacheFile, decode, encode cf = CacheFile("merged.hdf5") for ix, (data, attrs) in enumerate(cf): key = "onset_shift" old = decode(attrs[key]) print(f"Trace {ix} {key}:", old, end=" ") new = old + 5 attrs["onset_shift"] = encode(new) cf.set_trace_attrs(ix, attrs) test = decode(cf.get_trace_attrs(ix)["onset_shift"]) print("to", test) In this example, we shift all onsets by 5 samples to the right. Closing this, as we agreed a API approach is most suitable. Find updated documentations with examples for batch manipulation https://offline-inspect.readthedocs.io/en/latest/cachefile.html
2025-04-01T06:40:41.011711
2017-10-25T17:29:57
268483417
{ "authors": [ "jacktuck", "ryanstinson", "travisjeffery" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11460", "repo": "travisjeffery/jocko", "url": "https://github.com/travisjeffery/jocko/issues/69" }
gharchive/issue
Code Coverage Plugin @travisjeffery I would like to enable https://codecov.io/ on the repo. I don't think I can perform all the required steps to set it up but I can assist and add the required test setup. What do you think? Good idea, I've set it up. Let's get the coverage up before we add the badge to the README though :D @travisjeffery I messaged on gitter before looking at these issues - i will try and help with coverage too :)
2025-04-01T06:40:41.018532
2024-10-07T18:46:09
2571194352
{ "authors": [ "AdamRJensen", "trchudley" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:11461", "repo": "trchudley/pdemtools", "url": "https://github.com/trchudley/pdemtools/issues/31" }
gharchive/issue
x and y-units in mosaic and terrain example In this example https://pdemtools.readthedocs.io/en/latest/examples/mosaic_and_terrain.html there are some very nice figures. However, the x- and y- units are unknown. Would it be possible to show these as latitude/longitudes or something the reader can understand? Thanks Adam. The merging process I use to combined the mosaics from different tiles loses the CF-compliant coordinate attributes that the PDT data normally comes with. New addition of code below enforces the CF-compliant attributes and makes the mosaic data match the rest of the data. This is also reflected in the notebooks. dem['x'].attrs['axis'] = 'X' dem['x'].attrs['long_name'] = 'x coordinate of projection' dem['x'].attrs['standard_name'] = 'projection_x_coordinate' dem['x'].attrs['units'] = 'metre' dem['y'].attrs['axis'] = 'Y' dem['y'].attrs['long_name'] = 'y coordinate of projection' dem['y'].attrs['standard_name'] = 'projection_y_coordinate' dem['y'].attrs['units'] = 'metre'