added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T06:37:01.836553
2021-12-21T16:47:42
1086024722
{ "authors": [ "Gondolav", "samuelchassot" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1135", "repo": "Gondolav/pyfuncol", "url": "https://github.com/Gondolav/pyfuncol/pull/19" }
gharchive/pull-request
Fix subclasses bug and add support for frozensets Description In this PR: Support for arbitrary set, dict and list subclasses Support for frozenset Add filter_not for lists and sets Fixes #17 Type of change Please delete options that are not relevant. [x] Bug fix (non-breaking change which fixes an issue) [x] New feature (non-breaking change which adds functionality) [x] This change requires a documentation update Checklist [x] My code follows the style guidelines of this project [x] I have made corresponding changes to the documentation [x] I have added tests that prove my fix is effective or that my feature works You can review it now @samuelchassot ! Haha no I'll do them! Haha no I'll do them! hahah 👍 Actually it seems that some lines are not tested. Is it a bug? Should be fixed Great so let's merge :) Great job!
2025-04-01T06:37:01.844000
2015-03-05T21:42:59
60020912
{ "authors": [ "stevogotchi" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1136", "repo": "GoogleChrome/chrome-platform-analytics", "url": "https://github.com/GoogleChrome/chrome-platform-analytics/issues/32" }
gharchive/issue
Allow clients to explicitly reset client id Resetting client ID was requested as a feature during a recent (unrelated product) privacy review. No going to happen.
2025-04-01T06:37:01.869306
2015-08-31T07:26:12
104001501
{ "authors": [ "addyosmani", "beaufortfrancois", "jeffposnick" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1137", "repo": "GoogleChrome/samples", "url": "https://github.com/GoogleChrome/samples/issues/201" }
gharchive/issue
Add Array.prototype.includes sample @jeffposnick Do you mind if I take over this simple one? All yours! Thanks for taking this on!
2025-04-01T06:37:01.870838
2023-06-12T20:42:11
1753551371
{ "authors": [ "tunetheweb" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1138", "repo": "GoogleChrome/web-vitals", "url": "https://github.com/GoogleChrome/web-vitals/pull/355" }
gharchive/pull-request
Remove Internet Explorer from list of supported browsers As noted in #350 web-vitals no longer works in IE9 as of version 2.1.2. It probably still works in IE10 and IE11 but is no longer tested in either and so is at risk of breaking in either soon. With all versions of IE not officially unsupported I think it's best to remove any language of it's support from the README. Perhaps this should be a breaking change and saved for v4, but as noted v3 (and also the latest version of v3) already are not reflective of the actual situation so I'm tempted not to leave this for v4 and just merge this. Closing for now as not needed.
2025-04-01T06:37:01.880732
2017-10-16T14:21:38
265787820
{ "authors": [ "gauntface" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1139", "repo": "GoogleChrome/workbox", "url": "https://github.com/GoogleChrome/workbox/issues/904" }
gharchive/issue
Add definitions Glossary to Docs Taken from #640 From Kayce I'd avoid jargon-y terms like "precache" and "inject manifest" and just describe what it's doing in plain English. Or, if you do continue to use these terms, link to definitions. From Addy Workbox/Toolbox/Precache have historically opted to use these more jargon-y terms but if they're impeding new users from using our libraries, we should change that. I'm up for a glossary of terms or trying to stay away from using such jardon where we can. Created a place holder doc for this: https://docs.google.com/document/d/1IjDlBAsB4_SAx2XqPrNDGBj8tsR2UULLxCbJIYIk4NM/edit?usp=sharing Feel free to add content or suggestions and I'll ensure it lands in V3 release.
2025-04-01T06:37:01.887350
2023-08-31T06:58:53
1874903588
{ "authors": [ "mathiasbynens", "pwspot" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1140", "repo": "GoogleChromeLabs/chrome-for-testing", "url": "https://github.com/GoogleChromeLabs/chrome-for-testing/issues/48" }
gharchive/issue
It is unclear if chromedriver always matches the chrome version With chromedriver 114 and lower there was a dedicated URL to tell the latest chromedriver version: https://chromedriver.storage.googleapis.com/LATEST_RELEASE Here we have new links to tell the version, but it doesn't clearly specify if it is common for chrome and chromedriver: https://github.com/GoogleChromeLabs/chrome-for-testing#other-api-endpoints In https://googlechromelabs.github.io/chrome-for-testing/last-known-good-versions-with-downloads.json I can see that for Stable release there is a common version and I can assume that both chrome and chromedriver download links are based on that version. Could you please confirm if it is safe to assume that the versions always match? Nowhere in the docs it has been clearly stated. Here we have new links to tell the version, but it doesn't clearly specify if it is common for chrome and chromedriver: https://github.com/GoogleChromeLabs/chrome-for-testing#other-api-endpoints From the README: The set of “all CfT assets” for a given Chrome version is a matrix of supported binaries × platforms. The current list of supported binaries is: chrome a.k.a. Chrome for Testing (supported since v113.0.5672.0) chromedriver (supported since v115.0.5763.0) chrome-headless-shell (supported since v118.0.5944.0) The current list of supported platforms is: linux64 mac-arm64 mac-x64 win32 win64 Rest assured that any version number you obtain via the CfT JSON API or via the other endpoints are guaranteed to have all CfT assets available for that version.
2025-04-01T06:37:01.898133
2023-03-08T16:52:19
1615593382
{ "authors": [ "devAtQ", "sadym-chromium", "thiagowfx" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1141", "repo": "GoogleChromeLabs/chromium-bidi", "url": "https://github.com/GoogleChromeLabs/chromium-bidi/issues/518" }
gharchive/issue
Implement browsingContext.print (print to PDF) Implement Add browsingContext.print command by jgraham · Pull Request #363 · w3c/webdriver-bidi. CDP method: Page.printToPDF. Spec: https://w3c.github.io/webdriver-bidi/#command-browsingContext-print Tracking [x] Basic implementation [x] ~CDP does not have a shrinkToFit equivalent (c.f. comment)~ it is called preferCSSPageSize [x] Implement shrinkToFit [x] #577 [x] Issue 1430696: CDP Page.printToPDF crashes on tiny pages [ ] CDP Page.printToPDF throws invalid parameters exception on pages with tiny dimensions (e.g. 1x1 pixel) (cont.) [ ] Fix WPT tests /webdriver/tests/bidi/browsing_context/print/ [ ] (optional) Add E2E test: print to pdf -> open it in the browser (either data/pdf or file:///) -> take screenshot -> compare golden @sadym-chromium shrinkToFit does not have a CDP equivalent. What do we do in these situations? File a FR against CDP? What do we do in these situations? In general there are 3 options: Implement in CDP. Implement on top of existing CDP functionality + some extra logic. Roll back the spec part. Implement in CDP. How can we do so? This was also suggested by @whimboo on https://github.com/web-platform-tests/wpt/pull/38931#issuecomment-1468350079 Could you give me a pointer to the CDP repo? Implement in CDP. How can we do so? This was also suggested by @whimboo on web-platform-tests/wpt#38931 (comment) Could you give me a pointer to the CDP repo? sent PM @thiagowfx FYI there are bunch of new failing tests in WPT: https://github.com/GoogleChromeLabs/chromium-bidi/pull/625/files bidiMapper:mapperDebug:CDP sent ▸ { "id": 11, "method": "Page.navigate", "params": { "url": "data:application/pdf;base64,JVBERi0xLjQKJdPr6eEKMSAwIG9iago8PC9DcmVhdG9yIChDaHJvbWl1bSkKL1Byb2R1Y2VyIChTa2lhL1BERiBtMTE2KQovQ3JlYXRpb25EYXRlIChEOjIwMjMwNjEyMjEwNjM4KzAwJzAwJykKL01vZERhdGUgKEQ6MjAyMzA2MTIyMTA2MzgrMDAnMDAnKT4+CmVuZG9iagozIDAgb2JqCjw8L0xlbmd0aCAwPj4gc3RyZWFtCgplbmRzdHJlYW0KZW5kb2JqCjIgMCBvYmoKPDwvVHlwZSAvUGFnZQovUmVzb3VyY2VzIDw8L1Byb2NTZXQgWy9QREYgL1RleHQgL0ltYWdlQiAvSW1hZ2VDIC9JbWFnZUldPj4KL01lZGlhQm94IFswIDAgMjI2NzcuMTE5IDE3MDA4LjA4XQovQ29udGVudHMgMyAwIFIKL1N0cnVjdFBhcmVudHMgMAovUGFyZW50IDQgMCBSPj4KZW5kb2JqCjQgMCBvYmoKPDwvVHlwZSAvUGFnZXMKL0NvdW50IDEKL0tpZHMgWzIgMCBSXT4+CmVuZG9iago1IDAgb2JqCjw8L1R5cGUgL0NhdGFsb2cKL1BhZ2VzIDQgMCBSPj4KZW5kb2JqCnhyZWYKMCA2CjAwMDAwMDAwMDAgNjU1MzUgZiAKMDAwMDAwMDAxNSAwMDAwMCBuIAowMDAwMDAwMjAyIDAwMDAwIG4gCjAwMDAwMDAxNTUgMDAwMDAgbiAKMDAwMDAwMDM3NiAwMDAwMCBuIAowMDAwMDAwNDMxIDAwMDAwIG4gCnRyYWlsZXIKPDwvU2l6ZSA2Ci9Sb290IDUgMCBSCi9JbmZvIDEgMCBSPj4Kc3RhcnR4cmVmCjQ3OAolJUVPRgo=", "frameId": "A45327FAB30B9B82E7B22502D73B7520" }, "sessionId": "9371436D303A35CD0D25C2B137D48C64" } +0ms bidiMapper:mapperDebug:CDP received ◂ { "id": 11, "result": { "frameId": "A45327FAB30B9B82E7B22502D73B7520", "loaderId": "4C7ADCE66AA34C1EE27626ABB37C38EA", "errorText": "net::ERR_ABORTED" }, "sessionId": "9371436D303A35CD0D25C2B137D48C64" } +0ms FAILED [100%] When will this feature be released?
2025-04-01T06:37:01.900001
2023-10-25T13:31:19
1961438797
{ "authors": [ "thiagowfx" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1142", "repo": "GoogleChromeLabs/chromium-bidi", "url": "https://github.com/GoogleChromeLabs/chromium-bidi/pull/1477" }
gharchive/pull-request
ci: speed up prettier/eslint pre-commit hooks and use types instead of files c.f. https://github.com/pre-commit/mirrors-prettier/blob/main/.pre-commit-hooks.yaml and https://github.com/pre-commit/identify/blob/main/identify/extensions.py I had made a mistake. It's types_or, not types. Fixed.
2025-04-01T06:37:01.906768
2018-07-03T10:40:03
337842596
{ "authors": [ "jakearchibald" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1143", "repo": "GoogleChromeLabs/squoosh", "url": "https://github.com/GoogleChromeLabs/squoosh/issues/81" }
gharchive/issue
Vertical two-up Current mobile designs show the two-up working vertically, so this needs to be a feature of the two-up. Done!
2025-04-01T06:37:01.914903
2019-02-09T01:43:18
408383761
{ "authors": [ "markmandel", "nikkisingh0204", "pooneh-m", "thisisnotapril" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1144", "repo": "GoogleCloudPlatform/agones", "url": "https://github.com/GoogleCloudPlatform/agones/issues/577" }
gharchive/issue
Help us pick a new project logo! Hi all! We're thinking about getting a new logo for Agones, and we want your thoughts! The project founders have selected three options. We want you to vote and tell us which one you prefer! Here's the options: Option 1: Option 2: Option 3: Got a favorite? Now, how to vote! We're going to do this cheesy emoji style. For option 1, react to this post with the :+1: For option 2, react to this post with the :heart: For option 3, react to this post with the :tada: Voting will close on Feb 12! looking more attractive than the other two How about adding a little Kubernetes icon on the light blue controller of the first option to show that it is a controller for Kubernetes? How about adding a little Kubernetes icon on the light blue controller of the first option to show that it is a controller for Kubernetes? @markmandel @Kuqd what do you think of this idea? Honestly - nobody else does sub logos, so I'm personally not a fan. Better to be clean about it, and just have our own image. That all being said - looks like we have a clear winner :+1: Yep, sounds good! Now to design a t-shirt.... :) Since we've past the 12th, should we close this ticket? Winner winner is: :fireworks: :fireworks: :fireworks: :fireworks: :fireworks: :fireworks: :fireworks:
2025-04-01T06:37:01.916681
2024-07-10T21:39:22
2401788447
{ "authors": [ "Bslabe123" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1145", "repo": "GoogleCloudPlatform/ai-on-gke", "url": "https://github.com/GoogleCloudPlatform/ai-on-gke/pull/730" }
gharchive/pull-request
Jetstream Maxtext Deployment Module: All scale rules now in a single HPA Having multiple HPAs monitoring the same resource causes a race condition. Keeping all the rules in the same HPA fixes this. /gcbrun /gcbrun /gcbrun /gcbrun /gcbrun
2025-04-01T06:37:01.933561
2016-06-19T21:35:46
161093611
{ "authors": [ "garye", "sduskis" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1146", "repo": "GoogleCloudPlatform/cloud-bigtable-client", "url": "https://github.com/GoogleCloudPlatform/cloud-bigtable-client/pull/882" }
gharchive/pull-request
ResumingStreamingResultScanner should always close calls. ResumingStreamingResultScanner doesn't close calls when it throws an exception. Close the call for both types of exceptions. LGTM
2025-04-01T06:37:01.935484
2022-09-20T18:10:03
1379807137
{ "authors": [ "bjbloemker-google" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1147", "repo": "GoogleCloudPlatform/cloud-foundation-fabric", "url": "https://github.com/GoogleCloudPlatform/cloud-foundation-fabric/issues/829" }
gharchive/issue
compute-vm module does not support 'disk_encryption_key' property for compute instance templates Add disk_encryption_key field to google_compute_instance_template resource in compute-vm module (Supported when compute instance is made individually but not the instance template) Ref: https://registry.terraform.io/providers/hashicorp/google/latest/docs/resources/compute_instance_template#disk_encryption_key PR #830
2025-04-01T06:37:01.937534
2021-01-20T02:25:57
789568505
{ "authors": [ "bharathkkb" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1148", "repo": "GoogleCloudPlatform/cloud-foundation-toolkit", "url": "https://github.com/GoogleCloudPlatform/cloud-foundation-toolkit/pull/862" }
gharchive/pull-request
feat: enable comment bot all repos allows bot on all repos by default added deny list for disabling on specific modules This has been applied. Spot checked https://github.com/terraform-google-modules/terraform-google-network/pull/232 This has been applied. Spot checked https://github.com/terraform-google-modules/terraform-google-network/pull/232
2025-04-01T06:37:01.941063
2023-08-07T16:12:53
1839804696
{ "authors": [ "thomas-riccardi", "williamcruzme" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1149", "repo": "GoogleCloudPlatform/cloud-sql-proxy-operator", "url": "https://github.com/GoogleCloudPlatform/cloud-sql-proxy-operator/issues/402" }
gharchive/issue
Feature Request: support --quiet option Expected Behavior have a way to configure --quiet (or via env var): no useless logs increasing monitoring costs. cf https://github.com/GoogleCloudPlatform/cloud-sql-proxy/issues/1738 Actual Behavior currently we get lots of useless INFO logs with connections. Specifications Version: 1.0.2 Platform: gke +1
2025-04-01T06:37:01.945800
2018-01-22T11:54:12
290439254
{ "authors": [ "AthenaShi", "Tedezed" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1150", "repo": "GoogleCloudPlatform/cloudsql-proxy", "url": "https://github.com/GoogleCloudPlatform/cloudsql-proxy/issues/144" }
gharchive/issue
Cloudsql-Proxy in Kubernetes for cluster PostgreSQL Hi! I need to configure the proxy to attack to Postgres cluster, this has read and read-write nodes. Would the proxy support this configuration? Or you should deploy a proxy that manages this, such as crunchy-proxy Example diagrams: Cloudsql-Proxy managing read and read-write Crunchy-Poxy managing read and read-write and use Cloudsql-Proxy to connect with Google SQL THX! For graph one, do you need a LB that will automatic route to the master or the replica? If yes, then current proxy doesn't support this, go for graph two. If you just want one proxy to be able to setup the connection to both of the master and replica, and have the "Web Application" control which one to talk to, then graph one is supported. Hi, The solution that I take in my case is, deploy pgpool for sql management and set as master node and slave two cloudsql-proxy pointing to different set of servers. Pgpool managing read and read-write and use Cloudsql-Proxy to connect with Google SQL. THX! https://issuetracker.google.com/issues/37271935
2025-04-01T06:37:01.948245
2018-09-24T11:01:47
363098592
{ "authors": [ "illfelder", "rbalint" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1151", "repo": "GoogleCloudPlatform/compute-image-packages", "url": "https://github.com/GoogleCloudPlatform/compute-image-packages/pull/657" }
gharchive/pull-request
Improve support for compiling without C++11 This helps back-porting the packages to older releases using older compilers. I signed it! I signed it! Are the CLA issues resolved now? Can you please sync your branch and trigger the CLA checker again if your a member of the right group? Pull requests with multiple contributors breaks the CLA checker as far as I can tell. I created a test PR with only my commits and it does not pass the CLA test either: https://github.com/GoogleCloudPlatform/compute-image-packages/pull/719
2025-04-01T06:37:01.950160
2019-08-21T12:37:04
483394887
{ "authors": [ "gdzieleziesz" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1152", "repo": "GoogleCloudPlatform/deploymentmanager-samples", "url": "https://github.com/GoogleCloudPlatform/deploymentmanager-samples/issues/490" }
gharchive/issue
Unable to process unmanaged instance group with managed_instance_group.py managed_instance_group.py doesn't give an option to create unmanaged instnace group. In case of creation of unmanaged instnace group instance template should not be required. moved to cft Issue has been moved to the cft https://github.com/GoogleCloudPlatform/cloud-foundation-toolkit/issues/288
2025-04-01T06:37:01.952925
2021-02-19T17:03:26
812185966
{ "authors": [ "amanda-tarafa", "olavloite", "thiagotnunes" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1153", "repo": "GoogleCloudPlatform/dotnet-docs-samples", "url": "https://github.com/GoogleCloudPlatform/dotnet-docs-samples/pull/1263" }
gharchive/pull-request
fix: use current server time as version time for CreateBackup Use the current server time as version time instead of the local system time. The specs for the PITR samples indicate that the CreateBackup sample should set the current time as version time. By selecting the current timestamp from the server before actually creating the backup, the version time should be valid as it is after database creation and not in the future. Alternatively we could choose to remove the version time entirely from the sample (although that is not in line with the original intention for the sample). Fixes #1262 I think that the solution taken here is fair, but we have done something different in the Java Spanner tests, where we use the database earliest version time, instead of the server current timestamp (see https://github.com/googleapis/java-spanner/blob/master/google-cloud-spanner/src/test/java/com/google/cloud/spanner/it/ITPitrBackupAndRestore.java#L117). I would prefer if we did that unless you have concerns. I like that we are not showing the user where we are getting the date from. Failures are unrelated, merging...
2025-04-01T06:37:01.961889
2016-07-22T18:45:04
167107551
{ "authors": [ "daspecster", "dwmclary", "tseaver" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1154", "repo": "GoogleCloudPlatform/gcloud-python", "url": "https://github.com/GoogleCloudPlatform/gcloud-python/pull/2015" }
gharchive/pull-request
Adding UDF Resources to Queries The BigQuery client doesn't currently support adding UDF resources. This pull request adds a UDFResources object, and properties to set these resources on queries and QueryJobs. I signed it! See #2007 for an alternative. All commits were authored by me. I had a different email set as primary when the initial commit was made. I'm not entirely sure how that not works but you might have to rebase and change the user on those commits. On Friday, July 22, 2016, Daniel McClary<EMAIL_ADDRESS>wrote: All commits were authored by me. I had a different email set as primary when the initial commit was made. — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/GoogleCloudPlatform/gcloud-python/pull/2015#issuecomment-234679998, or mute the thread https://github.com/notifications/unsubscribe-auth/AALvyElSGhqFqGtucoId7um9jqo9fiHlks5qYVB7gaJpZM4JTBqe . -- Tom Schultz OK rebased to reflect the proper author email. @dwmclary Thanks for the patch! @dwmclary It appears that Github / the CLA bot do not think 1ff32523f5de4d9ecacf7294ace17a77b57211d0 was authored by you, but by another blind poet of the same name. Authored by me, and now painfully rebased to reflect it. OK, should be all fixed as soon as CLAbot looks at the authorship again. @googlebot can you check the CLA again? OK, this rebase stuff is silly, I'm going to close this PR and open a new one with the right authorship. @tseaver @daspecster Any idea why the CLA isn't flipping? At this point all commits and author emails should point to me. There doesn't appear to be any rebase magic I can perform beyond this. Worst-case, I can re-fork and resubmit if we can't make @googlebot behave. Re-opened here
2025-04-01T06:37:01.963019
2023-05-18T08:08:08
1715208954
{ "authors": [ "abhigupta1207", "junggil" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1155", "repo": "GoogleCloudPlatform/gcpdiag", "url": "https://github.com/GoogleCloudPlatform/gcpdiag/pull/67" }
gharchive/pull-request
CloudSQL New BP Ext Rule for maintenance window cloudsql/BP_EXT/2023_001 Add new rule for CloudSQL - cloudsql/BP_EXT/2023_001 LGTM. Thanks for the PR
2025-04-01T06:37:01.964892
2018-05-01T18:29:54
319282757
{ "authors": [ "tseaver", "tswast" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1156", "repo": "GoogleCloudPlatform/google-cloud-python", "url": "https://github.com/GoogleCloudPlatform/google-cloud-python/pull/5281" }
gharchive/pull-request
BigQuery: add UnknownJob type for redacted jobs. Fixes https://github.com/GoogleCloudPlatform/google-cloud-python/issues/5220 /cc @shollyman I can't see any benefit to the caller in getting back an instance of UnknownJob: I would certainly find it disconcerting and useless. ISTM that the caller's experience would be better if the backend should skip reporting jobs for which the caller does not access. As an alternative, Client.list_jobs() could just discard entries without any configuration. That said, I don't see an issue with the code here, so feel free to merge if the consensus is agin' me.
2025-04-01T06:37:01.966739
2017-01-24T19:49:06
202921561
{ "authors": [ "blowmage", "quartzmo" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1157", "repo": "GoogleCloudPlatform/google-cloud-ruby", "url": "https://github.com/GoogleCloudPlatform/google-cloud-ruby/issues/1205" }
gharchive/issue
Document Speech low-level API (GAPIC code) Types such as v1beta1/speech_api.rb are not currently published to the gh-pages documentation. [] Remove exclusions from .yardopts [] Add navigation to GAPIC/gRPC code Speech will need the functionality covered in #1180 in order to properly expose the GAPIC types. The GAPIC and Protobuf classes share a common namespace.
2025-04-01T06:37:01.972459
2018-05-24T00:36:06
325922852
{ "authors": [ "blowmage", "frankyn", "ryanmats" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1158", "repo": "GoogleCloudPlatform/google-cloud-ruby", "url": "https://github.com/GoogleCloudPlatform/google-cloud-ruby/issues/2103" }
gharchive/issue
Firestore - Passing Information out of Transaction Functions In the Firestore documentation we have code samples in the other languages for how to pass information out of Firestore Transactions back to the caller. (I/e a simple true/false boolean for whether population is updated). Is it possible to do this using the Ruby client? If so, how would this work? Thanks! Hi @ryanmats, thank you for the question and for using the Ruby Firestore client. It is possible to pass information back from a transaction, but to do this you should create a separate method context to pass values out. The Firestore transaction uses a Ruby closure which shares the Binding that the closure is created in. But the transaction closure may be executed more than once in response to errors on the server. So you should wait until the transaction is complete before using the value. Here is my take on the Python version of the documentation you referenced: require "google/cloud/firestore" firestore = Google::Cloud::Firestore.new # Get a document reference city_ref = firestore.doc "cities/SF" # define method that will create a transaction def update_in_transaction firestore, city_ref updated = nil firestore.transaction do |tx| snapshot = tx.get_all(city_ref).first new_population = snapshot.get(:population) + 1 if new_population < 1000000 tx.update city_ref, population: new_population updated = true else updated = false end end updated end # call method that will create a transaction result = update_in_transaction firestore, city_ref if result puts "Population updated" else puts "Sorry! Population is too big." end The Python code gets a DocumentSnapshot object from a DocumentReference object using a Transaction with this code: snapshot = city_ref.get(transaction=transaction). We don't have that in Ruby yet, which is why the line snapshot = tx.get_all(city_ref).first is used. I would prefer that this line be this instead: snapshot = city_ref.get transaction: tx, so I'll create an issue to implement that. @beccca @frankyn Any idea why the Ruby example is not present on this documentation? Hi @blowmage, by this documentation you mean the Firestore Transactions documentation. @ryanmats is writing samples for Ruby. Ah, gotcha. I see what's going on now. :) BTW, my previous recollection was off. The original trade-off was to allow Transaction#get to get either a DocumentSnapshot or fulfill a Query. So instead of stating snapshot = tx.get_all(city_ref).first you can just say snapshot = tx.get city_ref. @blowmage I think I'm good - thanks for all the help!
2025-04-01T06:37:01.973876
2016-06-30T17:02:06
163215137
{ "authors": [ "csells", "ivannaranjo" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1159", "repo": "GoogleCloudPlatform/google-cloud-visualstudio", "url": "https://github.com/GoogleCloudPlatform/google-cloud-visualstudio/issues/111" }
gharchive/issue
Hide zones in the GCE section of the Cloud Explorer by default We should make the zones part of the GCE tree in the explorer collapsible with an icon button (i.e. just show a flat list of VMs w/o the zone in the tree). Further, we should hide the zones by default (most folks don't want to dig through zones first before getting to their list of VMs). To provide the same data for VMs in multiple zones, we should build the zone name into the machine page, e.g. "machine - zone". This was fixed in 85f8119
2025-04-01T06:37:02.023479
2020-09-29T12:20:18
711057646
{ "authors": [ "poli44" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1160", "repo": "GoogleCloudPlatform/iot-smart-home-cloud", "url": "https://github.com/GoogleCloudPlatform/iot-smart-home-cloud/issues/21" }
gharchive/issue
what am I doing wrong poli@poli-Parallels-Virtual-Platform:~/Desktop/iot-smart-home-cloud-master/firebase/functions$ poli@poli-Parallels-Virtual-Platform:~/Desktop/iot-smart-home-cloud-master/firebase/functions$ firebase functions:config:set \ smarthome.id=$CLIENT_ID \ smarthome.secret=$CLIENT_SECRET ✔ Functions config updated. Please deploy your functions for the change to take effect by running firebase deploy --only functions poli@poli-Parallels-Virtual-Platform:~/Desktop/iot-smart-home-cloud-master/firebase/functions$ firebase functions:config:set \ smarthome.key="my-secret-string" ✔ Functions config updated. Please deploy your functions for the change to take effect by running firebase deploy --only functions poli@poli-Parallels-Virtual-Platform:~/Desktop/iot-smart-home-cloud-master/firebase/functions$ npm install audited 425 packages in 5.749s 33 packages are looking for funding run npm fund for details found 0 vulnerabilities poli@poli-Parallels-Virtual-Platform:~/Desktop/iot-smart-home-cloud-master/firebase/functions$ firebase deploy === Deploying to 'actionhome-4baaa'... i deploying database, storage, firestore, functions, hosting, remoteconfig Running command: npm --prefix "$RESOURCE_DIR" run lint functions@ lint /home/poli/Desktop/iot-smart-home-cloud-master/firebase/functions eslint . /home/poli/Desktop/iot-smart-home-cloud-master/firebase/functions/device-cloud/device-configuration.js 33:97 warning Expected to return a value at the end of arrow function consistent-return /home/poli/Desktop/iot-smart-home-cloud-master/firebase/functions/device-cloud/register-device.js 33:75 warning Expected to return a value at the end of arrow function consistent-return /home/poli/Desktop/iot-smart-home-cloud-master/firebase/functions/smart-home/device-model.js 72:2 error Unnecessary semicolon no-extra-semi 100:2 error Unnecessary semicolon no-extra-semi 132:2 error Unnecessary semicolon no-extra-semi /home/poli/Desktop/iot-smart-home-cloud-master/firebase/functions/smart-home/fulfillment.js 68:19 error Unexpected await inside a loop no-await-in-loop ✖ 6 problems (4 errors, 2 warnings) 3 errors and 0 warnings potentially fixable with the --fix option. npm ERR! code ELIFECYCLE npm ERR! errno 1 npm ERR! functions@ lint: eslint . npm ERR! Exit status 1 npm ERR! npm ERR! Failed at the functions@ lint script. npm ERR! This is probably not a problem with npm. There is likely additional logging output above. npm ERR! Which Firebase CLI features do you want to set up for this folder? Press Space to select features, then Enter to confirm your choices. Database: Deploy Firebase Realtime Database Rules, Firestore: D eploy rules and create indexes for Firestore, Functions: Configure and deploy Cloud Functions, Hosti ng: Configure and deploy Firebase Hosting sites, Storage: Deploy Cloud Storage security rules, Emula tors: Set up local emulators for Firebase features, Remote Config: Get, deploy, and rollback configu rations for Remote Config === Project Setup First, let's associate this project directory with a Firebase project. You can create multiple project aliases by running firebase use --add, but for now we'll just set up a default project. i Using project actionhome-4baaa (ActionHome) === Database Setup Firebase Realtime Database Rules allow you to define how your data should be structured and when your data can be read from and written to. ? What file should be used for Database Rules? database.rules.json ✔ Database Rules for actionhome-4baaa have been downloaded to database.rules.json. Future modifications to database.rules.json will update Database Rules when you run firebase deploy. === Firestore Setup Firestore Security Rules allow you to define how and when to allow requests. You can keep these rules in your project directory and publish them with firebase deploy. ? What file should be used for Firestore Rules? firestore.rules ? File firestore.rules already exists. Do you want to overwrite it with the Firestore Rules from the Firebase Console? Yes Firestore indexes allow you to perform complex queries while maintaining performance that scales with the size of the result set. You can keep index definitions in your project directory and publish them with firebase deploy. ? What file should be used for Firestore indexes? firestore.indexes.json ? File firestore.indexes.json already exists. Do you want to overwrite it with the Firestore Indexes from the Firebase Console? Yes === Functions Setup A functions directory will be created in your project with a Node.js package pre-configured. Functions can be deployed with firebase deploy. ? What language would you like to use to write Cloud Functions? JavaScript ? Do you want to use ESLint to catch probable bugs and enforce style? Yes ? File functions/package.json already exists. Overwrite? Yes ✔ Wrote functions/package.json ✔ Wrote functions/.eslintrc.json ? File functions/index.js already exists. Overwrite? Yes ✔ Wrote functions/index.js ✔ Wrote functions/.gitignore ? Do you want to install dependencies with npm now? Yes npm WARN deprecated<EMAIL_ADDRESS>CircularJSON is in maintenance only, flatted is its successor. <EMAIL_ADDRESS>postinstall /home/poli/Desktop/iot-smart-home-cloud-master/firebase/functions/node_modules/protobufjs node scripts/postinstall added 426 packages from 264 contributors and audited 426 packages in 48.911s 33 packages are looking for funding run npm fund for details found 12 low severity vulnerabilities run npm audit fix to fix them, or npm audit for details === Hosting Setup Your public directory is the folder (relative to your project directory) that will contain Hosting assets to be uploaded with firebase deploy. If you have a build process for your assets, use your build's output directory. ? What do you want to use as your public directory? public ? Configure as a single-page app (rewrite all urls to /index.html)? Yes ✔ Wrote public/index.html === Storage Setup Firebase Storage Security Rules allow you to define how and when to allow uploads and downloads. You can keep these rules in your project directory and publish them with firebase deploy. ? What file should be used for Storage Rules? storage.rules === Emulators Setup ? Which Firebase emulators do you want to set up? Press Space to select emulators, then Enter to con firm your choices. (Press to select, to toggle all, to invert selection)Functions Em ulator, Firestore Emulator, Hosting Emulator ? Which port do you want to use for the functions emulator? 5001 ? Which port do you want to use for the firestore emulator? 8080 ? Which port do you want to use for the hosting emulator? 5000 ? Would you like to enable the Emulator UI? Yes ? Which port do you want to use for the Emulator UI (leave empty to use any available port)? NaN ? Would you like to download the emulators now? Yes === Remoteconfig Setup ? What file should be used for your Remote Config template? remoteconfig.template.json i Writing configuration info to firebase.json... i Writing project information to .firebaserc... ✔ Firebase initialization complete! poli@poli-Parallels-Virtual-Platform:~/Desktop/iot-smart-home-cloud-master/firebase$ firebase functions:config:set \ cloudiot.region=$REGION \ cloudiot.registry=$REGISTRY ✔ Functions config updated. Please deploy your functions for the change to take effect by running firebase deploy --only functions poli@poli-Parallels-Virtual-Platform:~/Desktop/iot-smart-home-cloud-master/firebase$ firebase functions:config:set \ smarthome.id=$CLIENT_ID \ smarthome.secret=$CLIENT_SECRET ✔ Functions config updated. Please deploy your functions for the change to take effect by running firebase deploy --only functions poli@poli-Parallels-Virtual-Platform:~/Desktop/iot-smart-home-cloud-master/firebase$ firebase functions:config:set \ smarthome.key="my-secret-string" ✔ Functions config updated. Please deploy your functions for the change to take effect by running firebase deploy --only functions poli@poli-Parallels-Virtual-Platform:~/Desktop/iot-smart-home-cloud-master/firebase$ firebase deploy === Deploying to 'actionhome-4baaa'... i deploying database, storage, firestore, functions, hosting, remoteconfig Running command: npm --prefix "$RESOURCE_DIR" run lint functions@ lint /home/poli/Desktop/iot-smart-home-cloud-master/firebase/functions eslint . /home/poli/Desktop/iot-smart-home-cloud-master/firebase/functions/device-cloud/device-configuration.js 33:97 warning Expected to return a value at the end of arrow function consistent-return /home/poli/Desktop/iot-smart-home-cloud-master/firebase/functions/device-cloud/register-device.js 33:75 warning Expected to return a value at the end of arrow function consistent-return /home/poli/Desktop/iot-smart-home-cloud-master/firebase/functions/smart-home/device-model.js 72:2 error Unnecessary semicolon no-extra-semi 100:2 error Unnecessary semicolon no-extra-semi 132:2 error Unnecessary semicolon no-extra-semi /home/poli/Desktop/iot-smart-home-cloud-master/firebase/functions/smart-home/fulfillment.js 68:19 error Unexpected await inside a loop no-await-in-loop ✖ 6 problems (4 errors, 2 warnings) 3 errors and 0 warnings potentially fixable with the --fix option. npm ERR! code ELIFECYCLE npm ERR! errno 1 npm ERR! functions@ lint: eslint . npm ERR! Exit status 1 npm ERR! npm ERR! Failed at the functions@ lint script. npm ERR! This is probably not a problem with npm. There is likely additional logging output above. npm ERR! A complete log of this run can be found in: npm ERR! /home/poli/.npm/_logs/2020-09-29T12_10_09_161Z-debug.log Error: functions predeploy error: Command terminated with non-zero exit code1 poli@poli-Parallels-Virtual-Platform:~/Desktop/iot-smart-home-cloud-master/firebase$ ls database.rules.json firestore.rules public storage.rules firebase.json functions README.md firestore.indexes.json package-lock.json remoteconfig.template.json poli@poli-Parallels-Virtual-Platform:~/Desktop/iot-smart-home-cloud-master/firebase$ cd functions/ poli@poli-Parallels-Virtual-Platform:~/Desktop/iot-smart-home-cloud-master/firebase/functions$ npm install audited 426 packages in 5.93s 33 packages are looking for funding run npm fund for details found 12 low severity vulnerabilities run npm audit fix to fix them, or npm audit for details poli@poli-Parallels-Virtual-Platform:~/Desktop/iot-smart-home-cloud-master/firebase/functions$ firebase deploy === Deploying to 'actionhome-4baaa'... i deploying database, storage, firestore, functions, hosting, remoteconfig Running command: npm --prefix "$RESOURCE_DIR" run lint functions@ lint /home/poli/Desktop/iot-smart-home-cloud-master/firebase/functions eslint . /home/poli/Desktop/iot-smart-home-cloud-master/firebase/functions/device-cloud/device-configuration.js 33:97 warning Expected to return a value at the end of arrow function consistent-return /home/poli/Desktop/iot-smart-home-cloud-master/firebase/functions/device-cloud/register-device.js 33:75 warning Expected to return a value at the end of arrow function consistent-return /home/poli/Desktop/iot-smart-home-cloud-master/firebase/functions/smart-home/device-model.js 72:2 error Unnecessary semicolon no-extra-semi 100:2 error Unnecessary semicolon no-extra-semi 132:2 error Unnecessary semicolon no-extra-semi /home/poli/Desktop/iot-smart-home-cloud-master/firebase/functions/smart-home/fulfillment.js 68:19 error Unexpected await inside a loop no-await-in-loop ✖ 6 problems (4 errors, 2 warnings) 3 errors and 0 warnings potentially fixable with the --fix option. npm ERR! code ELIFECYCLE npm ERR! errno 1 npm ERR! functions@ lint: eslint . npm ERR! Exit status 1 npm ERR! npm ERR! Failed at the functions@ lint script. npm ERR! This is probably not a problem with npm. There is likely additional logging output above. npm ERR! A complete log of this run can be found in: npm ERR! /home/poli/.npm/_logs/2020-09-29T12_11_00_100Z-debug.log Error: functions predeploy error: Command terminated with non-zero exit code1 Having trouble? Try firebase [command] --help poli@poli-Parallels-Virtual-Platform:~/Desktop/iot-smart-home-cloud-master/firebase/functions$ npm install audited 426 packages in 6.197s 33 packages are looking for funding run npm fund for details found 12 low severity vulnerabilities run npm audit fix to fix them, or npm audit for details poli@poli-Parallels-Virtual-Platform:~/Desktop/iot-smart-home-cloud-master/firebase/functions$ cd .. poli@poli-Parallels-Virtual-Platform:~/Desktop/iot-smart-home-cloud-master/firebase$ npm install npm WARN saveError ENOENT: no such file or directory, open '/home/poli/Desktop/iot-smart-home-cloud-master/firebase/package.json' npm WARN enoent ENOENT: no such file or directory, open '/home/poli/Desktop/iot-smart-home-cloud-master/firebase/package.json' npm WARN firebase No description npm WARN firebase No repository field. npm WARN firebase No README data npm WARN firebase No license field. up to date in 1.166s found 0 vulnerabilities poli@poli-Parallels-Virtual-Platform:~/Desktop/iot-smart-home-cloud-master/firebase$ firebae deploy firebae: command not found poli@poli-Parallels-Virtual-Platform:~/Desktop/iot-smart-home-cloud-master/firebase$ firebase deploy === Deploying to 'actionhome-4baaa'... i deploying database, storage, firestore, functions, hosting, remoteconfig Running command: npm --prefix "$RESOURCE_DIR" run lint functions@ lint /home/poli/Desktop/iot-smart-home-cloud-master/firebase/functions eslint . /home/poli/Desktop/iot-smart-home-cloud-master/firebase/functions/device-cloud/device-configuration.js 33:97 warning Expected to return a value at the end of arrow function consistent-return /home/poli/Desktop/iot-smart-home-cloud-master/firebase/functions/device-cloud/register-device.js 33:75 warning Expected to return a value at the end of arrow function consistent-return /home/poli/Desktop/iot-smart-home-cloud-master/firebase/functions/smart-home/device-model.js 72:2 error Unnecessary semicolon no-extra-semi 100:2 error Unnecessary semicolon no-extra-semi 132:2 error Unnecessary semicolon no-extra-semi /home/poli/Desktop/iot-smart-home-cloud-master/firebase/functions/smart-home/fulfillment.js 68:19 error Unexpected await inside a loop no-await-in-loop ✖ 6 problems (4 errors, 2 warnings) 3 errors and 0 warnings potentially fixable with the --fix option. npm ERR! code ELIFECYCLE npm ERR! errno 1 npm ERR! functions@ lint: eslint . npm ERR! Exit status 1 npm ERR! npm ERR! Failed at the functions@ lint script. npm ERR! This is probably not a problem with npm. There is likely additional logging output above. npm ERR! A complete log of this run can be found in: npm ERR! /home/poli/.npm/_logs/2020-09-29T12_13_08_980Z-debug.log Error: functions predeploy error: Command terminated with non-zero exit code1 poli@poli-Parallels-Virtual-Platform:~/Desktop/iot-smart-home-cloud-master/firebase$ ok i delete 3 ";" in file smart-home\device-model.js delete await on line 68 smart-home\fulfillment.js and deploy result: 2 problems (0 errors, 2 warnings) is this the correct solution?
2025-04-01T06:37:02.033411
2015-03-10T02:42:51
60442651
{ "authors": [ "ghodss", "nikhiljindal" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1161", "repo": "GoogleCloudPlatform/kubernetes", "url": "https://github.com/GoogleCloudPlatform/kubernetes/pull/5226" }
gharchive/pull-request
Update README to include design overview To me, DESIGN.md is the most helpful document to actually understand what Kubernetes is. I think it belongs in the main README. LGTM, Thanks!
2025-04-01T06:37:02.039080
2015-03-17T06:04:12
62296824
{ "authors": [ "piosz", "roberthbailey", "satnam6502" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1162", "repo": "GoogleCloudPlatform/kubernetes", "url": "https://github.com/GoogleCloudPlatform/kubernetes/pull/5536" }
gharchive/pull-request
Removed guestbook.sh e2e test The test is rewritten in Go #5045 and is stable cc @jlowdermilk Was this test being kept around to cover some kubectl testing? @zmerlynn kubectl is now being sufficiently tested by the go tests. The shell test is being kept around until we are satisfied that the Go port is at least as good of an e2e test as the shell test was. For other transitions, we've waiting between 2 and 7 days to verify that we aren't losing test coverage during the migration. We've seen a few flakes in the new go test that have hopefully been fixed by #5595. I'm going to wait another day or two to merge this until we see the go test showing higher reliability than the shell test. Should wait until #5604 is fixed. @piosz Can you rebase? I'm going to take a look again today and see if we can finally nuke this shell test. Done. I've taken a detailed look at the test history for Shell tests that guestbook.sh passes and kubectl guestbook should create and stop a working application on both GCE and GKE. For Shell tests that guestbook.sh passes, I found that it was failing consistently on GKE until yesterday when #5749 was merged. Since then it is still a bit flaky on GKE (6 of 30 runs), both hitting timeouts and failing with Wrong entry received: {"data": ""}. It has been similarly flaky on GCE (also 6 in the last 30 runs but at different times), failing with similar errors. On the other hand kubectl guestbook should create and stop a working application has flaked a few times on both GKE and GCE over the last 30 runs, mainly due to the 60 second timeout which was extended to 3 minutes in #5845. Overall, the replacement test seems more stable than the shell test so it's time to remove the shell test. There is a slight decrease in test coverage removing the shell test because the shell test was using the public IP associated with the service to read/write from the guestbook whereas the go test is using the kubernetes master as a proxy to reach the service (so the test will pass even if the external LB creation fails). This gap is now covered by Services should be able to create a functioning external load balancer which was added in #5772. I'll merge on green. Shippable beat travis! Merging. Thanks!
2025-04-01T06:37:02.048124
2015-03-31T19:19:01
65526283
{ "authors": [ "brendandburns", "eparis", "jsafrane", "nzwulfin", "vmarmol" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1163", "repo": "GoogleCloudPlatform/kubernetes", "url": "https://github.com/GoogleCloudPlatform/kubernetes/pull/6237" }
gharchive/pull-request
example ansible setup repo This is a basic ansible repo that will do a couple of things set up an etcd node set up a master running apiserver, scheduler, controller-manager setup any number of nodes Hopefully this can be expanded to do things like set up skydns, set up a private docker repo, set up an overlay network (flannel) etc etc. But right now all it does is set up etcd and configure a master and nodes. @nzwulfin Today there exist 2 repos which try to configure kubernetes using ansible. https://github.com/eparis/kubernetes-ansible https://github.com/nzwulfin/ansible-atomic I'd like to see these merged, potentially IN the kubernetes tree. If @nzwulfin is not interested, we may want to move forward anyway. This is a stripped down version of both my repo and ideas from his repo. It does very very little. This is not a replacement for the salt-stack cluster setup. It only does some small portion of those things. @eparis any recommended reviewers? :) I'm going to say we blind LGTM this, as it is in contrib, and I'm not sure we have anyone who's expert enough in ansible to say anything deep. A README.md might be nice to help explain how to use it. @jsafrane could also be a reviewer (I believe @nzwulfin is out until next week) I miss some readme on how to use it. It's very Fedora specific. This is not bad, it's just not mentioned anywhere. Other distros don't have /etc/kubernetes, they have /etc/default with different option names. What is ansible/library/rpm_facts.py good for? All these facts can be acquired using Ansible itself. How do I use ansible to get installed (or running) rpms? I couldn't find a way other than my own module to collect those facts.... The same way as you do it in the python module: --- - shell: rpm -q iptables-service register: has_iptables yes, that's what i did do, I just hated having when has_iptables.rc == 0 throughout the code when custom facts lets me do when has_iptables guess it's just style, i'm not strongly attached. I do not have any strong preference either, I just find the extra python module to be overkill here. What about the other facts, like is_atomic? It's not trivial to get, you need 3 tasks to do it and still it's in yaml and not in python. You're probably right, I should move back into the playbook. @jsafrane What do you think of this? It looks good to me. @eparis I'm back around now. I think this is a good place to start some merging, I just need to check internally on the CLA for good corporate citizenship. I'm happy to merge this, once the CLA is set. erm, forgot it was eparis as the original author... merging.
2025-04-01T06:37:02.157765
2021-03-05T19:06:23
823342229
{ "authors": [ "aabmass", "alexvanboxel", "tbarker25" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1164", "repo": "GoogleCloudPlatform/opentelemetry-operations-go", "url": "https://github.com/GoogleCloudPlatform/opentelemetry-operations-go/issues/151" }
gharchive/issue
Span messages are prefixed with Span. (incorrect behaviour) I know this is done intentionally, but it makes the Stackdriver exporter qazi unusable. It is also not semantically correct as it does change the meaning. Example: If using java auto instrumentation and I want to search for a SELECT statement, I don't expect to need to search for Span.Client-SELECT ... This is the commit that introduced the change: https://github.com/GoogleCloudPlatform/opentelemetry-operations-go/commit/a1393affee9a13ed21974ae1d791da1b3189f4a0 If the change is to surface the kind in the UI, it would be better to change the UI and not the name of the span. Related https://github.com/GoogleCloudPlatform/opentelemetry-operations-go/issues/109 @alexvanboxel we've decided to remove the whole Span.<kind>., should be an easy fix This should be fixed by #155
2025-04-01T06:37:02.164768
2023-01-31T19:37:23
1564891493
{ "authors": [ "nehanene15" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1165", "repo": "GoogleCloudPlatform/professional-services-data-validator", "url": "https://github.com/GoogleCloudPlatform/professional-services-data-validator/issues/688" }
gharchive/issue
generate-table-partitions: Extend to support non-numeric partition keys Currently generate-table-partitions only supports numeric, monotonically increasing partition keys. We should extend support to non numeric keys with the use of NTILE. This issue would involve building a query using NTILE in the partition_row_builder and generate a new function (similar to _get_partition_key_filters) to run the query and build filters. Closed by #889
2025-04-01T06:37:02.169701
2023-06-09T14:21:28
1749994591
{ "authors": [ "iamkarlson", "mhoran" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1166", "repo": "GoogleCloudPlatform/prometheus-engine", "url": "https://github.com/GoogleCloudPlatform/prometheus-engine/issues/487" }
gharchive/issue
Prometheus UI is broken after adding authentication After implementing #242, Prometheus UI is broken. It doesn't pass any basicAuth headers to GMP. I solved it by adding some browser extension. Here's how: Configure Prometheus UI to use AUTH_USERNAME and AUTH_PASSWORD Run Prometheus UI: kubectl port-forward --namespace prometheus svc/infra-monitoring-prometheus-frontend-production 8080:9090 Open the browser at http://localhost:8080, and make sure it's failing to load metrics But Grafana access works fine. I found a way to overcome it: Install Modify Header Value addon to your browser. In addon, setup basic auth header with the values from your prometheus chart. Make sure your prometheus dashboard is now working (i.e. no erros, metrics can be searched, etc.). Are you running the latest frontend? I believe this was fixed in #339, which was released in v0.7.0 of the frontend.
2025-04-01T06:37:02.176780
2022-09-06T16:53:53
1363566087
{ "authors": [ "fmichaelobrien", "shaunmitchellve" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1167", "repo": "GoogleCloudPlatform/pubsec-declarative-toolkit", "url": "https://github.com/GoogleCloudPlatform/pubsec-declarative-toolkit/issues/93" }
gharchive/issue
As an arete create user I am running into a missing wait state between project creation and api enablement - running on an existing target project or running create twice is a workaround running arete create after manually deleting the CC cluster - cluster creation does not kick in without deleting the .arete cache running into existing .arete config - deleting - separate issue in https://github.com/GoogleCloudPlatform/pubsec-declarative-toolkit/issues/94 error - see end of details (timing/wait step between project creation and services enablement) 2:48PM INF Creating Config Controller Cluster.... 2:48PM FTL error="API [krmapihosting.googleapis.com] not enabled on project [153970848512]. Would you like to enable and retry (this will take a few minutes)? (y/N)? ERROR: (gcloud.anthos.config.controller.create) PERMISSION_DENIED: KRM API Hosting API has not been used in project<PHONE_NUMBER>12 before or it is disabled. Enable it by visiting https://console.developers.google.com/apis/api/krmapihosting.googleapis.com/overview?project=153970848512 then retry. If you enabled this API recently, wait a few minutes for the action to propagate to our systems and retry.- '@type': type.googleapis.com/google.rpc.Help links: - description: Google developers console API activation url: https://console.developers.google.com/apis/api/krmapihosting.googleapis.com/overview?project=153970848512- '@type': type.googleapis.com/google.rpc.ErrorInfo domain: googleapis.com metadata: consumer: projects/153970848512 service: krmapihosting.googleapis.com reason: SERVICE_DISABLED" details admin_root@cloudshell:~ (landing-zone-controller-w8hwa)$ arete create pdt-cno-kcc --region=northamerica-northeast1 --project=pubsec-declarative-toolkit-cno 2:37PM INF Config Controller setup complete admin_root@cloudshell:~ (landing-zone-controller-w8hwa)$ rm -rf .arete/ .bashrc .docker/ .npm/ .redhat/ .bash_history .cache/ gopath/ .profile .theia/ .bash_logout .config/ .kube/ README-cloudshell.txt wse_github/ admin_root@cloudshell:~ (landing-zone-controller-w8hwa)$ ls -la .arete/ total 20 drwxr--r-- 2 admin_root admin_root 4096 Sep 2 18:15 . drwxr-xr-x 12 admin_root admin_root 4096 Sep 6 14:37 .. -rw-r--r-- 1 admin_root admin_root 46 Aug 31 15:06 config.yaml -rw------- 1 admin_root admin_root 100 Aug 31 15:32 .create -rw-r--r-- 1 admin_root admin_root 1318 Sep 2 18:15 solutions.yaml admin_root@cloudshell:~ (landing-zone-controller-w8hwa)$ rm -rf .arete/ admin_root@cloudshell:~ (landing-zone-controller-w8hwa)$ arete create pdt-cno-kcc --region=northamerica-northeast1 --project=pubsec-declarative-toolkit-cno 2:40PM INF Enabling required services... 2:40PM INF Operation "operations/acat.p2-491974186555-2e6beaa9-f3df-4413-9a28-419db485c8e0" finished successfully. 2:41PM INF Creating Config Controller Cluster.... 2:41PM FTL error="ERROR: (gcloud.anthos.config.controller.create) ALREADY_EXISTS: Resource 'projects/pubsec-declarative-toolkit-cno/locations/northamerica-northeast1/krmApiHosts/pdt-cno-kcc' already exists- '@type': type.googleapis.com/google.rpc.ResourceInfo resourceName: projects/pubsec-declarative-toolkit-cno/locations/northamerica-northeast1/krmApiHosts/pdt-cno-kcc" deleting project - attempt to reuse may fail on 30 day deleted cache - will try admin_root@cloudshell:~ (landing-zone-controller-w8hwa)$ gcloud projects delete pubsec-declarative-toolkit-cno Your project will be deleted. Do you want to continue (Y/n)? y Deleted [https://cloudresourcemanager.googleapis.com/v1/projects/pubsec-declarative-toolkit-cno]. You can undo this operation for a limited period by running the command below. $ gcloud projects undelete pubsec-declarative-toolkit-cno See https://cloud.google.com/resource-manager/docs/creating-managing-projects for information on shutting down projects. admin_root@cloudshell:~ (landing-zone-controller-w8hwa)$ arete create pdt-cno-kcc --region=northamerica-northeast1 --project=pubsec-declarative-toolkit-cno ✔ My Billing Account - 019..3D ✔ nuage-cloud.org - 471..7 ✔ Folder Level ✔ pdt - 346..8 2:44PM FTL error="ERROR: (gcloud.projects.create) Project creation failed. The project ID you specified is already in use by another project. Please try an alternative ID." admin_root@cloudshell:~ (landing-zone-controller-w8hwa)$ arete create pdt-cno-kcc --region=northamerica-northeast1 --project=pubsec-declarative-toolkit-cno2 ✔ My Billing Account - 01..3D ✔ nuage-cloud.org -<PHONE_NUMBER>47 ✔ Folder Level ✔ pdt -<PHONE_NUMBER>68 2:45PM FTL error="ERROR: (gcloud.projects.create) argument PROJECT_ID: Bad value [pubsec-declarative-toolkit-cno2]: Project IDs are immutable and can be set only during project creation. They must start with a lowercase letter and can have lowercase ASCII letters, digits or hyphens. Project IDs must be between 6 and 30 characters.Usage: gcloud projects create [PROJECT_ID] [optional flags] optional flags may be --enable-cloud-apis | --folder | --help | --labels | --name | --organization | --set-as-defaultFor detailed information on this command and its flags, run: gcloud projects create --help" 30 char limit admin_root@cloudshell:~ (landing-zone-controller-w8hwa)$ arete create pdt-cno-kcc --region=northamerica-northeast1 --project=pubsec-declarative-tk-cno2 ✔ My Billing Account - 019952-0D0AAC-777E3D ✔ nuage-cloud.org -<PHONE_NUMBER>47 ✔ Folder Level ✔ pdt -<PHONE_NUMBER>68 2:48PM INF Create in progress for [https://cloudresourcemanager.googleapis.com/v1/projects/pubsec-declarative-tk-cno2].Waiting for [operations/cp.7885851846085518239] to finish.....done.Enabling service [cloudapis.googleapis.com] on project [pubsec-declarative-tk-cno2]...Operation "operations/acat.p2-153970848512-8ffc1200-8c5a-42fd-b142-e11cdaf69191" finished successfully.Updated property [core/project] to [pubsec-declarative-tk-cno2]. 2:48PM INF Creating Config Controller Cluster.... 2:48PM FTL error="API [krmapihosting.googleapis.com] not enabled on project [153970848512]. Would you like to enable and retry (this will take a few minutes)? (y/N)? ERROR: (gcloud.anthos.config.controller.create) PERMISSION_DENIED: KRM API Hosting API has not been used in project<PHONE_NUMBER>12 before or it is disabled. Enable it by visiting https://console.developers.google.com/apis/api/krmapihosting.googleapis.com/overview?project=153970848512 then retry. If you enabled this API recently, wait a few minutes for the action to propagate to our systems and retry.- '@type': type.googleapis.com/google.rpc.Help links: - description: Google developers console API activation url: https://console.developers.google.com/apis/api/krmapihosting.googleapis.com/overview?project=153970848512- '@type': type.googleapis.com/google.rpc.ErrorInfo domain: googleapis.com metadata: consumer: projects/153970848512 service: krmapihosting.googleapis.com reason: SERVICE_DISABLED" rerun on recently created project - or run on an existing project to avoid the service enablement missing wait timer arete create pdt-cno-kcc --region=northamerica-northeast1 --project=pubsec-declarative-tk-cno2 4:39PM INF Enabling required services... 4:40PM INF Operation "operations/acf.p2-153970848512-b3d4a2a6-fe02-4a5b-8f5d-d27d917f6527" finished successfully. 4:40PM INF Creating Network... ........................................................................................done.Created instance [pdt-cno-kcc].Fetching cluster endpoint and auth data.kubeconfig entrgenerated for krmapihost-pdt-cno-kcc. 5:09PM INF Add SA to roles/owner role... 5:09PM INF Config Controller setup complete reviewing https://cloud.google.com/anthos-config-management/docs/tutorials/landing-zone#removing_resources Breaking out periodic timeout issue (when we break the 30 min limit (we are usually within 22-25 min admin_root@cloudshell:~ (landing-zone-controller-w8hwa)$ arete create pdt-cno-kcc --region=northamerica-northeast1 --project=pubsec-declarative-tk-cno2 ✔ My Billing Account - 019952-0D0AAC-777E3D ✔ nuage-cloud.org -<PHONE_NUMBER>47 ✔ Folder Level ✔ pdt -<PHONE_NUMBER>68 2:48PM INF Create in progress for [https://cloudresourcemanager.googleapis.com/v1/projects/pubsec-declarative-tk-cno2].Waiting for [operations/cp.7885851846085518239] to finish.....done.Enabling service [cloudapis.googleapis.com] on project [pubsec-declarative-tk-cno2]...Operation "operations/acat.p2-153970848512-8ffc1200-8c5a-42fd-b142-e11cdaf69191" finished successfully.Updated property [core/project] to [pubsec-declarative-tk-cno2]. 2:48PM INF Creating Config Controller Cluster.... 2:48PM FTL error="API [krmapihosting.googleapis.com] not enabled on project [153970848512]. Would you like to enable and retry (this will take a few minutes)? (y/N)? ERROR: (gcloud.anthos.config.controller.create) PERMISSION_DENIED: KRM API Hosting API has not been used in project<PHONE_NUMBER>12 before or it is disabled. Enable it by visiting https://console.developers.google.com/apis/api/krmapihosting.googleapis.com/overview?project=153970848512 then retry. If you enabled this API recently, wait a few minutes for the action to propagate to our systems and retry.- '@type': type.googleapis.com/google.rpc.Help links: - description: Google developers console API activation url: https://console.developers.google.com/apis/api/krmapihosting.googleapis.com/overview?project=153970848512- '@type': type.googleapis.com/google.rpc.ErrorInfo domain: googleapis.com metadata: consumer: projects/153970848512 service: krmapihosting.googleapis.com reason: SERVICE_DISABLED" rerun on recently created project - or run on an existing project to avoid the service enablement missing wait timer arete create pdt-cno-kcc --region=northamerica-northeast1 --project=pubsec-declarative-tk-cno2 4:39PM INF Enabling required services... 4:40PM INF Operation "operations/acf.p2-153970848512-b3d4a2a6-fe02-4a5b-8f5d-d27d917f6527" finished successfully. 4:40PM INF Creating Network... ........................................................................................done.Created instance [pdt-cno-kcc].Fetching cluster endpoint and auth data.kubeconfig entrgenerated for krmapihost-pdt-cno-kcc. 5:09PM INF Add SA to roles/owner role... 5:09PM INF Config Controller setup complete @fmichaelobrien I'm a little confused on this one but I suspect what is going on is that you are getting stuck on the arete create cache here. arete will enable the APIs and it waits until that operation is complete before moving onto the next steps. I think the resolution here will be the addition to the arete create cache tracking that will bucket the cache per create command for multiple runs. Will add this feature and then if you can test again on that branch to determine if this solves this issue. All good, this one is not related to arete - it is the underlying anthoc config controller create under the covers - there are periodic timeouts occurring that we are working out the the CC team. Closing
2025-04-01T06:37:02.199229
2024-06-21T14:23:06
2366656429
{ "authors": [ "eeaton", "prabhu34" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1168", "repo": "GoogleCloudPlatform/terraform-google-cloud-functions", "url": "https://github.com/GoogleCloudPlatform/terraform-google-cloud-functions/issues/129" }
gharchive/issue
Cloud Functions broken by security fix for default Cloud Build SA Issue: Cloud Functions v2 relies on Cloud Build in the background to deploy AR resources, so it is silently impacted by the recent platform-wide security fix with the Cloud Build Service Account Change. Many customers also use the org policy iam.automaticIamGrantsForDefaultServiceAccounts, which is a recommended security best practice, and also now enforced by default for all new customer orgs. The combination of these two policies means that this terraform module will always fail. Deploying Cloud Functions depends on a Cloud Build job in the background, using the default compute SA, which must havepermissions like Storage Object Viewer to the the bucket gcf-v2-sources-$PROJECTNUMBER-$REGION. See example issue at https://github.com/terraform-google-modules/terraform-example-foundation/issues/1269 Recommend Fix: Document the dependency and troubleshooting guidance Expose a control for the user to specify which service account is used for the underlying cloud build job. (Not sure whether the Functions API itself already supports that, or if it would require an API-level fix Actually on further testing, this might be an issue with the upstream Functions gen 2 API rather than the terraform modules. Trying to deploy on console with both policies in place will consistently fail., with error messages like the following: This function has failed to deploy and will not work correctly. Please edit and redeploy. Cloud Run service projects/$PROJECTID/locations/us-central1/services/function-1 for the function was not found. The function will not work correctly. Please redeploy. Build failed with status: FAILURE and message: failed to Fetch: failed to download archive gs://gcf-v2-sources-$NUMBER-us-central1/function-1/function-source.zip: Access to bucket gcf-v2-sources-$NUMBER-us-central1 denied. You must grant Storage Object Viewer permission to<EMAIL_ADDRESS>. For more details see the logs at https://console.cloud.google.com/cloud-build/builds;region=us-central1/a5ed7bdf-db90-4578-97fe-1c54a713eb44?project=$NUMBER. Fixed in #132 @prabhu34 The fix unblocks the technical issue, but I think there's still a big challenge for usage and understanding. Going forward, this module will fail by default for all new customers (the overlapping result of changes to default IAM grants to compute service account + changes to default Cloud Build service account. So an extra and non-intuitive step is required to make this work, even if the user thinks that the principal running the module has all the necessary permissions, one of: add permissions to default compute SA override org policies to use legacy behavior for cloud build SA manually specify a build service account with sufficient privilege I've suggested #133 , WDYT?
2025-04-01T06:37:02.218525
2023-06-21T18:09:13
1768146553
{ "authors": [ "sdowell" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1169", "repo": "GoogleContainerTools/kpt-config-sync", "url": "https://github.com/GoogleContainerTools/kpt-config-sync/pull/703" }
gharchive/pull-request
test: stricter regex for gke operation filter Since we were not matching the end of the string, this filter could return operations for multiple clusters. The stricter regex should result in operations for a unique cluster name. /assign @haiyanmeng
2025-04-01T06:37:02.227146
2021-09-06T21:19:49
989441232
{ "authors": [ "HP-Nunes", "droot" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1170", "repo": "GoogleContainerTools/kpt", "url": "https://github.com/GoogleContainerTools/kpt/issues/2481" }
gharchive/issue
error: unknown command "cfg" for "kpt" Expected behavior I want to configure kpt setter values for the name, project, and location of a management cluster in Google Cloud. kpt cfg set -R . name "${MGMT_NAME}" kpt cfg set -R . gcloud.core.project "${MGMT_PROJECT}" kpt cfg set -R . location "${LOCATION}" Actual behavior I tried running the first line in bash, and got this in return: error: unknown command "cfg" for "kpt" Did you mean this? fn pkg Information I followed the installation instructions from ktp's website, using docker to install both ktp and ktp-gcloud. I'm running ktp version 1.0.0-beta.1 with kubectl version GitVersion:"v1.22.1" and kustomize/v4.3.0. I am following the Kubeflow's tutorial on deploying a management cluster on Google Cloud. I understand that someone has posted about this exact issue 4 days ago, but the issue was closed without a resolution. I am a beginner with ktp/Kubeflow, so apologies if the resolution is obvious. Kubeflow's tutorial has a note saying: Note: kpt v1.0.0-beta.1 or above doesn’t work due to a known issue: https://github.com/kubeflow/pipelines/issues/6100. Please downgrade gcloud or install kpt separately https://github.com/GoogleContainerTools/kpt/releases/tag/v0.39.2 for now. So looks like kubeflow hasn't migrated to kpt 1.0+ yet, so you will have to use https://github.com/GoogleContainerTools/kpt/releases/tag/v0.39.2 release. @droot that resolved my issue; many thanks! Just goes to show that you can get so absorbed into solving an issue that one fails to see the obvious.... I'm adding some additional details below for any n00b like myself in a similar predicament: I had the ktp v1.0+ binary installed, so I removed it running: sudo apt-get remove google-cloud-sdk-kpt Make sure to later run gcloud components update I reinstalled the v0.39 binary using the instructions from: https://cloud.google.com/service-mesh/docs/environment-setup Important: make sure to install ktp within the directory where you want it to execute. Within the context of the context of the Kubeflow's management setup tutorial, that would be within the gcp-blueprints folder: https://www.kubeflow.org/docs/distributions/gke/deploy/management-setup/ Verify that the install worked properly with ktp version The kpt cfg set command should now be working. As I understand it, this command is deprecated in the latest beta binary: https://kpt.dev/installation/migration?id=changes-to-setters
2025-04-01T06:37:02.257107
2022-01-23T21:22:04
1111989800
{ "authors": [ "Dabs-Rulez" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1171", "repo": "Goose-Nest/GT-RevertRebrand", "url": "https://github.com/Goose-Nest/GT-RevertRebrand/pull/76" }
gharchive/pull-request
Fixed 'Revert new Nitro icon by setting item' Fixed Nitro logo not hidden and added back old blurple for the text. Also changed text to white when tab is selected and made the text blurple when not selected @CanadaHonk Please check this PR
2025-04-01T06:37:02.312299
2024-01-03T20:31:06
2064609899
{ "authors": [ "ArakTaiRoth", "MikeS700", "Spark450", "twjeffery" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1172", "repo": "GovAlta/ui-components", "url": "https://github.com/GovAlta/ui-components/issues/1548" }
gharchive/issue
[Feature Request]: Breadcrumb component Is your feature request related to a problem? Please describe Currently in the Early Childhood Development System portal (Child Care Portal), there are a could different services (accessed through the sidenav) which feature breadcrumbing on their respective pages. Each breadcrumb patter is different. We will soon want all the services in this portal to utilize a consistent breadcrumb component/pattern. Describe the solution you'd like Would like to see a GoA breadcrumb component which can go down to at least 5 to 6 page levels Works on desktop/tablet and mobile breakpoints Is fully accessible Accounts for position/padding in relation to page templates or this first item of content below the breadcrumb on a page All states for the active links and active page in the breadcrumb Some guidelines around breadcrumb label character length and recommended naming convention if possible (brevity and matching the page title and considerate of the url if possible) Provide evidence this is a needed component Currently 3 of 6 Child Care Services in the Early Childhood Development System portal utilize breadcrumbing and all use a different pattern treatment. All services in the portal will use breadcrumbing in early 2024. Consistency is key as this is a single portal for our external users. The Design System team is already exploring breadcrumb component for their Design System website. https://www.figma.com/file/Vw1LNUBsNeFkToPTWqXAzR/Component---Breadcrumb?type=design&node-id=1401-3285&mode=design&t=7I1OkZM3WExILE2j-4 Describe alternatives you've considered Good article which may provide helpful when mapping out the component and it's guidelines: https://www.smashingmagazine.com/2022/04/breadcrumbs-ux-design/ Do you have anything already created for this that we can use? yes Are you currently using this proposal inside your own service yes Are you able to assist to bring the feature to reality? yes Additional context Here are some live breadcrumb examples from ECDS portal Hi @ArakTaiRoth do you need more context/info from me? Cheers Land Titles Office has a component designed and in development for this: Figma Design guidelines Development blueprint Service designer: @garnison-goa Development Developer: Daryl Chiew @Spark450 Gather the relevant info and convert into a design issue This needs to have a complete design definition before it can be worked on in development, see Jira Issue to follow its progress. A new issue will be created once the design work has been completed.
2025-04-01T06:37:02.365189
2023-03-07T07:47:14
1612907815
{ "authors": [ "Jongy", "mpozniak95" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1173", "repo": "Granulate/gprofiler", "url": "https://github.com/Granulate/gprofiler/pull/712" }
gharchive/pull-request
java: Add option to include method modifiers Description Added an option for java profiling: --java-include-method-modifiers, disabled by default. When flag provided we use: jvmti->GetMethodModifiers in async-profiler to get method modifiers. Related Issue https://github.com/Granulate/gprofiler/issues/570 How Has This Been Tested? Added test that tests this case Needs this: https://github.com/Granulate/async-profiler/pull/5 to be merged first, and then need to make a change in async_profiler_shared_build.sh Can update the commit, and we're good :) Pushed tag v2.9g6 Looks good, I will merge once tests pas..
2025-04-01T06:37:02.366380
2024-08-22T08:00:30
2480105552
{ "authors": [ "pomelo-nwu", "totoago" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1174", "repo": "GraphScope/portal", "url": "https://github.com/GraphScope/portal/issues/479" }
gharchive/issue
Interface is still pending, but the display is incorrect https://github.com/GraphScope/portal/pull/481
2025-04-01T06:37:02.370249
2015-04-30T08:36:00
72100388
{ "authors": [ "kbastani", "rbramley" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1175", "repo": "Graphify/graphify", "url": "https://github.com/Graphify/graphify/pull/21" }
gharchive/pull-request
Updated README.md to add in authentication This needs to be reflected on the gh-pages too Thanks for the pull request
2025-04-01T06:37:02.372981
2022-05-09T03:43:15
1229131981
{ "authors": [ "0HyperCube", "Keavon" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1176", "repo": "GraphiteEditor/Graphite", "url": "https://github.com/GraphiteEditor/Graphite/issues/630" }
gharchive/issue
Snapping system is not precise From here: It's hard to tell, but I think the handle points, when being dragged and snapped to things, are snapping a few pixels vertically offset from true. Try dragging a handle point and snapping it to the bottom of a neighboring shape and notice how it's below the bottom. This might also be the case for anchor points, but instead of vertically it's some other sort of offset. Try out the snapping of anchor and control points to other shapes then zoom in to see if they are precisely located in the same exact location (down to many decimal places before floating point error comes into play). Reply by @0HyperCube about this: Yes, if you drag a point with the path tool, it will snap your mouse position and then add the offset dictated by how far your mouse was originally away from the centre of the point. Also would cause more merge conflicts for the path refactor. So I'm marking this as blocked on #605. This has been fixed as the path tool now puts the centre of the square at the mouse position. @0HyperCube do you know which PR or commit fixed this, for the record?
2025-04-01T06:37:02.373962
2022-05-15T21:53:23
1236429454
{ "authors": [ "Keavon", "RahulHi" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1177", "repo": "GraphiteEditor/Graphite", "url": "https://github.com/GraphiteEditor/Graphite/issues/643" }
gharchive/issue
Viewport zoom/rotation while snapping doesn't show the snapped values In the viewport, the top right number inputs display the current precise values of the ongoing zoom/rotation, but they should show the snapped value which is what the viewport is rendering. I'll try to solve this.
2025-04-01T06:37:02.376336
2022-08-28T15:09:05
1353381736
{ "authors": [ "0HyperCube", "Keavon", "dchiasson" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1178", "repo": "GraphiteEditor/Graphite", "url": "https://github.com/GraphiteEditor/Graphite/issues/770" }
gharchive/issue
Rotating object via property panel eliminates shear I think the best way to prioritize this is to wait until we have the node-based transform system using the Transform node. Probably isn't worth the expendable effort of a temporary fix when that won't matter after we have nodes. Upon further thought, maybe this can be fixed without that much effort and it would probably be a quality of life improvement for users who run into this. Do you have any thoughts about this and its prioritization @0HyperCube? Upon further thought, maybe this can be fixed without that much effort and it would probably be a quality of life improvement for users who run into this. Do you have any thoughts about this and its prioritization @0HyperCube? This would probably be somewhat ugly because of how skew combines with rotation and scale in the affine2.
2025-04-01T06:37:02.379248
2023-10-23T01:22:47
1956170555
{ "authors": [ "NotThorny", "a-w-a-y" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1179", "repo": "Grasscutters/Cultivation", "url": "https://github.com/Grasscutters/Cultivation/issues/212" }
gharchive/issue
[Feature Request] Automatically Sync It would be nice if Cultivation could automatically sync with data from official Genshin servers, or at least showcased servers so maybe we could try out builds with new characters that we want to pull for Cultivation is just a launcher, it sounds like you're asking for something related to the server-side like Enka plugin (public showcase copying). There also is absolutely no way to auto sync official data as neither Grasscutter nor Cultivation have such access to official databases or data. Using InventoryKamera or other 3rd party scanning tools on your own PC, or Enka GC plugin for public showcase only is the closest you will get. Either way, that's something to do be done on your server-side (Grasscutter), not on Cultivation which is just a launcher.
2025-04-01T06:37:02.381865
2015-11-11T13:41:59
116337727
{ "authors": [ "michaltakac", "pmuens" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1180", "repo": "GravityProject/gravity", "url": "https://github.com/GravityProject/gravity/issues/44" }
gharchive/issue
Add visible reference to GitHub repository Add a reference to the GitHub repository so that all users can see it directly (maybe with the help of a bottom fixed div) in an unobtrusive way. But don't use the "Fork me on GitHub" banner. +1, maybe add footer text (copyright, made with <3 by etc. :D) to this fixed div on the bottom of layout and add link to github repo there. Fixed bottom div is great idea because it wont be burried down there under all posts :+1:
2025-04-01T06:37:02.470020
2024-02-26T19:31:00
2154957148
{ "authors": [ "andreww", "colinsauze" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1183", "repo": "GreenScheduler/cats", "url": "https://github.com/GreenScheduler/cats/issues/69" }
gharchive/issue
JOSS publication Submit CATS to the Journal of Open Source Software Brief update on this - I now have a very rough first draft (on google docs). I'll ask @sadielbartholomew and/or @Llannelongue to take a look before I migrate it to the repository (and to markdown) in a couple of weeks. In drafting, I've found a couple of issues / potential things we could improve. I'll open new issues for these. Oh, and we should also check the JOSS requirement checklist for anything else we need to do to the codebase.
2025-04-01T06:37:02.474350
2023-07-07T14:49:49
1793680963
{ "authors": [ "Llannelongue", "andreww", "tlestang" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1184", "repo": "GreenScheduler/cats", "url": "https://github.com/GreenScheduler/cats/pull/49" }
gharchive/pull-request
Update project structure The promised PR focusing on project structure, I've copied the figure below from the previous PR (the class names may be obsolete as I haven't changed any functions here but the target structure remains the same). Overall quite light changes as some improvements have now been merged into main already :) List of changes: Have a dedicated file to check the validity of all the arguments (more checks to be added in the future). Moved validation of --duration and --jobinfo to this file. The carbon intensity API calls is a separate step as part of main(). Starttime optimiser is also a separate step (rather than API and starttime being intertwined) Some other light renaming of functions/files as needed (looking at the files changed, git seems to have lost track of timeseries_conversion.py, but it was just renamed optimise_starttime.py, same content otherwise). As usual, thoughts/suggestions welcome! Thanks for the speedy review @tlestang! Looks good to me. I wonder if that diagram needs to make it into the documentation somewhere? Thanks @andreww and @tlestang for the reviews! I'll merge it into main now, and further edits can have their own PRs @Llannelongue Not a big deal, but next time we can squash those typos and fixup commits, and group related changes together. For big PRs like this it can really help keeping a readable history.
2025-04-01T06:37:02.554480
2018-07-31T10:19:11
346124572
{ "authors": [ "awb99", "dahlbyk" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1185", "repo": "GriddleGriddle/Griddle", "url": "https://github.com/GriddleGriddle/Griddle/issues/825" }
gharchive/issue
question: pagesize dynamic (rest of viewport) This is a question I could not find in the docs: I basically want Griddle to use the remaining screen that it can use, so that it will always display as many rows as it can. My page has a menu on top, so griddle can use viewport height- header pixels. I believe if there is lots of data to be displayed, then I want to show the maximum that I can to the user. Thanks! This very much depends on how your data is available, but I would look at combining LocalPlugin and PositionPlugin to manage a fixed-height viewport.
2025-04-01T06:37:02.568902
2015-08-27T15:24:55
103535051
{ "authors": [ "Grokzen", "jlec" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1186", "repo": "Grokzen/pykwalify", "url": "https://github.com/Grokzen/pykwalify/issues/20" }
gharchive/issue
Include tests Downstream projects like Linux distributions like to test the packages during installation. Please include the tests needed for the test suite in you releases at pypi. @jlec Uploaded version 1.4.1 to pypi, do you think that will be enough for testing? If you need any other files just ping Cool, that is perfectly fine.
2025-04-01T06:37:02.570963
2021-02-02T21:22:10
799700252
{ "authors": [ "BrucePerens", "akadusei" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1187", "repo": "GrottoPress/shield", "url": "https://github.com/GrottoPress/shield/issues/34" }
gharchive/issue
Shield won't build with Crystal 0.36.1 Shield won't build with the new Crystal release. This new version appears to detect duplicate named arguments better than before, OR duplicate named arguments are happening where they did not before. They appear to be duplicates in attribute lists that are passed to operations. You should be able to see the problems as soon as you attempt to build. You must be on Lucky master. Shield will support Crystal v0.36 when it is supported by Lucky, hopefully in the next version. I intend to begin work on this soon.
2025-04-01T06:37:02.576732
2018-02-13T05:43:40
296620610
{ "authors": [ "Grover-c13", "StacktrackSubmission" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1188", "repo": "Grover-c13/MultiAdmin", "url": "https://github.com/Grover-c13/MultiAdmin/issues/60" }
gharchive/issue
IndexOutOfRangeException: Array index is out of range. at PlayerStats.HurtPlayer (HitInfo info, UnityEngine.GameObject go) [0x00000] in :0 IndexOutOfRangeException: Array index is out of range. at PlayerStats.HurtPlayer (HitInfo info, UnityEngine.GameObject go) [0x00000] in <filename unknown>:0 at PlayerStats.Explode () [0x00000] in <filename unknown>:0 at AlphaWarheadDetonationController.ExplodePlayers () [0x00000] in <filename unknown>:0 at AlphaWarheadDetonationController.Explode () [0x00000] in <filename unknown>:0 at AlphaWarheadDetonationController.FixedUpdate () [0x00000] in <filename unknown>:0 == STACKTRACK STATS == Times seen: 1 Last reported: 2018-02-13 05:43:40.420612 Fixed in new game version
2025-04-01T06:37:02.608630
2024-05-23T21:34:33
2313889675
{ "authors": [ "iAmKevinMcKee", "lukas-frey" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1189", "repo": "GuavaCZ/filament-knowledge-base", "url": "https://github.com/GuavaCZ/filament-knowledge-base/issues/23" }
gharchive/issue
[Request]: Turn off Single Page Application mode What feature would you like to add? Right now spa mode is on by default, and I don't see any way to turn it off. I've tried KnowledgeBasePanel::configureUsing( fn (KnowledgeBasePanel $panel) => $panel->spa(false)); But that doesn't work. The problem when one panel has spa mode and others don't, you get weird behavior when going back and forth. When going from the knowledge base to another panel without spa mode on, the first time you navigate to the other panel, the global search is still the KB search until there is another page refresh. Notes No response Fixed in #24
2025-04-01T06:37:02.633462
2024-02-05T21:28:42
2119487267
{ "authors": [ "GuidaGG", "ah-bas-und-aer" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1190", "repo": "GuidaGG/vielfaeltig", "url": "https://github.com/GuidaGG/vielfaeltig/issues/26" }
gharchive/issue
sticky menu Feedback vielfältig: The sticky menus should not overlap the content – that means we avoid this making the content width smaller in all landscape modi ( and maybe for tablet portrait mode) For mobile here is the suggestion: Sticky menus are only visible on top, and is not fixed or fades out while scrolling (I quickly tried: https://xd.adobe.com/view/770b8a67-668a-40e0-93f5-65f260cbb802-2258/) what about the hamburger menu? in one slide is on top and another is on the bottom? The sticky menus should not overlap the content – that means we avoid this making the content width smaller in all landscape modi ( and maybe for tablet portrait mode) this means there is padding on the right? what about the hamburger menu? in one slide is on top and another is on the bottom? I don't understand exactly, but if you mean for mobile if you mean the sticky menu i thought of switching them in mobile: Auf's Ohr and Jobs to the top and the hamburger to the bottom, maybe we first have to ask client before getting active. The sticky menus should not overlap the content – that means we avoid this making the content width smaller in all landscape modi ( and maybe for tablet portrait mode) this means there is padding on the right? the padding left and right should be the same, but for some screens they can be wider. (Could teh size refer to the sticky elements? but how does this work on mobile? because then there will be a huge space on the left and right? Im not understanding 100% ok and the switching I dont make because it is still not sure? but how does this work on mobile? because then there will be a huge space on the left and right? Im not understanding 100% ok and the switching I dont make because it is still not sure? For mobile, the solution is to accept the overlapping and fade out the sticky header after scrolling/ not fix them i think this is not possible. We would have to build it from start being the middle of the content. I can put it stationary somewhere, but this still means it will be on top of content and it is even worse because it doesnt move. https://github.com/GuidaGG/vielfaeltig/assets/9657908/9a23e6ff-534d-47e3-bfd7-974a4f176791 At least when it is fixed, it sometimes is on top of content but when we scroll it goes away. this might be goog, but we have to gibe teh sticky menus a nice position in the beginning for mobile. And: They commited to the option with menu at the bottom. what about they keep stiky but on top and we give a top padding to content in mobile so we sure people can always see the content? Even there is hard to make it work very weel because there could be 1 or 2 or 3 sticky buttons. Otherwise I have to wait until I have an idea of how to implement that the content makes space for the sticky buttons. Otherwise I have to wait until I have an idea of how to implement that the content makes space for the sticky buttons. I don't understand, will this reduce the hight of the main content? The hamburger at the buttom and the sticky are not done yet, will you still edit this?
2025-04-01T06:37:02.675418
2023-10-27T22:02:08
1966243078
{ "authors": [ "MilanLempera" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1191", "repo": "H-edu-dev/upptime", "url": "https://github.com/H-edu-dev/upptime/issues/68" }
gharchive/issue
⚠️ Homepage has degraded performance In d306779, Homepage (https://h-edu.cz) experienced degraded performance: HTTP code: 200 Response time: 486 ms Resolved: Homepage performance has improved in 8fb9dbe after 29 minutes.
2025-04-01T06:37:02.718178
2016-09-08T19:04:29
175836234
{ "authors": [ "lipp", "zemirco" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1192", "repo": "HBM/react-components", "url": "https://github.com/HBM/react-components/issues/39" }
gharchive/issue
Chips: why are value entries objects with "text" key ...and not simply strings? @zemirco A single chip can have an icon as well. See https://github.com/HBM/react-components/pull/41 for an example.
2025-04-01T06:37:02.742855
2022-09-11T23:29:14
1369097373
{ "authors": [ "HDVinnie" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1196", "repo": "HDVinnie/TrackerHub", "url": "https://github.com/HDVinnie/TrackerHub/issues/1997" }
gharchive/issue
⚠️ PrivateHD has degraded performance In c67d66f, PrivateHD ($PHD) experienced degraded performance: HTTP code: 200 Response time: 3036 ms Resolved: PrivateHD performance has improved in 3adf634.
2025-04-01T06:37:02.747321
2022-07-05T15:00:17
1294420994
{ "authors": [ "HDVinnie" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1198", "repo": "HDVinnie/TrackerHub", "url": "https://github.com/HDVinnie/TrackerHub/issues/215" }
gharchive/issue
⚠️ SkipTheCommericals has degraded performance In 3d19e34, SkipTheCommericals ($STC) experienced degraded performance: HTTP code: 200 Response time: 2493 ms Resolved: SkipTheCommericals performance has improved in b270460.
2025-04-01T06:37:02.749437
2024-01-05T20:00:20
2067960634
{ "authors": [ "HDVinnie" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1199", "repo": "HDVinnie/TrackerHub", "url": "https://github.com/HDVinnie/TrackerHub/issues/25881" }
gharchive/issue
⚠️ TorrentLeech has degraded performance In abda01d, TorrentLeech ($TL) experienced degraded performance: HTTP code: 200 Response time: 5970 ms Resolved: TorrentLeech performance has improved in cd2f841 after 8 minutes.
2025-04-01T06:37:02.751897
2024-05-26T13:29:48
2317714996
{ "authors": [ "HDVinnie" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1200", "repo": "HDVinnie/TrackerHub", "url": "https://github.com/HDVinnie/TrackerHub/issues/27982" }
gharchive/issue
⚠️ Empornium has degraded performance In 0e6af8b, Empornium ($EMP) experienced degraded performance: HTTP code: 500 Response time: 1285 ms Resolved: Empornium performance has improved in 9a59654 after 40 minutes.
2025-04-01T06:37:02.754026
2024-07-28T08:24:23
2433813994
{ "authors": [ "HDVinnie" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1201", "repo": "HDVinnie/TrackerHub", "url": "https://github.com/HDVinnie/TrackerHub/issues/28731" }
gharchive/issue
⚠️ HawkeUno has degraded performance In 67e0e8b, HawkeUno ($UNO) experienced degraded performance: HTTP code: 403 Response time: 1342 ms Resolved: HawkeUno performance has improved in 310f691 after 6 minutes.
2025-04-01T06:37:02.760588
2023-01-17T15:36:42
1536592590
{ "authors": [ "HDVinnie" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1204", "repo": "HDVinnie/TrackerHub", "url": "https://github.com/HDVinnie/TrackerHub/issues/6081" }
gharchive/issue
⚠️ SceneTime has degraded performance In 54047b5, SceneTime ($ST) experienced degraded performance: HTTP code: 200 Response time: 3130 ms Resolved: SceneTime performance has improved in 8798c66.
2025-04-01T06:37:02.814923
2019-01-30T05:46:30
404612122
{ "authors": [ "jermwatt", "matiasdahl", "subhanjan21" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1207", "repo": "HIPS/autograd", "url": "https://github.com/HIPS/autograd/issues/470" }
gharchive/issue
Multiple variable derivation using autograd How to find a Jacobian using autograd for multiple variable. For example: R1 = x(1)*y(1) + 2y(2)x(2) R2 = -y(1) + y(2)(2)**2 R1 and R2 is a transpose matrix And I need to find [ [dR1/dx1 dR2/dx1], [dR1/dx2 dR2/dx2] ] Does Example 4 here help? @jermwatt Yes, it did give me a better concept. Thank you so much. Just a quick question. The code is so far working for this: f = lambda x,y: np.array([y[0]*x[0]**2, y[1]*x[0]+x[1]]) f_jac = jacobian(f, argnum = 0) But can you suggest me how to get the output for the jacobian: j = f_jac(np.array([[1.,1.], [2.,2.]])) Thanks. I was having the same problem. The above link helped alot. I tried to look at the tests for autograd.jacobian for hints about how to do this. Maybe one could add something similar the below? import autograd.numpy as np # Thinly-wrapped version of Numpy import autograd A = np.array([[1., 2.], [3., 4.]]) def f(x, y): """ Return vector function: f(x, y) = /1 2\ /x\ = /1x + 2y\ \3 4/ \y/ \3x + 4y/ """ return np.tensordot(A, np.array([x, y]), [[1], [0]]) import numpy as np0 # wrapped numpy does not have testing methods x0 = 1; y0 = 1 np0.testing.assert_array_equal(f(x0, y0), np.array([1*x0+2*y0, 3*x0 + 4*y0])) J_autograd = np.array([autograd.jacobian(f, k)(0., 0.) for k in [0, 1]]) np0.testing.assert_array_equal(J_autograd, A.T) In the doc string for jacobian: Should x (in the code) be argnum (mentioned in the doc string)? The code does not reference argnum (?) Or def f(*vector): """ Return vector function: f(x, y) = /1 2\ /x\ = /1x + 2y\ \3 4/ \y/ \3x + 4y/ """ return np.tensordot(A, np.array(vector), [[1], [0]])
2025-04-01T06:37:02.882147
2022-06-23T03:30:07
1281638875
{ "authors": [ "Bidaya0", "songzhibin97" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1208", "repo": "HXSecurity/DongTai", "url": "https://github.com/HXSecurity/DongTai/issues/722" }
gharchive/issue
Preflight Checklist [X] I agree to follow the Code of Conduct that this project adheres to. [X] I have searched the issue tracker for an issue that matches the one I want to file, without success. [X] I am not looking for support or already pursued the available support channels without success. Version now Installation Type Other (specify below) Service Name DongTai-WebAPI Describe the details of the bug and the steps to reproduce it Agent列表上报的指标数据与服务器不一致,请确认下是否数据源有误 Additional Information No response Logs No response fixed in newest version
2025-04-01T06:37:02.902700
2022-11-04T00:07:30
1435389257
{ "authors": [ "Verousca", "vkoldus" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1209", "repo": "HYF-Class19/home", "url": "https://github.com/HYF-Class19/home/issues/295" }
gharchive/issue
Verusca : Behavior, Strategy and Implementation, 3 Weeks Learning Objectives Priorities: 🥚, 🐣, 🐥, 🐔 (click to learn more) There is a lot to learn in this repository. If you can't master all the material at once, that's expected! Anything you don't master now will always be waiting for you to review when you need it. These 4 emojis will help you prioritize your study time and to measure your progress: 🥚: Understanding this material is required, it covers the base skills you'll need to move on. You do not need to finish all of them but should feel comfortable that you could with enough time. 🐣: You have started all of these exercises and feel you could complete them all if you just had more time. It may not be easy for you but with effort, you can make it through. 🐥: You have studied the examples and started some exercises if you had time. You should have a big-picture understanding of these concepts/skills, but may not be confident completing the exercises. 🐔: These concepts or skills are not necessary but are related to this module. If you are finished with 🥚, 🐣 and 🐥 you can use the 🐔 exercises to push yourself without getting distracted from the module's main objectives. 1. Remix Practice studying and remixing other people's solutions to coding challenges. Create your own solutions by mixing and matching pieces from other people's code. 🐣 Reconstructing: You can reconstruct a variety of solutions to the same challenge when they are presented as parsons problems. 🐣 Analyzing: You can analyze a function written at your level. This includes: [ ] Behavior: Write documentation, test cases and use cases to describe the function's behaviour. [ ] Strategy: Describe the function's strategy using plain English. [ ] Implementation: List the language features in a function and explain how each is used. [ ] Small Changes: You can think of 2+ changes to the function's implementation that would not change it's strategy. 🐣 Remixing You can analyze several solutions to the same challenge then ... [ ] Write: Your own solution by remixing the ones you studied. [ ] Explain: How the other solutions inspired yours; what ideas did you take from them? what ideas did you not take? [ ] Analyze: Complete a write-up of your own solution as though someone else wrote it. 2. Write 🐣 Function Design: [ ] Writing Tests: Given a working function, you can write passing test cases to describe it's behavior. [ ] Writing Functions: You can design multiple solutions to the same code challenge, keeping notes about different experiments you tried along the way. [ ] 🐣 Generating Documentation: You can write a JSDoc comment for your solutions and run a script to generate markdown documentation. [ ] 🐣 Fuzz Testing: You write solutions that pass randomly generated test cases. 🐥 Test Driven Development: You can solve open-ended, ambiguous coding challenges at your level: [ ] Reading Docs: You can understand what the function is supposed to do by reading it's JSDoc description. [ ] Writing Tests: You can write test cases before there is a function to test. [ ] Writing Functions: You can write one function that passes the test cases you have prepared (even if it's just 1 test case!). [ ] Refactoring: You can improve your function's implementation without failing any test cases that were passing. [ ] Iterative Development: You can repeat the TDD process until you are satisfied with your test cases and solution. [ ] 🐔 Code Golf: Write your solutions with the fewest characters possible! This won't help you write readable code, but it will make you think deeply about JS, your strategy and implementation. 3. Review 🐣 Continuous Integration: You can check your code's quality before pushing so your CI checks all pass. [ ] Formatting [ ] Linting [ ] Testing [ ] 🐣 Code Review: You can use a checklist to give a thorough, positive and constructive review of your classmates' solutions. [ ] 🐔 Code Coverage: You can explain what code coverage is, why it's important, and can write unit tests with 100% code coverage. Behavior, Strategy, and Implementation - week 1 [x] I have pushed my progress to my fork of exercises repo - template write-up I Need Help With: I solved an exercise but I couldn’t go backward after the final submission on Edabit. I re-wrote the solution on my VS Code but I couldn't check it on Edabit to be sure because it moved to the previous exercise. After solving an exercise on Edabit, is it possible to go backwards to copy the solution or see previous work done? General remarks I'm getting to understand a bit more about JS from this module. Sunday Prep Work I am working on the exercises on slack to be completed by Sunday Have you tried just opening the edabit challenge through the provided link again and going straight to the Code tab? If I do that, I can submit a solution again, even though I finished the challenge before. Thank you it worked Behavior, Strategy, and Implementation - week 1 [x] I have pushed my progress to my fork of exercises repo - template write-up I Need Help With: Nothing for now General remarks I'm getting to understand a bit more about JS from this module. Sunday Prep Work I am working on the easy exercises
2025-04-01T06:37:02.928677
2020-09-30T12:34:43
711907871
{ "authors": [ "afrouzhakim" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1211", "repo": "HackYourFuture-CPH/chattie", "url": "https://github.com/HackYourFuture-CPH/chattie/pull/185" }
gharchive/pull-request
getting last message of recent channels-backend Description get the last message and the title of the channels, from the recent channels for a specific user that defined in the endpoint. Fixes #4 How to test? run "npm run server" and go to "http://localhost:5000/api/channels-message/2" Checklist [x] I have performed a self-review of my own code [x] I have commented my code, particularly in hard-to-understand areas [x] I have made corresponding changes to the documentation [x] This PR is ready to be merged and not breaking any other functionality Hi Benjamin, thank you for your feedback. Yes you are right I must to have a better overview of my classmates work before start coding! I will try to use the API that you mentioned and complete my task on that.
2025-04-01T06:37:02.932784
2022-02-07T19:11:55
1126388226
{ "authors": [ "Divyajg", "santhoshboinapally" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1212", "repo": "HackYourFuture-CPH/fp-class19", "url": "https://github.com/HackYourFuture-CPH/fp-class19/pull/91" }
gharchive/pull-request
Frontend-ContactPage-Container Description Frontend Contact Page container , I have added Header ,Menu ,Contact us and Footer Components in this page .I removed these components from App.js and added Route path for contact us page. Fixes #71 How to test? npm run dev After you run this command it will show Contact page, Please remember it will not show header ,menu and footer in other pages because i removed components from App.js. Please provide a short summary how your changes can be tested? Checklist [x] I have performed a self-review of my own code [x] I have followed the name conventions for CSS Classnames and filenames, Components names and filenames, Style filenames, if you are in doubt check the the project README.MD and here https://github.com/HackYourFuture-CPH/curriculum/blob/master/review/review-checklist.md [x] I have commented my code, particularly in hard-to-understand areas, if you code was simple enough mark the box anyway [x] I have made corresponding changes to the documentation, if you code was simple enough mark the box anyway [x] This PR is ready to be merged and not breaking any other functionality It's looking nice on my machine as well. 👍
2025-04-01T06:37:02.948815
2018-10-18T09:35:17
371442380
{ "authors": [ "Aniket965", "iamcodeking" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1213", "repo": "Hacktoberfest-2018/Hello-world", "url": "https://github.com/Hacktoberfest-2018/Hello-world/pull/4714" }
gharchive/pull-request
Create java sum What does the program do ? sum In what programming language it is written? java This pull request do not contains files in specific folders. Please fix it and reopen for help see https://help.github.com/articles/moving-a-file-to-a-new-location/
2025-04-01T06:37:03.001635
2022-10-25T01:10:05
1421687493
{ "authors": [ "Googlproxer", "wf9a5m75" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1214", "repo": "HalleyAssist/ion-range-calendar", "url": "https://github.com/HalleyAssist/ion-range-calendar/pull/9" }
gharchive/pull-request
Update README.md It seems you renamed the class name Thanks for that. Sorry for the late reply, you got me just after I went on holiday.
2025-04-01T06:37:03.027416
2024-06-14T08:57:16
2352888605
{ "authors": [ "Hannibal046", "fsantosg" ], "license": "CC0-1.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1215", "repo": "Hannibal046/Awesome-LLM", "url": "https://github.com/Hannibal046/Awesome-LLM/pull/138" }
gharchive/pull-request
add evaluation frameworks LangSmith and Ragas Description Added a few frameworks (LangSmith and Ragas) to help on LLM evaluation tasks. Appreciate!
2025-04-01T06:37:03.048986
2024-05-06T11:21:59
2280649134
{ "authors": [ "HaotianZhangAI4Science", "zh2417" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1218", "repo": "HaotianZhangAI4Science/Delete", "url": "https://github.com/HaotianZhangAI4Science/Delete/issues/7" }
gharchive/issue
Questions about how to train the model Excellent work! I have some confusion. I would like to know if the final model obtained is a single "delete" model capable of executing all four subtasks of lead optimization (trained with seven masking strategies.?). Or does each subtask of lead optimization correspond to a separate "delete" model? (Each subtask model being trained with four masks, including three enhancement masks and one task-specific mask?) I noticed you mentioned only one checkpoint model. What subtask does this model correspond to?I look forward to your clarification. Thank you very much. Hi, Yes, I only provided a ckpt for usage, but it is recommended to use task-specific ckpt to generate molecules, but you can also use mixed-training for obtaining a general ckpt. I will open-source it as soon as possible! Thanks for your interest! Best, Odin So ,each subtask of lead optimization correspond to a separate "delete" model? (Each subtask model being trained with four masks, including three enhancement masks and one task-specific mask?) But the author, you only provided one model. Does this model correspond to the subtask of fragment growth? Could you clarify if the model generates atoms one by one? Is it an autoregressive model? Does the generation of the next atom depend on the previously generated atoms?
2025-04-01T06:37:03.108821
2022-11-16T11:39:05
1451429919
{ "authors": [ "Harfeur", "Sarvagon" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1222", "repo": "Harfeur/TwitchAlerts", "url": "https://github.com/Harfeur/TwitchAlerts/issues/1" }
gharchive/issue
Twitch Alerts bot is offline For several days now, the Twitch Alerts bot is offline and don't work anymore (error message "application is not working" and status down). Hi thank you for your report. I've restarted the server where the bot is running, but I forgot to restart the bot 😅
2025-04-01T06:37:03.112319
2019-02-21T05:47:38
412759077
{ "authors": [ "HarisIqbal88", "svenefftinge" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1223", "repo": "HarisIqbal88/PlotNeuralNet", "url": "https://github.com/HarisIqbal88/PlotNeuralNet/pull/15" }
gharchive/pull-request
Added live example This PR adds configuration to run and code on plot neural net in Gitpod, a free dev environment for GitHub we have been working on. The goal is to make contributions super easy by providing a single click to code experience. You can open gitpod workspaces on any github repository by prefixig the github URL with 'gitpod.io#/'. Depending on the URL it does the right thing. For instance, prefixing a pull request like this one will open it in code review mode. For your project I added a docker with the needed dependencies and made the start script so that it runs test/unet.py and opens the resulting pdf. Also the python editor is opened so that people can play around with it. I hope you find it useful, let me know we you have any questions. It seems that your code does not load docker created for this repository out of the box. Would be nice if you add some guidelines or make it run out of the box. How did you try it? If you start a gitpod on this branch (or my fork) the Dockerfile will be picked up: https://gitpod.io/#https://github.com/HarisIqbal88/PlotNeuralNet/pull/15
2025-04-01T06:37:03.121747
2021-02-28T16:14:59
818243526
{ "authors": [ "Crimson-Blade", "HarshKhandeparkar" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1224", "repo": "HarshKhandeparkar/rainbow-board", "url": "https://github.com/HarshKhandeparkar/rainbow-board/issues/9" }
gharchive/issue
Tool Icons We don't have an icon for the rainbow brush. The icon for the line tool only resembles the tool partially and is actually a graph. We need either custom-made icons or some icons from somewhere. I'll see what I can find Thanks
2025-04-01T06:37:03.123994
2023-10-18T15:39:03
1950051471
{ "authors": [ "HarshwardhanPatil07", "Jayu1214" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1225", "repo": "HarshwardhanPatil07/HactoberFest2023", "url": "https://github.com/HarshwardhanPatil07/HactoberFest2023/pull/170" }
gharchive/pull-request
Added a Number Guessing Game Hello There, I have added a Number Guessing game in this repository . This a game where the number will be selection by the device automatically between 1 to 100 . The user has to guess the correct number out of the 100 numbers. Hope this changes find you well . Thank you Regards. @Jayu1214 Follow code of conduct by Star the repo! too small commit
2025-04-01T06:37:03.127654
2016-04-28T01:24:57
151528317
{ "authors": [ "elbing", "rminnich" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1226", "repo": "Harvey-OS/harvey", "url": "https://github.com/Harvey-OS/harvey/pull/140" }
gharchive/pull-request
clang changes for rio. Signed-off-by: Ronald G. Minnich<EMAIL_ADDRESS> Rio will chase me until death. LGTM
2025-04-01T06:37:03.151229
2016-08-18T02:08:48
171802317
{ "authors": [ "Morganjackson", "alanhorizon", "kieusonlam", "sham3k" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1227", "repo": "Hashnode/mern-starter", "url": "https://github.com/Hashnode/mern-starter/issues/224" }
gharchive/issue
What is the right way to create a module and add new route? Hi everone, I just know the basic of React Redux, and i need your help :) I found MERN and I using it to make an small app to learn. I clone mern-start and use mern-cli to create a Test module. I just simple render a Hello World text. class Test extends Component { render() { return ( <div>Hello World</div> ); } } then i go to /client/route.js and add a new route under post route. <Route path="/test" getComponent={(nextState, cb) => { require.ensure([], require => { cb(null, require('./modules/Test/Test').default); }); }} /> restart from console and i got this error C:\Users\kieus\Desktop\MERN\client\modules\Auth\Auth.js:17 var _Auth = content.locals; ^ ReferenceError: content is not defined at Object.<anonymous> (Auth.js:6:1) at Module._compile (module.js:541:32) at loader (C:\Users\kieus\Desktop\MERN\node_modules\babel-register\lib\node.js:148:5) at Object.require.extensions.(anonymous function) [as .js] (C:\Users\kieus\Desktop\MERN\node_modules\babel-register\lib\node.js:158:7) at Module.load (module.js:458:32) at tryModuleLoad (module.js:417:12) at Function.Module._load (module.js:409:3) at Module.require (module.js:468:17) at require (internal/module.js:20:19) at Object.<anonymous> (routes.js:21:3) at Module._compile (module.js:541:32) at loader (C:\Users\kieus\Desktop\MERN\node_modules\babel-register\lib\node.js:148:5) at Object.require.extensions.(anonymous function) [as .js] (C:\Users\kieus\Desktop\MERN\node_modules\babel-register\lib\node.js:158:7) at Module.load (module.js:458:32) at tryModuleLoad (module.js:417:12) at Function.Module._load (module.js:409:3) [nodemon] app crashed - waiting for file changes before starting... am i missing something? What i have to do to make it right :( Also css background not working in development mode right? Is there any tutorial about MERN for beginer like: create basic module, jwt authentication? I know too many questions. Please help me Thanks. @kieusonlam Did you find a fix for this? I'm seeing the same thing after running mern init & merng module Editor using the latest mern-cli. Removing the usused style imports fixed this for me. import styles from './Editor.css'; Thanks @Morganjackson error gone away. :+1: wish I had found this issue before spending half a day debugging. I wonder what's ACTUALLY going on here. an empty styles import should not cause a cryptic "segmentation fault"-like error message like this. Downgrading to Webpack ^1.13.3 solved it. Strange bug, though.
2025-04-01T06:37:03.249002
2016-10-29T18:32:24
186093599
{ "authors": [ "Aylr", "levithatcher" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1231", "repo": "HealthCatalyst/healthcareai-py", "url": "https://github.com/HealthCatalyst/healthcareai-py/issues/52" }
gharchive/issue
If one chooses regression and pred col is binary, throw error This is the current unhelpful message: ValueError: Unable to parse string "N" at position 0 Set error (with helpful message) when classification is chosen and pred col is numeric While working on a fix for this problem I discovered a few related problems: a classification run on a non-binary column a classification run on a binary column who's unique values are not 'Y' and 'N' (for example: GenderFLG) a regression run on a column containing non-numeric data Relates to #265
2025-04-01T06:37:03.279156
2016-09-14T17:34:45
176971210
{ "authors": [ "kamarshad", "tristanhimmelman" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1232", "repo": "Hearst-DD/ObjectMapper", "url": "https://github.com/Hearst-DD/ObjectMapper/issues/574" }
gharchive/issue
Want to use the ObjectMapper for Swift 2.2 using Git submodule Hello Hearst-DD I want to use this library for swift 2.2 and like to integrate it using the submodule. Any help would be highly appreciated Hi there, you will need to use v.1.3.0 or earlier for swift 2.2. With regards to including the project as a submodule, you should be able to do this like you would with any other submodule. Hope this helps. Thanks @tristanhimmelman 👍
2025-04-01T06:37:03.428683
2021-04-08T14:22:45
853528587
{ "authors": [ "Helmasaur" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1234", "repo": "Helmasaur/Bioman", "url": "https://github.com/Helmasaur/Bioman/issues/159" }
gharchive/issue
Use a variable for the default channel before using it In commands and events, // instead of doing this everytime: client.channels.cache.get(config.defaultChannel).send(); // use this: const defaultChannel = client.channels.cache.get(config.defaultChannel); defaultChannel.send(); Maybe not needed if the channel isn't fetched from client each time.
2025-04-01T06:37:03.451574
2023-07-11T11:12:14
1798696513
{ "authors": [ "Markus-Goetz", "psteinb" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1235", "repo": "Helmholtz-AI-Energy/beamer-template", "url": "https://github.com/Helmholtz-AI-Energy/beamer-template/issues/31" }
gharchive/issue
problem building on mac Hi, I've been using this wonderful beamer template for some time. Rather recently, I discovered a problem when building on macOS: (./theme/beamerfontthemeHelmholtzAI.sty ! Package beamerhelmholtzai Error: Patching original frame title failed. See the beamerhelmholtzai package documentation for explanation. Type H <return> for immediate help. ... l.202 ...tching original frame title failed}\@ehc} ) (./theme/beamercolorthemeHelmholtzAI.sty) Any ideas? Disable this behaviour by setting HOMEBREW_NO_INSTALL_CLEANUP. Hide these hints with HOMEBREW_NO_ENV_HINTS (see `man brew`). $ make Rc files read: NONE Latexmk: This is Latexmk, John Collins, 7 Jan. 2023. Version 4.79. No existing .aux file, so I'll make a simple one, and require run of *latex. Latexmk: applying rule 'xelatex'... Rule 'xelatex': Reasons for rerun Category 'other': Rerun of 'xelatex' forced or previously required ------------ Run number 1 of rule 'xelatex' ------------ ------------ Running 'xelatex -no-pdf -synctex=1 -interaction=nonstopmode -recorder "talk.tex"' ------------ This is XeTeX, Version 3.141592653-2.6-0.999995 (TeX Live 2023/Homebrew) (preloaded format=xelatex) restricted \write18 enabled. entering extended mode (./talk.tex LaTeX2e <2022-11-01> patch level 1 L3 programming layer <2023-02-22> (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamer. cls Document Class: beamer 2023/02/20 v3.69 A class for typesetting presentations (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb asemodes.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/etoolbox/etool box.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb asedecode.sty)) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/iftex/iftex. sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb aseoptions.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/graphics/keyva l.sty)) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/geometry/geome try.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/iftex/ifvtex .sty)) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/pgf/math/pgfma th.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/pgf/utilities/ pgfrcs.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/utilitie s/pgfutil-common.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/utilitie s/pgfutil-latex.def) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/utilitie s/pgfrcs.code.tex (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/pgf.revi sion.tex))) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/pgf/utilities/ pgfkeys.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/utilitie s/pgfkeys.code.tex (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/utilitie s/pgfkeyslibraryfiltered.code.tex))) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/math/pgf math.code.tex (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/math/pgf mathutil.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/math/pgf mathparser.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/math/pgf mathfunctions.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/math/pgf mathfunctions.basic.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/math/pgf mathfunctions.trigonometric.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/math/pgf mathfunctions.random.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/math/pgf mathfunctions.comparison.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/math/pgf mathfunctions.base.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/math/pgf mathfunctions.round.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/math/pgf mathfunctions.misc.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/math/pgf mathfunctions.integerarithmetics.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/math/pgf mathcalc.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/math/pgf mathfloat.code.tex))) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/base/fleqn.clo ) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/base/size11.cl o) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/pgf/basiclayer /pgfcore.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/graphics/graph icx.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/graphics/graph ics.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/graphics/trig. sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/graphics-cfg/g raphics.cfg) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/graphics-def/x etex.def))) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/pgf/systemlaye r/pgfsys.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/systemla yer/pgfsys.code.tex (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/systemla yer/pgf.cfg) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/systemla yer/pgfsys-xetex.def (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/systemla yer/pgfsys-dvipdfmx.def (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/systemla yer/pgfsys-common-pdf.def)))) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/systemla yer/pgfsyssoftpath.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/systemla yer/pgfsysprotocol.code.tex)) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/xcolor/xcolor. sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/graphics-cfg/c olor.cfg) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/graphics/mathc olor.ltx)) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/basiclay er/pgfcore.code.tex (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/math/pgf int.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/basiclay er/pgfcorepoints.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/basiclay er/pgfcorepathconstruct.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/basiclay er/pgfcorepathusage.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/basiclay er/pgfcorescopes.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/basiclay er/pgfcoregraphicstate.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/basiclay er/pgfcoretransformations.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/basiclay er/pgfcorequick.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/basiclay er/pgfcoreobjects.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/basiclay er/pgfcorepathprocessing.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/basiclay er/pgfcorearrows.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/basiclay er/pgfcoreshade.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/basiclay er/pgfcoreimage.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/basiclay er/pgfcoreexternal.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/basiclay er/pgfcorelayers.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/basiclay er/pgfcoretransparency.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/basiclay er/pgfcorepatterns.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/basiclay er/pgfcorerdf.code.tex))) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/pgf/utilities/ xxcolor.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/base/atbegshi- ltx.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/hyperref/hyper ref.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/ltxcmds/ltxc mds.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pdftexcmds/p dftexcmds.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/infwarerr/in fwarerr.sty)) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/kvsetkeys/kvse tkeys.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/kvdefinekeys /kvdefinekeys.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pdfescape/pd fescape.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/hycolor/hycolo r.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/letltxmacro/le tltxmacro.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/auxhook/auxhoo k.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/hyperref/namer ef.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/refcount/refco unt.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/gettitlestri ng/gettitlestring.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/kvoptions/kvop tions.sty))) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/hyperref/pd1en c.def) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/intcalc/intc alc.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/etexcmds/ete xcmds.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/hyperref/puenc .def) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/url/url.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/bitset/bitse t.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/bigintcalc/b igintcalc.sty)) Package hyperref Message: Stopped early. ) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/hyperref/hxete x.def (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/stringenc/st ringenc.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/rerunfilecheck /rerunfilecheck.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/base/atveryend -ltx.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/uniquecounte r/uniquecounter.sty))) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb aserequires.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb asecompatibility.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb asefont.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/amsfonts/amssy mb.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/amsfonts/amsfo nts.sty)) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/sansmathaccent /sansmathaccent.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/koma-script/sc rlfile.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/koma-script/sc rlfile-hook.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/koma-script/sc rlogo.sty))))) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb asetranslator.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/translator/tra nslator.sty)) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb asemisc.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb asetwoscreens.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb aseoverlay.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb asetitle.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb asesection.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb aseframe.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb aseverbatim.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb aseframesize.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb aseframecomponents.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb asecolor.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb asenotes.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb asetoc.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb asetemplates.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb aseauxtemplates.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb aseboxes.sty))) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb aselocalstructure.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/tools/enumerat e.sty)) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb asenavigation.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb asenavigationsymbols.tex)) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb asetheorems.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/amsmath/amsmat h.sty For additional information on amsmath, use the `?' option. (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/amsmath/amstex t.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/amsmath/amsgen .sty)) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/amsmath/amsbsy .sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/amsmath/amsopn .sty)) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/amscls/amsthm. sty)) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerb asethemes.sty)) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamert hemedefault.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerf ontthemedefault.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamerc olorthemedefault.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beameri nnerthemedefault.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/beamer/beamero uterthemedefault.sty))) (./helmholtzai.sty (./theme/beamerthemeHelmholtzAI.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/booktabs/bookt abs.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/base/fontenc.s ty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/lm/t1lmss.fd)) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/fontspec/fonts pec.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/l3packages/xpa rse/xparse.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/l3kernel/expl3 .sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/l3backend/l3ba ckend-xetex.def))) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/fontspec/fonts pec-xetex.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/base/fontenc.s ty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/fontspec/fonts pec.cfg))) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/base/inputenc. sty Package inputenc Warning: inputenc package ignored with utf8 based engines. ) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/listings/listi ngs.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/listings/lstmi sc.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/listings/listi ngs.cfg)) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/pgf/frontendla yer/tikz.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/pgf/basiclayer /pgf.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/modules/ pgfmoduleshapes.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/modules/ pgfmoduleplot.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/pgf/compatibil ity/pgfcomp-version-0-65.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/pgf/compatibil ity/pgfcomp-version-1-18.sty)) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/pgf/utilities/ pgffor.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/utilitie s/pgffor.code.tex)) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/frontend layer/tikz/tikz.code.tex (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/librarie s/pgflibraryplothandlers.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/modules/ pgfmodulematrix.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/frontend layer/tikz/libraries/tikzlibrarytopaths.code.tex))) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/frontend layer/tikz/libraries/tikzlibrarycalc.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/frontend layer/tikz/libraries/tikzlibraryturtle.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/frontend layer/tikz/libraries/tikzlibrarypositioning.code.tex) (./theme/beamerfontthemeHelmholtzAI.sty ! Package beamerhelmholtzai Error: Patching original frame title failed. See the beamerhelmholtzai package documentation for explanation. Type H <return> for immediate help. ... l.202 ...tching original frame title failed}\@ehc} ) (./theme/beamercolorthemeHelmholtzAI.sty) (./theme/beamerinnerthemeHelmholtzAI.sty) (./theme/beamerouterthemeHelmholtzAI.sty))) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/fontawesome5/f ontawesome5.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/l3packages/l3k eys2e/l3keys2e.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/fontawesome5/f ontawesome5-utex-helper.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/fontawesome5/t ufontawesomefree.fd) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/fontawesome5/t ufontawesomebrands.fd))) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/natbib/natbib. sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/appendixnumber beamer/appendixnumberbeamer.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/pdfpages/pdfpa ges.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/base/ifthen.st y) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/tools/calc.sty ) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/eso-pic/eso-pi c.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/pdfpages/ppxet ex.def)) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/animate/animat e.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/oberdiek/ifdra ft.sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/media9/pdfbase .sty) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/zref/zref-absp age.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/zref/zref-base .sty)) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/ocgx2/ocgbase. sty)) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/librarie s/pgflibraryarrows.meta.code.tex) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/frontend layer/tikz/libraries/tikzlibraryarrows.code.tex (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/librarie s/pgflibraryarrows.code.tex)) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/frontend layer/tikz/libraries/tikzlibrarydecorations.text.code.tex (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/frontend layer/tikz/libraries/tikzlibrarydecorations.code.tex (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/modules/ pgfmoduledecorations.code.tex)) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/generic/pgf/librarie s/decorations/pgflibrarydecorations.text.code.tex)) Package hyperref Warning: Token not allowed in a PDF string (Unicode): (hyperref) removing `\noindent' on input line 38. Package hyperref Warning: Token not allowed in a PDF string (Unicode): (hyperref) removing `\inserttitle' on input line 43. Package hyperref Warning: Token not allowed in a PDF string (Unicode): (hyperref) removing `\insertsubtitle' on input line 43. (./talk.aux) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/base/ts1cmr.fd ) *geometry* driver: auto-detecting *geometry* detected driver: xetex Package hyperref Warning: Rerun to get /PageLabels entry. (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/translator/tra nslator-basic-dictionary-English.dict) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/translator/tra nslator-bibliography-dictionary-English.dict) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/translator/tra nslator-environment-dictionary-English.dict) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/translator/tra nslator-months-dictionary-English.dict) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/translator/tra nslator-numbers-dictionary-English.dict) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/translator/tra nslator-theorem-dictionary-English.dict) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/pdflscape/pdfl scape.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/pdflscape/pdfl scape-nometadata.sty (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/graphics/lscap e.sty))) No file talk.nav. [1] (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/amsfonts/umsa. fd) (/usr/local/Cellar/texlive/20230313_2/share/texmf-dist/tex/latex/amsfonts/umsb. fd) Overfull \vbox (2.91675pt too high) detected at line 105 Overfull \vbox (10.2738pt too high) has occurred while \output is active [2] Overfull \vbox (2.91675pt too high) detected at line 105 Overfull \vbox (10.2738pt too high) has occurred while \output is active [3] Overfull \vbox (2.91675pt too high) detected at line 105 Overfull \vbox (10.2738pt too high) has occurred while \output is active [4] Package natbib Warning: Citation `olah2020naturally' on page 5 undefined on inp ut line 120. Overfull \vbox (10.2738pt too high) has occurred while \output is active [5] Package natbib Warning: Citation `olah2020naturally' on page 6 undefined on inp ut line 120. Overfull \vbox (10.2738pt too high) has occurred while \output is active [6] Package natbib Warning: Citation `bekkers-lecture21' on page 7 undefined on inp ut line 166. Package natbib Warning: Citation `pmlr-v48-cohenc16' on page 7 undefined on inp ut line 166. Package natbib Warning: Citation `cesa2019' on page 7 undefined on input line 1 66. Overfull \vbox (3.60536pt too high) detected at line 166 Overfull \vbox (10.2738pt too high) has occurred while \output is active [7] Overfull \vbox (10.2738pt too high) has occurred while \output is active [8] Overfull \vbox (10.2738pt too high) has occurred while \output is active [9] Overfull \vbox (10.2738pt too high) has occurred while \output is active [10] Overfull \vbox (10.2738pt too high) has occurred while \output is active [11] Overfull \hbox (17.07181pt too wide) in paragraph at lines 238--238 [][] Overfull \vbox (22.37025pt too high) detected at line 238 Overfull \vbox (10.2738pt too high) has occurred while \output is active [12] Overfull \hbox (7.2992pt too wide) in paragraph at lines 238--238 [][] Overfull \vbox (22.37025pt too high) detected at line 238 Overfull \vbox (10.2738pt too high) has occurred while \output is active [13] Overfull \hbox (7.2992pt too wide) in paragraph at lines 238--238 [][] Overfull \vbox (22.37025pt too high) detected at line 238 Overfull \vbox (10.2738pt too high) has occurred while \output is active [14] Overfull \hbox (7.2992pt too wide) in paragraph at lines 238--238 [][] Overfull \vbox (22.37025pt too high) detected at line 238 Overfull \vbox (10.2738pt too high) has occurred while \output is active [15] Overfull \hbox (7.2992pt too wide) in paragraph at lines 238--238 [][] Overfull \vbox (10.2738pt too high) has occurred while \output is active [16] Overfull \vbox (21.55809pt too high) detected at line 251 Overfull \vbox (10.2738pt too high) has occurred while \output is active [17] Overfull \vbox (21.55809pt too high) detected at line 251 Overfull \vbox (10.2738pt too high) has occurred while \output is active [18] Overfull \vbox (10.2738pt too high) has occurred while \output is active [19] Overfull \vbox (10.2738pt too high) has occurred while \output is active [20] Overfull \vbox (10.2738pt too high) has occurred while \output is active [21] Overfull \vbox (10.2738pt too high) has occurred while \output is active [22] Overfull \vbox (10.2738pt too high) has occurred while \output is active [23] Overfull \vbox (10.2738pt too high) has occurred while \output is active [24] Overfull \vbox (10.2738pt too high) has occurred while \output is active [25] Overfull \vbox (10.2738pt too high) has occurred while \output is active [26] Overfull \vbox (10.2738pt too high) has occurred while \output is active [27] Overfull \vbox (10.2738pt too high) has occurred while \output is active [28] Overfull \vbox (10.2738pt too high) has occurred while \output is active [29] Overfull \vbox (10.2738pt too high) has occurred while \output is active [30] Overfull \vbox (10.2738pt too high) has occurred while \output is active [31] Overfull \vbox (10.2738pt too high) has occurred while \output is active [32] Overfull \vbox (10.2738pt too high) has occurred while \output is active [33] Overfull \vbox (10.2738pt too high) has occurred while \output is active [34] Overfull \vbox (10.2738pt too high) has occurred while \output is active [35] Overfull \vbox (10.2738pt too high) has occurred while \output is active [36] Overfull \vbox (10.2738pt too high) has occurred while \output is active [37] Overfull \vbox (10.2738pt too high) has occurred while \output is active [38] Overfull \vbox (10.2738pt too high) has occurred while \output is active [39] Overfull \vbox (10.2738pt too high) has occurred while \output is active [40] Overfull \vbox (10.2738pt too high) has occurred while \output is active [41] Overfull \vbox (10.2738pt too high) has occurred while \output is active [42] Overfull \vbox (10.2738pt too high) has occurred while \output is active [43] Overfull \vbox (10.2738pt too high) has occurred while \output is active [44] Overfull \vbox (10.2738pt too high) has occurred while \output is active [45] Overfull \vbox (10.2738pt too high) has occurred while \output is active [46] No file talk.bbl. Overfull \vbox (10.2738pt too high) has occurred while \output is active [47] Package hyperref Warning: Token not allowed in a PDF string (Unicode): (hyperref) removing `\translate ' on input line 368. LaTeX Warning: Hook 'shipout/lastpage' executed on wrong page (1 not 52). Rerun to correct this. (./talk.aux) LaTeX Warning: Label(s) may have changed. Rerun to get cross-references right. Package rerunfilecheck Warning: File `talk.out' has changed. (rerunfilecheck) Rerun to get outlines right (rerunfilecheck) or use package `bookmark'. ) (see the transcript file for additional information) Output written on talk.xdv (52 pages, 418208 bytes). SyncTeX written on talk.synctex.gz. Transcript written on talk.log. Latexmk: Missing input file 'talk.nav' (or dependence on it) from following: No file talk.nav. Latexmk: Missing bbl file 'talk.bbl' in following: No file talk.bbl. Latexmk: Getting log file 'talk.log' Latexmk: Examining 'talk.fls' Latexmk: Examining 'talk.log' Latexmk: References changed. Latexmk: References changed. Latexmk: References changed. Latexmk: Log file says output to 'talk.xdv' Latexmk: ====List of undefined refs and citations: Citation `olah2020naturally' on page 5 undefined on input line 120 Citation `olah2020naturally' on page 6 undefined on input line 120 Citation `bekkers-lecture21' on page 7 undefined on input line 166 Citation `pmlr-v48-cohenc16' on page 7 undefined on input line 166 Citation `cesa2019' on page 7 undefined on input line 166 Latexmk: If appropriate, the -f option can be used to get latexmk to try to force complete processing. Latexmk: Found bibliography file(s): ./references.bib Latexmk: Summary of warnings from last run of *latex: Latex failed to resolve 5 citation(s) Latexmk: Errors, so I did not complete making targets Collected error summary (may duplicate other messages): xelatex: Command for 'xelatex' gave return code 1 Refer to 'talk.log' and/or above output for details make: *** [talk.pdf] Error 12 section_end:1689066286:step_script section_start:1689066286:cleanup_file_variables Cleaning up project directory and file based variables section_end:1689066288:cleanup_file_variables ERROR: Job failed: Process exited with status 1  I now see this problem on linux as well: (/usr/local/texlive/2022/texmf-dist/tex/generic/pgf/frontendlayer/tikz/librarie s/tikzlibrarypositioning.code.tex) (./theme/beamerfontthemeHelmholtzAI.sty ! Package beamerhelmholtzai Error: Patching original frame title failed. See the beamerhelmholtzai package documentation for explanation. Type H <return> for immediate help. ... l.202 ...tching original frame title failed}\@ehc} ) (./theme/beamercolorthemeHelmholtzAI.sty) This is with texlive-full on fedora 38. Digging a bit further, I added the following statement: \ifpatchable*{\beamer@@frametitle}% {%true \PackageInfo{beamerhelmholtzai}{Patching beamer--frametitle is POSSIBLE}\@ehc }{%false \PackageError{beamerhelmholtzai}{Patching beamer--frametitle is impossible}\@ehc} Just before \patchcmd{\beamer@@frametitle}% in ./theme/beamerfontthemeHelmholtzAI.sty which produces the error stated above. The false path was triggered which hints to the fact why the entire patchcmd call will fail. Fun fact: all CI builds fail for me locally on fc38, the error is NOT triggered when I remove the ifpatchable call I am building with: xelatex -no-pdf -synctex=1 -interaction=nonstopmode --shell-escape -recorder "example.tex" at the version: $ xelatex --version XeTeX 3.141592653-2.6-0.999994 (TeX Live 2022/CVE-2023-32700 patched) kpathsea version 6.3.4 Copyright 2022 SIL International, Jonathan Kew and Khaled Hosny. There is NO warranty. Redistribution of this software is covered by the terms of both the XeTeX copyright and the Lesser GNU General Public License. For more information about these matters, see the file named COPYING and the XeTeX source. Primary author of XeTeX: Jonathan Kew. Compiled with ICU version 72.1; using 72.1 Compiled with zlib version 1.2.13; using 1.2.13 Compiled with FreeType2 version 2.13.0; using 2.13.0 Compiled with Graphite2 version 1.3.14; using 1.3.14 Compiled with HarfBuzz version 7.1.0; using 7.1.0 Compiled with libpng version 1.6.37; using 1.6.37 Compiled with pplib version v2.05 less toxic i hope Compiled with fontconfig version 2.14.2; using 2.14.2 It seems that new TexLive stacks are broken in combination with the template. It seems that only the legacy stacks are working. We do not now yet what the reason is.
2025-04-01T06:37:03.476026
2022-10-20T17:03:36
1416971207
{ "authors": [ "TommiNieminen", "claude-ws01" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1236", "repo": "Helsinki-NLP/OPUS-CAT", "url": "https://github.com/Helsinki-NLP/OPUS-CAT/issues/50" }
gharchive/issue
Garbage in translation (http & Trados & OmegaT) Hello, I'm getting "@ @ " garbage in the translations. I managed to hook Opus in OmegaT & Trados, get lots of "@ @ "... Tested with browser: same bad result. I tested with different versions of the plugin, and engine... always the same. Am I missing something ? Any ideas anyone ? Thank you. Deleting the "c_users_admin_local...opus folders" and restart resolved the issue... for now. (time will tell) ** Note: I wanted to save data in installation folder ("Store OPUS-CAT MT Engine data in..." UNCHECKED) and I got exception erros in both v.<IP_ADDRESS> & v1.2.0 did the same test with v1.2.3... no errors but previous en-es gone. exited OpusCATMTEngine_v1.2.3 then moved "c:\users\admin\local\opuscat" folder to the running opus folder, restarted opus and en-es was there tested HTTP again -> no garbage & perfect translation ** Suggestion improvements: - add, under "Store OPUS-CAT MT engine...", a field containing the current path of storage. - When changing the location of data folder (tick-untick "Store OPUS-CAT..."), if OPUS is not doing the move, display a warning/instruction message that the user can/must move the folder to maintain the list of model (with source path to new destination path). - in the translate model tab, beside the "translate" button, add a button/function to jump to browser & test the same string but with the "http://localhost...". It would allow to confirm the results are the same through HTTP (which relate to plugin results, if I'm not mistaking). Note: I know a lot of it is in the documentation... spreaded here and there (you have to admit), still these are details that are time savers if included as I suggest. Note 2: I (I) would have benefited from readiing your debug procedure/tools when it come to plugins. (ie, I was getting good results in the tranalstion tab of OPUS, but not in trados...)(how do you debug that???) That's it! Relieved that it now works... because when it works, hell it works nice. Opus is a awesome tool... be proud. :) Post note. My issue is resolved. "thank god" as would say my mexican friends. Kind regards Claude. I thought my issue was solved, but no. Finally, pinpointed the issue: the 2019 model files. The issue can be reproduced by adding a 2019 model. (at first I downloaded what I could find on opus, then went to tatoeba, that's why I had 2019 & 2020 model files) Furthermore, got a InvalidOperationException every time I attempted to "delete selected model", Opus crash, then I re-run, then "deleted" again the same model without error. But no issues at all deleting non-2019 models. Included the log file. opuscat_log_DELETING-MODEL.txt I leave the ticket open for you to see. cheers. Thanks for your thorough testing, I'll keep this open as an enhancement issue, since the UI fixes you mention should be fairly simple to implement. The root cause of the garbage output seems to be a fix I made in v<IP_ADDRESS> to get rid of batch file post-processing (this was causing problems to some users). Unfortunately this broke some older models, which used BPE subword segmentation. All the newer models use SentencePiece subword segmentation, and since I've only done testing on them recently, this bug went unnoticed. I will either fix it or remove the BPE models from the model download list. A workaround is to use only the newest models, since I think all language pairs have SentencePiece models available for them now. Thx for your reply, genuinely appreciated. Regarding downloading models... Some thoughts of Enhancements: size of the models a cancel "x" in/beside the progress bar (instead of having to exit Opus) (where I am, in a small village in the "3rd world", I'm limited to 3Mbps. With the DL size, I may decide for a more appropriate time to download.) BTW, really happy with Opus, it works really great, with great results. A jewel. :) Take care. Claude.
2025-04-01T06:37:03.479950
2023-09-29T07:48:41
1918772365
{ "authors": [ "FuckingToasters", "bjarneo" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1237", "repo": "HemmeligOrg/Hemmelig.app", "url": "https://github.com/HemmeligOrg/Hemmelig.app/issues/217" }
gharchive/issue
[OTHER] Change donwload files button text What would you like to share? I'd like to know, where i can change the text from the download button for files from hemmelig_files to something different like "Download Files" Regards, FuckingToasters Additional information No response If you are creating your own version of this application you can figure this out yourself. I am not tech support. I notice a trend with all your issues that it is more or less tech support questions. I do not mind answering, but, when it comes to things like this that people either can easily figure out themselves, or if they are forking this product, then I will not spend my time answering or solving this. If you are creating your own version of this application you can figure this out yourself. I am not tech support. I notice a trend with all your issues that it is more or less tech support questions. I do not mind answering, but, when it comes to things like this that people either can easily figure out themselves, or if they are forking this product, then I will not spend my time answering or solving this. Well these are Questions a user might have, you are the creator of it so you should know where the things are xd. I looked at every file but couldn't find it.
2025-04-01T06:37:03.510255
2017-09-18T19:07:04
258587632
{ "authors": [ "codecov-io", "fgbulsoni" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1238", "repo": "HewlettPackard/oneview-puppet", "url": "https://github.com/HewlettPackard/oneview-puppet/pull/164" }
gharchive/pull-request
Improvements for Server Profiles handling of Templates Description Improvements for Server Profiles handling of Templates SP resources which have a serverProfileTemplateUri declared now, will inherit all undeclared attributes from its template Added load_resource internal method to improve retrieval of different resources Added load_template method to common to be used by SP and SPTs at the moment <NOT READY YET, likely failing on tests and definitely coverage> Issues Resolved #156 Check List [ ] New functionality includes testing. [x] All tests pass ($ rake test). [ ] New functionality has been documented in the README if applicable. [ ] New functionality has been thoroughly documented in the examples (please include helpful comments). [ ] Changes are documented in the CHANGELOG. Codecov Report :exclamation: No coverage uploaded for pull request base (master@9039fd2). Click here to learn what that means. The diff coverage is 45.45%. @@ Coverage Diff @@ ## master #164 +/- ## ========================================= Coverage ? 97.33% ========================================= Files ? 128 Lines ? 2286 Branches ? 0 ========================================= Hits ? 2225 Misses ? 61 Partials ? 0 Impacted Files Coverage Δ ...ib/puppet/provider/oneview_server_profile/c7000.rb 100% <100%> (ø) lib/puppet/provider/common.rb 70.88% <21.42%> (ø) lib/puppet/provider/oneview_resource.rb 94.66% <85.71%> (ø) Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 9039fd2...d3772e8. Read the comment docs.
2025-04-01T06:37:03.511544
2017-08-18T17:04:45
251304626
{ "authors": [ "ffdarkpenguin", "ricardoas" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1239", "repo": "HewlettPackard/oneview-redfish-toolkit", "url": "https://github.com/HewlettPackard/oneview-redfish-toolkit/issues/16" }
gharchive/issue
Implement a schema loader As a dev, I access a class with pre-loaded schema for each redfish structure according to properties specified in a configuration file. The user should be able to specify the folder and the JSON file for each of these structures. Submitted a PR. Don´t know how to link this US with PR. A new US number 28 was created by ??? 'wafflebot'
2025-04-01T06:37:03.512583
2022-01-10T14:30:44
1097969735
{ "authors": [ "EliasBoulharts", "Sispheor" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1240", "repo": "HewlettPackard/squest", "url": "https://github.com/HewlettPackard/squest/issues/317" }
gharchive/issue
The instance details page is not displayed when the spoc is not defined. Done and merged in dev branch
2025-04-01T06:37:03.540527
2023-11-10T20:25:35
1988337594
{ "authors": [ "HiDeoo", "Kotsuha" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1241", "repo": "HiDeoo/toggler-vscode", "url": "https://github.com/HiDeoo/toggler-vscode/pull/21" }
gharchive/pull-request
Add toggle-reverse described in issue 16 Describe the pull request Implement the feature described in issue #16. Screenshots If applicable, add screenshots to help explain what is being modified. This is draft. For example, I see there are test files. I did not write test for this new feature. Feel free to modify anything. Thanks for the contribution, super appreciated :raised_hands: This looks pretty straightforward and working as expected, I just added the following changes: Refactor to use an explicit direction (forward or backward) instead of a boolean reverse flag. Add a test to ensure it's working and properly loop back to the end of the list. Update the CHANGELOG file. Thanks again. Sorry for the late reply, did not see your message earlier. I think there is a difference between your case and the one from this PR you linked. The direction explicitely had 2 options that be easily described (forward and backward) whereas in your case, it may not be that explicit with the use of a password or not. It's difficult to say without knowing or writing the code, but my first guess would be, if this happens more than a few times, to not use RequestOpenSession(true) and maybe have an extra wrapper function, e.g. RequestOpenSessionWithPassword() which would be way more explicit for the reader of the code that would call the underlying function with the correct parameter.
2025-04-01T06:37:03.561514
2018-08-20T11:06:44
352088010
{ "authors": [ "HidekiKawahara", "kalenkovich" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1242", "repo": "HidekiKawahara/legacy_STRAIGHT", "url": "https://github.com/HidekiKawahara/legacy_STRAIGHT/issues/1" }
gharchive/issue
Size mismatch in the morphing example I am trying to run the morphing example from the docs. When I get to morphing in the STRAIGHT and run the following line: mObject3 = directSTRAIGHTmorphing(neutralHai,angryHai,0.5,'linear'); I get the following error: Subscripted assignment dimension mismatch. Error in directSTRAIGHTmorphing (line 45) ap(1:nr1,1:nc1) = (1-mRate)*mObject1.aperiodicityIndex; Debugging shows that nr1 = 2049 while size(mObject1.aperiodicityIndex, 1) = 1025. I am confused as to what might cause this. Thank you for pointing out this. This problem is caused by the mismatch in default parameter setting in exstraightsource.m and exstraightspec.m . Please edit the line 230 of exstraightsource.m by changeing 40 to 80 prm.F0defaultWindowLength = 80; % default frame length for pitch extraction (ms) This fixes the problem. I will check the other side effects. If it is OK, then, I will update the code. I checked side effects and found it is OK.
2025-04-01T06:37:03.566031
2017-02-21T05:22:45
209045337
{ "authors": [ "HiltonGiesenow" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1243", "repo": "HiltonGiesenow/PoShMon", "url": "https://github.com/HiltonGiesenow/PoShMon/issues/110" }
gharchive/issue
Link Directly to Issue / Section Provide a link (in email output) directly to the section, where applicable. For instance, in SharePoint database monitoring, provide a link to Central Admin's database page Done, in next release Still needs some further tests
2025-04-01T06:37:03.585791
2017-10-18T14:27:11
266513937
{ "authors": [ "cropd", "macrozone" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1244", "repo": "HippoAR/react-native-arkit", "url": "https://github.com/HippoAR/react-native-arkit/pull/93" }
gharchive/pull-request
Expose the camera direction This adds the direction of the camera to the readCamera function #70 and might be useful for the expose camera rotation issue #67. cool! thx
2025-04-01T06:37:03.597854
2018-05-22T16:30:29
325372472
{ "authors": [ "HoLLy-HaCKeR", "PR4GM4" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1245", "repo": "HoLLy-HaCKeR/EazFixer", "url": "https://github.com/HoLLy-HaCKeR/EazFixer/issues/9" }
gharchive/issue
Latest version 2018.2 isnt supported? Executing memory patches... Initializing modules... Processing... Cleanup... Applied patches: StringFixer: Failed (Exception has been thrown by the target of an invocation.) ResourceResolver: Failed (Init error: Could not find resolver type) AssemblyResolver: Failed (Init error: Could not find resolver type) Writing new assembly... Done. Could you double-check that the program you're deobfuscating is indeed using EazFuscator, and then upload your sample? I can confirm its Eaz 2018.2 (Latest Release) Upload: https://www105.zippyshare.com/v/KOwGRCis/file.html Found that issue was some naive comparisons on my part. I'm using SigComparer to compare methods to the expected string decrypter, but one of the methods in your target assembly has this code: \u0008\u2007\u2000.\u0002(1996868267) + \u0006\u2005\u2000.\u0002(16) My code thinks both of these are encrypted strings, but the second one isnt (even though it has the same signature and method name). Pushing a fix in a few minutes.
2025-04-01T06:37:03.607974
2023-01-24T13:01:32
1554943143
{ "authors": [ "Hoffelhas", "ShayHill", "moorsey" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1246", "repo": "Hoffelhas/autodoist", "url": "https://github.com/Hoffelhas/autodoist/issues/34" }
gharchive/issue
"You can only submit max 100 commands pr. sync operation. Please split them up in smaller chunks" Getting the following when running python3 autodoist.py -a *** -l next -hf 2 2023-01-24 12:56:47 INFO You are running with the following functionalities: Next action labelling mode: Enabled Regenerate sub-tasks mode: Disabled Shifted end-of-day mode: Disabled 2023-01-24 12:56:49 INFO Autodoist has successfully connected to Todoist! 2023-01-24 12:56:50 INFO SQLite DB has successfully initialized! 2023-01-24 12:57:04 ERROR Error trying to sync with Todoist API: 400 Client Error: Bad Request for url: https://api.todoist.com/sync/v9/sync Traceback (most recent call last): File "/srv/dev-disk-by-uuid-12b417ae-a39e-46cb-aae6-35bf23871f11/dockerdata/autodoist/autodoist/autodoist.py", line 521, in sync response.raise_for_status() File "/usr/local/lib/python3.9/dist-packages/requests/models.py", line 1021, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 400 Client Error: Bad Request for url: https://api.todoist.com/sync/v9/sync Hm, that's indeed strange. I'll give it a look too this weekend. Looks like the API key has some issues when passed to the sync API. I was not able to recreate the issue, so I've added a few additional debug logs and pushed it to a new branch '34_extra_logs'. Could you please run this version with the --debug flag on, and send me a copy with the sensitive information removed? I'm only interested in the last part where it caches the sync API errors. Many thanks. Looks like it could be a limit issue? 2023-02-11 11:51:56 DEBUG response: { "error": "You can only submit max 100 commands pr. sync operation. Please split them up in smaller chunks", "error_code": 36, "error_extra": { "event_id": "1eab52c58829436faa9d0d2a319b6052", "retry_after": 7 }, "error_tag": "LIMITS_REACHED_COMMANDS", "http_code": 400 } Hey @Hoffelhas Just checking in, if you have any thoughts on this one? Many thanks! Dug a little into this, as I understand, the following is the code that sends the batched requests off and is what will fail if there are more than 100. data = 'sync_token=' + api.sync_token + \ '&commands=' + json.dumps(api.queue) Many people will not hit this, but as I have quite a lot of tasks and not been able to use autodoist for some time, there are now too many requests to send Having a look to see if I can figure out how to only send 100 at a time and will contribute back if I manage it! Found code on splitting lists into chunks, but not fully sure how to integrate with these lines without making a mess Hi there, sorry it took a while for me to respond. Todoist indeed got quite a bit stricter with the max. amount of syncs you're allowed to send in a given period. Basically the documentation says the following: For each user, you can make a maximum of 450 partial sync requests within a 15 minute period. For each user, you can make a maximum of 45 full sync requests within a 15 minute period. The maximum number of commands is 100 per request. When adding new over time tasks, it's indeed difficult to reach these numbers. However if you have a project with >100 items, and you would activate or change labelling on the project level, then you indeed would get a batch that's too big. However, you're thinking in the right direction; it should then be split up in multiple batches with max. 100. This should be relatively simple: when we enter the 'if api.queue' at line 1523, we have to check if api.queue>100, if so, split it up and run each block through sync(api) currently at line 1524. However do note that if you reach the 450 changes within 15 minutes, then Todoist will hard block your connection. So even we implement this work-around, you should not label and un-label your project with >100 items more than a few times per hour. Yes, I saw that other limitation. I don't think that would really be an issue normally, even the 100 commands per request is a bit unique to initial syncs I think, or as you say, large parallel processed projects Appreciate the reply! OK, I've come up with: # Sync all queued up changes if len(api.queue) < 100: sync(api) else: start = 0 end = len(api.queue) step = 100 for i in range(start, end, step): x = i sync(api.queue[x:x+step]) But I don't think the last line is right, doesn't work when testing at least anyway. Looks like it needs to be "sync(api)", but needs to be called on the list batch Will keep trying anyway! I've done something similar to moorsey on my own project: _COMMAND_CHUNK_SIZE = 99 def _write_some_changes( headers: CaseInsensitiveDict[str], commands: list[Command] ) -> str: """Write changes to the Todoist API. :param headers: Headers for the request (produced by headers.get_headers) :param commands: list of dictionaries (commands) to add to the API :return: sync_token from the API """ resp = requests.post( SYNC_URL, headers=headers, data=json.dumps({"commands": commands}) ) resp.raise_for_status() return str(resp.json()["sync_token"]) def write_changes( sync_token: str, headers: CaseInsensitiveDict[str], commands: list[Command] ) -> str: """Write the changes to the Todoist API, one chunk at a time. :param sync_token: current sync_token, will be updated if any commands are sent :param headers: Headers for the request (produced by headers.get_headers) :param commands: list of dictionaries (commands) to add to the API :return: sync_token from the API I don't know what the soft limit is, but I get lot of bad request errors if I send 1000 commands at once. """ if not commands: return sync_token try: sync_token = _write_some_changes(headers, commands[:_COMMAND_CHUNK_SIZE]) except Exception: # give up and start the whole main loop over return "*" time.sleep(1) return write_changes(sync_token, headers, commands[_COMMAND_CHUNK_SIZE:]) It's not isolated enough to be pasted into autodoist unfortunately, but it might give clues to someone in the thread. It does work. I hit the limit often when hiding / unhiding large projects with autotagging. Just put some code together for this also. Disclaimer, had help from my friend Google Bard on this. We muddled our way through together! First pull request after being on the internet for some time, hoping @Hoffelhas is well and able to look through the other contributions soon Hoping to get my GTD game back in order after a few years lost in the ocean, now hopefully have my next action labelling back! Adding "learn python" to my projects list!
2025-04-01T06:37:03.623088
2023-10-18T15:25:14
1950025145
{ "authors": [ "ArnoSchiller", "simutaroman" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1247", "repo": "HomagGroup/Blazor3D", "url": "https://github.com/HomagGroup/Blazor3D/issues/32" }
gharchive/issue
How to select object via code Hey, thanks for this great package! I've got a 3D view and an external button to delete the selected object (like in your examples). When the object is deleted i want to preselect the nearest object in the scene. Is there a way to trigger selection (highlighting) in code? Thanks! Hi See Example1 private async Task OnSelectObjectByUUIDClick() { if (selectObjectGuid.HasValue) { await View3D1.SelectByUuidAsync(selectObjectGuid.Value); } } Ohh sorry, I overlooked that. Thank you very much!
2025-04-01T06:37:03.627873
2024-07-13T10:47:23
2406843215
{ "authors": [ "0100101101001011", "HomeSpan", "ronny-antoon" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1248", "repo": "HomeSpan/HomeSpan", "url": "https://github.com/HomeSpan/HomeSpan/issues/885" }
gharchive/issue
ESP32 Arduino Smart Home Project With HomeSpan(HomeKit protocol) VS. MATTER protocol Hi there! First and foremost, I'd like to express my gratitude for creating such an amazing project! I am currently working on a smart home project involving smart lights with Apple Home support. I am contemplating the pros and cons of integrating an esp32 Arduino project with Apple Home using either HomeKit or Matter protocol. Could you please provide some insights into the advantages and disadvantages of these two protocols for this specific integration? HomeSpan better in all things, but their is one problem that forced me to move to Matter, the MDNS issue that randomly happens. Matter its slow than HomeSpan, hard than HomeSpan, But never "no Response" even that they have a memory leak. I think that's right. MDNS is not universally supported the same way by every router, so in some cases it's just too unpredictable. A good case study is the commercial company Rachio - they build really nice irrigation controllers and automatic water hose valves. A few years ago they announced HomeKit compatibility, but ran into the same problems. It works really well for some people, but not very well for others. They eventually decided to stop supporting HomeKit, though they retained the software and the pairing codes so you can still use it if you'd like (just don't call them if it doesn't work). In my own home I originally used a Linksys Velop mesh network for a year and it worked fine. Then the Velop started having problems as a result of firmware updates. At first the problems did not impact HomeSpan stuff, but with each new update I had more and more problems with HomeSpan as well as commercial HomeKit devices. I eventually threw the Velop system in the garbage and switched to a NetGear Orbi mesh network and it's been rock solid for 3+ years. Not sure why some routers work better than others. I recommend starting with HomeSpan. If you load example 1 and it is stable for a few weeks, that likely means your routers fully support the MDNS protocol (at least the version used by Espressif). If unstable (using Example 1), then you may want to try Matter. If you are a developer person, and have a good experience, i will advise you to go for espressif framework and not arduino. You can control everything in espressif. But in arduino you will get some compiled files that cant really change and sone flags that cant change. So i recommend using esp_matter over some arduino matter. But in overall HomeSpan the better simple faster for development and use.
2025-04-01T06:37:03.640883
2021-02-27T02:40:23
817801172
{ "authors": [ "jonchang", "wangliguang517" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1249", "repo": "Homebrew/brew", "url": "https://github.com/Homebrew/brew/issues/10721" }
gharchive/issue
sshfs.c:15:10: fatal error: 'fuse_lowlevel.h' file not found brew config output when brew install datawire/blackbird/telepresence sshfs.c:15:10: fatal error: 'fuse_lowlevel.h' file not found #include <fuse_lowlevel.h> ^~~~~~~~~~~~~~~~~ 1 error generated. make[1]: *** [sshfs-sshfs.o] Error 1 make[1]: *** Waiting for unfinished jobs.... make: *** [install-recursive] Error 1 Do not report this issue to Homebrew/brew or Homebrew/core! Please create pull requests instead of asking for help on Homebrew's GitHub, Twitter or any other official channels. brew doctor output sshfs.c:15:10: fatal error: 'fuse_lowlevel.h' file not found #include <fuse_lowlevel.h> ^~~~~~~~~~~~~~~~~ 1 error generated. make[1]: *** [sshfs-sshfs.o] Error 1 make[1]: *** Waiting for unfinished jobs.... make: *** [install-recursive] Error 1 Do not report this issue to Homebrew/brew or Homebrew/core! Please create pull requests instead of asking for help on Homebrew's GitHub, Twitter or any other official channels. [X] The brew doctor above contains no "Warning" lines. What were you trying to do (and why)? sshfs.c:15:10: fatal error: 'fuse_lowlevel.h' file not found #include <fuse_lowlevel.h> ^~~~~~~~~~~~~~~~~ 1 error generated. make[1]: *** [sshfs-sshfs.o] Error 1 make[1]: *** Waiting for unfinished jobs.... make: *** [install-recursive] Error 1 Do not report this issue to Homebrew/brew or Homebrew/core! Please create pull requests instead of asking for help on Homebrew's GitHub, Twitter or any other official channels. What happened (include all command output)? sshfs.c:15:10: fatal error: 'fuse_lowlevel.h' file not found #include <fuse_lowlevel.h> ^~~~~~~~~~~~~~~~~ 1 error generated. make[1]: *** [sshfs-sshfs.o] Error 1 make[1]: *** Waiting for unfinished jobs.... make: *** [install-recursive] Error 1 Do not report this issue to Homebrew/brew or Homebrew/core! Please create pull requests instead of asking for help on Homebrew's GitHub, Twitter or any other official channels. What did you expect to happen? sshfs.c:15:10: fatal error: 'fuse_lowlevel.h' file not found #include <fuse_lowlevel.h> ^~~~~~~~~~~~~~~~~ 1 error generated. make[1]: *** [sshfs-sshfs.o] Error 1 make[1]: *** Waiting for unfinished jobs.... make: *** [install-recursive] Error 1 Do not report this issue to Homebrew/brew or Homebrew/core! Please create pull requests instead of asking for help on Homebrew's GitHub, Twitter or any other official channels. Step-by-step reproduction instructions (by running brew commands) sshfs.c:15:10: fatal error: 'fuse_lowlevel.h' file not found #include <fuse_lowlevel.h> ^~~~~~~~~~~~~~~~~ 1 error generated. make[1]: *** [sshfs-sshfs.o] Error 1 make[1]: *** Waiting for unfinished jobs.... make: *** [install-recursive] Error 1 Do not report this issue to Homebrew/brew or Homebrew/core! Please create pull requests instead of asking for help on Homebrew's GitHub, Twitter or any other official channels. Do not report this issue to Homebrew/brew or Homebrew/core! Please report this to whomever runs datawire/blackbird.
2025-04-01T06:37:03.650647
2023-02-15T03:12:27
1585140777
{ "authors": [ "Bo98", "MikeMcQuaid", "apainintheneck", "carlocab" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1250", "repo": "Homebrew/brew", "url": "https://github.com/Homebrew/brew/issues/14634" }
gharchive/issue
brew irb errors immediately when HOMEBREW_PRY=1 brew doctor output Your system is ready to brew. Verification [X] My "brew doctor output" above says Your system is ready to brew. and am still able to reproduce my issue. [X] I ran brew update twice and am still able to reproduce my issue. [X] This issue's title and/or description do not reference a single formula e.g. brew install wget. If they do, open an issue at https://github.com/Homebrew/homebrew-core/issues/new/choose instead. brew config output HOMEBREW_VERSION: 3.6.21-132-g724e3e6 ORIGIN: https://github.com/Homebrew/brew HEAD: 724e3e646abf04474b083f581a963a52e39fe47d Last commit: 34 minutes ago Core tap origin: https://github.com/Homebrew/homebrew-core Core tap HEAD: 0bbb89420e74756a5a5c145ed7efa4a32f7e7e7c Core tap last commit: 2 days ago Core tap branch: master Core tap JSON: 15 Feb 03:07 UTC HOMEBREW_PREFIX: /home/linuxbrew/.linuxbrew HOMEBREW_CASK_OPTS: [] HOMEBREW_GITHUB_API_TOKEN: set HOMEBREW_MAKE_JOBS: 2 HOMEBREW_PRY: set Homebrew Ruby: 2.6.8 => /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/portable-ruby/2.6.8_1/bin/ruby CPU: dual-core 64-bit skylake Clang: N/A Git: 2.39.1 => /bin/git Curl: 7.81.0 => /bin/curl Kernel: Linux 5.4.0-1103-azure x86_64 GNU/Linux OS: Ubuntu 22.04.1 LTS (jammy) Host glibc: 2.35 /usr/bin/gcc: 11.3.0 /usr/bin/ruby: N/A glibc: N/A gcc@11: N/A gcc: N/A xorg: N/A What were you trying to do (and why)? Use HOMEBREW_PRY=1 brew irb to programmatically examine the contents of formulae. What happened (include all command output)? linuxbrew@codespaces-07ca68:/workspaces/brew$ HOMEBREW_PRY=1 brew irb Error: cannot load such file -- method_source Please report this issue: https://docs.brew.sh/Troubleshooting /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.6.0/gems/pry-0.14.2/lib/pry/pry_instance.rb:3:in `require' /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.6.0/gems/pry-0.14.2/lib/pry/pry_instance.rb:3:in `<top (required)>' /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.6.0/gems/pry-0.14.2/lib/pry.rb:61:in `require' /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.6.0/gems/pry-0.14.2/lib/pry.rb:61:in `<top (required)>' /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/dev-cmd/irb.rb:54:in `require' /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/dev-cmd/irb.rb:54:in `irb' /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/brew.rb:93:in `<main>' What did you expect to happen? For a Ruby REPL to start. Step-by-step reproduction instructions (by running brew commands) HOMEBREW_PRY=1 brew irb I think this is caused by how we remove vendored gems via git. Workaround is to delete your local pry gem in vendor/bundle/ruby/... and re-trigger the install Hmm, yea, makes sense: 39c6f7d6fbaba464798751c2c42f8f881da00e80 is the first bad commit commit 39c6f7d6fbaba464798751c2c42f8f881da00e80 Author: apainintheneck<EMAIL_ADDRESS>Date: Sun Feb 5 20:54:15 2023 -0800 Stop including `*flight` block source in cask API Originally we were going to try and load the *flight blocks from the API but we ended up going with downloading the caskfile for the subset of casks that need this functionality for consisty's sake. This reverts the following commits from most recent to oldest: - ffc74a51fb32b66a4cd8bc41dbd076dd23d9100e - e5616e94fe42505434c330be35eeafef2739944f - d1490c3d5c087d00f2bca1787cce331202b195c5 - 7ca5a5d9a71a73f21bbb8555a38048f027bee89b - 2d5d132713d0701d02d5ff33e9918812d13d2a83 It also changes how *flight blocks are handled in `.to_h`. Essentially, when *flight blocks exist they are just included as a hash of the artifact to nil to indicate that they exist. More information isn't necessary since we don't evaluate the current source code in the *flight artifacts that we get from the API. .gitignore | 1 + Library/Homebrew/Gemfile | 1 - Library/Homebrew/Gemfile.lock | 1 - .../cask/artifact/abstract_flight_block.rb | 6 - Library/Homebrew/cask/cask.rb | 3 +- Library/Homebrew/cask/cask_loader.rb | 6 +- Library/Homebrew/test/cask/cask_spec.rb | 48 ------- .../fixtures/cask/Casks/conditional-flight.rb | 21 --- .../gems/method_source-1.0.0/lib/method_source.rb | 141 ------------------- .../lib/method_source/code_helpers.rb | 154 --------------------- .../lib/method_source/source_location.rb | 138 ------------------ .../lib/method_source/version.rb | 3 - 12 files changed, 4 insertions(+), 519 deletions(-) delete mode 100644 Library/Homebrew/test/support/fixtures/cask/Casks/conditional-flight.rb delete mode 100644 Library/Homebrew/vendor/bundle/ruby/2.6.0/gems/method_source-1.0.0/lib/method_source.rb delete mode 100644 Library/Homebrew/vendor/bundle/ruby/2.6.0/gems/method_source-1.0.0/lib/method_source/code_helpers.rb delete mode 100644 Library/Homebrew/vendor/bundle/ruby/2.6.0/gems/method_source-1.0.0/lib/method_source/source_location.rb delete mode 100644 Library/Homebrew/vendor/bundle/ruby/2.6.0/gems/method_source-1.0.0/lib/method_source/version.rb Ok, that worked. Thanks for the tip! Would be nice for brew to handle this better somehow, but not really sure how that would work. I suppose having pry a part of the Gemfile probably would improve things here, though I don't think there was a universal agreement to the approach (I still think gem groups makes sense for stuff like that). There's been various other issues caused by vendor gem changes over the years (including brew doctor failures in homebrew-core), but it seems we've ironed that out a little. I'd be game to start including all these sorts of development gems that are installed optionally and/or we remove those that aren't really used. @apainintheneck given your recent PR: is this still a problem? This is something completely different than what was covered in my PR. I actually ran into the exact same error as @carlocab when working with pry and this thread was very helpful. I agree that it'd be nice to have a better way of handling dev dependencies but it's a relatively easy fix in this case. This was fixed in https://github.com/Homebrew/brew/commit/d69c3ef3df7c332fdc646e0533b898b661f5fccf by vendoring pry. Thanks for cleaning this up @apainintheneck 🙇🏻
2025-04-01T06:37:03.659263
2024-04-27T13:17:23
2267012052
{ "authors": [ "MikeMcQuaid", "dduugg", "reitermarkus" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1251", "repo": "Homebrew/brew", "url": "https://github.com/Homebrew/brew/issues/17167" }
gharchive/issue
Documentation is replaced by auto-generated RBI comments. brew doctor output N/A Verification [X] My "brew doctor output" above says Your system is ready to brew. and am still able to reproduce my issue. [X] I ran brew update twice and am still able to reproduce my issue. [X] This issue's title and/or description do not reference a single formula e.g. brew install wget. If they do, open an issue at https://github.com/Homebrew/homebrew-core/issues/new/choose instead. brew config output N/A What were you trying to do (and why)? Documentation for e.g. RuboCops only shows auto-generated comments from RBI files: From https://rubydoc.brew.sh/RuboCop/Cop/Homebrew/Blank.html: DO NOT EDIT MANUALLY This is an autogenerated file for dynamic methods in RuboCop::Cop::Homebrew::Blank. Please instead update this file by running bin/tapioca dsl RuboCop::Cop::Homebrew::Blank. What happened (include all command output)? I assume the reason is that yard-sorbet is run last. It should be changed to skip any comments in RBI files. What did you expect to happen? Original documentation should be shown. Step-by-step reproduction instructions (by running brew commands) 1. `brew rubydoc --open` 1. Look at `RuboCop::Cop::Homebrew::Blank` documentation. I opened a PR upstream: https://github.com/Shopify/tapioca/pull/1885 🤞 Thanks for issues @reitermarkus and @dduugg! We can also patch Tapioca if upstream doesn't accept the patch (early feedback isn't positive) 🙈. Something like: https://github.com/Homebrew/brew/compare/tapioca-patch?expand=1 We can also patch Tapioca if upstream doesn't accept the patch (early feedback isn't positive) 🙈. Something like: tapioca-patch?expand=1 (compare) @dduugg This would work fine for me, monkey-patching like this is IMO fairly inconsequential when we're not exposing it to end-users. @dduugg, fine by me. We can always revert it once it is fixed upstream. Upstream PR was accepted, this will be fixed in the next tapioca release
2025-04-01T06:37:03.669442
2016-10-13T11:26:51
182763939
{ "authors": [ "DomT4", "MikeMcQuaid", "UniqMartin", "apjanke", "bfontaine", "dunn", "mistydemeo", "retokromer", "scpeters", "tschoonj", "vladshablinsky", "vszakats" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1252", "repo": "Homebrew/brew", "url": "https://github.com/Homebrew/brew/pull/1283" }
gharchive/pull-request
All things must come to an end eventually. [x] Have you followed the guidelines in our Contributing document? [x] Have you checked to ensure there aren't other open Pull Requests for the same change? [x] Have you successfully run brew tests with your changes locally? 😢 Thanks for all your contributions and all the help you provided me! Why not as former maintainer? Why not as former maintainer? @retokromer He will be going there whether he likes it or not 😉 Yes, thanks for your contributions! Thank you @DomT4 for your dedicated and professional work on Homebrew, it will be missed. Not to forget all the help and support you gave along my PRs. ❤️ @DomT4 I've appreciated all of your work very much. Thank you! ❤️ Wasn't quite expecting so many people to notice the PR 😅. Left a message in the team chat for the maintainer side of things, but thank you to the folks who commented here, it is very much appreciated. I guess people always say that, but it remains true to me ❤️. Getting to work with some of you & being involved in the Homebrew community in general has been a pleasure, sometimes rather frustrating as FOSS can be, but Homebrew is one of the best open-source communities you can stumble into and contribute towards, especially if you're new or returning to the whole open-source thing and want to help out but aren't really sure where to start. I have no idea really why Mike & the team that was here at the time, a few of which remain active, trusted me with maintaining Homebrew however long ago it was now, but I managed to not completely burn the metaphorical house down so I guess it didn't work out too awfully. Folks are welcome to ping me if there's stuff you stumble across later and think "Why?", this account will remain active even if I haven't decided how much it'll be used going forwards and your pings will get filtered into my GH mailbox. Just don't ping me to tell me you think what I did was terrible 😉. Up to @MikeMcQuaid when this gets merged. I'll leave that in your hands. And belatedly, congratulations on becoming a maintainer @tschoonj. Your work on the GNOME stuff has always been 💯 and I've no idea how you consistently find the patience for it. @DomT4 As I stated in Slack but will restate here: it's been a pleasure working with you and you should be proud of all the great work you've done. @tschoonj @reitermarkus @jawshooah Just a note that I used this chance to add you to the README/manpage too. :sob: @MikeMcQuaid Thanks Mike. Appreciate all the support you've given me around Homebrew over my time here. Fairly frequently you had more faith in me than I had in myself. Let me know if I need to sign or post anything to step down from the SFC PLC. @bfontaine ❤️❤️. Thanks for all your work Dom! It's been a pleasure working with you. Thanks for everything Dom, from the amazing and mind-blowing amount work you contributed to Homebrew to the pleasure of working with you and the discussions we had! You've certainly played a key role in me getting more involved with Homebrew and feeling truly welcome here. :heart: 😔 @DomT4 I didn't have too much time to work with you, however I remember you finding some tricky bugs in my code and me fixing them, and it stays as a good memory for me. Thanks for all of your work on Homebrew and best of luck on all your endeavors. @apjanke Thanks Andrew. My sincere apologies that my departure likely leaves the way Homebrew handles Java sandboxing heavily on your shoulders 🙈. Maybe not the best leaving gift in the world. @UniqMartin I might have lasted longer if I hadn't tried to do as much of absolutely everything as I could heh, which is perhaps advice I pass on to any new maintainer in future months or years who stumbles upon this thread. Sometimes you need a break from things and I'm not very good at giving myself them. Thank you for the many discussions we had; I'm glad and thankful for any influence I had in making you feel welcome around Homebrew and a part of the team. @vladshablinsky I learnt quite a bit from keeping an eye on your PRs. Occasionally finding interesting ways to break things was something I formed a habit of doing 😆, but that being useful is always a relief. Although we didn't get the chance to do much work together thrilled to see you join the maintainers team, Homebrew is lucky to have you around.
2025-04-01T06:37:03.671417
2024-06-15T00:20:58
2354287268
{ "authors": [ "Bo98", "BrewTestBot", "MikeMcQuaid" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1253", "repo": "Homebrew/brew", "url": "https://github.com/Homebrew/brew/pull/17513" }
gharchive/pull-request
sorbet: Update RBI files. Autogenerated by the sorbet workflow. @Bo98 guess this is related to the bootsnap changes recently. Yeah bit of an oversight on my part. Tapioca can't autogen the stuff in Portable Ruby, but we only use one method here so will just provide that manually (and more strongly typed than what Tapioca provides anyway).
2025-04-01T06:37:03.682526
2016-05-13T21:23:32
154803091
{ "authors": [ "Blaisorblade", "mikemcquaid", "xu-cheng" ], "license": "BSD-2-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0000.json.gz:1254", "repo": "Homebrew/brew", "url": "https://github.com/Homebrew/brew/pull/242" }
gharchive/pull-request
Synchronize counts for link and unlink [x] Have you followed the guidelines in our Contributing document? [x] Have you checked to ensure there aren't other open Pull Requests for the same change? [x] Have you added an explanation of what your changes do and why you'd like us to include them? [ ] Have you written new tests for your changes? Here's an example. [x] Have you successfully run brew tests with your changes locally? The output of linking and unlinking a formula can give different counts in some cases, as mentioned in #239: $ brew link neovim Linking /usr/local/Cellar/neovim/0.1.4... 40 symlinks created $ brew unlink neovim Unlinking /usr/local/Cellar/neovim/0.1.4... 66 symlinks removed Investigation confirms that the links created and removed are the same, but unlink counts also directories among symlinks removed. This PR improves the situation to synchronize link and unlink, getting both to talk about 40 symlinks. Verbose logs confirm this is the amount of links created/removed (see below). This PR also takes care to not affect statistics used by other commands, which specifically mention directories. If desired, one could also log the count of created and removed directories in link/unlink's output. Furthermore, this PR restores logging of mkdir operation during verbose linking, which was removed in https://github.com/Homebrew/legacy-homebrew/commit/f899878220668c7c7f0fcf43c6d294a52b7e79ed without a specific rationale. Logging both creations and removals appears more symmetric. In previous discussion in #239, @mikemcquaid explained the issue was known and low-priority, while @xu-cheng claimed this was not an issue and offered an alternative explanation. I investigated the alternative explanation in more detail, and was unable to find confirming evidence, at least in this instance. $ brew -v unlink neovim Unlinking /usr/local/Cellar/neovim/0.1.4... rm /usr/local/bin/nvim rm /usr/local/share/locale/af/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/ca/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/cs/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/cs.cp1250/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/de/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/en_GB/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/eo/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/es/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/fi/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/fr/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/ga/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/it/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/ja/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/ja.euc-jp/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/ja.sjis/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/ko/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/ko.UTF-8/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/nb/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/nl/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/no/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/pl/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/pl.UTF-8/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/pl.cp1250/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/pt_BR/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/ru/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/ru.cp1251/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/sk/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/sk.cp1250/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/sv/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/uk/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/uk.cp1251/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/vi/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/zh_CN/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/zh_CN.UTF-8/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/zh_CN.cp936/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/zh_TW/LC_MESSAGES/nvim.mo rm /usr/local/share/locale/zh_TW.UTF-8/LC_MESSAGES/nvim.mo rm /usr/local/share/man/man1/nvim.1 rm /usr/local/share/nvim rmdir /usr/local/share/locale/zh_TW.UTF-8/LC_MESSAGES rmdir /usr/local/share/locale/zh_TW.UTF-8 rmdir /usr/local/share/locale/zh_CN.cp936/LC_MESSAGES rmdir /usr/local/share/locale/zh_CN.cp936 rmdir /usr/local/share/locale/zh_CN.UTF-8/LC_MESSAGES rmdir /usr/local/share/locale/zh_CN.UTF-8 rmdir /usr/local/share/locale/uk.cp1251/LC_MESSAGES rmdir /usr/local/share/locale/uk.cp1251 rmdir /usr/local/share/locale/sk.cp1250/LC_MESSAGES rmdir /usr/local/share/locale/sk.cp1250 rmdir /usr/local/share/locale/ru.cp1251/LC_MESSAGES rmdir /usr/local/share/locale/ru.cp1251 rmdir /usr/local/share/locale/pl.cp1250/LC_MESSAGES rmdir /usr/local/share/locale/pl.cp1250 rmdir /usr/local/share/locale/pl.UTF-8/LC_MESSAGES rmdir /usr/local/share/locale/pl.UTF-8 rmdir /usr/local/share/locale/no/LC_MESSAGES rmdir /usr/local/share/locale/no rmdir /usr/local/share/locale/ko.UTF-8/LC_MESSAGES rmdir /usr/local/share/locale/ko.UTF-8 rmdir /usr/local/share/locale/ja.sjis/LC_MESSAGES rmdir /usr/local/share/locale/ja.sjis rmdir /usr/local/share/locale/ja.euc-jp/LC_MESSAGES rmdir /usr/local/share/locale/ja.euc-jp rmdir /usr/local/share/locale/cs.cp1250/LC_MESSAGES rmdir /usr/local/share/locale/cs.cp1250 40 symlinks removed $ brew -v link neovim Linking /usr/local/Cellar/neovim/0.1.4... ln -s ../Cellar/neovim/0.1.4/bin/nvim nvim ln -s ../../../../Cellar/neovim/0.1.4/share/locale/af/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../../Cellar/neovim/0.1.4/share/locale/ca/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../../Cellar/neovim/0.1.4/share/locale/cs/LC_MESSAGES/nvim.mo nvim.mo mkdir /usr/local/share/locale/cs.cp1250 mkdir /usr/local/share/locale/cs.cp1250/LC_MESSAGES ln -s ../../../../Cellar/neovim/0.1.4/share/locale/cs.cp1250/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../../Cellar/neovim/0.1.4/share/locale/de/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../../Cellar/neovim/0.1.4/share/locale/en_GB/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../../Cellar/neovim/0.1.4/share/locale/eo/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../../Cellar/neovim/0.1.4/share/locale/es/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../../Cellar/neovim/0.1.4/share/locale/fi/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../../Cellar/neovim/0.1.4/share/locale/fr/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../../Cellar/neovim/0.1.4/share/locale/ga/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../../Cellar/neovim/0.1.4/share/locale/it/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../../Cellar/neovim/0.1.4/share/locale/ja/LC_MESSAGES/nvim.mo nvim.mo mkdir /usr/local/share/locale/ja.euc-jp mkdir /usr/local/share/locale/ja.euc-jp/LC_MESSAGES ln -s ../../../../Cellar/neovim/0.1.4/share/locale/ja.euc-jp/LC_MESSAGES/nvim.mo nvim.mo mkdir /usr/local/share/locale/ja.sjis mkdir /usr/local/share/locale/ja.sjis/LC_MESSAGES ln -s ../../../../Cellar/neovim/0.1.4/share/locale/ja.sjis/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../../Cellar/neovim/0.1.4/share/locale/ko/LC_MESSAGES/nvim.mo nvim.mo mkdir /usr/local/share/locale/ko.UTF-8 mkdir /usr/local/share/locale/ko.UTF-8/LC_MESSAGES ln -s ../../../../Cellar/neovim/0.1.4/share/locale/ko.UTF-8/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../../Cellar/neovim/0.1.4/share/locale/nb/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../../Cellar/neovim/0.1.4/share/locale/nl/LC_MESSAGES/nvim.mo nvim.mo mkdir /usr/local/share/locale/no mkdir /usr/local/share/locale/no/LC_MESSAGES ln -s ../../../../Cellar/neovim/0.1.4/share/locale/no/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../../Cellar/neovim/0.1.4/share/locale/pl/LC_MESSAGES/nvim.mo nvim.mo mkdir /usr/local/share/locale/pl.UTF-8 mkdir /usr/local/share/locale/pl.UTF-8/LC_MESSAGES ln -s ../../../../Cellar/neovim/0.1.4/share/locale/pl.UTF-8/LC_MESSAGES/nvim.mo nvim.mo mkdir /usr/local/share/locale/pl.cp1250 mkdir /usr/local/share/locale/pl.cp1250/LC_MESSAGES ln -s ../../../../Cellar/neovim/0.1.4/share/locale/pl.cp1250/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../../Cellar/neovim/0.1.4/share/locale/pt_BR/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../../Cellar/neovim/0.1.4/share/locale/ru/LC_MESSAGES/nvim.mo nvim.mo mkdir /usr/local/share/locale/ru.cp1251 mkdir /usr/local/share/locale/ru.cp1251/LC_MESSAGES ln -s ../../../../Cellar/neovim/0.1.4/share/locale/ru.cp1251/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../../Cellar/neovim/0.1.4/share/locale/sk/LC_MESSAGES/nvim.mo nvim.mo mkdir /usr/local/share/locale/sk.cp1250 mkdir /usr/local/share/locale/sk.cp1250/LC_MESSAGES ln -s ../../../../Cellar/neovim/0.1.4/share/locale/sk.cp1250/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../../Cellar/neovim/0.1.4/share/locale/sv/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../../Cellar/neovim/0.1.4/share/locale/uk/LC_MESSAGES/nvim.mo nvim.mo mkdir /usr/local/share/locale/uk.cp1251 mkdir /usr/local/share/locale/uk.cp1251/LC_MESSAGES ln -s ../../../../Cellar/neovim/0.1.4/share/locale/uk.cp1251/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../../Cellar/neovim/0.1.4/share/locale/vi/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../../Cellar/neovim/0.1.4/share/locale/zh_CN/LC_MESSAGES/nvim.mo nvim.mo mkdir /usr/local/share/locale/zh_CN.UTF-8 mkdir /usr/local/share/locale/zh_CN.UTF-8/LC_MESSAGES ln -s ../../../../Cellar/neovim/0.1.4/share/locale/zh_CN.UTF-8/LC_MESSAGES/nvim.mo nvim.mo mkdir /usr/local/share/locale/zh_CN.cp936 mkdir /usr/local/share/locale/zh_CN.cp936/LC_MESSAGES ln -s ../../../../Cellar/neovim/0.1.4/share/locale/zh_CN.cp936/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../../Cellar/neovim/0.1.4/share/locale/zh_TW/LC_MESSAGES/nvim.mo nvim.mo mkdir /usr/local/share/locale/zh_TW.UTF-8 mkdir /usr/local/share/locale/zh_TW.UTF-8/LC_MESSAGES ln -s ../../../../Cellar/neovim/0.1.4/share/locale/zh_TW.UTF-8/LC_MESSAGES/nvim.mo nvim.mo ln -s ../../../Cellar/neovim/0.1.4/share/man/man1/nvim.1 nvim.1 ln -s ../Cellar/neovim/0.1.4/share/nvim nvim 40 symlinks created Saving this output to ~/foo and counting link creations and removal with grep confirms they're both indeed 40, and that both directory creations and removals are here 26. This fits the explanation that "66 symlinks removed" counted both symlinks and directories. $ grep '^ln ' ~/foo|wc -l 40 $ grep '^rm ' ~/foo|wc -l 40 $ grep '^rmdir ' ~/foo|wc -l 26 $ grep '^mkdir ' ~/foo|wc -l 26 Nice work here. A suggestion but otherwise 👍. Thanks for jumping on this! LGTM. Will wait for any other maintainer thoughts and otherwise 🚢 LGTM. but FYI, this won't fix all the synchronized number problem. Because depending on your other files in the prefix, how the symlink will be created will be varied.(See Keg#resolve_any_conflicts for more detail. Because depending on your other files in the prefix, how the symlink will be created will be varied.(See Keg#resolve_any_conflicts for more detail. Ah, thanks for the pointer! I think I see what you mean, but it seems that cannot keep happening if I keep repeating link/unlink as in #239. Just to make sure I get it: For each conflict, brew link foo will replace a symlink to a directory pertaining to another formula to symlinks to individual files; brew unlink foo will not remove symlinks for other formulas, and it will not "merge" the symlinks to a directory back into one to a folder, so we arrive at a slightly different situation (which is fine). However, it appears that repeating immediately brew link foo/brew unlink foo will not find again the same conflicts, so there the counts should match. And that's good: otherwise, just repeated linking and unlinking (certainly an odd operation) would "leak" inodes. That's why I opened #239. Anyway: I'm satisfied with the result after this PR. Thanks for your contribution to Homebrew! You rock!