added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
created
timestamp[us]date
2001-10-09 16:19:16
2025-01-01 03:51:31
id
stringlengths
4
10
metadata
dict
source
stringclasses
2 values
text
stringlengths
0
1.61M
2025-04-01T04:34:39.728860
2024-08-27T05:12:05
2488293206
{ "authors": [ "Dev-Khant", "PranavPuranik" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8318", "repo": "mem0ai/mem0", "url": "https://github.com/mem0ai/mem0/pull/1756" }
gharchive/pull-request
get_all returns dictionary Description Small fix on readme - m.get_all() returns dict. Fixes #1755 Type of change Please delete options that are not relevant. [ ] Bug fix (non-breaking change which fixes an issue) [ ] New feature (non-breaking change which adds functionality) [ ] Breaking change (fix or feature that would cause existing functionality to not work as expected) [ ] Refactor (does not change functionality, e.g. code style improvements, linting) [x] Documentation update How Has This Been Tested? Please describe the tests that you ran to verify your changes. Provide instructions so we can reproduce. Please also list any relevant details for your test configuration Please delete options that are not relevant. [ ] Unit Test [x] Test Script (used example) Checklist: [ ] My code follows the style guidelines of this project [ ] I have performed a self-review of my own code [ ] I have commented my code, particularly in hard-to-understand areas [ ] I have made corresponding changes to the documentation [ ] My changes generate no new warnings [ ] I have added tests that prove my fix is effective or that my feature works [ ] New and existing unit tests pass locally with my changes [ ] Any dependent changes have been merged and published in downstream modules [ ] I have checked my code and corrected any misspellings Maintainer Checklist [ ] closes #xxxx (Replace xxxx with the GitHub issue number) [ ] Made sure Checks passed @PranavPuranik Thanks for fixing this!
2025-04-01T04:34:39.740425
2024-06-14T09:37:43
2352966124
{ "authors": [ "hal-eisen-MG", "kgolubic" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8319", "repo": "memgraph/documentation", "url": "https://github.com/memgraph/documentation/pull/842" }
gharchive/pull-request
[main < typo-fix-20240614] Fix typos Description Fixed typos across the repo Pull request type Please check what kind of PR this is: [ ] Fix or improvement of an existing page Related PRs and issues PR this doc page is related to: (especially necessary if the PR is related to a release) Closes: (paste the link to the issue it closes) Checklist: [ ] Check all content with Grammarly [ ] Perform a self-review of my code [ ] Make corresponding changes to the rest of the documentation (consult with the DX team) [ ] The build passes locally [ ] My changes generate no new warnings or errors [ ] Add a corresponding label [ ] If release-related, add a product and version label [ ] If release-related, add release note on product PR @hal-eisen-MG I've applied all of the suggestions. Merging to main. @hal-eisen-MG I've applied all of the suggestions. Hm. I'm confused. I'm not seeing any subsequent changes. Should we jump on a quick call? @hal-eisen-MG I've applied all of the suggestions. Hm. I'm confused. I'm not seeing any subsequent changes. Should we jump on a quick call? Sorry. I forgot to push the changes from the local branch. The changes should be on the GitHub branch as well.
2025-04-01T04:34:39.782793
2019-08-02T11:49:21
476140843
{ "authors": [ "Adam-Dupaski", "ConnorLand" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8323", "repo": "mendix/docs", "url": "https://github.com/mendix/docs/pull/1851" }
gharchive/pull-request
Add Leroy's microflow RN I just wanted you to have eyes on this to make sure it was going in the right place. Thanks! @ConnorLand Not sure why you didn't merge this Friday? Please familiarize with my 3 changes upon return. @ConnorLand let's discuss this PR upon your return
2025-04-01T04:34:39.783801
2022-07-19T11:16:19
1309366050
{ "authors": [ "MarkvanMents" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8324", "repo": "mendix/docs", "url": "https://github.com/mendix/docs/pull/4831" }
gharchive/pull-request
Update offline installation instructions This addresses issues #4788 and #4803 which identify issues with the offline installation instructions with later versions of Mendix Studio Pro 9 Looks good from R&D PoV. Will merge now and pick up later if there are any further issues.
2025-04-01T04:34:39.786104
2024-07-30T10:04:48
2437435787
{ "authors": [ "MariaShaposhnikova", "atennapel" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8325", "repo": "mendix/docs", "url": "https://github.com/mendix/docs/pull/8107" }
gharchive/pull-request
WIP: [RUN-4179] Add information about the deletion of anonymous users I only added section 3.4 I have no idea why github indicates all the other changes. Might be a newline issue, but I do have autocrlf set to true in git. @atennapel , thank you for your PR! Does it have a specific version it should be merged for? No, this information is relevant for Mx 10 and 9. But I will still update this PR with more information about sessions. @MariaShaposhnikova New PR (without all the unnecessary changes): https://github.com/mendix/docs/pull/8127
2025-04-01T04:34:39.791173
2024-11-05T08:17:35
2634744061
{ "authors": [ "leonardomendix", "uicontent" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8326", "repo": "mendix/web-widgets", "url": "https://github.com/mendix/web-widgets/pull/1318" }
gharchive/pull-request
fix(deps): update dependency mendix to ^10.16.49747 This PR contains the following updates: Package Type Update Change mendix dependencies minor ^10.15.46408 -> ^10.16.49747 Configuration 📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied. ♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 🔕 Ignore: Close this PR and you won't be reminded about this update again. This PR has been generated by Renovate Bot. Let's merge #1320 first Merged!
2025-04-01T04:34:39.796698
2018-09-13T05:26:18
359747309
{ "authors": [ "Andries-Smit", "ChengMaoDSM" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8327", "repo": "mendixlabs/charts", "url": "https://github.com/mendixlabs/charts/issues/58" }
gharchive/issue
add a direct label to the bar chart. I wanted to have something like https://plot.ly/javascript/bar-charts/#bar-chart-with-direct-labels. I do not see any field that can be used for this number. Dear ChengMaoDSM, At the moment there is no way to move the label inside the bar. The addition json settings could not provided this. It is a feature that need to be build into the chart widget, like we did with the heatmap. I wil discuss this request with the product owner and will get back to you. With kind regards, Andries If the text is show stopper you could consider building the chart from scratch with the Any chart module. https://appstore.home.mendix.com/link/app/106517/ https://docs.mendix.com/refguide/charts-any-configuration Sample: https://docs.mendix.com/refguide/charts-any-cheat-sheet#2-6-column-chart Or use a building block from the module. Cheers Will not be implemented in the standard chart widgets, could be solved with any chart Will not be implemented in the standard chart widgets, could be solved with any chart
2025-04-01T04:34:39.805369
2019-05-16T10:46:23
444884928
{ "authors": [ "JelleDekker", "codecov-io" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8328", "repo": "mendixlabs/google-maps", "url": "https://github.com/mendixlabs/google-maps/pull/35" }
gharchive/pull-request
Marker labels Added support for non persistent objects and marker labels. This allows markers on the map to be easily identified if the map is displayed next to a ListView for example. Codecov Report Merging #35 into master will decrease coverage by 0.39%. The diff coverage is 100%. @@ Coverage Diff @@ ## master #35 +/- ## ========================================= - Coverage 83.43% 83.04% -0.4% ========================================= Files 3 3 Lines 169 171 +2 Branches 44 45 +1 ========================================= + Hits 141 142 +1 - Misses 28 29 +1 Impacted Files Coverage Δ src/components/Map.ts 81.69% <ø> (ø) :arrow_up: src/components/Marker.ts 90% <100%> (-10%) :arrow_down: Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 0550e59...6d76cc8. Read the comment docs. Example of the added functionality in use:
2025-04-01T04:34:39.808794
2022-03-19T00:08:58
1174119314
{ "authors": [ "hungpham3112", "mengelbrecht" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8329", "repo": "mengelbrecht/lightline-bufferline", "url": "https://github.com/mengelbrecht/lightline-bufferline/issues/103" }
gharchive/issue
How to hide a buffer in bufferline filetype: terminal When opening terminal, a buffer will show the name of terminal to. How to disable this only filetype: Terminal? Thanks in advanced. I will have a look at it soon. I added support for a custom buffer filter function in branch feature/custom-filter-function, which I plan to merge after some testing. You can use this to hide all neovim terminal buffers: function LightlineBufferlineFilter(buffer) return getbufvar(a:buffer, '&buftype') !=# 'terminal' endfunction let g:lightline#bufferline#buffer_filter = "LightlineBufferlineFilter" I'm using vim-plug. I added this line but still can't use your snippet in this branch. Plug 'https://github.com/mengelbrecht/lightline-bufferline.git', {'branch': 'feature/custom-filter-function'} I merged the feature branch. The code should now work without switching branches. I merged the feature branch. The code should now work without switching branches. Please reopen this issue if it does not work. Thanks, it worked very well.
2025-04-01T04:34:39.810718
2020-11-26T02:03:01
751224408
{ "authors": [ "adrian5", "mengelbrecht" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8330", "repo": "mengelbrecht/lightline-bufferline", "url": "https://github.com/mengelbrecht/lightline-bufferline/issues/76" }
gharchive/issue
Spacing / margins? Maybe I'm not seeing the obvious, but is there a straight-forward way to make tabs wider? I.e.: | foo.vim | context.js | style.css | @adrian5 you can try the feature/margins branch (PR #78) and set the following properties: let g:lightline#bufferline#margin_left = 4 let g:lightline#bufferline#margin_right = 4 Thanks, that does it! @adrian5 I merged the branch
2025-04-01T04:34:39.831783
2020-07-02T07:45:58
649713386
{ "authors": [ "codecov-commenter", "whatasoda" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8331", "repo": "mercari/proto-to-type", "url": "https://github.com/mercari/proto-to-type/pull/1" }
gharchive/pull-request
chore: prepare release introduce semantic-release update REAMDE Codecov Report :exclamation: No coverage uploaded for pull request base (master@6ca6c3e). Click here to learn what that means. The diff coverage is n/a. @@ Coverage Diff @@ ## master #1 +/- ## ========================================= Coverage ? 93.28% ========================================= Files ? 7 Lines ? 253 Branches ? 65 ========================================= Hits ? 236 Misses ? 17 Partials ? 0 Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 6ca6c3e...3cc0387. Read the comment docs.
2025-04-01T04:34:39.834532
2022-07-01T16:44:10
1291619576
{ "authors": [ "Jeeppler", "de-jcup" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8332", "repo": "mercedes-benz/sechub", "url": "https://github.com/mercedes-benz/sechub/issues/1448" }
gharchive/issue
Upgrade dependencies to Spring Boot 2.7.1 Upgrade to Spring Boot 2.7.1: https://github.com/spring-projects/spring-boot/releases/tag/v2.7.1 We did upgrade to Spring Boot 2.7.1, but we did not upgrade the dependencies. See: #915 We should just ignore/ close this one and do directly #1484 @de-jcup I already did this and it was reviewed by @winzj. Just forgot to close it.
2025-04-01T04:34:39.851556
2023-07-20T11:15:08
1813745776
{ "authors": [ "luke88jones", "simoneb" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8333", "repo": "mercurius-js/mercurius-gateway", "url": "https://github.com/mercurius-js/mercurius-gateway/pull/81" }
gharchive/pull-request
fix: apply updated schema on refresh fixes #80 update the gateway refresh function to call replaceSchema on the GQL instance @luke88jones I'm seeing a consistent failure on Node 18 due to a timeout. Because this hasn't happened in the past and because the new behavior you introduced may have interactions with the event loop, I'm tempted to think that this is a legitimate failure, although I'm not sure why it would happen just on Node 18 @simoneb Sorry for the silence on this I've been on paternity leave. It looks like one of the polling interval tests is timing out. I'll see what I can find. @simoneb I've just checked out the main branch locally I have the same failure. I'm trying to track down the issue but it looks to be something to do with closing the userService in this test in the tests/pollingInterval.js file "Polling schemas (if service is down, schema shouldn't be changed)"
2025-04-01T04:34:39.858748
2018-09-04T17:30:13
356906653
{ "authors": [ "axoplasm", "sanjuroj" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8334", "repo": "mercycorps/TolaActivity", "url": "https://github.com/mercycorps/TolaActivity/issues/583" }
gharchive/issue
"Date collected" field in Add Results form auto-populates The "Date collected" field in the Add Results modal form auto-populates with the current date. This triggers a check to see if the date is valid given the period that is selected, which it probably won't be, which results in an error message on the "Measure against target" field. Need to eliminate the auto-population of the "Date collected" field. It should be blank until the user has entered a date. This will have the side-effect of eliminating the unwanted error message on the "Measure against target" field. So what's happening here: Datepickers need a date by default so they know what month to show when they are summoned. If they don't get a date, they default to "today." We initialize this datepicker invisibly on page load because it's secretly an inline element (hidden on page load) adjacent to the real input field So when it gets initialized, it gets a default date and automagically writes that to the real date field This was creating a secondary bug no one noticed where the datepicker was overriding the date field even if that field already had a value The "Measure against target" field was running a validation function against the value of the "Date Collected" field & returning an error if that field was not empty. Which it never was thanks to the magic described above So my fix(es) are like this. When the "add results" modal appears: Grab the value of the Date Collected field (which might be null for a new result) Initialize the invisible datepicker as above, setting the default date to the value in 1. (As a side effect of the JQuery datepicker, if the value is '' (null) it will use today) If the value in 1 was empty (i.e. an empty datefield in a new results form), erase the default field Run the validation on Measure Against Target And as an extra fun bonus, the datepicker isn't positioned absolutely for some highly irritating reason! Will
2025-04-01T04:34:39.899598
2021-11-12T13:18:14
1051959226
{ "authors": [ "joncodo" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8335", "repo": "merico-dev/lake", "url": "https://github.com/merico-dev/lake/issues/733" }
gharchive/issue
Lead_time_minutes in the domain layer Description Github lead time should be in minutes. Rename column in domain layer to lead_time_minutes duplicate of #722
2025-04-01T04:34:39.902992
2024-12-04T06:09:05
2716653639
{ "authors": [ "NKarolak", "merill" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8336", "repo": "merill/bluesky", "url": "https://github.com/merill/bluesky/issues/1118" }
gharchive/issue
[Bug] Error: User has not verified their email Bug details I've just registered, verified my email by clicking on the mail link, connected my Bsky profile, entered what I post about... All fine. But when I try to verify that I'm an MVP (https://mvp.microsoft.com/en-us/MVP/profile/0789d8ed-d3d1-480b-9fee-5bc01628baa2), it keeps saying "Error: User has not verified their email". Could it be related to that my email address countains a "+"? Solved by visiting the homepage again, then Register, then Sign in. Although I was signed in all the time to enter my data 🤷🏻‍♀️ Thanks @NKarolak There was a related issue that others saw too and I was able to get to the bottom of it and fix it.
2025-04-01T04:34:39.905285
2024-12-12T09:18:12
2735330917
{ "authors": [ "manekinekko", "merill" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8337", "repo": "merill/bluesky", "url": "https://github.com/merill/bluesky/issues/1127" }
gharchive/issue
Missing JavaScript-related topics Bug details Hi, fellow FTE here 👋 I appreciate the time and effort you are putting into building this platform! Thank you ❤️ While selecting the topics I write about, I could not find those related to my work. May I request adding: Developer - JavaScript Developer - Node.js Developer - WebAssembly Done! Thanks for requesting @manekinekko
2025-04-01T04:34:39.931171
2020-09-23T04:56:44
707047425
{ "authors": [ "HJesusH", "merqlove" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8338", "repo": "merqlove/do_snapshot", "url": "https://github.com/merqlove/do_snapshot/issues/23" }
gharchive/issue
The certificate's owner does not match hostname The certificate's owner does not match hostname 'assets.merqlove.ru.s3.amazonaws.com' This occurs when i use the next command in a linux/debian server wget https://assets.merqlove.ru.s3.amazonaws.com/do_snapshot/do_snapshot.tgz Fixed
2025-04-01T04:34:39.940111
2019-12-02T17:18:35
531314408
{ "authors": [ "girishranganathan", "iambami" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8339", "repo": "meshery/meshery-nsm", "url": "https://github.com/meshery/meshery-nsm/issues/45" }
gharchive/issue
Support installing NSM and its sample apps using helm Current Description The adapter is using helm to install NSM and its sample apps. But we are using a cloned version of the helm repo to use it. Expected Behavior We want to be able to use the official helm repos to be able to install NSM and its sample apps. This will prevent us from using edge version all the time and also provide us the ability to install specific versions of NSM and sample apps. Hi, @girishranganathan is it okay to close this issue? or does it need to be assigned to someone?
2025-04-01T04:34:39.947854
2022-08-17T20:49:40
1342279047
{ "authors": [ "leecalcote" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8340", "repo": "meshery/meshery", "url": "https://github.com/meshery/meshery/issues/6054" }
gharchive/issue
[mesheryctl] app onboard receives 400 response code See mesheryctl Command Tracker for current status of commands. Current Behavior mesheryctl app onboard -s "Kubernetes Manifest" -f https://raw.githubusercontent.com/kubernetes/website/main/content/en/examples/application/job/cronjob.yaml Error: Response Status Code 400, possible Server Error Server Response: map[string]interface {}{"apiVersion":"batch/v1", "kind":"CronJob", "metadata":map[string]interface {}{"name":"hello"}, "spec":map[string]interface {}{"jobTemplate":map[string]interface {}{"spec":map[string]interface {}{"template":map[string]interface {}{"spec":map[string]interface {}{"containers":[]interface {}{map[string]interface {}{"command":[]interface {}{"/bin/sh", "-c", "date; echo Hello from the Kubernetes cluster"}, "image":"busybox:1.28", "imagePullPolicy":"IfNotPresent", "name":"hello"}}, "restartPolicy":"OnFailure"}}}}, "schedule":"* * * * *"}} DEBU[1954] Application size: 1507 INFO[1954] attempting to save application to remote provider INFO[1954] application successfully sent to remote provider: [{"id":"5e6cec58-8db0-418e-a145-2153c6885be9","name":"cronjob.yaml","user_id":"48368c3b-5134-477f-95bb-4a1ba56dd908","application_file":"name: Autogenerated\nservices:\n faf7c427-102f-4042-bf85-f3c622bc52dc:\n name: hello\n type: CronJob.K8s\n namespace: default\n settings:\n spec:\n jobTemplate:\n spec:\n template:\n spec:\n containers:\n - command:\n - /bin/sh\n - -c\n - date; echo Hello from the Kubernetes cluster\n image: busybox:1.28\n imagePullPolicy: IfNotPresent\n name: hello\n restartPolicy: OnFailure\n schedule: '* * * * *'\n","location":{"branch":"","host":"https://raw.githubusercontent.com/kubernetes/website/main/content/en/examples/application/job/cronjob.yaml","path":"","type":"http"},"type":{"String":"Kubernetes Manifest","Valid":true},"SourceContent":null,"created_at":"2022-08-17T20:43:50.483192Z","updated_at":"2022-08-17T20:43:50.483215Z"}] DEBU[1954] Application Content size 418 INFO[1954] application source successfully uploaded to remote provider INFO[1954] attempting to fetch applications from cloud DEBU[1954] constructed applications url: https://meshery.layer5.io/applications?page=0&page_size=10 map[string]interface {}{"name":"Autogenerated", "services":map[string]interface {}{"faf7c427-102f-4042-bf85-f3c622bc52dc":map[string]interface {}{"name":"hello", "namespace":"default", "settings":map[string]interface {}{"spec":map[string]interface {}{"jobTemplate":map[string]interface {}{"spec":map[string]interface {}{"template":map[string]interface {}{"spec":map[string]interface {}{"containers":[]interface {}{map[string]interface {}{"command":[]interface {}{"/bin/sh", "-c", "date; echo Hello from the Kubernetes cluster"}, "image":"busybox:1.28", "imagePullPolicy":"IfNotPresent", "name":"hello"}}, "restartPolicy":"OnFailure"}}}}, "schedule":"* * * * *"}}, "type":"CronJob.K8s"}}} INFO[1954] applications successfully retrieved from remote provider Screenshots/Logs Environment Host OS: Mac Meshery Server Version: stable-v0.6.0-rc-6fd Meshery Client Version: stable-v0.6.0-rc-6fd Platform: Kubernetes Contributor Guides and Handbook 📚 mesheryctl Contributing Guide 📺 Beginner's guide to contributing to Meshery and mesheryctl ⌨️ mesheryctl Command Tracker ⌨️ Meshery CLI Commands and Documentation 🙋🏾🙋🏼 Questions: Discussion Forum and Community Slack @Revolyssup is this another issue caused by source type? @Revolyssup @alphaX86 ?
2025-04-01T04:34:39.966257
2022-05-28T18:53:47
1251701929
{ "authors": [ "Revolyssup", "codecov-commenter", "leecalcote", "sudo-NithishKarthik" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8341", "repo": "meshery/meshery", "url": "https://github.com/meshery/meshery/pull/5549" }
gharchive/pull-request
Replace kubeopenapi-jsonschema based component generation with CUE Description It makes the process at least 3 times faster(I actually measured this by taking multiple measurements 😝 ) Notes for Reviewers For more information, please refer meshkit pr Signed commits [x] Yes, I signed my commits. @sudo-NithishKarthik Wouldn't we first want the meshkit PR to be merged with a release? Yes MeshKit PR ready for merge and release? Codecov Report Merging #5549 (6ad5552) into master (2a22fe8) will decrease coverage by 4.34%. The diff coverage is n/a. @@ Coverage Diff @@ ## master #5549 +/- ## ========================================== - Coverage 17.47% 13.12% -4.35% ========================================== Files 107 107 Lines 9993 9993 ========================================== - Hits 1746 1312 -434 - Misses 7821 8401 +580 + Partials 426 280 -146 Flag Coverage Δ gointegrationtests ? unittests 13.12% <ø> (ø) Flags with carried forward coverage won't be shown. Click here to find out more. Impacted Files Coverage Δ mesheryctl/internal/cli/root/system/update.go 1.05% <0.00%> (-58.95%) :arrow_down: mesheryctl/internal/cli/root/system/start.go 1.88% <0.00%> (-55.19%) :arrow_down: mesheryctl/internal/cli/root/system/status.go 1.14% <0.00%> (-55.18%) :arrow_down: mesheryctl/internal/cli/root/system/restart.go 1.92% <0.00%> (-48.08%) :arrow_down: mesheryctl/internal/cli/root/system/stop.go 2.00% <0.00%> (-46.67%) :arrow_down: mesheryctl/internal/cli/root/system/logs.go 3.26% <0.00%> (-36.96%) :arrow_down: mesheryctl/internal/cli/root/system/check.go 1.03% <0.00%> (-23.03%) :arrow_down: mesheryctl/internal/cli/root/system/dashboard.go 1.70% <0.00%> (-9.41%) :arrow_down: mesheryctl/pkg/utils/healthcheck.go 8.82% <0.00%> (-4.42%) :arrow_down: Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 1bf2635...6ad5552. Read the comment docs.
2025-04-01T04:34:39.970893
2023-02-21T04:52:00
1592786350
{ "authors": [ "Chadha93", "VIKAS0264", "l5io" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8342", "repo": "meshery/meshery", "url": "https://github.com/meshery/meshery/pull/7134" }
gharchive/pull-request
fixed jquery error Notes for Reviewers This PR fixes #7133 jquery throws unrecognised expression error because of incorrect jquery selector it is fixed in this PR. Signed commits [x] Yes, I signed my commits. 🚀 Preview for commit 67da861da156ee219ee9b5ada3485bf5d1c446d6 at: https://63f44ede9e215904edaeca88--meshery-docs-preview.netlify.app 🚀 Preview for commit 44a87313a7797fdeed6eaa19c1cce1fb69c1a52a at: https://63f45d7037dfa1112757f77f--meshery-docs-preview.netlify.app @VIKAS0264 Let's discuss this on the Meshery Dev call. Please add this as an agenda item in the meeting minutes if you would. :)
2025-04-01T04:34:39.976093
2022-11-17T20:51:09
1453975621
{ "authors": [ "Chadha93", "HM-8", "leecalcote" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8343", "repo": "meshery/play", "url": "https://github.com/meshery/play/issues/30" }
gharchive/issue
[Docs] Add project setup guidelines with commands in contributing.md Current State The contributing.md is missing project setup command instructions. Desired State Add project setup guidelines with commands in contributing.md Refer to the Makefile to get setup commands. Contributor Guides and Resources 📚 Instructions for contributing to documentation Meshery documentation site and source 🛠 Meshery Build & Release Strategy 🎨 Wireframes and designs for Meshery UI in Figma 🙋🏾🙋🏼 Questions: Layer5 Discussion Forum and Layer5 Community Slack Hi @Chadha93. Interested in working on this issue. Would love to get some guidance. @HM-8, after you familiarize with the make targets to build this site, please add a few instructions on how to build the site for other contributors. Here is a reference - https://github.com/layer5io/layer5/blob/master/CONTRIBUTING.md Okay @leecalcote. I'll get on it. Thanks for the reference.
2025-04-01T04:34:40.108343
2022-09-19T17:12:04
1378271090
{ "authors": [ "rugeli" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8344", "repo": "mesoscope/cellpack", "url": "https://github.com/mesoscope/cellpack/pull/66" }
gharchive/pull-request
Feature/reorganize ingredient data Problem the 4th task in Migrate old ingredients into new "objects dict" #57 Solution Created a partners object to convert ingredients that should be moved into this object. Wrote tests for _convert_to_objects Type of change Please delete options that are not relevant. New feature (non-breaking change which adds functionality) merged to staging/ and deleted feature/ branch, grateful for helps!
2025-04-01T04:34:40.118586
2017-10-25T19:02:26
268512049
{ "authors": [ "elezar", "nickbp" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8346", "repo": "mesosphere/dcos-commons", "url": "https://github.com/mesosphere/dcos-commons/pull/1771" }
gharchive/pull-request
INFINITY-2584/2596: Rename stop to pause, and followup fixes from PR 1763 Docs fixes and whatever else comes up following PR #1763 TODO: need to also rename start to resume TODO: need to also rename start to resume Is still think || and |> are the way to go. Or even: ⏯
2025-04-01T04:34:40.126089
2017-12-01T21:32:44
278602863
{ "authors": [ "mpereira" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8347", "repo": "mesosphere/dcos-commons", "url": "https://github.com/mesosphere/dcos-commons/pull/1903" }
gharchive/pull-request
Make default placement constraint values be valid stringified JSON. Values should be a JSON-stringified array of constraints (which are expressed as arrays of strings). Example: "hostname:MAX_PER:1" becomes "[[\"hostname\", \"MAX_PER\", \"1\"]]" Empty strings become JSON-stringified empty arrays, "" -> "[]". Chat with Andrew: Murilo Pereira [12:14 PM] for values that used to default to `"hostname:MAX_PER:1"`, does it matter whether the JSON equivalent has the `1` as a string or number? e.g. `"[[\"hostname\", \"MAX_PER\", \"1\"]]"` for string Andrew Hoskins [12:21 PM] Hey! Make it a string Murilo Pereira [12:21 PM] cool. and empty defaults just `""` right? Andrew Hoskins [12:26 PM] Empty state should actually be `"[]"` [12:26] Because the meme type is <…>+json, so it expects a valid JSON value One of the things I was struggling with (after finding out about the need for the propagating mustache triple bracket on friday) was the need for double-escaping quotes in config.json files, e.g. https://github.com/mesosphere/dcos-commons/pull/1903/files#diff-56550f345e72e9ddaffbc774bb87a7a9R124 Another was missing test config files with hard-coded constraints like https://github.com/mesosphere/dcos-commons/pull/1903/files#diff-9c0a8da33f4e191b89245b7045401adfL4 The other things was that making [] be the default value instead caused the constraint parser to treat it as a empty constraint instead of an empty list of constraints: testcase com.mesosphere.sdk.elastic.scheduler.ServiceTest > testSpec: java.io.IOException: Invalid number of entries in rule. Expected 2 or 3, got 0: [] at com.mesosphere.sdk.offer.evaluate.placement.MarathonConstraintParser.parseRow(MarathonConstraintParser.java:106) which is addressed by https://github.com/mesosphere/dcos-commons/pull/1903/commits/b8662cfffe13ca8cf585f7b86a9cb8675bf2c017 I just noticed https://github.com/mesosphere/dcos-commons/blob/INFINITY-2620-placement-constraint-defaults/sdk/scheduler/src/test/java/com/mesosphere/sdk/offer/evaluate/placement/MarathonConstraintParserTest.java#L296-L299 so it's possible we don't need to double-escape double quotes in config.json files and use single quotes instead, I'll push a commit trying that after the checks for the current code pass. To recap: config.json files need stringified JSON to double-escape double quotes "placement_constraint": { "type": "string", "description": "The Marathon-style placement constraint string controlling node placement", "default": "[[\\\"hostname\\\", \\\"UNIQUE\\\"]]", "media": { "type": "application/x-zone-constraints+json" } } marathon.json.mustache files have triple-brace mustache templates "PLACEMENT_CONSTRAINT": "{{{nodes.placement_constraint}}}", expands into: "PLACEMENT_CONSTRAINT": "[[\"hostname\", \"UNIQUE\"]]", svc.yml files need the single quote around the triple braces placement: '{{{PLACEMENT_CONSTRAINT}}}' expands into: placement: '[["hostname", "UNIQUE"]]' Maybe we could make it possible to just use a regular array value in config.json in the future, but for now this should work. @benclarkwood https://jira.mesosphere.com/browse/INFINITY-2769
2025-04-01T04:34:40.130129
2015-02-11T05:24:38
57277978
{ "authors": [ "ConnorDoyle", "drexin", "sttts" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8348", "repo": "mesosphere/marathon", "url": "https://github.com/mesosphere/marathon/issues/1178" }
gharchive/issue
Potential deadlock in MarathonSchedulerActor In MarathonSchedulerActor.performAsyncWithLockFor calls Semaphore.acquire when in blocking mode. The acquire method blocks the entire thread. It appears that depending on how concurrent acquisition is interleaved, deadlock could occur. Yes, I also came across this, see: https://github.com/mesosphere/marathon/pull/988 I seem to have forgotten to create a follow up ticket. Thanks Connor! What about this patch? If Set[PathId] is the way to go, this better shouldn't be merged last-minute.
2025-04-01T04:34:40.164181
2015-08-26T15:21:24
103295337
{ "authors": [ "aameek", "aquamatthias", "gkleiman", "jasongilanfarr", "jgarcia-mesosphere", "kopax", "mcclurmc", "meichstedt", "mingqi", "nashasha1" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8349", "repo": "mesosphere/marathon", "url": "https://github.com/mesosphere/marathon/issues/2101" }
gharchive/issue
Could not modify Group with key In 0.8.2 We create an app, marathon return success. But it never really create it. And it not in deployment nor in queue. So we upgrade to 0.9.2 It shows error: Could not modify Group with key: root:2015-07-31T08:50:18.342Z We remove some zookeeper log/snapshot. It is ok. But after we recover some apps, the error appear again. Could you please provide some more information from the marathon logfile? There should be an exception logged with that message. Aug 26 09:21:17 ip----* marathon[27448]: [2015-08-26 09:21:17,252] WARN (mesosphere.marathon.api.MarathonExceptionMapper:30) Aug 26 09:21:17 ip----* marathon[27448]: mesosphere.marathon.StoreCommandFailedException: Could not fetch AppDefinition with key: 5ecb8fba-afdf-4691-9f2f-f0111106b4ed Aug 26 09:21:17 ip----* marathon[27448]: #011at mesosphere.marathon.state.MarathonStore$$anonfun$mesosphere$marathon$state$MarathonStore$$exceptionTransform$1.applyOrElse(MarathonStore.scala:75) Aug 26 09:21:17 ip----* marathon[27448]: #011at mesosphere.marathon.state.MarathonStore$$anonfun$mesosphere$marathon$state$MarathonStore$$exceptionTransform$1.applyOrElse(MarathonStore.scala:74) Aug 26 09:21:17 ip----* marathon[27448]: #011at scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:36) Aug 26 09:21:17 ip----* marathon[27448]: #011at scala.util.Failure$$anonfun$recover$1.apply(Try.scala:215) Aug 26 09:21:17 ip----* marathon[27448]: #011at scala.util.Try$.apply(Try.scala:191) Aug 26 09:21:17 ip----* marathon[27448]: #011at scala.util.Failure.recover(Try.scala:215) Aug 26 09:21:17 ip----* marathon[27448]: #011at scala.concurrent.Future$$anonfun$recover$1.apply(Future.scala:324) Aug 26 09:21:17 ip----* marathon[27448]: #011at scala.concurrent.Future$$anonfun$recover$1.apply(Future.scala:324) Aug 26 09:21:17 ip----* marathon[27448]: #011at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32) Aug 26 09:21:17 ip----* marathon[27448]: #011at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) Aug 26 09:21:17 ip----* marathon[27448]: #011at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) Aug 26 09:21:17 ip----* marathon[27448]: #011at java.lang.Thread.run(Thread.java:745) Aug 26 09:21:17 ip----* marathon[27448]: Caused by: mesosphere.marathon.StoreCommandFailedException: Could not load key app:5ecb8fba-afdf-4691-9f2f-f0111106b4ed Aug 26 09:21:17 ip----* marathon[27448]: #011at mesosphere.util.state.zk.ZKStore$$anonfun$exceptionTransform$1.applyOrElse(ZKStore.scala:83) Aug 26 09:21:17 ip----* marathon[27448]: #011at mesosphere.util.state.zk.ZKStore$$anonfun$exceptionTransform$1.applyOrElse(ZKStore.scala:82) Aug 26 09:21:17 ip----* marathon[27448]: #011... 10 more Aug 26 09:21:17 ip----* marathon[27448]: Caused by: org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = ConnectionLoss for /marathon/state/app:5ecb8fba-afdf-4691-9f2f-f0111106b4ed Aug 26 09:21:17 ip----* marathon[27448]: #011at org.apache.zookeeper.KeeperException.create(KeeperException.java:99) Aug 26 09:21:17 ip----* marathon[27448]: #011at org.apache.zookeeper.KeeperException.create(KeeperException.java:51) Aug 26 09:21:17 ip----* marathon[27448]: #011at com.twitter.zk.AsyncCallbackPromise$class.process(AsyncCallbackPromise.scala:18) Aug 26 09:21:17 ip----* marathon[27448]: #011at com.twitter.zk.DataCallbackPromise.process(AsyncCallbackPromise.scala:53) Aug 26 09:21:17 ip----* marathon[27448]: #011at com.twitter.zk.DataCallbackPromise.processResult(AsyncCallbackPromise.scala:56) Aug 26 09:21:17 ip----* marathon[27448]: #011at org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:564) Aug 26 09:21:17 ip----* marathon[27448]: #011at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:498) Aug 26 09:21:17 ip----* marathon[27448]: [2015-08-26 09:21:17,252] WARN Deployment failed for change: 2015-08-26T09:21:04.810Z (mesosphere.marathon.state.GroupManager:146) Aug 26 09:21:17 ip----* marathon[27448]: mesosphere.marathon.StoreCommandFailedException: Could not modify Group with key: root:2015-07-30T00:07:26.834Z Aug 26 09:21:17 ip----* marathon[27448]: #011at mesosphere.marathon.state.MarathonStore$$anonfun$mesosphere$marathon$state$MarathonStore$$exceptionTransform$1.applyOrElse(MarathonStore.scala:75) Aug 26 09:21:17 ip----* marathon[27448]: #011at mesosphere.marathon.state.MarathonStore$$anonfun$mesosphere$marathon$state$MarathonStore$$exceptionTransform$1.applyOrElse(MarathonStore.scala:74) Aug 26 09:21:17 ip----* marathon[27448]: #011at scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:36) Aug 26 09:21:17 ip---**-* marathon[27448]: #011at scala.util.Failure$$anonfun$recover$1.apply(Try.scala:215) we found this when the number of our containers is larger than 1000. So did marathon have some unusual behaviour when it manages 1000+ containers. I and @nashasha1 from same company. I post another logging with better format below. The exception happened when App was created or updated. The exception says Can not create entity group:root:2015-07-31T08:50:18.342Z and Could not modify Group with key: root:2015-07-31T08:50:18.342Z. App creating and updating happened on yesterday, that is 2015-08-26. It's very strange that Marathon tried to create and modify group with 2015-07-31 version. Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,210] WARN (mesosphere.marathon.api.MarathonExceptionMapper:30) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: mesosphere.marathon.StoreCommandFailedException: Could not modify Group with key: root:2015-07-31T08:50:18.342Z Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at mesosphere.marathon.state.MarathonStore$$anonfun$mesosphere$marathon$state$MarathonStore$$exceptionTransform$1.applyOrElse(MarathonStore.scala:75) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at mesosphere.marathon.state.MarathonStore$$anonfun$mesosphere$marathon$state$MarathonStore$$exceptionTransform$1.applyOrElse(MarathonStore.scala:74) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:36) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.util.Failure$$anonfun$recover$1.apply(Try.scala:215) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.util.Try$.apply(Try.scala:191) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.util.Failure.recover(Try.scala:215) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.concurrent.Future$$anonfun$recover$1.apply(Future.scala:324) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.concurrent.Future$$anonfun$recover$1.apply(Future.scala:324) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at java.lang.Thread.run(Thread.java:745) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: Caused by: mesosphere.marathon.StoreCommandFailedException: Can not create entity group:root:2015-07-31T08:50:18.342Z Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at mesosphere.util.state.zk.ZKStore$$anonfun$exceptionTransform$1.applyOrElse(ZKStore.scala:83) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at mesosphere.util.state.zk.ZKStore$$anonfun$exceptionTransform$1.applyOrElse(ZKStore.scala:82) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011... 10 more Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: Caused by: org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = ConnectionLoss for /marathon/state/group:root:2015-07-31T08:50:18.342Z Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at org.apache.zookeeper.KeeperException.create(KeeperException.java:99) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at org.apache.zookeeper.KeeperException.create(KeeperException.java:51) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at com.twitter.zk.AsyncCallbackPromise$class.process(AsyncCallbackPromise.scala:18) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at com.twitter.zk.StringCallbackPromise.process(AsyncCallbackPromise.scala:23) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at com.twitter.zk.StringCallbackPromise.processResult(AsyncCallbackPromise.scala:25) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:605) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:498) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,210] WARN Deployment failed for change: 2015-08-26T10:04:58.043Z (mesosphere.marathon.state.GroupManager:146) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: mesosphere.marathon.StoreCommandFailedException: Could not modify Group with key: root:2015-07-31T08:50:18.342Z Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at mesosphere.marathon.state.MarathonStore$$anonfun$mesosphere$marathon$state$MarathonStore$$exceptionTransform$1.applyOrElse(MarathonStore.scala:75) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at mesosphere.marathon.state.MarathonStore$$anonfun$mesosphere$marathon$state$MarathonStore$$exceptionTransform$1.applyOrElse(MarathonStore.scala:74) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:36) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.util.Failure$$anonfun$recover$1.apply(Try.scala:215) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.util.Try$.apply(Try.scala:191) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.util.Failure.recover(Try.scala:215) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.concurrent.Future$$anonfun$recover$1.apply(Future.scala:324) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.concurrent.Future$$anonfun$recover$1.apply(Future.scala:324) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at java.lang.Thread.run(Thread.java:745) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: Caused by: mesosphere.marathon.StoreCommandFailedException: Can not create entity group:root:2015-07-31T08:50:18.342Z Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at mesosphere.util.state.zk.ZKStore$$anonfun$exceptionTransform$1.applyOrElse(ZKStore.scala:83) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at mesosphere.util.state.zk.ZKStore$$anonfun$exceptionTransform$1.applyOrElse(ZKStore.scala:82) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011... 10 more Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: Caused by: org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = ConnectionLoss for /marathon/state/group:root:2015-07-31T08:50:18.342Z Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at org.apache.zookeeper.KeeperException.create(KeeperException.java:99) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at org.apache.zookeeper.KeeperException.create(KeeperException.java:51) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at com.twitter.zk.AsyncCallbackPromise$class.process(AsyncCallbackPromise.scala:18) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at com.twitter.zk.StringCallbackPromise.process(AsyncCallbackPromise.scala:23) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at com.twitter.zk.StringCallbackPromise.processResult(AsyncCallbackPromise.scala:25) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:605) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:498) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [INFO] [08/26/2015 10:05:09.214] [marathon-akka.actor.default-dispatcher-4] [akka://marathon/user/$b] POSTing to all endpoints. Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,215] INFO <IP_ADDRESS> - - [26/Aug/2015:10:04:58 +0000] "POST /v2/apps/ HTTP/1.1" 500 76 "-" "curl/7.37.1" (mesosphere.chaos.http.ChaosRequestLog:15) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,320] INFO Opening socket connection to server <IP_ADDRESS>/<IP_ADDRESS>:2181. Will not attempt to authenticate using SASL (unknown error) (org.apache.zookeeper.ClientCnxn:975) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,320] INFO Socket connection established to <IP_ADDRESS>/<IP_ADDRESS>:2181, initiating session (org.apache.zookeeper.ClientCnxn:852) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,322] INFO Session establishment complete on server <IP_ADDRESS>/<IP_ADDRESS>:2181, sessionid = 0x14f68ce878b000f, negotiated timeout = 1800000 (org.apache.zookeeper.ClientCnxn:1235) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,328] INFO Setting framework ID to 20150713-183426-1459822508-5050-27691-0000 (mesosphere.marathon.MarathonSchedulerService:86) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [INFO] [08/26/2015 10:05:09.328] [pool-2-thread-1] [akka://marathon/user/$b] Sending POST to:http://<IP_ADDRESS>:8080/v1/apps/events/ Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,328] INFO Setting framework ID to 20150713-183426-1459822508-5050-27691-0000 (mesosphere.marathon.MarathonSchedulerService:86) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [INFO] [08/26/2015 10:05:09.329] [pool-2-thread-1] [akka://marathon/user/$b] Sending POST to:https://puck.alauda.club:8443/v1/slark Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [INFO] [08/26/2015 10:05:09.329] [pool-2-thread-1] [akka://marathon/user/$b] Sending POST to:https://puck.alauda.club:8443/v1/mirana Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,330] INFO <IP_ADDRESS> - - [26/Aug/2015:10:05:08 +0000] "GET /v2/info HTTP/1.1" 200 977 "-" "ELB-HealthChecker/1.0" (mesosphere.chaos.http.ChaosRequestLog:15) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,330] INFO <IP_ADDRESS> - - [26/Aug/2015:10:05:09 +0000] "GET /v2/apps//c30495b3-bf6c-4693-a13a-a0805a78804f HTTP/1.1" 200 1625 "http://<IP_ADDRESS>:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_10_0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/44.0.2403.89 Safari/537.36" (mesosphere.chaos.http.ChaosRequestLog:15) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,331] INFO <IP_ADDRESS> - - [26/Aug/2015:10:05:08 +0000] "GET /v2/info HTTP/1.1" 200 977 "-" "ELB-HealthChecker/1.0" (mesosphere.chaos.http.ChaosRequestLog:15) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,398] INFO <IP_ADDRESS> - - [26/Aug/2015:10:05:09 +0000] "GET /v2/deployments HTTP/1.1" 200 2 "http://<IP_ADDRESS>:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_10_0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/44.0.2403.89 Safari/537.36" (mesosphere.chaos.http.ChaosRequestLog:15) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,429] INFO started processing 1 offers, launching at most 1 tasks per offer and 1000 tasks in total (mesosphere.marathon.tasks.IterativeOfferMatcher$:129) Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,429] INFO Offer [20150720-062308-1577787308-5050-21885-O36819081]. Decline with default filter refuseSeconds (use --decline_offer_duration to configure) (mesosphere.marathon.tasks.IterativeOfferMatcher$:228) Another odd thing is the group version is always 2015-07-13T20:47:41.034Z which query by /v2/groups. I look into the content of Zookeeper node /marathon/state/group:root, the version indeed is 2015-07-13T20:47:41.034. It's seem the version hadn't changes from first created. If the problem is related to big group. we have thousand Apps in root group. The size of zookeeper node /marathon/state/group:root is more than 500K now. is this problem related to big group? actually we don't use Marathon's group feature. Should we create separate group for each app to prevent large group? thanks any idea about this? In general, 500k is not a size where I'd expect this kind of problem. See information on jute.maxbuffer in the ZooKeeper Admin Doc But yes, we've run into problems with big node sizes – the problem, however, is not the size of the node itself, but the packet size when fetching the node. When trying to delete such a node via the zkCli, the output looks like this: Welcome to ZooKeeper! 2015-08-13 10:47:41,229 [myid:] - INFO [main-SendThread(localhost:2181):ClientCnxn$SendThread@975] - Opening socket connection to server localhost/<IP_ADDRESS>:2181. Will not attempt to authenticate using SASL (unknown error) JLine support is enabled 2015-08-13 10:47:41,299 [myid:] - INFO [main-SendThread(localhost:2181):ClientCnxn$SendThread@852] - Socket connection established to localhost/<IP_ADDRESS>:2181, initiating session 2015-08-13 10:47:41,308 [myid:] - INFO [main-SendThread(localhost:2181):ClientCnxn$SendThread@1235] - Session establishment complete on server localhost/<IP_ADDRESS>:2181, sessionid = 0x14f25f4aa2200df, negotiated timeout = 30000 WATCHER:: WatchedEvent state:SyncConnected type:None path:null [zk: localhost:2181(CONNECTED) 0] rmr /universe 2015-08-13 10:47:45,908 [myid:] - WARN [main-SendThread(localhost:2181):ClientCnxn$SendThread@1102] - Session 0x14f25f4aa2200df for server localhost/<IP_ADDRESS>:2181, unexpected error, closing socket connection and attempting reconnect java.io.IOException: Packet len4198383 is out of range! at org.apache.zookeeper.ClientCnxnSocket.readLength(ClientCnxnSocket.java:112) at org.apache.zookeeper.ClientCnxnSocketNIO.doIO(ClientCnxnSocketNIO.java:79) at org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:366) at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1081) WATCHER:: WatchedEvent state:Disconnected type:None path:null Exception in thread "main" org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = ConnectionLoss for /universe/marathon-demo/state at org.apache.zookeeper.KeeperException.create(KeeperException.java:99) at org.apache.zookeeper.KeeperException.create(KeeperException.java:51) at org.apache.zookeeper.ZooKeeper.getChildren(ZooKeeper.java:1472) at org.apache.zookeeper.ZooKeeper.getChildren(ZooKeeper.java:1500) at org.apache.zookeeper.ZKUtil.listSubTreeBFS(ZKUtil.java:114) at org.apache.zookeeper.ZKUtil.deleteRecursive(ZKUtil.java:49) at org.apache.zookeeper.ZooKeeperMain.processZKCmd(ZooKeeperMain.java:703) at org.apache.zookeeper.ZooKeeperMain.processCmd(ZooKeeperMain.java:588) at org.apache.zookeeper.ZooKeeperMain.executeLine(ZooKeeperMain.java:360) at org.apache.zookeeper.ZooKeeperMain.run(ZooKeeperMain.java:323) at org.apache.zookeeper.ZooKeeperMain.main(ZooKeeperMain.java:282) matthias@master-zuob ~ $ Note the Packet len4198383 is out of range! in the stack trace – I have not seen that in the one you provided, so I don't think that you're running into node size problems (at least not yet). However, it seems to be a zk-specific problem. From a marathon point of view, storing state fails, which means that deploying must fail. Questions Have you updated any AppDefinitions since the root group's version timestamp? Can you try to manually (non-destructively) modify the zNode via the zkCli? Any update/ideas on this? We are seeing a similar issue though our group size or packet size is not that big... curl --silent -X GET http:///v2/groups | python -mjson.tool | wc 39871 65801 1626350 Hey @mingqi if the group reaches a size of more than 500KB, any update to any app or group will fail with a version of Marathon before 0.13.0. For every change of an AppDefinition we create and store a Deployment in ZK which stores the group before the change and the group after the change. If your group is greater than 500KB, the deployment object will be 2 times this size. The default jute.maxbuffer is 1MB and will exceed this threshold. For that reason we introduced ZKCompression, which is available in Marathon 0.13. We successfully started 2000 Apps simultaneously with this version. Can you try, if this version solves your problem? @aameek the group is stored as protobuf. The size of the json object is very different. Can you look into zk, how big the object is? Just adding my :+1: to this issue. My team and I worked with @jgarcia-mesosphere over the last two weeks to narrow down a problem we were seeing in our dev cluster to this bug. The issue is that Marathon stores group defs in ZK, and the root group seems to store all the data for all child groups. When we got to a point where our /marathon/group:root znode went over about 500kb, Marathon wouldn't be able to process new deployments (we'd see either "conflicting deployment" or "futures timed out" errors in the logs). The max for any znode is 1mb, but Marathon is updating the znode with old group + deployment metadata + updated group, so the current znode doesn't even have to approach 1mb before you start seeing this problem. Once we understood that Marathon was trying to update the group:root znode in a way that would surpass the hard limit of 1mb per znode, we were able to fiddle with our apps enough to destroy enough of them to get us below 500kb for that group:root znode. That got our dev cluster stable, but we have an upper limit on the number of app groups we can run concurrently now. It looks like Marathon is scalable up to thousands of tasks, but those tasks need to have small definitions. We've built a "Heroku-like" system on top of Marathon for our devs, and that means that we've got a high number of apps per group, and each app has around 100 environment variables. So when your devs create multiple groups with this kind of configuration, you'll run into scalability issues sooner than you'd expect. For instance, we currently have 323 apps defined in our dev Marathon (only 120 of them actually running), but our group:root znode is over 400kb. We're going to be rearchitecting our applications to fetch their environments themselves, instead of injecting them via the env section of the app definition. It would be good if Marathon would at least catch the exceptional case where it attempts to stuff more in the group:root znode than it can handle, and long-term to break up these potentially large znodes into parent/child znodes, or use a different data store, or something even more clever that I haven't thought of yet. For anyone experiencing this issue: try deleting your smallest app or app group, then the next smallest, until you have control of your cluster again. That got us moving again. @mcclurmc thanks for the info. We are very aware of this limitation. We introduced a command line parameter --max_apps, so you get an error, if you try to create more apps than that number. This configuration is mostly bandaid to fail fast and does not take the zk node size into account (depending on how much data you store on the app level). Step 1) we will refactor how the group is stored (only references to the app) Step 2) we will support more capable persistent stores. To check for this, pull Marathon statistics from http://<marathon>/metrics and check for the "service.mesosphere.marathon.state.MarathonStore.DeploymentPlan.write-data-size" key to see the largest deploy your Marathon has tried to write. Is this fixed ? @kopax not yet Will it be fixed ? There is a new persistence layer in 1.4-SNAPSHOT that is currently enabled by default that stores in ZK in a much more scalable fashion. I was able to store 2,500 apps (easily scales more) with 450,000 tasks. The theoretical limit is about a million "objects" of a given type. Fixed by #4178
2025-04-01T04:34:40.167002
2016-07-26T15:25:24
167636148
{ "authors": [ "aquamatthias", "jdef" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8350", "repo": "mesosphere/marathon", "url": "https://github.com/mesosphere/marathon/issues/4146" }
gharchive/issue
docker containerizer now allows relative containerPath starting with mesos 1.0 this simplifies the volume API and we should update the validators (and docs) in marathon to allow non-absolute (relative) containerPath Review is here: https://phabricator.mesosphere.com/D399
2025-04-01T04:34:40.167962
2019-12-16T04:53:50
538184140
{ "authors": [ "rishabh96b" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8351", "repo": "mesosphere/universe", "url": "https://github.com/mesosphere/universe/pull/2457" }
gharchive/pull-request
Revert "[DCOS-61537] Change container type to MESOS in replicator" Reverts mesosphere/universe#2453 as the customer faced a broken replicator when running it in non VN mode. Yes it was tested but I'll investigate why it broke on their side.
2025-04-01T04:34:40.173466
2022-07-28T18:09:03
1321309031
{ "authors": [ "spitko", "this-username-is-taken" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8352", "repo": "messari/subgraphs", "url": "https://github.com/messari/subgraphs/pull/686" }
gharchive/pull-request
Aave v3: Fix stable mode borrow balance, close borrower position on liquidation Also changed aToken transfer logic so that transfer volume is not included in market/protocol deposit and withdraw volume. Nice update!
2025-04-01T04:34:40.192072
2024-02-02T16:19:22
2115278634
{ "authors": [ "Simone-RM", "jwinarske" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8353", "repo": "meta-flutter/meta-flutter", "url": "https://github.com/meta-flutter/meta-flutter/issues/395" }
gharchive/issue
Building on Linux for Arm (aarch64) fails Hello, I've added the meta-flutter layer to my yocto configuration (running on Ubuntu 22.04.3). When I try "bitbake flutter-engine" (or any other app, eg. flutter-animated-background), I get the following error: WARNING: flutter-engine-git-r0 do_fetch: Failed to fetch URL gn://github.com/flutter/engine.git;name=flutter, attempting MIRRORS if available ERROR: flutter-engine-git-r0 do_fetch: Fetcher failure: Fetch command export PSEUDO_DISABLED=1; export DBUS_SESSION_BUS_ADDRESS="unix:path=/run/user/1000/bus"; export PATH="/home/simone/yocto/engicam-kirkstone/build/tmp/sysroots-uninative/aarch64-linux/usr/bin:/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot-native/usr/bin/python3-native:/home/simone/yocto/engicam-kirkstone/sources/poky/scripts:/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot-native/usr/bin/aarch64-poky-linux:/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot/usr/bin/crossscripts:/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot-native/usr/sbin:/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot-native/usr/bin:/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot-native/sbin:/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot-native/bin:/home/simone/yocto/engicam-kirkstone/sources/poky/bitbake/bin:/home/simone/yocto/engicam-kirkstone/build/tmp/hosttools"; export HOME="/home/simone"; export PATH="/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot-native/usr/share/depot_tools:/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot-native/usr/share/depot_tools/bootstrap-2@3.8.9.chromium.14_bin/python/bin:${PATH}"; export DEPOT_TOOLS_UPDATE=0; export GCLIENT_PY3=0; export CURL_CA_BUNDLE=/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot-native/etc/ssl/certs/ca-certificates.crt; gclient.py config --spec 'solutions = [ { "managed": False, "name": "src/flutter", "url": "https://github.com/flutter/engine.git", "custom_vars": { "download_android_deps": False, "download_windows_deps": False, "download_linux_deps": False, }, "custom_deps": {} } ]' && gclient.py sync --shallow --no-history -R -D --revision f40e976bedff57e69e1b3d89a7c2a3c617a03dad -j 8 -v failed with exit code 127, output: /usr/bin/env: ‘python’: No such file or directory any idea how to fix it? Python3 is installed on my host (via the python-is-python3 package): $ /usr/bin/env python Python 3.10.12 (main, Nov 20 2023, 15:14:05) [GCC 11.4.0] on linux I'm using the latest kirkstone branch (commit f4f58f952482793a3ddb1ea3685733e48da45b8a) Replacing /usr/bin/env python with /usr/bin/python in tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot-native/usr/share/depot_tools/gclient.py seems to help (it's downloading a lot of stuff now). It's not ideal though. @Simone-RM Hi. This layer is in production in a number of places without issue. So the likelihood is that you have a host configuration issue. Please check that your Yocto system per-requisites are met: https://docs.yoctoproject.org/4.0.15/ref-manual/system-requirements.html I think I have all the prerequisites, and I've been building images (without Flutter) on this setup for a while now. One other bit of information: I'm running Ubuntu on aarch64 (Apple Silicon with Parallels). Has anyone every tried that? I'm starting to think that is the real problem... changing gclient's shebang allows it to download everything, but then compilation fails: Generating GN files in: out/linux_debug_arm64 NOTE: is_debug = false use_goma = false goma_dir = "None" full_dart_sdk = false enable_unittests = false skia_enable_flutter_defines = true skia_use_dng_sdk = false skia_use_sfntly = false skia_enable_pdf = false skia_use_x11 = false skia_use_wuffs = true skia_use_expat = true skia_use_fontconfig = true skia_use_icu = true is_official_build = true android_full_debug = false is_clang = true skia_gl_standard = "" use_clang_static_analyzer = false enable_coverage = false enable_lto = true host_os = "linux" host_cpu = "arm64" target_os = "linux" target_cpu = "arm64" dart_target_arch = "arm64" flutter_use_fontconfig = true dart_component_kind = "static_library" embedder_for_target = true dart_lib_export_symbols = false flutter_runtime_mode = "debug" dart_version_git_info = true dart_runtime_mode = "develop" enable_desktop_embeddings = true enable_backtrace = true bssl_use_clang_integrated_as = true target_sysroot = "/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot" custom_sysroot = "/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot" custom_toolchain = "/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/src/buildtools/linux-x64/clang" custom_target_triple = "aarch64-unknown-linux-gnu" skia_use_gl = true skia_use_vulkan = true skia_use_vma = false shell_enable_vulkan = true disable_hidden_visibility = false dart_include_wasm_opt = false dart_platform_sdk = true build_glfw_shell = false build_embedder_examples = false stripped_symbols = false dart_use_mallinfo2 = true impeller_enable_3d = true impeller_enable_vulkan = true impeller_enable_opengles = true angle_build_all = false angle_has_astc_encoder = false angle_use_x11 = false angle_use_wayland = false angle_enable_vulkan_validation_layers = false angle_vulkan_headers_dir = "//third_party/vulkan-deps/vulkan-headers/src" angle_vulkan_loader_dir = "//third_party/vulkan-deps/vulkan-loader/src" angle_vulkan_tools_dir = "//third_party/vulkan-deps/vulkan-tools/src" arm_tune = "armv8a+crc+crypto" Unknown architecture (aarch64) -- unable to run ninja. No prebuilt ninja binary was found for this system. Try building your own binary by doing: cd ~ git clone https://github.com/ninja-build/ninja.git -b v1.8.2 cd ninja && ./configure.py --bootstrap Then add ~/ninja/ to your PATH. WARNING: exit code 1 from a shell command. which is a pity (also considering I do have a prebuilt ninja installed as /usr/bin/ninja) @Simone-RM Good call on adding the details and renaming the issue. I have never built a Poky Distro on an Aarch64 machine. Largely because all the machines are so largely under-powered, and I would prefer to watch grass grow :) I build on an AMD Thread Ripper 32+ HW threads, Intel 32 HW threads, AGL builds run on self hosted High Availability servers, and others use Extra Large Virtual CPU instances (AWS). AWS is now offering Yocto image builds as a service. Are you building a full OS image, or just trying to build Flutter engine artifacts? If full OS image can you build it short of the meta-flutter recipes? If just flutter engine artifacts I would suggest a standalone command line build with my flutter-engine patches. You might also try updating depot-tools to tip of tree, and switching it to python3. It's using an older commit with python2, as this was a requirement for flutter-engine; may have changed in the last six months. You can also try adding ninja-native to DEPENDS, keep in mind ninja is part of depot-tools; which might be an issue with your newer combo (Ubuntu running on apple silicon). I tried rolling depot-tools to tip of tree main, and it's complaining about missing python3-httplib2. I don't have time to dig into this at the moment. I have never built a Poky Distro on an Aarch64 machine. Largely because all the machines are so largely under-powered, and I would prefer to watch grass grow :) I build on an AMD Thread Ripper 32+ HW threads, Intel 32 HW threads, AGL builds run on self hosted High Availability servers, and others use Extra Large Virtual CPU instances (AWS). AWS is now offering Yocto image builds as a service. I was using an AWS machine for my builds, but having everything local on my machine is more practical. And building on a new Mac is not that much slower than on the 32-cores AWS server. ;-) Are you building a full OS image, or just trying to build Flutter engine artifacts? full image. I'm trying to add flutter to a working image, just to see if it works/how it performs on the boards we're using. * If full OS image can you build it short of the meta-flutter recipes? yes You might also try updating depot-tools to tip of tree, and switching it to python3. It's using an older commit with python2, as this was a requirement for flutter-engine; may have changed in the last six months. You can also try adding ninja-native to DEPENDS, keep in mind ninja is part of depot-tools; which might be an issue with your newer combo (Ubuntu running on apple silicon). I'll try something like that. I've seen that there are several scripts that only know about x64 arch as far as Linux is concerned. I've also tried building an image on Ubuntu 20 on an AWS server, including meta-flutter-animated-background-example: it built correctly, but I suspect that it didn't include everything needed, as I couldn't find a way to launch the app... I tried building flutter-auto, but it causes the compiler to crash :( | FAILED: shell/CMakeFiles/flutter-auto.dir/backend/wayland_vulkan.cc.o | /yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/recipe-sysroot-native/usr/bin/aarch64-poky-linux/aarch64-poky-linux-g++ --sysroot=/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/recipe-sysroot/usr -DASIO_DISABLE_HANDLER_TYPE_REQUIREMENTS -DASIO_HEADER_ONLY -DASIO_NO_DEPRECATED -DASIO_NO_TYPEID -DASIO_STANDALONE -DBUILD_BACKEND_WAYLAND_VULKAN -DCRASH_HANDLER_DSN=\"\" -DCRASH_HANDLER_RELEASE=\"\" -DEGL_NO_X11 -DENABLE_AGL_CLIENT -DENABLE_DART_VM_LOGGING -DENABLE_PLUGIN_ACCESSIBILITY -DENABLE_PLUGIN_COMP_REGION -DENABLE_PLUGIN_COMP_SURF -DENABLE_PLUGIN_DESKTOP_WINDOW -DENABLE_PLUGIN_GOOGLE_SIGN_IN -DENABLE_PLUGIN_GO_ROUTER -DENABLE_PLUGIN_ISOLATE -DENABLE_PLUGIN_KEYBOARD_MANAGER -DENABLE_PLUGIN_KEY_EVENT -DENABLE_PLUGIN_LOGGING -DENABLE_PLUGIN_MOUSE_CURSOR -DENABLE_PLUGIN_PACKAGE_INFO -DENABLE_PLUGIN_PLATFORM -DENABLE_PLUGIN_RESTORATION -DENABLE_PLUGIN_TEXT_INPUT -DENABLE_PLUGIN_URL_LAUNCHER -DENABLE_XDG_CLIENT -DENV64BIT -DFILAMENT_SUPPORTS_WAYLAND -DGIT_BRANCH=\"agl\" -DGIT_HASH=\"3114782\" -DHAVE_MEMFD_CREATE -DHAVE_STRCHRNUL -DLINUX -DMESA_EGL_NO_X11_HEADERS -DNDEBUG -DPATH_PREFIX=\"/usr\" -DWL_EGL_PLATFORM -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/build/shell -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/shell -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/shell/.. -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/shell/logging -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/shell/../third_party/spdlog-1.12.0/include -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/shell/../third_party/asio-1-28-1/asio/include -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/third_party/filament/libs/bluevk/include -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/third_party/. -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/third_party/flutter -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/third_party/flutter/third_party -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/third_party/flutter/third_party/rapidjson/include -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/third_party/flutter/shell/platform/common/public -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/third_party/flutter/shell/platform/common/client_wrapper/include -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/third_party/spdlog-1.12.0/include -march=armv8-a+crc+crypto -fstack-protector-strong -O2 -D_FORTIFY_SOURCE=2 -Wformat -Wformat-security -Werror=format-security --sysroot=/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/recipe-sysroot -O2 -pipe -g -feliminate-unused-debug-types -fmacro-prefix-map=/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0=/usr/src/debug/flutter-auto/git-r0 -fdebug-prefix-map=/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0=/usr/src/debug/flutter-auto/git-r0 -fdebug-prefix-map=/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/recipe-sysroot= -fdebug-prefix-map=/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/recipe-sysroot-native= -fvisibility-inlines-hidden -fno-rtti -DNDEBUG -flto -fno-fat-lto-objects -fPIE -fvisibility-inlines-hidden -Wtrigraphs -Wchar-subscripts -Wcomment -Wreturn-type -Wsequence-point -Wswitch -Wuninitialized -Wunused -Wswitch-bool -Wformat -Wformat-security -Wconversion -Wcast-align -Wunused-parameter -Winvalid-offsetof -fsigned-char -O2 -std=c++17 -MD -MT shell/CMakeFiles/flutter-auto.dir/backend/wayland_vulkan.cc.o -MF shell/CMakeFiles/flutter-auto.dir/backend/wayland_vulkan.cc.o.d -o shell/CMakeFiles/flutter-auto.dir/backend/wayland_vulkan.cc.o -c /yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/shell/backend/wayland_vulkan.cc | /yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/shell/backend/wayland_vulkan.cc: In member function 'virtual FlutterRendererConfig WaylandVulkanBackend::GetRenderConfig()': | /yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/shell/backend/wayland_vulkan.cc:66:8: internal compiler error: in reshape_init_class, at cp/decl.c:6470 | 66 | }}; | | ^ | 0x186e23e internal_error(char const*, ...) | ???:0 | 0x6a3710 fancy_abort(char const*, int, char const*) | ???:0 | 0x714ad9 reshape_init(tree_node*, tree_node*, int) | ???:0 | 0x6c10dc can_convert_arg(tree_node*, tree_node*, tree_node*, int, int) | ???:0 | 0x6c10dc can_convert_arg(tree_node*, tree_node*, tree_node*, int, int) | ???:0 | 0x6b6423 perform_implicit_conversion_flags(tree_node*, tree_node*, int, int) | ???:0 | 0x86661a convert_for_initialization(tree_node*, tree_node*, tree_node*, int, impl_conv_rhs, tree_node*, int, int) | ???:0 | 0x867480 check_return_expr(tree_node*, bool*) | ???:0 | 0x82f663 finish_return_stmt(tree_node*) | ???:0 | 0x7eb604 c_parse_file() | ???:0 | 0x8be425 c_common_parse_file() | ???:0 | Please submit a full bug report, | with preprocessed source if appropriate. | Please include the complete backtrace with any bug report. | See <https://gcc.gnu.org/bugs/> for instructions. | ninja: build stopped: subcommand failed. | WARNING: exit code 1 from a shell command. ERROR: Task (/yocto/engicam/sources/meta-flutter/recipes-graphics/toyota/flutter-auto_git.bb:do_compile) failed with exit code '1' I have an M2 I rarely use. I've ran Ubuntu on it, and perf was not so great. I will try Ubuntu as a QEMU image with my QEMU launcher app to see if perf is any better: https://github.com/meta-flutter/virt_qemu That said I also exclusively use a raid10 SSD card on my workstations. Aside from high core count CPU, fast disk IO is the next bottleneck on Yocto builds. You can build a nice Yocto build workstation for ~$5k US. It pays for itself in a few months of regular use. I would start with standalone building the flutter engine directly on Ubuntu on m2. With that working and if there are a clear set of patches (or none), then doing the Yocto build will be much clearer. I moved depot-tools to tip of tree. See if that helps thanks, it does help. Now flutter_engine compiles with a simple bbappend to fix clang's path: CLANG_PATH = "${WORKDIR}/src/buildtools/linux-arm64/clang" I'll have to fix the install job too somehow (all the 'clang_x64' instances should be replaced with 'clang_arm64') to work. flutter-auto compilation crashes just like on the Intel server though. What value is assigned to HOST_ARCH? I'll remove the DISTRO_FEATURES filter for vulkan backend selection. HOST_ARCH is aarch64 https://github.com/meta-flutter/meta-flutter/pull/409/files Passes building flutter-gallery on Mac M2 -> UTM virtualization -> Ubuntu 22 merged let me know if that unblocks you I can confirm that this issue can be closed, thanks!
2025-04-01T04:34:40.197426
2023-10-08T20:11:14
1932051687
{ "authors": [ "jmikedupont2" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8354", "repo": "meta-introspector/call-auto-gpt", "url": "https://github.com/meta-introspector/call-auto-gpt/issues/7" }
gharchive/issue
Issues are disabled for this repo mockopenai_1 | github.GithubException.GithubException: 410 {"message": "Issues are disabled for this repo", "documentation_url": "https://docs.github.com/v3/issues/"} mockopenai_1 | <IP_ADDRESS> - - [08/Oct/2023 20:04:52] "POST /v1/chat/completions HTTP/1.1" 500 - /agent-action test /agent-action test2 /agent-action test23 /agent-action test23 /agent-action test23 test tst of comment filter /agent-action do it /no-agent-action tst /agent-action pls /agent-action please run again test comment /agent-action lets dance /agent-action please
2025-04-01T04:34:40.202147
2024-04-07T13:26:23
2229772951
{ "authors": [ "jmikedupont2" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8355", "repo": "meta-introspector/meta-meme", "url": "https://github.com/meta-introspector/meta-meme/issues/139" }
gharchive/issue
The mythical bridge Paraphrase: The bridge between the knowable and unknowable is created by the myth or narrative, it allows the making of the meaning, the motivation, the inspiration, the world view or perspective of a fundamental observation about bravery, confidence and respect for forces greater than us source: New Books Network: Wole Talabi, "Convergence Problems" (Astra Publishing House, 2024) Starting from: 00:28:34 Episode webpage: https://newbooksnetwork.com/convergence-problems claude : You make an insightful observation about the role of myth and narrative in bridging the knowable and unknowable, and providing meaning, motivation, and perspective. Myths and narratives allow us to contextualize and find significance in the fundamental mysteries and forces of the universe that lie beyond our full comprehension. By weaving stories and archetypes around primal experiences like courage in the face of the unknown, or humility before vastly powerful natural or cosmic phenomena, myths provide a symbolic framework for grappling with the great unknowns of existence. They transform abstract universal truths into resonant tales that speak to the human psyche. In this way, myths don't just entertain, but illuminate different worldviews and philosophies about our place in the grand scheme. They suggest moral guideposts and inspiring ideals, while leaving room for the ineffable. A potent myth about bravery in confronting the unexplained can instill strength, just as one about nature's majesty can seed awe and respect. Rather than reducing the unknowable to cold facts, myths preserve a sense of wonderful mystery. They provide a conceptual bridge between our finite minds and the infinite unknowns we intuively sense lurking just beyond the veil of the known world. Myths dignify the quest to understand through imagination and symbolic storytelling. So in encapsulating fundamental truths about the human condition and our relationship to the unknowable through narrative, myths play a vital role in our psychological, cultural and spiritual understandings of the cosmos and our own existence within it. They are profoundly meaningful creations born from our universal experience of the known and unknown realms. gemini : Myths and narratives weave a tapestry that bridges the chasm between the knowable and the unknowable. They take our fragmentary understanding of the world and use it to craft stories that illuminate the profound mysteries that surround us. These stories, passed down through generations, become a repository of human experience, offering explanations for natural phenomena, moral codes for navigating complex social interactions, and a sense of purpose in a vast and often uncaring universe. Myths inspire us with tales of bravery and resilience in the face of the unknown, and cultivate a sense of awe and respect for forces beyond our control. They serve as a springboard for our own creativity, prompting us to question, explore, and forge our own unique understanding of the world.
2025-04-01T04:34:40.374543
2021-07-10T12:15:42
941238843
{ "authors": [ "Encrylize", "giarve" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8356", "repo": "metacall/core", "url": "https://github.com/metacall/core/pull/166" }
gharchive/pull-request
Add function call support to Wasm loader Description This PR implements support for function calls in the Wasm loader. Note that multiple return values are not yet supported. Type of change [X] New feature (non-breaking change which adds functionality) Checklist: [X] I have performed a self-review of my own code. [X] I have commented my code, particularly in hard-to-understand areas. [X] I have made corresponding changes to the documentation. [ ] My changes generate no new warnings. [X] I have added tests/screenshots (if any) that prove my fix is effective or that my feature works. [X] I have tested the tests implicated (if any) by my own code and they pass (make test or ctest -VV -R <test-name>). [X] I have tested my code with OPTION_BUILD_SANITIZER and OPTION_TEST_MEMORYCHECK. [X] I have run make clang-format in order to format my code and my code follows the style guidelines. Note that memcheck reports some memory leaks. However, these seem to be caused by Wasmtime and are probably not our fault. I will investigate these further once the loader is more complete and decide whether they should be suppressed or a fix is warranted. Nice work!
2025-04-01T04:34:40.376603
2017-07-17T08:10:59
243322041
{ "authors": [ "blackillzone", "retr0h" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8357", "repo": "metacloud/molecule", "url": "https://github.com/metacloud/molecule/issues/905" }
gharchive/issue
Sharing completion file for fish Hi everyone, Here is my completion script for Molecule 1.25, adapted for Fish Shell: https://gist.github.com/blackillzone/ef9962ba652b80a8c7952aae3f8a53dc Didn't knew where sharing this, so I created a new issue, my apologies if it was the wrong way. Once I'll migrate to version 2.0 of Molecule, I'll check to do an updated version also. Feel free to use it ;). Hi @blackillzone thanks for the info. While we are not actively maintaining Molecule v1 any longer, we appreciate your note. I'm going to close this issue for others to reference. Molecule v2s uses click's internal mechanism for command completion.
2025-04-01T04:34:40.390438
2021-04-06T09:52:19
851256551
{ "authors": [ "fsteeg", "katauber" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8358", "repo": "metafacture/metafacture-playground", "url": "https://github.com/metafacture/metafacture-playground/pull/13" }
gharchive/pull-request
Generate Link to share API calls Popup with link to api call of current workflow For an independent review, this PR should have had its base set to 6-improveUI instead of containing the changes from 6-improveUI and being based on main. But we can now leave it as it is, since it should show a reviewable diff after https://github.com/metafacture/metafacture-playground/pull/12 is merged. But we can now leave it as it is, since it should show a reviewable diff after #12 is merged. Hm, no, that did not work. There is one little conflict, but all other changes still show up. Maybe it would work to rebase 9-generateLink on the current main? I think this looks better now :)
2025-04-01T04:34:40.391621
2019-09-10T14:27:26
491709071
{ "authors": [ "desandro", "starchild" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8359", "repo": "metafizzy/infinite-scroll", "url": "https://github.com/metafizzy/infinite-scroll/issues/893" }
gharchive/issue
Show loading image Not a bug as such, but just wondering how to show a loading image if you reach the end of the container and a next page exists? Use .infinite-scroll-request status element to display content when Infinite Scroll is requesting a page.
2025-04-01T04:34:40.437100
2018-09-01T16:55:43
356217230
{ "authors": [ "aqrln", "tshemsedinov" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8360", "repo": "metarhia/globalstorage", "url": "https://github.com/metarhia/globalstorage/issues/186" }
gharchive/issue
Large datasets navigation Sometimes query may return large datasets, for example gs.select({ category: 'Person' }) called from client will send request to the server, then spread requests to all servers storing mentioned category then start receiving data by chunks in lazy mode. So client-side cursor will receive first chunk and will generate on('data') event and first 100 records will be available at the client-side to be iterated by cursor. But we may not want to transfer more before user navigate to below grid position. Is it ok for ours gui rendering console if cursor will have no all data at once and data will arrive chunk by chunk pushing to dataset? Also how GUI can inform cursor to get next chunk? @aqrln Here is an example how cursors may fetch data @aqrln I think we cant use just cursor.dataset: Array, we need something lige fetch(callback(err)) or to load all records to cursor and then it will be available in cursor.dataset or can use event to get by chunks on('data', callback) @aqrln I forgot a link to mentioned example: https://github.com/metarhia/globalstorage/pull/194 Cursor and work in two modes: (1) dataset holder, (2) data transformation from parent cursor not holding datasets. So cursors are chained but we can materialize datasets at any step. I am going to add cursor.materialize() that will receive all data from parent cursor and save it it's own dataset. This allow us to minimize data copies in memory. Do we need to copy data again to redux store? Or we will not use cursor materialization and will collect aggregated/transformed data just in store? @tshemsedinov store is immutable and can only be changed with pure functions (reducers) in response to actions, so if something changes, we need to copy it again. See https://github.com/metarhia/globalstorage/issues/195
2025-04-01T04:34:40.443167
2021-07-05T07:21:45
936777841
{ "authors": [ "LachlanStuart", "sergii-mamedov" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8361", "repo": "metaspace2020/metaspace", "url": "https://github.com/metaspace2020/metaspace/pull/901" }
gharchive/pull-request
Update Lithops version We are updating to avoid the problems described here. copy that Retroactive LGTM :+1: (we agreed it was fine to merge on Slack)
2025-04-01T04:34:40.444581
2016-01-04T21:38:47
124845067
{ "authors": [ "DominikGuzei", "darko-mijic" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8362", "repo": "meteor-space/vo-monetary", "url": "https://github.com/meteor-space/vo-monetary/pull/5" }
gharchive/pull-request
introduce better floating point precision for money VOs This PR adds improved handling of floating point numbers for Money VOs. The number required decimal places (precision) is calculated on creation of the VO and saved together with the "base" number which is an integer version of the float by multiplication of 10 * decimal places. Is it OK if I review this in the morning? Of course, im also going to bed now :wink:
2025-04-01T04:34:40.470276
2017-03-27T20:04:55
217359320
{ "authors": [ "sebhoss" ], "license": "CC0-1.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8363", "repo": "metio/yosql", "url": "https://github.com/metio/yosql/issues/30" }
gharchive/issue
setup CI/CD blocked by Bazel install on build.metio.wtf: https://github.com/metio/build.metio.wtf/issues/1 now using github actions
2025-04-01T04:34:40.479978
2017-11-30T16:36:20
278187236
{ "authors": [ "MoseleyS", "carolinejjones" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8364", "repo": "metoppv/improver", "url": "https://github.com/metoppv/improver/pull/355" }
gharchive/pull-request
318 falling snow level Reference issue #318 Testing: [X] Ran tests and they passed OK [X] Added new tests for the new feature(s) I have also added a small change to how I find the x and y coordinate. This PR also fails some CLI tests because the test data are more recent than this branch's upstream/master. This does not prevent the PR being completed or merged and could be overcome by rebasing the branch if we wanted to.
2025-04-01T04:34:40.498525
2020-01-21T22:42:06
553186843
{ "authors": [ "svartalf", "tobz" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8365", "repo": "metrics-rs/metrics", "url": "https://github.com/metrics-rs/metrics/issues/66" }
gharchive/issue
Consuming external metrics with absolute values It is very common to expose some external metrics as counters created outside of the current application, for example, by collecting system statistics (ex. total amount of bytes sent via the network interface) or fetching them from other application (as various Prometheus exporters do). The problem is that these metrics has absolute values and with current API it is possible to increment counters only and there is no way to replace the value. From my understanding, something like Constant Metrics can be added to allow one-time loading of the metric values; this idea better to be discussed further. Naive way would be to add Recorder::set_counter method, which change the value of the inner AtomicU64. Obviously, it is not the best solution possible, yet, it can work as a temporary workaround till better idea will land the metrics crate; as an addition, it can be feature-gated with something like #[cfg(feature = "unstable")]. Extra links: https://github.com/tikv/rust-prometheus/pull/303 https://www.robustperception.io/setting-a-prometheus-counter https://github.com/prometheus-net/prometheus-net/issues/73 There's some renewed activity around this topic in #172 based on improvements we've made in metrics 0.13 that allow us to inject logic to the pipeline without having to necessarily extend the macros themselves.
2025-04-01T04:34:40.499828
2021-02-25T13:31:46
816439827
{ "authors": [ "0nkery", "tobz" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8366", "repo": "metrics-rs/metrics", "url": "https://github.com/metrics-rs/metrics/pull/180" }
gharchive/pull-request
Observer: Use BTreeMap as metrics storage to sort by key This change allows preserving alphabetical order in the observer interface. Nice! Looks good to me. Thanks for your contribution! Thanks for merging!
2025-04-01T04:34:40.504502
2019-06-19T22:44:45
458257970
{ "authors": [ "queer" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8367", "repo": "mewna/catnip", "url": "https://github.com/mewna/catnip/issues/334" }
gharchive/issue
Emit a warning event on high heartbeat latency Self-explanatory; also makes it so users don't have to measure it themselves. Added in 98e1a8ed
2025-04-01T04:34:40.515577
2020-10-05T07:10:09
714581002
{ "authors": [ "Farnazmdi", "aarmey", "codecov-commenter", "jclagarde" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8368", "repo": "meyer-lab/tHMM", "url": "https://github.com/meyer-lab/tHMM/pull/646" }
gharchive/pull-request
Added Supp. Figure Captions I'm not quite sure where the text explanation that was previously in the caption section should go. I added it above the figure though. @JC Lagarde I have seen many papers without explaining their supplementary figures, and only providing a caption. Besides, we have mentioned these figures in the main results where it was relevant. @aarmey Do we want a paragraph explaining each supplementary figure? Codecov Report Merging #646 into master will not change coverage. The diff coverage is n/a. @@ Coverage Diff @@ ## master #646 +/- ## ======================================= Coverage 88.50% 88.50% ======================================= Files 21 21 Lines 1505 1505 ======================================= Hits 1332 1332 Misses 173 173 Flag Coverage Δ #unittests 88.50% <ø> (ø) Flags with carried forward coverage won't be shown. Click here to find out more. Continue to review full report at Codecov. Legend - Click here to learn more Δ = absolute <relative> (impact), ø = not affected, ? = missing data Powered by Codecov. Last update 5d25984...0f9f558. Read the comment docs. @Farnazmdi you're right that the captions likely should do this. Optionally, we could include a supplementary methods section with a more detailed description.
2025-04-01T04:34:40.544260
2020-04-30T10:04:46
609782066
{ "authors": [ "makeclean", "tzanio", "v-dobrev" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8369", "repo": "mfem/mfem", "url": "https://github.com/mfem/mfem/issues/1452" }
gharchive/issue
Compile fail on nurb_ex1p Hi Was just compiling mfem (master) on my new machine; gcc - v gcc version 9.3.1 20200408 (Red Hat 9.3.1-2) (GCC) mpich 3.3 cmake -DHYPRE_DIR=$WORKDIR/hypre -DMETIS_DIR=$WORKDIR/metis -DMFEM_ENABLE_EXAMPLES=ON -DMFEM_ENABLE_MINIAPPS=ON -DMFEM_THREAD_SAFE=ON -DMFEM_USE_MPI=ON -DMFEM_USE_OPENMP=ON -DCMAKE_INSTALL_PREFIX=.. Build for HYPRE & METIS went fine, all of mfem with the exception of this example compiles, the failure message is /home/adavis/opt/mfem/miniapps/nurbs/nurbs_ex1p.cpp:116:27: error: cannot convert ‘mfem::Vector [nd]’ to ‘mfem::Vector&’ 116 | el.CalcShape(ip, shape); | ^~~~~ | | | mfem::Vector [nd] In file included from /home/adavis/opt/mfem/mesh/triangle.hpp:16, from /home/adavis/opt/mfem/mesh/mesh_headers.hpp:21, from /home/adavis/opt/mfem/mfem.hpp:42, from /home/adavis/opt/mfem/bld/mfem.hpp:3, from /home/adavis/opt/mfem/miniapps/nurbs/nurbs_ex1p.cpp:44: /home/adavis/opt/mfem/mesh/../fem/fe.hpp:344:35: note: initializing argument 2 of ‘virtual void mfem::FiniteElement::CalcShape(const mfem::IntegrationPoint&, mfem::Vector&) const’ 344 | Vector &shape) const = 0; | ~~~~~~~~^~~~~ /home/adavis/opt/mfem/miniapps/nurbs/nurbs_ex1p.cpp:128:40: error: ‘shape’ cannot be used as a function 128 | elmat(i, j) += w*shape(i)*laplace(j); | ^ make[2]: *** [miniapps/nurbs/CMakeFiles/nurbs_ex1p.dir/build.make:83: miniapps/nurbs/CMakeFiles/nurbs_ex1p.dir/nurbs_ex1p.cpp.o] Error 1 make[1]: *** [CMakeFiles/Makefile2:3517: miniapps/nurbs/CMakeFiles/nurbs_ex1p.dir/all] Error 2 make: *** [Makefile:161: all] Error 2 I suppose I could comment out this example, but I was actually interested in playing with it. There is a typo here: https://github.com/mfem/mfem/blob/7690eca85041e0b69afe81a063c71015184e82a3/miniapps/nurbs/nurbs_ex1p.cpp#L78 The line should be: Vector shape(nd); Unfortunately, we missed it because we don't test regularly with -DMFEM_THREAD_SAFE=ON. Thanks for reporting the issue. We will fix it soon in master. theres a copy-pasta in nurbs_ex1.cpp too Thanks for reporting this @makeclean Discussion continued in #1455
2025-04-01T04:34:40.550571
2022-07-25T20:57:23
1317360385
{ "authors": [ "prathamgit", "tzanio", "v-dobrev" ], "license": "BSD-3-Clause", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8370", "repo": "mfem/mfem", "url": "https://github.com/mfem/mfem/issues/3116" }
gharchive/issue
Minor correction in Mfem_Assert call in sparsrMat class Should it be B.width in the line << B.height << ", B.width = " << width in SparseMatric class ? https://github.com/mfem/mfem/blob/670b7b1ae32f2126df8f4539a238b0aa4b8e33fe/linalg/sparsemat.cpp#L3059 Looks correct to me, the full line with some reformatting is essentially cerr << "Mismatch of this matrix size and rhs. " << "This height = " << height << ", width = " << width << ", B.height = " << B.height << ", B.width = " << width; @tzanio, I think the point is to replace ", B.width = " << width --> ", B.width = " << B.width. Oh, that makes sense, sorry. It should be indeed << ", B.width = " << B.width in the last line. @prathamgit, do you want to propose a simple PR to fix it? @tzanio i did not understand what is "simple PR fix" ? Oh I just meant you can put the fix in a branch and issue a pull request to fix it in master. Oh okay. Yes I can do that. Thanks! Hi @prathamgit, do you still plan to make a pull request for this? Oh I have, a few days ago. Can you post a link to it here? Right sorry. Here is the link. https://github.com/mfem/mfem/pull/3161 I hope this link works. Addressed in https://github.com/mfem/mfem/pull/3161
2025-04-01T04:34:40.555851
2023-02-12T20:32:40
1581399712
{ "authors": [ "Apollon77" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8371", "repo": "mfucci/node-matter", "url": "https://github.com/mfucci/node-matter/pull/247" }
gharchive/pull-request
Really check for Statuscode errors and return Error classes This PR uses the new Error classes to instoduce StatusError classes that are thrown when an unexpected StatusCode is returned. This also fixes the "waitForSucces" to really wait for a success reswponse and not for "any" response Ok, I checked more in deoth the throwIfError and waitForSuccess and optimized more ... hopefully now less "duplicate" code @mfucci Ready for re-reveiw ... I adjusted ... you were right regarding duplicste throwIfError Ready for re-review
2025-04-01T04:34:40.569971
2015-05-13T12:36:50
75970501
{ "authors": [ "alexandrujuncu", "mgax" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8372", "repo": "mgax/jekyllcms", "url": "https://github.com/mgax/jekyllcms/issues/2" }
gharchive/issue
Template editor Feature request: a way of editing the template online. For now, I think GitHub's web editor is adequate for this, I'd like to focus on content editing.
2025-04-01T04:34:40.591801
2016-06-06T20:49:00
158771416
{ "authors": [ "TheDonDope", "gvsdan", "hank-ehly" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8373", "repo": "mgechev/angular2-seed", "url": "https://github.com/mgechev/angular2-seed/issues/971" }
gharchive/issue
Enable SCSS #952 Integration is not injecting the npm dependencies i'm trying to use the scss #952 integration in development mode is working fine but when i use the prod mode is not creating the '.css' for all the npm dependencies injected in the 'project.config.ts' @gvsdan, thank you for your comment. I'll take a look. @hank-ehly Thanks to you for making the sass support Fixed by #976, thank you very much Hank! @hank-ehly Everything is working fine but now the external dependencies in production are broken.
2025-04-01T04:34:40.592806
2017-09-23T11:42:13
260002218
{ "authors": [ "arielhenryson", "wKoza" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8374", "repo": "mgechev/codelyzer", "url": "https://github.com/mgechev/codelyzer/issues/420" }
gharchive/issue
no-access-missing-member not finding super class method When "no-access-missing-member": true, and one of the template try to access member that was inherited from superclasses the error is thrown Please, take a look at the existing issues.
2025-04-01T04:34:40.594099
2022-02-13T16:43:17
1136115520
{ "authors": [ "neclimdul" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8375", "repo": "mglaman/phpstan-drupal", "url": "https://github.com/mglaman/phpstan-drupal/pull/327" }
gharchive/pull-request
Provide a stub for twig_escape_filter twig_escape_filter is a hidden function inside a class file so not discovered though normal discovery or autoloading so providing scan stub allows its use to be analyzed. Fixes #310 I couldn't recreate it outside the Drupal repository either but this fixes it.
2025-04-01T04:34:40.620528
2011-12-15T15:54:50
2568548
{ "authors": [ "benjaminoakes", "mgsnova" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8378", "repo": "mgsnova/feature", "url": "https://github.com/mgsnova/feature/issues/1" }
gharchive/issue
Documentation: Rails example Hi mgsnova, Thanks for writing feature. It makes sense to have a standard library for something like this. We have something similar that looks at a MySQL database (which we may contribute to feature sometime). In the meantime, we have a simple Rails app that doesn't have a MySQL database, so your gem fits the bill nicely. To say thanks, I wrote up some docs about how I made it work and made this pull request. Happy holidays! Ben Hi Ben, thanks for using the gem and giving feedback. Can you please remove the files added with commit 4c2b1bb. I think they are not really needed now in the code. I will merge the pull request then. Best regards, mgsnova Hi @mgsnova, Happy to. I actually didn't mean for them to show up in this pull request, but they did because I pushed to my fork. Sorry about that. Next time, I guess I should do the pull request from a branch. :) Best wishes, Ben To keep it simple, I just did git revert 4c2b1bb04c2f22d87947e1afb8f822cf33f51ba5. Best wishes and thanks again! Ben
2025-04-01T04:34:40.640948
2017-03-08T23:27:05
212888602
{ "authors": [ "coveralls", "femtotrader", "mhallsmoore", "ryankennedyio" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8381", "repo": "mhallsmoore/qstrader", "url": "https://github.com/mhallsmoore/qstrader/pull/190" }
gharchive/pull-request
Live trade session Minor refactor of the original backtest.py class to also allow for live trading. Live trading only differs in that it's "end-condition" is the passing of time, until and end_session_time. All loop logic remains the same. This branch is a dependancy of https://github.com/mhallsmoore/qstrader/pull/186. I created a new branch here as the logic has no overlap with IB integration. I'm suggesting we don't do the IBService setup/teardown in the TradingSession -- we must do it in the setup scripts where the user defines their strategy, symbols, handlers, etc. Otherwise we need to create a TradingSession for each service, or create an abstraction where we take in a list of services (unnecessarily complex). As it stands, the TradingSession simply takes in a set of QSTrader components (price, risk, execution handlers, etc) -- each of these components may have services/external system dependancies, but the trading loop does not -and should not- care. Would appreciate prompt integration with master so I can get cracking with live market data in the IB-PriceHandler Coverage increased (+0.04%) to 57.143% when pulling fdb7a55e09b112e2a43136cd539c9b54b459e87b on ryankennedyio:live_trade_session into ca05a06cf8429f571d566a39409f674142676af3 on mhallsmoore:master. I wonder if such a class can handle backtest live trade but also paper trader (live price feed but simulated execution) I really think that an abstract class AbstractTradingSession and 3 inherited classes BacktestSession, LiveTradeSession, PaperTradeSession is a cleaner approach. But renaming simulate_trading method to start_trading (or simply start or run) is a good idea. No problem - promptly done! ;-) If you feel you need to change some of the logic in the previous (to this merge) Backtest object (which is now TradingSession), go ahead. I modified it to make the API cleaner, such that we didn't end up with these rather large ***_backtest.py files containing all of the instantiations of the respective components. I'm definitely open to trying to find ways to make this simpler for new users, but at the same time extensible for those who wish to tinker a bit more. Thanks @mhallsmoore @femtotrader with regard to paper trade, don't all brokers offer a separate paper trading account ? I.e. from the system point of view, live is the same as paper? I think even if we did do that, the trading loop logic should be the same; it's just the ExecutionHandler and Portfolio classes wouldn't have feeds to the broker for live execution/portfolio stats? Just IMO more abstraction here probably isn't going to be used too often? I think we should have abstract base classes where we expect that users will implement their own inherited classes "fairly" often. I guess if we find people often need to override the trading sessions, we can add the abstraction in.
2025-04-01T04:34:40.679033
2021-08-24T13:14:01
978086703
{ "authors": [ "mheob" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8382", "repo": "mheob/vscode-snippets", "url": "https://github.com/mheob/vscode-snippets/issues/4" }
gharchive/issue
[FR] Add next.js snippets To improve the extension, there should be the most common next.js snippets like for example: getStaticProps getServerSideProps … Fixed in #14
2025-04-01T04:34:40.691291
2024-04-15T15:42:24
2244014619
{ "authors": [ "akhiljns", "mhewedy" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8383", "repo": "mhewedy/vermin", "url": "https://github.com/mhewedy/vermin/issues/66" }
gharchive/issue
not able to run commands from shell script after provisioning vm Provisioning vm_01 : invalid optionline 3: set: - set: usage: set [-abefhkmnptuvxBCHP] [-o option-name] [--] [arg ...] /tmp/master.sh: line 4: $'\r': command not found /tmp/master.sh: line 8: $'\r': command not found Could not set property: Invalid pretty host name '<IP_ADDRESS> ' /tmp/master.sh: line 12: $'\r': command not found : No such file or directory /tmp/master.sh: line 14: $'\r': command not found Warning: apt-key output should not be parsed (stdout is not a terminal) Executing: /tmp/apt-key-gpghome.TIP5MpsQmS/gpg.1.sh --keyserver keyserver.ubuntu.com --recv-keys CC86BB64 " not a key ID: skipping Usage: add-apt-repository <sourceline> add-apt-repository is a script for adding apt sources.list entries. It can be used to add any repository and also provides a shorthand syntax for adding a Launchpad PPA (Personal Package Archive) repository. <sourceline> - The apt repository source line to add. This is one of: a complete apt line in quotes, a repo url and areas in quotes (areas defaults to 'main') a PPA shortcut. a distro component Examples: apt-add-repository 'deb http://myserver/path/to/repo stable myrepo' apt-add-repository 'http://myserver/path/to/repo myrepo' apt-add-repository 'https://packages.medibuntu.org free non-free' apt-add-repository http://extras.ubuntu.com/ubuntu apt-add-repository ppa:user/repository apt-add-repository ppa:user/distro/repository apt-add-repository multiverse If --remove is given the tool will remove the given sourceline from your sources.list add-apt-repository: error: no such option: - E: Command line option ' ' [from -y ] is not understood in combination with the other options. E: Command line option ' ' [from -y ] is not understood in combination with the other options. sudo: yq: command not found sudo: yq: command not found sudo: yq: command not found sudo: yq: command not found sudo: yq: command not found usage: /usr/sbin/netplan [-h] [--debug] ... /usr/sbin/netplan : error: argument : invalid choice: 'apply\r' (choose from 'help', 'apply', 'generate', 'get', 'info', 'ip', 'set', 'rebind', 'try') sleep: invalid time interval ‘1\r’ Try 'sleep --help' for more information. /tmp/master.sh: line 28: $'\r': command not found Warning: apt-key output should not be parsed (stdout is not a terminal) "pg: invalid option "- ' invalid /tmp/master.sh: line 34: deb [arch=amd64] https://download.docker.com/linux/ubuntu \ \bionic : No such file or directory Warning: apt-key output should not be parsed (stdout is not a terminal) "pg: invalid option "- /tmp/master.sh: line 78: warning: here-document at line 36 delimited by end-of-file (wanted `EOF') deb https://apt.kubernetes.io/ kubernetes-xenial main EOF sudo apt-get update sudo apt-get install -y \ containerd.io=1.2.13-1 \ \docker-ce=5:19.03.8~3-0~ubuntu-bionic docker-ce-cli=5:19.03.8~3-0~ubuntu-bionic -00 kubectl=1.18.0ll -y kubelet=1.18.0 sudo apt-mark hold docker-ce kubelet kubeadm kubectl echo "net.bridge.bridge-nf-call-iptables=1" | sudo tee -a /etc/sysctl.conf sudo sysctl -p ## configure docker cat <<EOF | sudo tee /etc/docker/daemon.json { "exec-opts": ["native.cgroupdriver=systemd"], "log-driver": "json-file", "log-opts": { "max-size": "100m" }, "storage-driver": "overlay2" } EOF @mhewedy so turns out this is not the problem of vermin the sh file i'm trying to run is master.sh inside /etc/samples-provision/k8s/master.sh the vagrant image i'm using is ubuntu/focal64 I tried running the file from inside the vm and i'm getting same issues @mhewedy I'm seeing strange behaviour when i'm trying to run the script using ./master.sh it fails but individual commands are running fine, this might be a problem of CRLF debugging more Ok upon debugging CRLF was the issue, but there are other issues as well with the master.sh the yq commands you have used are older and v4 version supports new syntax, i'm trying to make this script so that it supports newer builds Ok upon debugging CRLF was the issue, but there are other issues as well with the master.sh the yq commands you have used are older and v4 version supports new syntax, i'm trying to make this script so that it supports newer builds Good analysis, thanks for your efforts
2025-04-01T04:34:40.697579
2017-11-24T22:00:30
276703804
{ "authors": [ "jorgesca", "mhinz" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8384", "repo": "mhinz/neovim-remote", "url": "https://github.com/mhinz/neovim-remote/issues/68" }
gharchive/issue
Option NOT to start a neovim process? I don't know if this makes sense, here's the issue I have setup and IDE to start neovim for editing files. I use the terminal version of neovim. If the server has not been started when I ask the IDE to open a file, nvr will start a neovim instance, which actually blocks the IDE itself. Is there a way to avoid this? If I start a neovim server first, all works fine: $ NVIM_LISTEN_ADDRESS=/tmp/nvim_quartus nvim The IDE is configured to do this for opening files: /usr/bin/nvim --servername /tmp/nvim_quartus --remote +%l %f (where %l gets replaced by the line number and %f by the file name) Oh, I see. This was meant for people who alias nvr to nvim. I'll add a --nostart option! Excellent, thanks! I'll upload a new version together with some other fixes later today. Thanks for reporting!
2025-04-01T04:34:40.728135
2023-05-08T16:50:45
1700589161
{ "authors": [ "miacias" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8385", "repo": "miacias/brew-buddies", "url": "https://github.com/miacias/brew-buddies/issues/38" }
gharchive/issue
GoogleMaps: Create test map make a map on a React page, such as Map.js center the map on a location (set location to a variable so the user's home zip code can be used by default, else a standard address in USA) rendered with "@googlemaps/react-wrapper"
2025-04-01T04:34:40.737551
2020-02-26T00:19:28
570949485
{ "authors": [ "HewmEvor", "sajalmandrekar" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8386", "repo": "micahflee/torbrowser-launcher", "url": "https://github.com/micahflee/torbrowser-launcher/issues/454" }
gharchive/issue
SIGNATURE VERIFICATION FAILED Just following the instruction attached to failed verification to report this error here. Process used (can be duplicated (many times)) Use sudo to install TOR takes 10 minutes to start the d/l and install Launcher Open Launcher, and 13 minutes to start d/l actual package. .10 seconds to start Dl and install Verifying starts, and 13 minutes later, the generic error pops open. I have repeated the key resets, purge install, reboot, and then re-try. This has been tried several times using different sites different suggestions - nothing has worked. RIG: Intel i3-3110M CPU@ 2.40GHZ Memory 6G Ubuntu 18.04.3 LTS Tried install from: Software Boutique sudo app-get github.com/micahflee/torbrowser-launcher/issues/ Yes - this may be "similar to 4 existing issues" but not same. - it seems that same matter with this issue. It may be due to your Tor being outdated. Try updating it and see if it works.This repository provides a latest version. Run the following commands on terminal: sudo add-apt-repository ppa:micahflee/ppa sudo apt install torbrowser-launcher Or check out this thread: https://github.com/micahflee/torbrowser-launcher/issues/410
2025-04-01T04:34:40.739814
2015-04-01T19:32:03
65761659
{ "authors": [ "michael-lazar", "yskmt" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8387", "repo": "michael-lazar/rtv", "url": "https://github.com/michael-lazar/rtv/pull/59" }
gharchive/pull-request
comment/reply edit and delete functions added There are some common parts in add_comment, edit_comment and delete_comment, which can be refactored. Also, sometimes the comment deletion and edition are not immediately updated. Is it a PRAW problem? My guess would be that Reddit is caching the submission on the server side and only updates the cache every so often. I agree there is some work that could be done to refactor this. There is a lot of code that will be shared between the SubmissionPage and the SubredditPage. For deleting, how about a confirmation prompt Are you sure you want to delete this post? (y/n): In order to rebase to the newest master, I will close this pull request.
2025-04-01T04:34:40.752280
2020-11-21T20:32:58
748098084
{ "authors": [ "MaazArfi", "akkiiiiii" ], "license": "apache-2.0", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8388", "repo": "michael-rapp/ChromeLikeTabSwitcher", "url": "https://github.com/michael-rapp/ChromeLikeTabSwitcher/issues/33" }
gharchive/issue
Save instance I am making a browser app and I want to save the instance of each and every tab but I don't know how if someone knows how I can do this please tell me because I see in the issue section there are a lot of guys talking about they also used this library in their browsers so please tell me how can I save the state of each and every tab by using this library. there no inbuild method i guess you need write your own logic by using : room or sqlite
2025-04-01T04:34:40.767990
2015-12-01T01:54:29
119623340
{ "authors": [ "hut8", "izabelavmelo", "michaelbromley", "nathan-mickler" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8389", "repo": "michaelbromley/angularUtils", "url": "https://github.com/michaelbromley/angularUtils/issues/269" }
gharchive/issue
When calling $compile on a code containing dirPagination, registered pagination-id is not found I'm dynamically pulling in HTML code and running it through $compile after adding it to my DOM and everything seems to be working except that I'm now getting the following error message: pagination directive: the itemsPerPage id argument (id: locations) does not match a registered pagination-id It appears that the dir-paginate directive is getting compiled before the dir-pagination-controls directive and is therefore unable to find the registered pagination-id. How do I get around this? Here is a simple plunkr demonstrating the problem. The same data set is pulled in two times (on top: not working, and on bottom: working) http://plnkr.co/edit/t7YNiM Hi, The way the directive itself works involved dynamically calling $compile() after doing some manipulation - adding an ngRepeat directive etc. I can imagine there is some conflict arising perhaps to do with the order of execution of the calls to $compile. In your plunk I do not see the error you reported above - I get uncaught Error: [$compile:ctreq] Controller 'ngInclude', required by directive 'ngInclude', can't be found!. I can't say when I'll be able to put in some time to investigating this - I have limited free time to work on issues here at the moment. I will label it as a bug to fix though. I may have accidentally updated that plunkr since then with other attempts to get this working...I intended to create a new plunkr with updates. If you find out another way I can inject the code dynamically, please let me know. Thanks. On Tuesday, December 1, 2015, Michael Bromley<EMAIL_ADDRESS>wrote: Hi, The way the directive itself works involved dynamically calling $compile() after doing some manipulation - adding an ngRepeat directive etc. I can imagine there is some conflict arising perhaps to do with the order of execution of the calls to $compile. In your plunk I do not see the error you reported above - I get uncaught Error: [$compile:ctreq] Controller 'ngInclude', required by directive 'ngInclude', can't be found!. I can't say when I'll be able to put in some time to investigating this - I have limited free time to work on issues here at the moment. I will label it as a bug to fix though. — Reply to this email directly or view it on GitHub https://github.com/michaelbromley/angularUtils/issues/269#issuecomment-161072810 . I figured out a solution by editing dirPagination.js (see: http://plnkr.co/edit/KOzZGt) If I modify line #236 as follows, then it works to dynamically inject the content. var paginationId = scope.paginationId || attrs.paginationId || DEFAULT_ID; ...to... var paginationId = attrs.paginationId || DEFAULT_ID; Are there side effects that this will have that I'm nor aware of? Ok, thanks. I suspect that there is a specific reason why I wrote it that way, but perhaps it covers a case you don't need. I'll run it through the test suite when I get time and see what happens. Thanks for sharing your fix, that's very helpful in the mean time. Hi, Finally got round to looking at this again. Making your suggested change and running through the tests, I get a failure of this test dirPagination directive dynamic pagination ids should allow independent pagination FAILED Expected [ '1', '2', '3' ] to equal [ '4', '5' ]. So it's not viable for me to make that change in the source. I believe this is causing a problem I'm having with using this package with angular-ui-router. I'm able to reproduce it like this: Hit a page with the pagination directive. It works fine. Navigate to a different angular-ui-router state which also contains the directive (without reloading the entire page) It breaks I have limited time on this project so I can't look into a fix right now, sorry :frowning: Any ideas what a change might be that could fix it? We have the same problem in our project. We used more than 1 pagination in the same page, and this is being dynamically loaded (according to the number passed by the user). The scope in "dirPaginationCompileFn" is not accessible. So, when the pagination-id is dynamically passed, it can't evaluate "$parse(attrs.paginationId)(scope)"
2025-04-01T04:34:40.771266
2021-07-07T07:10:24
938565527
{ "authors": [ "michaelbrooks", "sandeepsn1997" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8390", "repo": "michaelbrooks/rq-win", "url": "https://github.com/michaelbrooks/rq-win/pull/11" }
gharchive/pull-request
fix for issue #9 => prepare_job_execution() takes 2 positional arguments but 3 were given "prepare_job_execution" takes 2 input param including self, but the calling statement passes 3 with additional "heartbeat_ttl" which is not used by the method, I have removed the extra input param and issue seems to be solved for me. Thank you! I see this interface was changed in rq here: https://github.com/rq/rq/pull/1372 Could you update setup.py to version 0.4.0 and make the minimum rq version 1.8.1? I'm not sure that this will behave as expected with older rq versions. I have updated setup.py to version 0.4.0 and made the minimum rq version 1.8.1 Thank you!
2025-04-01T04:34:40.881008
2015-12-18T08:05:58
122901507
{ "authors": [ "C-Pro", "michaelpq" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8391", "repo": "michaelpq/michaelpq.github.io", "url": "https://github.com/michaelpq/michaelpq.github.io/issues/2" }
gharchive/issue
Future dates on past presentations Hi! Finally fixed!
2025-04-01T04:34:40.889466
2024-10-18T19:27:58
2598240726
{ "authors": [ "michaelrsweet" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8392", "repo": "michaelrsweet/pappl", "url": "https://github.com/michaelrsweet/pappl/issues/370" }
gharchive/issue
Bring back CUPS 2.x support - 2.5 or later... Since CUPS 2.5 has the necessary APIs to support PAPPL 2.0, bring back the CUPS 3 -> 2.5 wrappers. [master b494f1d] Restore ability to build against CUPS 2.5.
2025-04-01T04:34:40.935436
2016-12-06T22:30:50
193904282
{ "authors": [ "lizzzp1", "orrybaram" ], "license": "mit", "license_source": "bigquery", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8393", "repo": "micnews/article-json-to-amp", "url": "https://github.com/micnews/article-json-to-amp/pull/38" }
gharchive/pull-request
Ads ADS ADS ADS More ads mean more 💰💰💰 so we're adding ads back into ⚡⚡⚡AMP⚡⚡⚡ WHO's GONNA LOOK? how about one of the previous contributors? [ ] @kesla [ ] @iefserge [ ] @ellell or maybe someone else? [ ] @stankud [ ] @ryanscottaudio [ ] @danmakenoise [x] @lizzzp1 LGTM what does ava do out of curiosity? Ava is a test runner, similar to tape but it runs the tests asynchronous. On Tue, Dec 6, 2016, 5:45 PM Liz<EMAIL_ADDRESS>wrote: LGTM what does ava do out of curiosity? — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/micnews/article-json-to-amp/pull/38#issuecomment-265297473, or mute the thread https://github.com/notifications/unsubscribe-auth/AAsKey5i3qeVG4bop7qy5Nbh1PT1PZS5ks5rFeVxgaJpZM4LF90j .
2025-04-01T04:34:40.948385
2023-03-03T12:56:26
1608584266
{ "authors": [ "AjinJayan", "htex0189" ], "license": "Apache-2.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8394", "repo": "micro-ROS/micro_ros_mbed", "url": "https://github.com/micro-ROS/micro_ros_mbed/issues/34" }
gharchive/issue
Creating Multiple publisher Hi, I run the example with one publisher and it is working. I can read the message with micro_ros agent. But when adding another publisher, the new publisher is not registered with micro_ros agent. rclc_publisher_init_default returns error at second publisher. So How to create multiple publisher. And is it necessary to use executors with timer to publish or publish messages directly without using executors? Thanks Is the meta config for rmw_microxrcedds for mbed is not yet developed to configure max publisher,subscriber,node for mbed? I changed the colcon meta config - increased the number of publisher subscribers. I still cannot create two publisher. Please could you tell what should be done after changing colcon meta file Thanks that was really helpful
2025-04-01T04:34:40.967993
2023-12-15T18:30:34
2044186955
{ "authors": [ "mbthornton-lbl", "turbomam" ], "license": "CC0-1.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8395", "repo": "microbiomedata/nmdc-schema", "url": "https://github.com/microbiomedata/nmdc-schema/issues/1578" }
gharchive/issue
Create before/after diagram of Workflow-related schema changes Add a tab to the existing diagram: https://app.diagrams.net/#G1Ufsblf98rGzRhJMDt7ZMgosioywnGJIU Create a before and after high-level schema diagrams and expected workflow-related changes Could we please use an automated diagram generator for the after state?
2025-04-01T04:34:40.984169
2023-01-28T00:42:15
1560600947
{ "authors": [ "HyperTesto", "gRegorLove" ], "license": "CC0-1.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8396", "repo": "microformats/php-mf2", "url": "https://github.com/microformats/php-mf2/issues/246" }
gharchive/issue
Properties not parsed when using Tailwind height class I managed to get a minimal example that happened to me while working with Tailwind CSS, but i guess it's more generic to any class that starts with h- prefix. If i use Tailwind CSS h-<number> class on some elements that have other microformat classes like in this minimal example with u-logo and u-nickname: <article class="h-card"> <h1> TITLE </h1> <img class="h-36 photo u-logo" alt="An example alt title" src="https://example.com/img.png" /> <p class="h-21 u-nickname"> John Doe </p> </article> The parser skips mf properties and produces : { "items": [ { "type": [ "h-card" ], "properties": { "name": [ "TITLE An example alt title\nJohn Doe" ] } } ], "rels": {}, "rel-urls": {}, "debug": { "package": "https://packagist.org/packages/mf2/mf2", "source": "https://github.com/indieweb/php-mf2", "version": "v0.5.0", "note": [ "This output was generated from the php-mf2 library available at https://github.com/indieweb/php-mf2", "Please file any issues with the parser at https://github.com/indieweb/php-mf2/issues", "Using the Masterminds HTML5 parser" ] } } Instead, if I remove h- classes, it's working as expected: <article class="h-card"> <h1> TITLE </h1> <img class="photo u-logo" alt="An example alt title" src="https://example.com/img.png" /> <p class="u-nickname"> John Doe </p> </article> with output: { "items": [ { "type": [ "h-card" ], "properties": { "logo": [ { "value": "https://example.com/img.png", "alt": "An example alt title" } ], "nickname": [ "John Doe" ], "name": [ "TITLE An example alt title\nJohn Doe" ] } } ], "rels": {}, "rel-urls": {}, "debug": { "package": "https://packagist.org/packages/mf2/mf2", "source": "https://github.com/indieweb/php-mf2", "version": "v0.5.0", "note": [ "This output was generated from the php-mf2 library available at https://github.com/indieweb/php-mf2", "Please file any issues with the parser at https://github.com/indieweb/php-mf2/issues", "Using the Masterminds HTML5 parser" ] } } This scenario seems to work using the Go implementation Good catch! The specification is to skip classes with numbers, but it shouldn't be skipping the entire element. Will look into it.
2025-04-01T04:34:41.161848
2019-12-21T17:19:17
541349774
{ "authors": [ "Ba4bes", "johnpapa" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8397", "repo": "microsoft/25-days-of-serverless", "url": "https://github.com/microsoft/25-days-of-serverless/issues/364" }
gharchive/issue
[CHALLENGE SUBMISSION] - day 21 Please provide the following information for your submission: What is your name? (First, Last) Barbara Forbes Where is your GitHub Repo ?(With your challenge solution) https://github.com/Ba4bes/25daysofserverless/tree/master/Daya21GiftRegistries What Challenge is this for? (Challenge number between 1 and 25) 21 (Optional) Anything else we should know? e.g., language used, location, blog post? PowerShell Function App and CosmosDB The unthinkable has happened: after weeks of community members all over the globe chipping in to help solve people's problems with serverless technology, the evil grinch's heart is starting to soften! He's still not quite ready to give up his server-stealing plot, but our whole Azure Cloud Advocacy team is flying up to the grinch's lair to help convince him to change his ways. As such, the team behind 25 Days of Serverless is OOF (Out Of Facility/Office) between Dec 24, 2019 and January 1, 2020. We'll not review Issues in this period of time, but encourage you to catch up on any challenges you haven't had the time to work on yet. 😄 Our Discord Server is a perfect place to reach out to for community support during this time. After our return, we'll add contributors who submitted a working solution as an Issue to CONTRIBUTORS.md file until January 15, 2020.
2025-04-01T04:34:41.167236
2024-04-17T08:01:18
2247646413
{ "authors": [ "AskeHolst", "akoniecki", "florian-d" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8398", "repo": "microsoft/AL-Go", "url": "https://github.com/microsoft/AL-Go/issues/1041" }
gharchive/issue
IncrementVersionNumber - the new version number is less than the old version number Hi, I have the following problem with the "IncrementVersionNumber" workflow: [IncrementVersionNumber] Unexpected error when running action. Error Message: The new version number (1.10) is less than the old version number (1.9). The version number must be incremented., StackTrace: at Set-VersionInSettingsFile, D:\a\_actions\microsoft\AL-Go-Actions\v5.0\IncrementVersionNumber\IncrementVersionNumber.psm1: line 108 <- at <ScriptBlock>, D:\a\_actions\microsoft\AL-Go-Actions\v5.0\IncrementVersionNumber\IncrementVersionNumber.ps1: line 46 <- at <ScriptBlock>, D:\a\_temp\989f38ee-08d1-4139-9f8b-28d0b976b3b7.ps1: line 4 <- at <ScriptBlock>, <No file>: line 1 I couldn't find the exact reason in the code, but it seems the System.Version cast is not applied correctly to variables before ($newValue -lt $oldValue) comparison. These are the steps to recreate the issue: Create new repo from AL-Go-PTE template Wait until Initial Build is done Run "Create new app" workflow with sample code parameter Run "Increment Version Number" workflow with Updated Version Number parameter: 1.9 Run "Increment Version Number" workflow with Updated Version Number parameter: +0.1 Here is my public repo with the issue recreated: https://github.com/akoniecki/versionincrement-debugging/actions/runs/8718133126/job/23914709863#step:7:60 We just got the exact same issue. Error: Unexpected error when running action. Error Message: The new version number (2.10) is less than the old version number (2.9). The version number must be incremented., StackTrace: at Set-VersionInSettingsFile, /home/runner/work/_actions/microsoft/AL-Go/f2f452940bc16f36a4c59f1fd1982c566f16c03c/Actions/IncrementVersionNumber/IncrementVersionNumber.psm1: line 108 <- at <ScriptBlock>, /home/runner/work/_actions/microsoft/AL-Go/f2f452940bc16f36a4c59f1fd1982c566f16c03c/Actions/IncrementVersionNumber/IncrementVersionNumber.ps1: line 46 <- at <ScriptBlock>, /home/runner/work/_temp/a3b6a4b4-9bc3-4c1d-ab35-7258a178aaaf.ps1: line 4 <- at <ScriptBlock>, <No file>: line 1 After updating AL-GO, we encountered the same error during the ‘Create Release’ process. We’re also using AL-Go-PTE. Unexpected error when running action. Error Message: The new version number (<IP_ADDRESS>) is less than the old version number (<IP_ADDRESS>). The version number must be incremented., StackTrace: at Set-VersionInSettingsFile, D:\a\_actions\microsoft\AL-Go-Actions\v5.0\IncrementVersionNumber\IncrementVersionNumber.psm1: line 108 <- at <ScriptBlock>, D:\a\_actions\microsoft\AL-Go-Actions\v5.0\IncrementVersionNumber\IncrementVersionNumber.psm1: line 154 <- at Set-VersionInAppManifests, D:\a\_actions\microsoft\AL-Go-Actions\v5.0\IncrementVersionNumber\IncrementVersionNumber.psm1: line 150 <- at <ScriptBlock>, D:\a\_actions\microsoft\AL-Go-Actions\v5.0\IncrementVersionNumber\IncrementVersionNumber.ps1: line 53 <- at <ScriptBlock>, D:\a\_temp\d26634c8-a3cc-4341-991e-cf830077b236.ps1: line 4 <- at <ScriptBlock>, <No file>: line 1
2025-04-01T04:34:41.175046
2024-12-02T08:56:06
2711109815
{ "authors": [ "dibyanugraha", "freddydk", "frottke" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8399", "repo": "microsoft/AL-Go", "url": "https://github.com/microsoft/AL-Go/issues/1344" }
gharchive/issue
[Bug]: Experimental feature "git submodules" seems to be a breaking change (in some ways) AL-Go version v6.1 Describe the issue We have identified a potential bug in version 6.1: all submodules, regardless of their location within the repository, are being downloaded during the git checkout process. To provide some context, our repository’s root folder contains a Git submodule for internal scripts, which serves as a convenient way to update them globally. Unfortunately, the workflow attempts to download this submodule, despite not having the necessary permissions to do so. I truly appreciate all the new features, especially experimental ones. However, would it be possible to "hide" such features behind a feature flag or check? If needed, I would be happy to grant access to a repository where this issue can be reproduced. Expected behavior We would have expected that only app/project paths are checked for Git submodules, rather than the entire repository. Alternatively, it would be helpful if authentication could be configured specifically for submodules. Steps to reproduce first variation: create a submodule that is not in an configured app/project path then run CI/CD or PR Build. See that the git submodules is cloned. second variation: create a submodule where git (cli, github_token) has no permissions to access the repository. It looks like that GHTOKENWORKFLOW is not used. Additional context (logs, screenshots, etc.) No response Got it. We cannot really use the GhTokenWorkflow for this as we need to clone the repo in order to get to the settings / secrets. But... - I see that this change will break you when having submodules, which the standard github token cannot access. We will investigate what we can do about this - in the meantime, you would have to revert back to v6.0 - then we will have a look at what we can do for this. Thank you Freddy for the, as always, quick response! :) We’ve already rolled back to v6.0, knowing this would be not a fast fix. One more remark: Isn't every repository other than the current one out of scope for the default github_token? In other words, how should a submodule repository be configured to make it accessible? Currently, I don't see any other option except granting additional permissions to the checkout action (GitHub token). How do you structure your repos? If they reside in the same organization, you can use relative path instead of full url path in your submodule. My apologies I have to refer to outside blog, but this blog gives a good example how to change your submodule into relative path (given with above circumstances): https://www.damirscorner.com/blog/posts/20210423-ChangingUrlsOfGitSubmodules.html @dibyanugraha - I don't think the github token from one repository would be able to pull the submodule of another repository even with relative urls. The default github token has very limited permissions I think Just tested this - for private repositories, we need an auth token on the actions/checkout - the default github.token won't suffice. Currently, we do not have a token for use at that time, so we will have to refactor the support for submodules if we are to support that. Yes, we can pass token to the checkout itself, but unfortunately the github_token workflow won't be suffice. It must be a security token variable in GitHub that has 'Read' access to submodule repositories. I will check if there is an easy way to support submodules with this knowledge - else we will have to pull the experimental support at this time.
2025-04-01T04:34:41.177301
2021-01-20T08:39:14
789763887
{ "authors": [ "JesperSchulz", "VMotin" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8400", "repo": "microsoft/ALAppExtensions", "url": "https://github.com/microsoft/ALAppExtensions/issues/10637" }
gharchive/issue
[Request-for-external] Codeunit 90 Purch.-Post - CalcInvoiceDiscountPosting Hi, I was wondering if it would be possible to change the procedure CalcInvoiceDiscountPosting Codeunit 90 from local to external. My team and I use this procedure a lot in our code, and changing it to external would be a huge help. Thanks for your work, have a nice day Thanks for reporting this. We agree, and we’ll publish a fix asap, either in an update for the current version or in the next major release. We will update this issue with information about availability. Please do not reply to this, as we do not monitor closed issues. If you have follow-up questions or requests, please create a new issue where you reference this one.
2025-04-01T04:34:41.180160
2021-08-19T13:13:16
974650639
{ "authors": [ "JesperSchulz", "fridrichovsky" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8401", "repo": "microsoft/ALAppExtensions", "url": "https://github.com/microsoft/ALAppExtensions/issues/14049" }
gharchive/issue
[Event Request] Codeunit 6620 "Copy Document Mgt".InitSalesLineFields Please add new event local procedure InitSalesLineFields(var ToSalesLine: Record "Sales Line") //----------------------------------------OnInitSalesLineFieldsOnBeforeSetJob:BEGIN var IsHandled: boolean; //----------------------------------------OnInitSalesLineFieldsOnBeforeSetJob:END begin OnBeforeInitSalesLineFields(ToSalesLine); if ToSalesLine."Document Type" <> ToSalesLine."Document Type"::Order then begin ToSalesLine."Prepayment %" := 0; ToSalesLine."Prepayment VAT %" := 0; ToSalesLine."Prepmt. VAT Calc. Type" := "Tax Calculation Type"::"Normal VAT"; ToSalesLine."Prepayment VAT Identifier" := ''; ToSalesLine."Prepayment VAT %" := 0; ToSalesLine."Prepayment Tax Group Code" := ''; ToSalesLine."Prepmt. Line Amount" := 0; ToSalesLine."Prepmt. Amt. Incl. VAT" := 0; end; ToSalesLine."Prepmt. Amt. Inv." := 0; ToSalesLine."Prepmt. Amount Inv. (LCY)" := 0; ToSalesLine."Prepayment Amount" := 0; ToSalesLine."Prepmt. VAT Base Amt." := 0; ToSalesLine."Prepmt Amt to Deduct" := 0; ToSalesLine."Prepmt Amt Deducted" := 0; ToSalesLine."Prepmt. Amount Inv. Incl. VAT" := 0; ToSalesLine."Prepayment VAT Difference" := 0; ToSalesLine."Prepmt VAT Diff. to Deduct" := 0; ToSalesLine."Prepmt VAT Diff. Deducted" := 0; ToSalesLine."Prepmt. Amt. Incl. VAT" := 0; ToSalesLine."Prepmt. VAT Amount Inv. (LCY)" := 0; ToSalesLine."Quantity Shipped" := 0; ToSalesLine."Qty. Shipped (Base)" := 0; ToSalesLine."Return Qty. Received" := 0; ToSalesLine."Return Qty. Received (Base)" := 0; ToSalesLine."Quantity Invoiced" := 0; ToSalesLine."Qty. Invoiced (Base)" := 0; ToSalesLine."Reserved Quantity" := 0; ToSalesLine."Reserved Qty. (Base)" := 0; ToSalesLine."Qty. to Ship" := 0; ToSalesLine."Qty. to Ship (Base)" := 0; ToSalesLine."Return Qty. to Receive" := 0; ToSalesLine."Return Qty. to Receive (Base)" := 0; ToSalesLine."Qty. to Invoice" := 0; ToSalesLine."Qty. to Invoice (Base)" := 0; ToSalesLine."Qty. Shipped Not Invoiced" := 0; ToSalesLine."Return Qty. Rcd. Not Invd." := 0; ToSalesLine."Shipped Not Invoiced" := 0; ToSalesLine."Return Rcd. Not Invd." := 0; ToSalesLine."Qty. Shipped Not Invd. (Base)" := 0; ToSalesLine."Ret. Qty. Rcd. Not Invd.(Base)" := 0; ToSalesLine."Shipped Not Invoiced (LCY)" := 0; ToSalesLine."Return Rcd. Not Invd. (LCY)" := 0; //----------------------------------------OnInitSalesLineFieldsOnBeforeSetJob:BEGIN OnInitSalesLineFieldsOnBeforeSetJob(ToSalesLine,IsHandled) if not IsHandled then begin //----------------------------------------OnInitSalesLineFieldsOnBeforeSetJob:END ToSalesLine."Job No." := ''; ToSalesLine."Job Task No." := ''; ToSalesLine."Job Contract Entry No." := 0; //----------------------------------------OnInitSalesLineFieldsOnBeforeSetJob:BEGIN end; //----------------------------------------OnInitSalesLineFieldsOnBeforeSetJob:END OnAfterInitSalesLineFields(ToSalesLine); end; //----------------------------------------OnInitSalesLineFieldsOnBeforeSetJob:BEGIN [IntegrationEvent(false, false)] local procedure OnInitSalesLineFieldsOnBeforeSetJob(var ToSalesLine: Record "Sales Line"; var IsHandled: boolean) begin end; //----------------------------------------OnInitSalesLineFieldsOnBeforeSetJob:END Thanks for reporting this. We agree, and we’ll publish a fix asap, either in an update for the current version or in the next major release. Please do not reply to this, as we do not monitor closed issues. If you have follow-up questions or requests, please create a new issue where you reference this one.
2025-04-01T04:34:41.182146
2021-10-13T20:50:06
1025683915
{ "authors": [ "JesperSchulz", "ajkauffmann" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8402", "repo": "microsoft/ALAppExtensions", "url": "https://github.com/microsoft/ALAppExtensions/issues/14766" }
gharchive/issue
[Event parameter request] codeunit 5431 "Calc. Item Plan - Plan Wksh." - OnCodeOnAfterSetReqLineFilters Please add CurrTemplateName and CurrWorksheetName to event OnCodeOnAfterSetReqLineFilters in codeunit 5431 "Calc. Item Plan - Plan Wksh." Original code: OnCodeOnAfterSetReqLineFilters(ReqLineExtern); [IntegrationEvent(false, false)] local procedure OnCodeOnAfterSetReqLineFilters(var ReqLineExtern: Record "Requisition Line") begin end; Modified code: OnCodeOnAfterSetReqLineFilters(ReqLineExtern, CurrTemplateName, CurrWorksheetName); [IntegrationEvent(false, false)] local procedure OnCodeOnAfterSetReqLineFilters(var ReqLineExtern: Record "Requisition Line"; CurrTemplateName: Code[10]; CurrWorksheetName: Code[10]) begin end; Thanks for reporting this. We agree, and we’ll publish a fix asap, either in an update for the current version or in the next major release. Please do not reply to this, as we do not monitor closed issues. If you have follow-up questions or requests, please create a new issue where you reference this one.
2025-04-01T04:34:41.183767
2021-11-22T14:34:59
1060230973
{ "authors": [ "JesperSchulz", "iltererdogan" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8403", "repo": "microsoft/ALAppExtensions", "url": "https://github.com/microsoft/ALAppExtensions/issues/15285" }
gharchive/issue
[Event Request] table 5902 "Service Line" - OnBeforeCalcInvDiscToInvoice Hi, Is it possible to add OnBeforeCalcInvDiscToInvoice publisher to service line? procedure InitQtyToInvoice() begin "Qty. to Invoice" := MaxQtyToInvoice; "Qty. to Invoice (Base)" := MaxQtyToInvoiceBase; "VAT Difference" := 0; // <<New OnBeforeCalcInvDiscToInvoice(Rec, CurrFieldNo); // >> New CalcInvDiscToInvoice; OnAfterInitQtyToInvoice(Rec, CurrFieldNo); end; // <<New [IntegrationEvent(false, false)] local procedure OnBeforeCalcInvDiscToInvoice(var ServLine: Record "Service Line"; CallingFieldNo: Integer) begin end; // >> New Thanks for reporting this. We agree, and we’ll publish a fix asap, either in an update for the current version or in the next major release. Please do not reply to this, as we do not monitor closed issues. If you have follow-up questions or requests, please create a new issue where you reference this one.
2025-04-01T04:34:41.186089
2021-12-11T11:35:14
1077511652
{ "authors": [ "JesperSchulz", "auzhhv" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8404", "repo": "microsoft/ALAppExtensions", "url": "https://github.com/microsoft/ALAppExtensions/issues/15510" }
gharchive/issue
[Event request] table 38 "Purchase Header" - OnAfterTestNoSeries Can you add PurchSetup to event OnAfterTestNoSeries in table 38 "Purchase Header"? procedure TestNoSeries() var IsHandled: Boolean; begin GetPurchSetup(); IsHandled := false; OnBeforeTestNoSeries(Rec, IsHandled); if not IsHandled then case "Document Type" of "Document Type"::Quote: PurchSetup.TestField("Quote Nos."); "Document Type"::Order: PurchSetup.TestField("Order Nos."); "Document Type"::Invoice: begin PurchSetup.TestField("Invoice Nos."); PurchSetup.TestField("Posted Invoice Nos."); end; "Document Type"::"Return Order": PurchSetup.TestField("Return Order Nos."); "Document Type"::"Credit Memo": begin PurchSetup.TestField("Credit Memo Nos."); PurchSetup.TestField("Posted Credit Memo Nos."); end; "Document Type"::"Blanket Order": PurchSetup.TestField("Blanket Order Nos."); end; // >>>>>>>>>>>>>>> //OnAfterTestNoSeries(Rec); OnAfterTestNoSeries(Rec, PurchSetup); // <<<<<<<<<<<<<<< end; Thanks for reporting this. We agree, and we’ll publish a fix asap, either in an update for the current version or in the next major release. Please do not reply to this, as we do not monitor closed issues. If you have follow-up questions or requests, please create a new issue where you reference this one.
2025-04-01T04:34:41.196360
2024-04-09T11:22:15
2233216701
{ "authors": [ "JesperSchulz", "pri-kise" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8405", "repo": "microsoft/ALAppExtensions", "url": "https://github.com/microsoft/ALAppExtensions/issues/26251" }
gharchive/issue
[Event Change Request] codeunit 73 "Purch.-Explode BOM" - OnBeforeInsertExplodedPurchLine Describe the request local procedure ExplodeBOMCompLines(PurchLine: Record "Purchase Line") var PreviousPurchLine: Record "Purchase Line"; Resource: Record Resource; InsertLinesBetween: Boolean; SkipComponent: Boolean; begin ... if PurchHeader."Language Code" = '' then ToPurchLine.Description := FromBOMComp.Description else if not ItemTranslation.Get(FromBOMComp."No.", FromBOMComp."Variant Code", PurchHeader."Language Code") then ToPurchLine.Description := FromBOMComp.Description; //+Change //-OnBeforeInsertExplodedPurchLine(ToPurchLine, PurchLine, FromBOMComp); OnBeforeInsertExplodedPurchLine(ToPurchLine, PurchLine, FromBOMComp, PurchHeader); //-Change ToPurchLine.Insert(); OnAfterInsertExplodedPurchLine(ToPurchLine, PurchLine, FromBOMComp); ... end; [IntegrationEvent(false, false)] //+Change //-local procedure OnBeforeInsertExplodedPurchLine(var ToPurchaseLine: Record "Purchase Line"; PurchaseLine: Record "Purchase Line"; FromBOMComp: Record "BOM Component") local procedure OnBeforeInsertExplodedPurchLine(var ToPurchaseLine: Record "Purchase Line"; PurchaseLine: Record "Purchase Line"; FromBOMComp: Record "BOM Component"; PurchaseHeader: Record "Purchase Header";) //-Change begin end; Additional context We need Access to a field of the header record in this event. Internal work item: AB#524721 Thanks for reporting this. We agree, and we’ll publish a fix asap, either in an update for the current version or in the next major release. Please do not reply to this, as we do not monitor closed issues. If you have follow-up questions or requests, please create a new issue where you reference this one. Build ID: 18489.
2025-04-01T04:34:41.206416
2018-11-13T14:20:04
380248292
{ "authors": [ "BertDeTemmerman", "wbor-nvt" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8406", "repo": "microsoft/ALAppExtensions", "url": "https://github.com/microsoft/ALAppExtensions/issues/342" }
gharchive/issue
[Event-Request] REP2000005 File SEPA Payments We want to create the (NON-)SEPA payments file automatically. Therefore we would like an event/function that allows us to set the parameters of the requestpage on the file payments reports and also another to hide that requestpage. When clicking the ExportPaymentLines action on the PAG2000001 EB Payment Journal, the following code gets executed: In TAB2000005 Export Protocol ExportPaymentLines(VAR PmtJnlLine : Record "Payment Journal Line") IF CheckPaymentLines(PmtJnlLine) THEN BEGIN TESTFIELD("Export Object ID"); PmtJnlLineToExport.COPY(PmtJnlLine); PmtJnlLineToExport.SETRANGE(Status,PmtJnlLineToExport.Status::Created); PmtJnlLineToExport.SETRANGE("Export Protocol Code",Code); PmtJnlLineToExport.SETRANGE("Journal Batch Name",PmtJnlLine."Journal Batch Name"); PmtJnlLineToExport.SETRANGE("Journal Template Name",PmtJnlLine."Journal Template Name"); IF "Export Object Type"="Export Object Type"::Report THEN ////Here we would need something to hide the requestpage ////Such as: OnBeforeRunReportExportPaymentLines(VAR Rec, VAR ShowRequestpage, VAR PmtJnlLineToExport) REPORT.RUNMODAL("Export Object ID",TRUE,FALSE,PmtJnlLineToExport) ELSE BEGIN IF PmtJnlLine."Exported To File" THEN IF NOT CONFIRM(ExportAgainQst) THEN EXIT; In both File payment reports (REP2000005 File SEPA Payments and REP2000006 File Non Euro SEPA Payments), we would need an event that's let us fill in the requestpage. Something like OnAfterInit(VAR GenJnlLine, VAR AutomaticPosting, VAR IncludeDimText, VAR ExecutionDate, VAR FileName) Could these be provided please? :) Version BE Dynamics NAV 13.0 (25789) We are also working on a solution that requires running REP2000005 & REP2000006 with no Request Page. TAB2000005 has already received the necessary OnBeforeExportPaymentLinesOnRunReport event. There is still no way to alter the Global variables set in the Request page however. Is this still planned for future update? Ideal would be an OnBeforePreReport trigger OnBeforePreReport(VAR GenJnlLine, VAR AutomaticPosting, VAR IncludeDimText, VAR ExecutionDate, VAR FileName) @BertDeTemmerman Did you get your requested event in the meanwhile, to set the request page parameters, such as ExecutionDate? Something like OnAfterInit(VAR GenJnlLine, VAR AutomaticPosting, VAR IncludeDimText, VAR ExecutionDate, VAR FileName) Short: No. Long: No. But we can skip the 2nd report now ... I think we just copied those reports or didn't implement this. local procedure PostPaymentJournal(var GenJnlLine: Record "Gen. Journal Line"; var PaymentJournalLine: Record "Payment Journal Line"; BalancingPostingDate: Date) var PaymentJournalPost: Report "Payment Journal Post"; IsHandled: Boolean; begin IsHandled := false; OnBeforePostPaymentJournal(GenJnlLine, PaymentJournalLine, AutomaticPosting, BalancingPostingDate, IsHandled); if IsHandled then exit; PaymentJournalPost.SetParameters(GenJnlLine, AutomaticPosting, REPORT::"File SEPA Payments", BalancingPostingDate); PaymentJournalPost.SetTableView(PaymentJournalLine); PaymentJournalPost.RunModal; end;
2025-04-01T04:34:41.208936
2020-07-23T11:09:58
664388705
{ "authors": [ "bc-ghost", "osteeger" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8407", "repo": "microsoft/ALAppExtensions", "url": "https://github.com/microsoft/ALAppExtensions/issues/7963" }
gharchive/issue
[Event Request] codeunit 481 "Check Dimensions" - Event for procedure CheckPurchDimLines Can we get the following event for the "Check Dimensions" Codeunit: OnTempPurchLineRepeat local procedure CheckPurchDimLines(PurchHeader: Record "Purchase Header"; var TempPurchLine: Record "Purchase Line" temporary) begin with TempPurchLine do begin Reset; SetFilter(Type, '<>%1', Type::" "); if FindSet then repeat if (PurchHeader.Receive and ("Qty. to Receive" <> 0)) or (PurchHeader.Invoice and ("Qty. to Invoice" <> 0)) or (PurchHeader.Ship and ("Return Qty. to Ship" <> 0)) then begin CheckPurchDimCombLine(TempPurchLine); CheckPurchDimValuePostingLine(TempPurchLine); OnTempPurchLineRepeat(TempPurchLine); //NEW end until Next = 0; end; end; [IntegrationEvent(false, false)] local procedure OnTempPurchLineRepeat(var TempPurchLine: Record "Purchase Line") begin end; Thanks for reporting this. We agree, and we’ll publish a fix asap, either in an update for the current version or in the next major release. We will update this issue with information about availability.
2025-04-01T04:34:41.212484
2020-08-19T07:14:27
681606280
{ "authors": [ "RebeccaAnne", "animia" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8408", "repo": "microsoft/AdaptiveCards", "url": "https://github.com/microsoft/AdaptiveCards/issues/4614" }
gharchive/issue
Disabled button is not supported. What platform is your feature request related to? (Delete other platforms). .NET HTML Android iOS (Swift/Objective C) JavaScript Is your feature request related to a problem? Please describe. When user click any button, we want to change button into disabled status. Describe the solution you'd like When user click any button, we want to change button into disabled button in local. It means button is changed into dimmed status without server's interaction. Describe alternatives you've considered I think the alternative is dimmed button image which cannot be clicked. Additional context This will be introduced with our 1.5 release. See #4670
2025-04-01T04:34:41.230653
2019-12-25T22:26:58
542390571
{ "authors": [ "MSNev", "Malvoz" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8409", "repo": "microsoft/ApplicationInsights-JS", "url": "https://github.com/microsoft/ApplicationInsights-JS/issues/1148" }
gharchive/issue
Set a long max-age for immutable resources (I'm personally not a user of the Application Insights SDK, but as I stumbled on this I thought I'd file an issue, I'm hopefully in the right place.) Considering not all browsers support the immutable directive, these resources should typically include a long max-age, conventionally: max-age=31536000. https://az416426.vo.msecnd.net/scripts/a/ai.0.js returns: Cache-Control: public, max-age=86400, immutable and: https://github.com/microsoft/ApplicationInsights-JS/blob/d0184962d439fcd448f18cfa8400e70a4cf57f5e/AISKU/scripts/uploadFilesToAzureStorage.ps1#L2 Hi Malvoz, These short times are actually required as ai.0.js will return the current version of the script. With these set to a day (86400), this means that if we happen to publish a breaking change then browsers will only keep this cached for a max of 24 hours. If we set this to 1 year (365 days) as suggested that we would break ever user for an entire year AND we could also never publish any update within the same sliding window (as it's relative to when it's uploaded and/or (depending on browsers / proxies that are caching it) first fetched by an end user. While i agree that 1 day is not ideal (as we don't publish that often) it is a trade off and to correctly handle this we would need another trade off, by including an additional fetch. Something like -> Request to ai.0.js => causes a 302 or client side redirect script to fetch the true versioned (with a hash) -> This hashed file can then have the 1 year (or more policy) as it should never change (based on the hashed content). The real downside of this is for first time users / uncached requests, there would now be 2 requests. This requires some infrastructure changes which will need to be addressed as part of a future work item and some analysis on the impacts. For now I'm going to close this issue as I don't believe that it will be addressed in the near term -- but I will be working to have this addressed in the longer term -- just not for the legacy version (ai.0.js).
2025-04-01T04:34:41.232449
2020-05-26T18:09:55
625083432
{ "authors": [ "anthturner" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8410", "repo": "microsoft/AuthJanitor", "url": "https://github.com/microsoft/AuthJanitor/pull/63" }
gharchive/pull-request
Switch to SecureString for RegeneratedSecret @ericmaino suggested using SecureString for the RegeneratedSecret so that we don't accidentally serialize sensitive information. I totally agree, and since System.Security.SecureString is now available in netstandard2.1 it's usable. Unfortunately the support for it in other libraries leaves a lot to be desired, so this PR also includes extension methods to transform back and forth between a SecureString and a string, understanding that doing so compromises in-memory confidentiality of the string. This is old enough that it was easier to just recreate the commit. Closing.
2025-04-01T04:34:41.241318
2023-12-20T07:56:36
2050038535
{ "authors": [ "keerthiduvvuri", "kupatkar99" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8411", "repo": "microsoft/AzureStorageExplorer", "url": "https://github.com/microsoft/AzureStorageExplorer/issues/7615" }
gharchive/issue
A11y_Azure Tools Storage Explorer_Disk_DiskToolbar_Usable: No Status message is Conveyed to User that refresh completed. Preflight Checklist [X] I have installed the latest version of Storage Explorer. [X] I have checked existing resources, including the troubleshooting guide and the release notes. [X] I have searched for similar issues. Storage Explorer Version 1.32.1(96) Regression From NO Architecture x64 Storage Explorer Build Number 20231114.10 Platform macOS OS Version MAC Version 13.6.1 Bug Description GitHub Tags: #A11yUsable;#A11yTCS;#MAC;#DesktopApp;#A11ySev4;#BM_AzureToolsStorageExplorer_MAC_Dec2023;#Azure Tools Storage Explorer;#FTP;#Voiceover; Environment Details: Application: Microsoft Azure Storage Explore version: 1.32.1 (96) Screen Reader: Voice Over Steps to Reproduce Launch Storage Explorer Tab Navigate to "Open Connect dialog" pane and Sign into Azure. Tab Navigate to "Manage Accounts" pane and Select the subscriptions you will use. Once selected, click on "Apply" button. Tab Navigate to "Disks" of the selected subscription. Tab Navigate to "A11ytest" and click on it. Tab Navigate to Toolbar. Turn on Voice Over and Observe that No Status message is Conveyed to User that refresh completed. Actual Experience No Status message is Conveyed to User that refresh completed. Expected Experience Status message should be Conveyed to User that refresh completed. Additional Context User Impact: Screen reader users will face difficulty if No Status message is Conveyed to User that refresh completed. Attachment: https://github.com/microsoft/AzureStorageExplorer/assets/98728920/56106b73-4d7b-4b8d-aadb-ba5865f399c9 Verified the issue in below environment: Azure Tools Storage Explorer: Version: 1.33.0-dev (96) AzCopy Version: 10.22.2 Platform: macOS Build Number: 20240111.7 Issue is fixed. Hence Closing the bug Please find added attachment for reference. #Closed;
2025-04-01T04:34:41.243483
2023-01-12T12:34:38
1530638267
{ "authors": [ "tamirkamara" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8412", "repo": "microsoft/AzureTRE", "url": "https://github.com/microsoft/AzureTRE/pull/3075" }
gharchive/pull-request
Add Pylance extension and fix [str | None] issue What is being addressed Mismatch types in Python can cause runtime issues. We currently don't use any tool to alert us from those kind of problems. How is this addressed Add Pylance VSCode extension Fix the most frequent issue of assigning default None to optional str params /test-extended /test-extended /test-extended
2025-04-01T04:34:41.244873
2020-06-02T05:53:33
628921696
{ "authors": [ "yilongli" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8413", "repo": "microsoft/AzureTipsAndTricks", "url": "https://github.com/microsoft/AzureTipsAndTricks/issues/94" }
gharchive/issue
Tip 226: is Azure Accelerated Networking based on FPGA-based SmartNIC or Mellanox NIC I am a bit confused because both FPGA-based SmartNIC and Mellanox CX3 are mentioned in the article: https://microsoft.github.io/AzureTipsAndTricks/blog/tip226.html. Thank you! Never mind. I found the answer: the FPGA is integrated with a Mellanox NIC in a bump-in-the-wire fashion.
2025-04-01T04:34:41.246484
2020-02-17T18:42:45
566465198
{ "authors": [ "a-patel", "mbcrump", "vaheminasyan2" ], "license": "CC-BY-4.0", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8414", "repo": "microsoft/AzureTipsAndTricks", "url": "https://github.com/microsoft/AzureTipsAndTricks/pull/69" }
gharchive/pull-request
Update tip158 Create Thumbnail Images with Azure Functions and Azure Storage fixed broken link Thanks @a-patel ! Hey @vaheminasyan2 can you add him to our list of contributors. @mbcrump, added. Thanks
2025-04-01T04:34:41.250460
2019-06-25T10:09:06
460332646
{ "authors": [ "CoHealer", "jwiley84", "prakash9707" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8415", "repo": "microsoft/BotBuilder-Samples", "url": "https://github.com/microsoft/BotBuilder-Samples/issues/1595" }
gharchive/issue
18.bot authentication, Does not allows to enter the credentials in oauth prompt I used azure active directory v2 connection and I created a azure application with allows users to login with multiple organization. After this when I runs this code it displays a empty window and takes the account which is previously logged in. For the first time it allows me to enter the microsoft email and password. After when I type "logout" then I does not allow me to enter the username and password. I think it does not logout the user properly. My azure application redirect url or reply url is "https://token.botframework.com/.auth/web/redirect" and logout url is empty. In the azure active directory v2, I had given my tenant id as "common". I think the problem is with the azure connections. when I press login in oauth card, I got the below response and after a couple of seconds it will disappear automatically. What may be the issue? [bug] @jwiley84, please investigate This seems to be a behaviour limited to the emulator, as that is the only channel I can reproduce it in. Let me reach out to a member of their team. The standard webchat (not the Azure's 'Test in WebChat') allows for 're-logging', if you will. Hi @prakash9707! I got a response from the emulator team. They've recently implemented some changes in how ngrok is being used with OAuth prompts. The details are here. Try to update your emulator and give it another shot. @prakash9707 , did this resolve your issue? Closing issue.
2025-04-01T04:34:41.254230
2020-03-23T07:10:58
585982385
{ "authors": [ "VanyLaw", "zhixzhan" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8416", "repo": "microsoft/BotFramework-Composer", "url": "https://github.com/microsoft/BotFramework-Composer/issues/2336" }
gharchive/issue
lu file can't save when changing in dialog page. Describe the bug After create a interruptionSample, changed the getProfile dialog into luis recognizer, show error. I change the lu as below, but it can't save in that panel. it can only save in lu page. - my name is {personName:userName} - {personName:userName} - {age:userAge} - I'm {age:userAge} years old @ prebuilt personName userName @ prebuilt age userAge Version master branch Browser [x] Chrome [ ] Safari [ ] Firefox [ ] Edge OS [ ] macOS [x] Windows [ ] Ubuntu it's kind of known issue inline lu editor can't save with error. talked on #2141 #2267, and will be fixed in #2272
2025-04-01T04:34:41.258425
2021-06-01T07:55:09
908013658
{ "authors": [ "johnataylor", "riftek", "tonyanziano" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8417", "repo": "microsoft/BotFramework-Composer", "url": "https://github.com/microsoft/BotFramework-Composer/issues/7967" }
gharchive/issue
Azure bot framework composer new deployment authentication when adding a new deployment. on authentication to deploy, composer cant find my email unless I put a false name and enter another email. of which authenticates but throws an error of the previous nonexistent name not being found. @tonyanziano can you comment @riftek can you please include the following in your issue: Screenshots of what you are seeing Logs / Error messages Reproduction steps (if possible) @riftek can you please follow these steps to enable debugging output in Composer, and then try to login again and reproduce this issue? Open a shell / cmd terminal Navigate to Composer's installation directory run the following command set DEBUG=composer* Start composer from the terminal "Bot Framework Composer.exe" You should now see a lot of output in the terminal window while Composer starts up. Please try to login and then once it fails, copy all of that output and paste it here. Thank you. we eventually sorted it out by creating a new bot and resources.. and we have to authenticate and deploy with a private microsoft account that needs to be linked with our work active directory. mainly we could not get old azure bot to work with new azure. Great, glad to hear you got it resolved.
2025-04-01T04:34:41.265058
2023-04-05T04:17:03
1654907517
{ "authors": [ "dkulkarni-ms", "mrk1989", "stevkan" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8418", "repo": "microsoft/BotFramework-WebChat", "url": "https://github.com/microsoft/BotFramework-WebChat/issues/4691" }
gharchive/issue
Getting console error in using WebChat in portal Is it an issue related to Adaptive Cards? No Is this an accessibility issue? No What version of Web Chat are you using? Latest production Which distribution are you using Web Chat from? Bundle (webchat.js) Which hosting environment does this issue primarily affect? Web apps Which browsers and platforms do the issue happened? Browser: Edge (latest) Which area does this issue affect? Telemetry What is the public URL for the website? No response Please describe the bug We are trying to use Azure Bot Service to create a chatbot and using Webchat to integrate in Portal. However when we integrate in Azure portal, we see console errors related to Cors policy. We have enabled the origin in WebApp Service still getting this Cors policy issue. How can this be fixed? Bot loads fine and works fine. Its only that error messages in console needs to be addressed. Also we are using Iframe to show the out of box webchat. Do you see any errors in console log? Yes uploaded the screenshot above. How to reproduce the issue? Navigate to ... Click on ... Type "..." in the send box What do you expect? Either fix or way to work around this issue and not see the errors in console logs What actually happened? As explained above, errors in console logs Do you have any screenshots or recordings to repro the issue? No response Adaptive Card JSON No response Additional context No response using Webchat to integrate in Portal Can you explain what you mean by this statement? Do you mean you have simply created an Azure Web App to host the iframe page? Or are you trying to do something else? And, for the above, can you explain how you are trying to implement or code this? The error exaplain too.. you have to add your domain in section CORS.
2025-04-01T04:34:41.269828
2019-12-12T09:14:00
536848187
{ "authors": [ "compulim", "coveralls" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8419", "repo": "microsoft/BotFramework-WebChat", "url": "https://github.com/microsoft/BotFramework-WebChat/pull/2719" }
gharchive/pull-request
Add Object.is polyfill Fixes #2718. Changelog Entry Fixed Fixes #2718. Add Object.is polyfill for IE11, by @compulim in PR #2719 Description Usage of Object.is was added in #2710 but the polyfill for ES5 browsers was not added. Specific Changes Add corejs/features/object/is to ES5 bundle [x] Testing Added Manual test steps Load ES5 bundle in IE11, the UI should show up Coverage remained the same at 61.108% when pulling 89f814e80b8cba7d838e1f7aaafa65ace994a451 on compulim:fix-2718 into 9e43da0b413efc369f823e4278e2444e3fb0b5b3 on microsoft:master.
2025-04-01T04:34:41.271648
2021-08-21T08:09:42
976078169
{ "authors": [ "98973991", "CBL-Mariner-Bot", "anphel31", "thcrain-msft" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8420", "repo": "microsoft/CBL-Mariner", "url": "https://github.com/microsoft/CBL-Mariner/pull/1316" }
gharchive/pull-request
[AUTOPATCHER] Patches CVE-2021-38593 for qt5-qtbase Patches CVE-2021-38593 for qt5-qtbase. This patch and PR was automatically generated and needs human verification before merging. This has picked up 3 identical patch files from different upstream branches. We only need 1. None of the patches apply to our version of the code Ok Closing- does not apply for our configuration of qt5-qtbase.
2025-04-01T04:34:41.278639
2021-05-24T15:55:49
899792590
{ "authors": [ "christopherco" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8421", "repo": "microsoft/CBL-Mariner", "url": "https://github.com/microsoft/CBL-Mariner/pull/974" }
gharchive/pull-request
kernel: Add key into trusted keyring Merge Checklist All boxes should be checked before merging the PR (just tick any boxes which don't apply to this PR) [ ] The toolchain has been rebuilt successfully (or no changes were made to it) [ ] The toolchain/worker package manifests are up-to-date [ ] Any updated packages successfully build (or no packages were changed) [ ] All package sources are available [ ] cgmanifest files are up-to-date and sorted (./cgmanifest.json, ./toolkit/tools/cgmanifest.json, ./toolkit/scripts/toolchain/cgmanifest.json, .github/workflows/cgmanifest.json) [ ] LICENSE-MAP files are up-to-date (./SPECS/LICENSES-AND-NOTICES/data/licenses.json, ./SPECS/LICENSES-AND-NOTICES/LICENSES-MAP.md, ./SPECS/LICENSES-AND-NOTICES/LICENSE-EXCEPTIONS.PHOTON) [ ] All source files have up-to-date hashes in the *.signatures.json files [ ] sudo make go-tidy-all and sudo make go-test-coverage pass [ ] Documentation has been updated to match any changes to the build system [ ] Ready to merge Summary Add key to default system keyring to support signed loadable kernel modules. Signed-off-by: Chris Co<EMAIL_ADDRESS>Change Log Add mariner key to kernel keyring Rev kernel related specs Does this affect the toolchain? YES Associated issues Links to CVEs Test Methodology Local build, pipeline build https://github.com/microsoft/CBL-Mariner/pull/1000
2025-04-01T04:34:41.297437
2019-05-20T17:59:49
446240953
{ "authors": [ "ctstone" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8422", "repo": "microsoft/ComputerVision", "url": "https://github.com/microsoft/ComputerVision/issues/172" }
gharchive/issue
Is a DSVM required? Consider documenting the steps to run the notebooks on a vanilla Ubuntu VM @jiata should we close this issue? I see a PR was linked and merged...
2025-04-01T04:34:41.298558
2020-03-16T20:54:13
582593189
{ "authors": [ "kipergil", "lokijota" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8423", "repo": "microsoft/CosmicClone", "url": "https://github.com/microsoft/CosmicClone/issues/15" }
gharchive/issue
Add support to copy multiple containers Or, at the very least, when you get to the final screen after a copy, enable Previous button so that you can go back to the previous screens and just edit the names of the containers. Apart from that, excellent tool. This could/should be added to ADF, pre-creating containers when copying CosmosDB to CosmosDb is a waste of time. This feature should be implemented. It should be must have feature.
2025-04-01T04:34:41.305953
2024-06-24T11:25:19
2369988580
{ "authors": [ "AArnott", "Nuklon", "Zoraver" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8424", "repo": "microsoft/CsWin32", "url": "https://github.com/microsoft/CsWin32/issues/1218" }
gharchive/issue
"Pointers cannot reference marshaled structures. Use ByRef instead" caused by VariableLengthInlineArray Actual behavior Pointers cannot reference marshaled structures. Use ByRef instead. Only happens on .NET Framework 472. If you remove the public global::Windows.Win32.VariableLengthInlineArray<char> HardwareID; field from SP_DRVINFO_DETAIL_DATA_W it works correctly. It worked correctly in CsWin32 0.3.49. Expected behavior Work correctly. Repro steps NativeMethods.txt content: SetupDiGetDriverInfoDetail NativeMethods.json content (if present): Any of your own code that should be shared? Context CsWin32 version: 0.3.106 Win32Metadata version (if explicitly set by project): 61.0.15-preview Target Framework: net472 If you include System.Memory package, the error instead becomes: Error (active) CS0117 'Unsafe' does not contain a definition for 'SkipInit' I am running into a similar issue when attempting to call CM_Register_Notification on .NET Framework 4.8 using CsWin32 0.3.106. This does not occur when targeting .NET 8. I believe that this issue is effectively a duplicate of #1184. I can't get this to repro using the instructions given in this issue. But I agree with @zoraver that this is likely a duplicate of one that I can repro. So I'm closing this. Sorry, I forgot to add instructions on how to reproduce. I retried it and it's still present in latest daily. You need to call SetupDiGetDriverInfoDetail with the details from the device: uint driverInfoDetailDataSize = (uint)sizeof(SP_DRVINFO_DETAIL_DATA_W) + (PInvoke.MAX_DEVICE_ID_LEN * 2); IntPtr driverInfoDetailDataHandle = Marshal.AllocHGlobal((IntPtr)driverInfoDetailDataSize); try { var driverInfoDetailData = (SP_DRVINFO_DETAIL_DATA_W*)driverInfoDetailDataHandle; driverInfoDetailData->cbSize = (uint)sizeof(SP_DRVINFO_DETAIL_DATA_W); uint requiredSize = 0; // This call throws the above exception. if (PInvoke.SetupDiGetDriverInfoDetail(deviceInfoList, &deviceInfoData, &driverInfoData, driverInfoDetailData, driverInfoDetailDataSize, &requiredSize)) { } Thanks for that, @nuklon. My fix for the believed duplicate bug is in version 0.3.149, for which no daily build exists yet due to a temporary service outage during our last daily build. I've re-kicked it off. If you could please test your scenario with that and report back, I'd appreciate it. Thanks, I can confirm that it's working fine now with 0.3.149. Another workaround down 😁
2025-04-01T04:34:41.316479
2024-09-09T10:13:20
2513531003
{ "authors": [ "Marv51", "Sergio0694", "manodasanW" ], "license": "MIT", "license_source": "github-api", "license_type": "permissive", "provenance": "gharchive-dolma-0004.json.gz:8425", "repo": "microsoft/CsWinRT", "url": "https://github.com/microsoft/CsWinRT/issues/1739" }
gharchive/issue
InvalidCastException for ICommand from dotnet class library Describe the bug My app roughly uses an MVVM architecture. My ViewModel classes are all in an external class library. I am getting an InvalidCastException for ICommand-properties of a ViewModel class, if I bind to it from a WinUI3 button. These exceptions seem to be not fatal but they mean that the ICommand never gets called with AOT breaking functionality badly. Note: I am using System.Windows.Input.ICommand and relying the mapping of that to the equivalent WinRT interface. The exact output is: Exception thrown: 'System.InvalidCastException' in System.Private.CoreLib.dll Exception thrown at 0x00007FF909F16D9A (KernelBase.dll) in App7.exe: WinRT originate error - 0x80004002 : 'Specified cast is not valid.'. To Reproduce My class library project: public class SimpleCommand(Action action) : System.Windows.Input.ICommand { public event EventHandler? CanExecuteChanged; public bool CanExecute(object? parameter) => true; public void Execute(object? parameter) => action(); } public class ViewModel : INotifyPropertyChanged { private readonly SimpleCommand _nextQuote; public event PropertyChangedEventHandler? PropertyChanged; public ICommand NextQuote => _nextQuote; public string Quote { get; set; } = "Knock knock."; public ViewModel() { _nextQuote = new SimpleCommand(() => { Quote = "Who's there?"; PropertyChanged?.Invoke(this, new PropertyChangedEventArgs(nameof(Quote))); }); } } I use it like this: public sealed partial class MainWindow : Window { public ViewModel ViewModel { get; set; } = new(); public MainWindow() { this.InitializeComponent(); } } <?xml version="1.0" encoding="utf-8"?> <Window x:Class="App7.MainWindow" xmlns="http://schemas.microsoft.com/winfx/2006/xaml/presentation" xmlns:x="http://schemas.microsoft.com/winfx/2006/xaml" xmlns:local="using:App7" xmlns:d="http://schemas.microsoft.com/expression/blend/2008" xmlns:mc="http://schemas.openxmlformats.org/markup-compatibility/2006" mc:Ignorable="d"> <StackPanel Orientation="Horizontal" HorizontalAlignment="Center" VerticalAlignment="Center"> <TextBlock Text="{x:Bind ViewModel.Quote, Mode=OneWay}"></TextBlock> <Button Command="{x:Bind ViewModel.NextQuote, Mode=OneWay}">Refresh</Button> </StackPanel> </Window> Expected behavior It works if PublishAot is false, so expect it to either still work with AOT or it should show a warning message. Version Info <WindowsSdkPackageVersion>10.0.22621.41</WindowsSdkPackageVersion> <PublishAot>true</PublishAot> <!-- ..... --> <PackageReference Include="Microsoft.Windows.CsWinRT" Version="2.1.1" /> <PackageReference Include="Microsoft.Windows.SDK.BuildTools" Version="10.0.26100.1" /> <PackageReference Include="Microsoft.WindowsAppSDK" Version="1.6.240829007" /> Additional context When I change the ViewModel property to a concrete class, instead of the ICommand interface, that resolves the issue. This works: public SimpleCommand NextQuote => _nextQuote; The better solution I found for this issue is to manually add the ModulInitializer below to the WinUI project. It took me quite a while to understand how and why this is necessary. internal static class CswinrtInitalizer { [System.Runtime.CompilerServices.ModuleInitializer] internal static void InitializeVtableLookup() { ComWrappersSupport.RegisterTypeComInterfaceEntriesLookup(new Func<Type, ComWrappers.ComInterfaceEntry[]>(LookupVtableEntries)); ComWrappersSupport.RegisterTypeRuntimeClassNameLookup(new Func<Type, string>(LookupRuntimeClassName)); } private static ComWrappers.ComInterfaceEntry[] LookupVtableEntries(Type type) { return type.ToString() switch { "ClassLibrary1.SimpleCommand" => [ new() { IID = ABI.System.Windows.Input.ICommandMethods.IID, Vtable = ABI.System.Windows.Input.ICommandMethods.AbiToProjectionVftablePtr }, ], _ => default }; } private static string LookupRuntimeClassName(Type type) { string typeName = type.ToString(); return default; } } I think the generator should be able to automatically generate this code for my classes. An alternative workaround is to add at least one property anywhere that uses the class directly instead of the interface. That triggers the generator to create the necessary code automatically. If you add a second button in my example above and this one calls: public SimpleCommand AlternativeNextQuote => _nextQuote; That is already enough to also fix the ICommand-button. "I think the generator should be able to automatically generate this code for my classes." It does, but you need to have the generator enabled and mark the type as partial. Add an explicit package reference for CsWinRT 2.1.1 in your project, that's needed until the next .NET SDK update that will enable the generator by default. @Marv51 to confirm, your library is targeting the .NET Windows OS version TFM like net8.0-windows10.0.22621.0 right? @manodasanW My library is targeting plain .net 8 like this: <Project Sdk="Microsoft.NET.Sdk"> <PropertyGroup> <TargetFramework>net8.0</TargetFramework> <ImplicitUsings>enable</ImplicitUsings> <Nullable>enable</Nullable> </PropertyGroup> </Project> Sounds like that is not supported? @Sergio0694, you are right, when I add CsWinRT that does seem to fix my issue and generate the needed code automatically. <Project Sdk="Microsoft.NET.Sdk"> <PropertyGroup> <TargetFramework>net8.0</TargetFramework> <ImplicitUsings>enable</ImplicitUsings> <Nullable>enable</Nullable> <CsWinRTWindowsMetadata>10.0.22621.0</CsWinRTWindowsMetadata> </PropertyGroup> <ItemGroup> <PackageReference Include="Microsoft.Windows.CsWinRT" Version="2.1.3" /> </ItemGroup> </Project> I must have misunderstood the documentation a bit, I did not consider that plain .net libraries also need a CsWinRT reference, thanks for the help everybody.