added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T04:34:39.728860
| 2024-08-27T05:12:05
|
2488293206
|
{
"authors": [
"Dev-Khant",
"PranavPuranik"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8318",
"repo": "mem0ai/mem0",
"url": "https://github.com/mem0ai/mem0/pull/1756"
}
|
gharchive/pull-request
|
get_all returns dictionary
Description
Small fix on readme - m.get_all() returns dict.
Fixes #1755
Type of change
Please delete options that are not relevant.
[ ] Bug fix (non-breaking change which fixes an issue)
[ ] New feature (non-breaking change which adds functionality)
[ ] Breaking change (fix or feature that would cause existing functionality to not work as expected)
[ ] Refactor (does not change functionality, e.g. code style improvements, linting)
[x] Documentation update
How Has This Been Tested?
Please describe the tests that you ran to verify your changes. Provide instructions so we can reproduce. Please also list any relevant details for your test configuration
Please delete options that are not relevant.
[ ] Unit Test
[x] Test Script (used example)
Checklist:
[ ] My code follows the style guidelines of this project
[ ] I have performed a self-review of my own code
[ ] I have commented my code, particularly in hard-to-understand areas
[ ] I have made corresponding changes to the documentation
[ ] My changes generate no new warnings
[ ] I have added tests that prove my fix is effective or that my feature works
[ ] New and existing unit tests pass locally with my changes
[ ] Any dependent changes have been merged and published in downstream modules
[ ] I have checked my code and corrected any misspellings
Maintainer Checklist
[ ] closes #xxxx (Replace xxxx with the GitHub issue number)
[ ] Made sure Checks passed
@PranavPuranik Thanks for fixing this!
|
2025-04-01T04:34:39.740425
| 2024-06-14T09:37:43
|
2352966124
|
{
"authors": [
"hal-eisen-MG",
"kgolubic"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8319",
"repo": "memgraph/documentation",
"url": "https://github.com/memgraph/documentation/pull/842"
}
|
gharchive/pull-request
|
[main < typo-fix-20240614] Fix typos
Description
Fixed typos across the repo
Pull request type
Please check what kind of PR this is:
[ ] Fix or improvement of an existing page
Related PRs and issues
PR this doc page is related to:
(especially necessary if the PR is related to a release)
Closes:
(paste the link to the issue it closes)
Checklist:
[ ] Check all content with Grammarly
[ ] Perform a self-review of my code
[ ] Make corresponding changes to the rest of the documentation (consult with the DX team)
[ ] The build passes locally
[ ] My changes generate no new warnings or errors
[ ] Add a corresponding label
[ ] If release-related, add a product and version label
[ ] If release-related, add release note on product PR
@hal-eisen-MG I've applied all of the suggestions. Merging to main.
@hal-eisen-MG I've applied all of the suggestions.
Hm. I'm confused. I'm not seeing any subsequent changes. Should we jump on a quick call?
@hal-eisen-MG I've applied all of the suggestions.
Hm. I'm confused. I'm not seeing any subsequent changes. Should we jump on a quick call?
Sorry. I forgot to push the changes from the local branch. The changes should be on the GitHub branch as well.
|
2025-04-01T04:34:39.782793
| 2019-08-02T11:49:21
|
476140843
|
{
"authors": [
"Adam-Dupaski",
"ConnorLand"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8323",
"repo": "mendix/docs",
"url": "https://github.com/mendix/docs/pull/1851"
}
|
gharchive/pull-request
|
Add Leroy's microflow RN
I just wanted you to have eyes on this to make sure it was going in the right place. Thanks!
@ConnorLand Not sure why you didn't merge this Friday? Please familiarize with my 3 changes upon return.
@ConnorLand let's discuss this PR upon your return
|
2025-04-01T04:34:39.783801
| 2022-07-19T11:16:19
|
1309366050
|
{
"authors": [
"MarkvanMents"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8324",
"repo": "mendix/docs",
"url": "https://github.com/mendix/docs/pull/4831"
}
|
gharchive/pull-request
|
Update offline installation instructions
This addresses issues #4788 and #4803 which identify issues with the offline installation instructions with later versions of Mendix Studio Pro 9
Looks good from R&D PoV. Will merge now and pick up later if there are any further issues.
|
2025-04-01T04:34:39.786104
| 2024-07-30T10:04:48
|
2437435787
|
{
"authors": [
"MariaShaposhnikova",
"atennapel"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8325",
"repo": "mendix/docs",
"url": "https://github.com/mendix/docs/pull/8107"
}
|
gharchive/pull-request
|
WIP: [RUN-4179] Add information about the deletion of anonymous users
I only added section 3.4 I have no idea why github indicates all the other changes. Might be a newline issue, but I do have autocrlf set to true in git.
@atennapel , thank you for your PR! Does it have a specific version it should be merged for?
No, this information is relevant for Mx 10 and 9. But I will still update this PR with more information about sessions.
@MariaShaposhnikova New PR (without all the unnecessary changes): https://github.com/mendix/docs/pull/8127
|
2025-04-01T04:34:39.791173
| 2024-11-05T08:17:35
|
2634744061
|
{
"authors": [
"leonardomendix",
"uicontent"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8326",
"repo": "mendix/web-widgets",
"url": "https://github.com/mendix/web-widgets/pull/1318"
}
|
gharchive/pull-request
|
fix(deps): update dependency mendix to ^10.16.49747
This PR contains the following updates:
Package
Type
Update
Change
mendix
dependencies
minor
^10.15.46408 -> ^10.16.49747
Configuration
📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined).
🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied.
♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox.
🔕 Ignore: Close this PR and you won't be reminded about this update again.
This PR has been generated by Renovate Bot.
Let's merge #1320 first
Merged!
|
2025-04-01T04:34:39.796698
| 2018-09-13T05:26:18
|
359747309
|
{
"authors": [
"Andries-Smit",
"ChengMaoDSM"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8327",
"repo": "mendixlabs/charts",
"url": "https://github.com/mendixlabs/charts/issues/58"
}
|
gharchive/issue
|
add a direct label to the bar chart.
I wanted to have something like https://plot.ly/javascript/bar-charts/#bar-chart-with-direct-labels.
I do not see any field that can be used for this number.
Dear ChengMaoDSM,
At the moment there is no way to move the label inside the bar. The addition json settings could not provided this. It is a feature that need to be build into the chart widget, like we did with the heatmap.
I wil discuss this request with the product owner and will get back to you.
With kind regards, Andries
If the text is show stopper you could consider building the chart from scratch with the Any chart module. https://appstore.home.mendix.com/link/app/106517/
https://docs.mendix.com/refguide/charts-any-configuration
Sample:
https://docs.mendix.com/refguide/charts-any-cheat-sheet#2-6-column-chart
Or use a building block from the module.
Cheers
Will not be implemented in the standard chart widgets, could be solved with any chart
Will not be implemented in the standard chart widgets, could be solved with any chart
|
2025-04-01T04:34:39.805369
| 2019-05-16T10:46:23
|
444884928
|
{
"authors": [
"JelleDekker",
"codecov-io"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8328",
"repo": "mendixlabs/google-maps",
"url": "https://github.com/mendixlabs/google-maps/pull/35"
}
|
gharchive/pull-request
|
Marker labels
Added support for non persistent objects and marker labels. This allows markers on the map to be easily identified if the map is displayed next to a ListView for example.
Codecov Report
Merging #35 into master will decrease coverage by 0.39%.
The diff coverage is 100%.
@@ Coverage Diff @@
## master #35 +/- ##
=========================================
- Coverage 83.43% 83.04% -0.4%
=========================================
Files 3 3
Lines 169 171 +2
Branches 44 45 +1
=========================================
+ Hits 141 142 +1
- Misses 28 29 +1
Impacted Files
Coverage Δ
src/components/Map.ts
81.69% <ø> (ø)
:arrow_up:
src/components/Marker.ts
90% <100%> (-10%)
:arrow_down:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 0550e59...6d76cc8. Read the comment docs.
Example of the added functionality in use:
|
2025-04-01T04:34:39.808794
| 2022-03-19T00:08:58
|
1174119314
|
{
"authors": [
"hungpham3112",
"mengelbrecht"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8329",
"repo": "mengelbrecht/lightline-bufferline",
"url": "https://github.com/mengelbrecht/lightline-bufferline/issues/103"
}
|
gharchive/issue
|
How to hide a buffer in bufferline filetype: terminal
When opening terminal, a buffer will show the name of terminal to. How to disable this only filetype: Terminal? Thanks in advanced.
I will have a look at it soon.
I added support for a custom buffer filter function in branch feature/custom-filter-function, which I plan to merge after some testing. You can use this to hide all neovim terminal buffers:
function LightlineBufferlineFilter(buffer)
return getbufvar(a:buffer, '&buftype') !=# 'terminal'
endfunction
let g:lightline#bufferline#buffer_filter = "LightlineBufferlineFilter"
I'm using vim-plug. I added this line but still can't use your snippet in this branch.
Plug 'https://github.com/mengelbrecht/lightline-bufferline.git', {'branch': 'feature/custom-filter-function'}
I merged the feature branch. The code should now work without switching branches.
I merged the feature branch. The code should now work without switching branches. Please reopen this issue if it does not work.
Thanks, it worked very well.
|
2025-04-01T04:34:39.810718
| 2020-11-26T02:03:01
|
751224408
|
{
"authors": [
"adrian5",
"mengelbrecht"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8330",
"repo": "mengelbrecht/lightline-bufferline",
"url": "https://github.com/mengelbrecht/lightline-bufferline/issues/76"
}
|
gharchive/issue
|
Spacing / margins?
Maybe I'm not seeing the obvious, but is there a straight-forward way to make tabs wider? I.e.:
| foo.vim | context.js | style.css |
@adrian5 you can try the feature/margins branch (PR #78) and set the following properties:
let g:lightline#bufferline#margin_left = 4
let g:lightline#bufferline#margin_right = 4
Thanks, that does it!
@adrian5 I merged the branch
|
2025-04-01T04:34:39.831783
| 2020-07-02T07:45:58
|
649713386
|
{
"authors": [
"codecov-commenter",
"whatasoda"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8331",
"repo": "mercari/proto-to-type",
"url": "https://github.com/mercari/proto-to-type/pull/1"
}
|
gharchive/pull-request
|
chore: prepare release
introduce semantic-release
update REAMDE
Codecov Report
:exclamation: No coverage uploaded for pull request base (master@6ca6c3e). Click here to learn what that means.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #1 +/- ##
=========================================
Coverage ? 93.28%
=========================================
Files ? 7
Lines ? 253
Branches ? 65
=========================================
Hits ? 236
Misses ? 17
Partials ? 0
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 6ca6c3e...3cc0387. Read the comment docs.
|
2025-04-01T04:34:39.834532
| 2022-07-01T16:44:10
|
1291619576
|
{
"authors": [
"Jeeppler",
"de-jcup"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8332",
"repo": "mercedes-benz/sechub",
"url": "https://github.com/mercedes-benz/sechub/issues/1448"
}
|
gharchive/issue
|
Upgrade dependencies to Spring Boot 2.7.1
Upgrade to Spring Boot 2.7.1: https://github.com/spring-projects/spring-boot/releases/tag/v2.7.1
We did upgrade to Spring Boot 2.7.1, but we did not upgrade the dependencies.
See: #915
We should just ignore/ close this one and do directly
#1484
@de-jcup I already did this and it was reviewed by @winzj. Just forgot to close it.
|
2025-04-01T04:34:39.851556
| 2023-07-20T11:15:08
|
1813745776
|
{
"authors": [
"luke88jones",
"simoneb"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8333",
"repo": "mercurius-js/mercurius-gateway",
"url": "https://github.com/mercurius-js/mercurius-gateway/pull/81"
}
|
gharchive/pull-request
|
fix: apply updated schema on refresh
fixes #80
update the gateway refresh function to call replaceSchema on the GQL instance
@luke88jones I'm seeing a consistent failure on Node 18 due to a timeout. Because this hasn't happened in the past and because the new behavior you introduced may have interactions with the event loop, I'm tempted to think that this is a legitimate failure, although I'm not sure why it would happen just on Node 18
@simoneb Sorry for the silence on this I've been on paternity leave.
It looks like one of the polling interval tests is timing out. I'll see what I can find.
@simoneb I've just checked out the main branch locally I have the same failure. I'm trying to track down the issue but it looks to be something to do with closing the userService in this test in the tests/pollingInterval.js file
"Polling schemas (if service is down, schema shouldn't be changed)"
|
2025-04-01T04:34:39.858748
| 2018-09-04T17:30:13
|
356906653
|
{
"authors": [
"axoplasm",
"sanjuroj"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8334",
"repo": "mercycorps/TolaActivity",
"url": "https://github.com/mercycorps/TolaActivity/issues/583"
}
|
gharchive/issue
|
"Date collected" field in Add Results form auto-populates
The "Date collected" field in the Add Results modal form auto-populates with the current date. This triggers a check to see if the date is valid given the period that is selected, which it probably won't be, which results in an error message on the "Measure against target" field.
Need to eliminate the auto-population of the "Date collected" field. It should be blank until the user has entered a date. This will have the side-effect of eliminating the unwanted error message on the "Measure against target" field.
So what's happening here:
Datepickers need a date by default so they know what month to show when they are summoned. If they don't get a date, they default to "today."
We initialize this datepicker invisibly on page load because it's secretly an inline element (hidden on page load) adjacent to the real input field
So when it gets initialized, it gets a default date and automagically writes that to the real date field
This was creating a secondary bug no one noticed where the datepicker was overriding the date field even if that field already had a value
The "Measure against target" field was running a validation function against the value of the "Date Collected" field & returning an error if that field was not empty. Which it never was thanks to the magic described above
So my fix(es) are like this. When the "add results" modal appears:
Grab the value of the Date Collected field (which might be null for a new result)
Initialize the invisible datepicker as above, setting the default date to the value in 1. (As a side effect of the JQuery datepicker, if the value is '' (null) it will use today)
If the value in 1 was empty (i.e. an empty datefield in a new results form), erase the default field
Run the validation on Measure Against Target
And as an extra fun bonus, the datepicker isn't positioned absolutely for some highly irritating reason! Will
|
2025-04-01T04:34:39.899598
| 2021-11-12T13:18:14
|
1051959226
|
{
"authors": [
"joncodo"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8335",
"repo": "merico-dev/lake",
"url": "https://github.com/merico-dev/lake/issues/733"
}
|
gharchive/issue
|
Lead_time_minutes in the domain layer
Description
Github lead time should be in minutes. Rename column in domain layer to lead_time_minutes
duplicate of #722
|
2025-04-01T04:34:39.902992
| 2024-12-04T06:09:05
|
2716653639
|
{
"authors": [
"NKarolak",
"merill"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8336",
"repo": "merill/bluesky",
"url": "https://github.com/merill/bluesky/issues/1118"
}
|
gharchive/issue
|
[Bug] Error: User has not verified their email
Bug details
I've just registered, verified my email by clicking on the mail link, connected my Bsky profile, entered what I post about... All fine.
But when I try to verify that I'm an MVP (https://mvp.microsoft.com/en-us/MVP/profile/0789d8ed-d3d1-480b-9fee-5bc01628baa2), it keeps saying "Error: User has not verified their email".
Could it be related to that my email address countains a "+"?
Solved by visiting the homepage again, then Register, then Sign in.
Although I was signed in all the time to enter my data 🤷🏻♀️
Thanks @NKarolak
There was a related issue that others saw too and I was able to get to the bottom of it and fix it.
|
2025-04-01T04:34:39.905285
| 2024-12-12T09:18:12
|
2735330917
|
{
"authors": [
"manekinekko",
"merill"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8337",
"repo": "merill/bluesky",
"url": "https://github.com/merill/bluesky/issues/1127"
}
|
gharchive/issue
|
Missing JavaScript-related topics
Bug details
Hi, fellow FTE here 👋
I appreciate the time and effort you are putting into building this platform! Thank you ❤️
While selecting the topics I write about, I could not find those related to my work. May I request adding:
Developer - JavaScript
Developer - Node.js
Developer - WebAssembly
Done!
Thanks for requesting @manekinekko
|
2025-04-01T04:34:39.931171
| 2020-09-23T04:56:44
|
707047425
|
{
"authors": [
"HJesusH",
"merqlove"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8338",
"repo": "merqlove/do_snapshot",
"url": "https://github.com/merqlove/do_snapshot/issues/23"
}
|
gharchive/issue
|
The certificate's owner does not match hostname
The certificate's owner does not match hostname 'assets.merqlove.ru.s3.amazonaws.com'
This occurs when i use the next command in a linux/debian server
wget https://assets.merqlove.ru.s3.amazonaws.com/do_snapshot/do_snapshot.tgz
Fixed
|
2025-04-01T04:34:39.940111
| 2019-12-02T17:18:35
|
531314408
|
{
"authors": [
"girishranganathan",
"iambami"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8339",
"repo": "meshery/meshery-nsm",
"url": "https://github.com/meshery/meshery-nsm/issues/45"
}
|
gharchive/issue
|
Support installing NSM and its sample apps using helm
Current Description
The adapter is using helm to install NSM and its sample apps. But we are using a cloned version of the helm repo to use it.
Expected Behavior
We want to be able to use the official helm repos to be able to install NSM and its sample apps. This will prevent us from using edge version all the time and also provide us the ability to install specific versions of NSM and sample apps.
Hi, @girishranganathan is it okay to close this issue? or does it need to be assigned to someone?
|
2025-04-01T04:34:39.947854
| 2022-08-17T20:49:40
|
1342279047
|
{
"authors": [
"leecalcote"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8340",
"repo": "meshery/meshery",
"url": "https://github.com/meshery/meshery/issues/6054"
}
|
gharchive/issue
|
[mesheryctl] app onboard receives 400 response code
See mesheryctl Command Tracker for current status of commands.
Current Behavior
mesheryctl app onboard -s "Kubernetes Manifest" -f https://raw.githubusercontent.com/kubernetes/website/main/content/en/examples/application/job/cronjob.yaml
Error: Response Status Code 400, possible Server Error
Server Response:
map[string]interface {}{"apiVersion":"batch/v1", "kind":"CronJob", "metadata":map[string]interface {}{"name":"hello"}, "spec":map[string]interface {}{"jobTemplate":map[string]interface {}{"spec":map[string]interface {}{"template":map[string]interface {}{"spec":map[string]interface {}{"containers":[]interface {}{map[string]interface {}{"command":[]interface {}{"/bin/sh", "-c", "date; echo Hello from the Kubernetes cluster"}, "image":"busybox:1.28", "imagePullPolicy":"IfNotPresent", "name":"hello"}}, "restartPolicy":"OnFailure"}}}}, "schedule":"* * * * *"}}
DEBU[1954] Application size: 1507
INFO[1954] attempting to save application to remote provider
INFO[1954] application successfully sent to remote provider: [{"id":"5e6cec58-8db0-418e-a145-2153c6885be9","name":"cronjob.yaml","user_id":"48368c3b-5134-477f-95bb-4a1ba56dd908","application_file":"name: Autogenerated\nservices:\n faf7c427-102f-4042-bf85-f3c622bc52dc:\n name: hello\n type: CronJob.K8s\n namespace: default\n settings:\n spec:\n jobTemplate:\n spec:\n template:\n spec:\n containers:\n - command:\n - /bin/sh\n - -c\n - date; echo Hello from the Kubernetes cluster\n image: busybox:1.28\n imagePullPolicy: IfNotPresent\n name: hello\n restartPolicy: OnFailure\n schedule: '* * * * *'\n","location":{"branch":"","host":"https://raw.githubusercontent.com/kubernetes/website/main/content/en/examples/application/job/cronjob.yaml","path":"","type":"http"},"type":{"String":"Kubernetes Manifest","Valid":true},"SourceContent":null,"created_at":"2022-08-17T20:43:50.483192Z","updated_at":"2022-08-17T20:43:50.483215Z"}]
DEBU[1954] Application Content size 418
INFO[1954] application source successfully uploaded to remote provider
INFO[1954] attempting to fetch applications from cloud
DEBU[1954] constructed applications url: https://meshery.layer5.io/applications?page=0&page_size=10
map[string]interface {}{"name":"Autogenerated", "services":map[string]interface {}{"faf7c427-102f-4042-bf85-f3c622bc52dc":map[string]interface {}{"name":"hello", "namespace":"default", "settings":map[string]interface {}{"spec":map[string]interface {}{"jobTemplate":map[string]interface {}{"spec":map[string]interface {}{"template":map[string]interface {}{"spec":map[string]interface {}{"containers":[]interface {}{map[string]interface {}{"command":[]interface {}{"/bin/sh", "-c", "date; echo Hello from the Kubernetes cluster"}, "image":"busybox:1.28", "imagePullPolicy":"IfNotPresent", "name":"hello"}}, "restartPolicy":"OnFailure"}}}}, "schedule":"* * * * *"}}, "type":"CronJob.K8s"}}}
INFO[1954] applications successfully retrieved from remote provider
Screenshots/Logs
Environment
Host OS: Mac
Meshery Server Version: stable-v0.6.0-rc-6fd
Meshery Client Version: stable-v0.6.0-rc-6fd
Platform: Kubernetes
Contributor Guides and Handbook
📚 mesheryctl Contributing Guide
📺 Beginner's guide to contributing to Meshery and mesheryctl
⌨️ mesheryctl Command Tracker
⌨️ Meshery CLI Commands and Documentation
🙋🏾🙋🏼 Questions: Discussion Forum and Community Slack
@Revolyssup is this another issue caused by source type?
@Revolyssup @alphaX86 ?
|
2025-04-01T04:34:39.966257
| 2022-05-28T18:53:47
|
1251701929
|
{
"authors": [
"Revolyssup",
"codecov-commenter",
"leecalcote",
"sudo-NithishKarthik"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8341",
"repo": "meshery/meshery",
"url": "https://github.com/meshery/meshery/pull/5549"
}
|
gharchive/pull-request
|
Replace kubeopenapi-jsonschema based component generation with CUE
Description
It makes the process at least 3 times faster(I actually measured this by taking multiple measurements 😝 )
Notes for Reviewers
For more information, please refer meshkit pr
Signed commits
[x] Yes, I signed my commits.
@sudo-NithishKarthik Wouldn't we first want the meshkit PR to be merged with a release?
Yes
MeshKit PR ready for merge and release?
Codecov Report
Merging #5549 (6ad5552) into master (2a22fe8) will decrease coverage by 4.34%.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #5549 +/- ##
==========================================
- Coverage 17.47% 13.12% -4.35%
==========================================
Files 107 107
Lines 9993 9993
==========================================
- Hits 1746 1312 -434
- Misses 7821 8401 +580
+ Partials 426 280 -146
Flag
Coverage Δ
gointegrationtests
?
unittests
13.12% <ø> (ø)
Flags with carried forward coverage won't be shown. Click here to find out more.
Impacted Files
Coverage Δ
mesheryctl/internal/cli/root/system/update.go
1.05% <0.00%> (-58.95%)
:arrow_down:
mesheryctl/internal/cli/root/system/start.go
1.88% <0.00%> (-55.19%)
:arrow_down:
mesheryctl/internal/cli/root/system/status.go
1.14% <0.00%> (-55.18%)
:arrow_down:
mesheryctl/internal/cli/root/system/restart.go
1.92% <0.00%> (-48.08%)
:arrow_down:
mesheryctl/internal/cli/root/system/stop.go
2.00% <0.00%> (-46.67%)
:arrow_down:
mesheryctl/internal/cli/root/system/logs.go
3.26% <0.00%> (-36.96%)
:arrow_down:
mesheryctl/internal/cli/root/system/check.go
1.03% <0.00%> (-23.03%)
:arrow_down:
mesheryctl/internal/cli/root/system/dashboard.go
1.70% <0.00%> (-9.41%)
:arrow_down:
mesheryctl/pkg/utils/healthcheck.go
8.82% <0.00%> (-4.42%)
:arrow_down:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 1bf2635...6ad5552. Read the comment docs.
|
2025-04-01T04:34:39.970893
| 2023-02-21T04:52:00
|
1592786350
|
{
"authors": [
"Chadha93",
"VIKAS0264",
"l5io"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8342",
"repo": "meshery/meshery",
"url": "https://github.com/meshery/meshery/pull/7134"
}
|
gharchive/pull-request
|
fixed jquery error
Notes for Reviewers
This PR fixes #7133
jquery throws unrecognised expression error because of incorrect jquery selector it is fixed in this PR.
Signed commits
[x] Yes, I signed my commits.
🚀 Preview for commit 67da861da156ee219ee9b5ada3485bf5d1c446d6 at: https://63f44ede9e215904edaeca88--meshery-docs-preview.netlify.app
🚀 Preview for commit 44a87313a7797fdeed6eaa19c1cce1fb69c1a52a at: https://63f45d7037dfa1112757f77f--meshery-docs-preview.netlify.app
@VIKAS0264 Let's discuss this on the Meshery Dev call. Please add this as an agenda item in the meeting minutes if you would. :)
|
2025-04-01T04:34:39.976093
| 2022-11-17T20:51:09
|
1453975621
|
{
"authors": [
"Chadha93",
"HM-8",
"leecalcote"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8343",
"repo": "meshery/play",
"url": "https://github.com/meshery/play/issues/30"
}
|
gharchive/issue
|
[Docs] Add project setup guidelines with commands in contributing.md
Current State
The contributing.md is missing project setup command instructions.
Desired State
Add project setup guidelines with commands in contributing.md
Refer to the Makefile to get setup commands.
Contributor Guides and Resources
📚 Instructions for contributing to documentation
Meshery documentation site and source
🛠 Meshery Build & Release Strategy
🎨 Wireframes and designs for Meshery UI in Figma
🙋🏾🙋🏼 Questions: Layer5 Discussion Forum and Layer5 Community Slack
Hi @Chadha93. Interested in working on this issue. Would love to get some guidance.
@HM-8, after you familiarize with the make targets to build this site, please add a few instructions on how to build the site for other contributors. Here is a reference - https://github.com/layer5io/layer5/blob/master/CONTRIBUTING.md
Okay @leecalcote. I'll get on it. Thanks for the reference.
|
2025-04-01T04:34:40.108343
| 2022-09-19T17:12:04
|
1378271090
|
{
"authors": [
"rugeli"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8344",
"repo": "mesoscope/cellpack",
"url": "https://github.com/mesoscope/cellpack/pull/66"
}
|
gharchive/pull-request
|
Feature/reorganize ingredient data
Problem
the 4th task in Migrate old ingredients into new "objects dict" #57
Solution
Created a partners object to convert ingredients that should be moved into this object.
Wrote tests for _convert_to_objects
Type of change
Please delete options that are not relevant.
New feature (non-breaking change which adds functionality)
merged to staging/ and deleted feature/ branch, grateful for helps!
|
2025-04-01T04:34:40.118586
| 2017-10-25T19:02:26
|
268512049
|
{
"authors": [
"elezar",
"nickbp"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8346",
"repo": "mesosphere/dcos-commons",
"url": "https://github.com/mesosphere/dcos-commons/pull/1771"
}
|
gharchive/pull-request
|
INFINITY-2584/2596: Rename stop to pause, and followup fixes from PR 1763
Docs fixes and whatever else comes up following PR #1763
TODO: need to also rename start to resume
TODO: need to also rename start to resume
Is still think || and |> are the way to go. Or even: ⏯
|
2025-04-01T04:34:40.126089
| 2017-12-01T21:32:44
|
278602863
|
{
"authors": [
"mpereira"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8347",
"repo": "mesosphere/dcos-commons",
"url": "https://github.com/mesosphere/dcos-commons/pull/1903"
}
|
gharchive/pull-request
|
Make default placement constraint values be valid stringified JSON.
Values should be a JSON-stringified array of constraints (which are expressed as arrays of strings).
Example: "hostname:MAX_PER:1" becomes "[[\"hostname\", \"MAX_PER\", \"1\"]]"
Empty strings become JSON-stringified empty arrays, "" -> "[]".
Chat with Andrew:
Murilo Pereira [12:14 PM]
for values that used to default to `"hostname:MAX_PER:1"`, does it matter
whether the JSON equivalent has the `1` as a string or number? e.g.
`"[[\"hostname\", \"MAX_PER\", \"1\"]]"` for string
Andrew Hoskins [12:21 PM]
Hey! Make it a string
Murilo Pereira [12:21 PM]
cool. and empty defaults just `""` right?
Andrew Hoskins [12:26 PM]
Empty state should actually be `"[]"`
[12:26]
Because the meme type is <…>+json, so it expects a valid JSON value
One of the things I was struggling with (after finding out about the need for the propagating mustache triple bracket on friday) was the need for double-escaping quotes in config.json files, e.g. https://github.com/mesosphere/dcos-commons/pull/1903/files#diff-56550f345e72e9ddaffbc774bb87a7a9R124
Another was missing test config files with hard-coded constraints like https://github.com/mesosphere/dcos-commons/pull/1903/files#diff-9c0a8da33f4e191b89245b7045401adfL4
The other things was that making [] be the default value instead caused the constraint parser to treat it as a empty constraint instead of an empty list of constraints:
testcase com.mesosphere.sdk.elastic.scheduler.ServiceTest >
testSpec: java.io.IOException: Invalid number of entries in rule. Expected 2 or 3, got 0: []
at com.mesosphere.sdk.offer.evaluate.placement.MarathonConstraintParser.parseRow(MarathonConstraintParser.java:106)
which is addressed by https://github.com/mesosphere/dcos-commons/pull/1903/commits/b8662cfffe13ca8cf585f7b86a9cb8675bf2c017
I just noticed https://github.com/mesosphere/dcos-commons/blob/INFINITY-2620-placement-constraint-defaults/sdk/scheduler/src/test/java/com/mesosphere/sdk/offer/evaluate/placement/MarathonConstraintParserTest.java#L296-L299 so it's possible we don't need to double-escape double quotes in config.json files and use single quotes instead, I'll push a commit trying that after the checks for the current code pass.
To recap:
config.json files need stringified JSON to double-escape double quotes
"placement_constraint": {
"type": "string",
"description": "The Marathon-style placement constraint string controlling node placement",
"default": "[[\\\"hostname\\\", \\\"UNIQUE\\\"]]",
"media": {
"type": "application/x-zone-constraints+json"
}
}
marathon.json.mustache files have triple-brace mustache templates
"PLACEMENT_CONSTRAINT": "{{{nodes.placement_constraint}}}",
expands into:
"PLACEMENT_CONSTRAINT": "[[\"hostname\", \"UNIQUE\"]]",
svc.yml files need the single quote around the triple braces
placement: '{{{PLACEMENT_CONSTRAINT}}}'
expands into:
placement: '[["hostname", "UNIQUE"]]'
Maybe we could make it possible to just use a regular array value in config.json in the future, but for now this should work.
@benclarkwood https://jira.mesosphere.com/browse/INFINITY-2769
|
2025-04-01T04:34:40.130129
| 2015-02-11T05:24:38
|
57277978
|
{
"authors": [
"ConnorDoyle",
"drexin",
"sttts"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8348",
"repo": "mesosphere/marathon",
"url": "https://github.com/mesosphere/marathon/issues/1178"
}
|
gharchive/issue
|
Potential deadlock in MarathonSchedulerActor
In MarathonSchedulerActor.performAsyncWithLockFor calls Semaphore.acquire when in blocking mode. The acquire method blocks the entire thread. It appears that depending on how concurrent acquisition is interleaved, deadlock could occur.
Yes, I also came across this, see: https://github.com/mesosphere/marathon/pull/988
I seem to have forgotten to create a follow up ticket. Thanks Connor!
What about this patch? If Set[PathId] is the way to go, this better shouldn't be merged last-minute.
|
2025-04-01T04:34:40.164181
| 2015-08-26T15:21:24
|
103295337
|
{
"authors": [
"aameek",
"aquamatthias",
"gkleiman",
"jasongilanfarr",
"jgarcia-mesosphere",
"kopax",
"mcclurmc",
"meichstedt",
"mingqi",
"nashasha1"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8349",
"repo": "mesosphere/marathon",
"url": "https://github.com/mesosphere/marathon/issues/2101"
}
|
gharchive/issue
|
Could not modify Group with key
In 0.8.2 We create an app, marathon return success. But it never really create it. And it not in deployment nor in queue.
So we upgrade to 0.9.2
It shows error:
Could not modify Group with key: root:2015-07-31T08:50:18.342Z
We remove some zookeeper log/snapshot. It is ok.
But after we recover some apps, the error appear again.
Could you please provide some more information from the marathon logfile? There should be an exception logged with that message.
Aug 26 09:21:17 ip----* marathon[27448]: [2015-08-26 09:21:17,252] WARN (mesosphere.marathon.api.MarathonExceptionMapper:30)
Aug 26 09:21:17 ip----* marathon[27448]: mesosphere.marathon.StoreCommandFailedException: Could not fetch AppDefinition with key: 5ecb8fba-afdf-4691-9f2f-f0111106b4ed
Aug 26 09:21:17 ip----* marathon[27448]: #011at mesosphere.marathon.state.MarathonStore$$anonfun$mesosphere$marathon$state$MarathonStore$$exceptionTransform$1.applyOrElse(MarathonStore.scala:75)
Aug 26 09:21:17 ip----* marathon[27448]: #011at mesosphere.marathon.state.MarathonStore$$anonfun$mesosphere$marathon$state$MarathonStore$$exceptionTransform$1.applyOrElse(MarathonStore.scala:74)
Aug 26 09:21:17 ip----* marathon[27448]: #011at scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:36)
Aug 26 09:21:17 ip----* marathon[27448]: #011at scala.util.Failure$$anonfun$recover$1.apply(Try.scala:215)
Aug 26 09:21:17 ip----* marathon[27448]: #011at scala.util.Try$.apply(Try.scala:191)
Aug 26 09:21:17 ip----* marathon[27448]: #011at scala.util.Failure.recover(Try.scala:215)
Aug 26 09:21:17 ip----* marathon[27448]: #011at scala.concurrent.Future$$anonfun$recover$1.apply(Future.scala:324)
Aug 26 09:21:17 ip----* marathon[27448]: #011at scala.concurrent.Future$$anonfun$recover$1.apply(Future.scala:324)
Aug 26 09:21:17 ip----* marathon[27448]: #011at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32)
Aug 26 09:21:17 ip----* marathon[27448]: #011at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
Aug 26 09:21:17 ip----* marathon[27448]: #011at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
Aug 26 09:21:17 ip----* marathon[27448]: #011at java.lang.Thread.run(Thread.java:745)
Aug 26 09:21:17 ip----* marathon[27448]: Caused by: mesosphere.marathon.StoreCommandFailedException: Could not load key app:5ecb8fba-afdf-4691-9f2f-f0111106b4ed
Aug 26 09:21:17 ip----* marathon[27448]: #011at mesosphere.util.state.zk.ZKStore$$anonfun$exceptionTransform$1.applyOrElse(ZKStore.scala:83)
Aug 26 09:21:17 ip----* marathon[27448]: #011at mesosphere.util.state.zk.ZKStore$$anonfun$exceptionTransform$1.applyOrElse(ZKStore.scala:82)
Aug 26 09:21:17 ip----* marathon[27448]: #011... 10 more
Aug 26 09:21:17 ip----* marathon[27448]: Caused by: org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = ConnectionLoss for /marathon/state/app:5ecb8fba-afdf-4691-9f2f-f0111106b4ed
Aug 26 09:21:17 ip----* marathon[27448]: #011at org.apache.zookeeper.KeeperException.create(KeeperException.java:99)
Aug 26 09:21:17 ip----* marathon[27448]: #011at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
Aug 26 09:21:17 ip----* marathon[27448]: #011at com.twitter.zk.AsyncCallbackPromise$class.process(AsyncCallbackPromise.scala:18)
Aug 26 09:21:17 ip----* marathon[27448]: #011at com.twitter.zk.DataCallbackPromise.process(AsyncCallbackPromise.scala:53)
Aug 26 09:21:17 ip----* marathon[27448]: #011at com.twitter.zk.DataCallbackPromise.processResult(AsyncCallbackPromise.scala:56)
Aug 26 09:21:17 ip----* marathon[27448]: #011at org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:564)
Aug 26 09:21:17 ip----* marathon[27448]: #011at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:498)
Aug 26 09:21:17 ip----* marathon[27448]: [2015-08-26 09:21:17,252] WARN Deployment failed for change: 2015-08-26T09:21:04.810Z (mesosphere.marathon.state.GroupManager:146)
Aug 26 09:21:17 ip----* marathon[27448]: mesosphere.marathon.StoreCommandFailedException: Could not modify Group with key: root:2015-07-30T00:07:26.834Z
Aug 26 09:21:17 ip----* marathon[27448]: #011at mesosphere.marathon.state.MarathonStore$$anonfun$mesosphere$marathon$state$MarathonStore$$exceptionTransform$1.applyOrElse(MarathonStore.scala:75)
Aug 26 09:21:17 ip----* marathon[27448]: #011at mesosphere.marathon.state.MarathonStore$$anonfun$mesosphere$marathon$state$MarathonStore$$exceptionTransform$1.applyOrElse(MarathonStore.scala:74)
Aug 26 09:21:17 ip----* marathon[27448]: #011at scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:36)
Aug 26 09:21:17 ip---**-* marathon[27448]: #011at scala.util.Failure$$anonfun$recover$1.apply(Try.scala:215)
we found this when the number of our containers is larger than 1000.
So did marathon have some unusual behaviour when it manages 1000+ containers.
I and @nashasha1 from same company. I post another logging with better format below. The exception happened when App was created or updated. The exception says Can not create entity group:root:2015-07-31T08:50:18.342Z and Could not modify Group with key: root:2015-07-31T08:50:18.342Z. App creating and updating happened on yesterday, that is 2015-08-26. It's very strange that Marathon tried to create and modify group with 2015-07-31 version.
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,210] WARN (mesosphere.marathon.api.MarathonExceptionMapper:30)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: mesosphere.marathon.StoreCommandFailedException: Could not modify Group with key: root:2015-07-31T08:50:18.342Z
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at mesosphere.marathon.state.MarathonStore$$anonfun$mesosphere$marathon$state$MarathonStore$$exceptionTransform$1.applyOrElse(MarathonStore.scala:75)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at mesosphere.marathon.state.MarathonStore$$anonfun$mesosphere$marathon$state$MarathonStore$$exceptionTransform$1.applyOrElse(MarathonStore.scala:74)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:36)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.util.Failure$$anonfun$recover$1.apply(Try.scala:215)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.util.Try$.apply(Try.scala:191)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.util.Failure.recover(Try.scala:215)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.concurrent.Future$$anonfun$recover$1.apply(Future.scala:324)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.concurrent.Future$$anonfun$recover$1.apply(Future.scala:324)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at java.lang.Thread.run(Thread.java:745)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: Caused by: mesosphere.marathon.StoreCommandFailedException: Can not create entity group:root:2015-07-31T08:50:18.342Z
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at mesosphere.util.state.zk.ZKStore$$anonfun$exceptionTransform$1.applyOrElse(ZKStore.scala:83)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at mesosphere.util.state.zk.ZKStore$$anonfun$exceptionTransform$1.applyOrElse(ZKStore.scala:82)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011... 10 more
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: Caused by: org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = ConnectionLoss for /marathon/state/group:root:2015-07-31T08:50:18.342Z
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at org.apache.zookeeper.KeeperException.create(KeeperException.java:99)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at com.twitter.zk.AsyncCallbackPromise$class.process(AsyncCallbackPromise.scala:18)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at com.twitter.zk.StringCallbackPromise.process(AsyncCallbackPromise.scala:23)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at com.twitter.zk.StringCallbackPromise.processResult(AsyncCallbackPromise.scala:25)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:605)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:498)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,210] WARN Deployment failed for change: 2015-08-26T10:04:58.043Z (mesosphere.marathon.state.GroupManager:146)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: mesosphere.marathon.StoreCommandFailedException: Could not modify Group with key: root:2015-07-31T08:50:18.342Z
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at mesosphere.marathon.state.MarathonStore$$anonfun$mesosphere$marathon$state$MarathonStore$$exceptionTransform$1.applyOrElse(MarathonStore.scala:75)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at mesosphere.marathon.state.MarathonStore$$anonfun$mesosphere$marathon$state$MarathonStore$$exceptionTransform$1.applyOrElse(MarathonStore.scala:74)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:36)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.util.Failure$$anonfun$recover$1.apply(Try.scala:215)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.util.Try$.apply(Try.scala:191)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.util.Failure.recover(Try.scala:215)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.concurrent.Future$$anonfun$recover$1.apply(Future.scala:324)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.concurrent.Future$$anonfun$recover$1.apply(Future.scala:324)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at java.lang.Thread.run(Thread.java:745)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: Caused by: mesosphere.marathon.StoreCommandFailedException: Can not create entity group:root:2015-07-31T08:50:18.342Z
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at mesosphere.util.state.zk.ZKStore$$anonfun$exceptionTransform$1.applyOrElse(ZKStore.scala:83)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at mesosphere.util.state.zk.ZKStore$$anonfun$exceptionTransform$1.applyOrElse(ZKStore.scala:82)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011... 10 more
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: Caused by: org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = ConnectionLoss for /marathon/state/group:root:2015-07-31T08:50:18.342Z
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at org.apache.zookeeper.KeeperException.create(KeeperException.java:99)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at com.twitter.zk.AsyncCallbackPromise$class.process(AsyncCallbackPromise.scala:18)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at com.twitter.zk.StringCallbackPromise.process(AsyncCallbackPromise.scala:23)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at com.twitter.zk.StringCallbackPromise.processResult(AsyncCallbackPromise.scala:25)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:605)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: #011at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:498)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [INFO] [08/26/2015 10:05:09.214] [marathon-akka.actor.default-dispatcher-4] [akka://marathon/user/$b] POSTing to all endpoints.
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,215] INFO <IP_ADDRESS> - - [26/Aug/2015:10:04:58 +0000] "POST /v2/apps/ HTTP/1.1" 500 76 "-" "curl/7.37.1" (mesosphere.chaos.http.ChaosRequestLog:15)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,320] INFO Opening socket connection to server <IP_ADDRESS>/<IP_ADDRESS>:2181. Will not attempt to authenticate using SASL (unknown error) (org.apache.zookeeper.ClientCnxn:975)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,320] INFO Socket connection established to <IP_ADDRESS>/<IP_ADDRESS>:2181, initiating session (org.apache.zookeeper.ClientCnxn:852)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,322] INFO Session establishment complete on server <IP_ADDRESS>/<IP_ADDRESS>:2181, sessionid = 0x14f68ce878b000f, negotiated timeout = 1800000 (org.apache.zookeeper.ClientCnxn:1235)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,328] INFO Setting framework ID to 20150713-183426-1459822508-5050-27691-0000 (mesosphere.marathon.MarathonSchedulerService:86)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [INFO] [08/26/2015 10:05:09.328] [pool-2-thread-1] [akka://marathon/user/$b] Sending POST to:http://<IP_ADDRESS>:8080/v1/apps/events/
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,328] INFO Setting framework ID to 20150713-183426-1459822508-5050-27691-0000 (mesosphere.marathon.MarathonSchedulerService:86)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [INFO] [08/26/2015 10:05:09.329] [pool-2-thread-1] [akka://marathon/user/$b] Sending POST to:https://puck.alauda.club:8443/v1/slark
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [INFO] [08/26/2015 10:05:09.329] [pool-2-thread-1] [akka://marathon/user/$b] Sending POST to:https://puck.alauda.club:8443/v1/mirana
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,330] INFO <IP_ADDRESS> - - [26/Aug/2015:10:05:08 +0000] "GET /v2/info HTTP/1.1" 200 977 "-" "ELB-HealthChecker/1.0" (mesosphere.chaos.http.ChaosRequestLog:15)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,330] INFO <IP_ADDRESS> - - [26/Aug/2015:10:05:09 +0000] "GET /v2/apps//c30495b3-bf6c-4693-a13a-a0805a78804f HTTP/1.1" 200 1625 "http://<IP_ADDRESS>:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_10_0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/44.0.2403.89 Safari/537.36" (mesosphere.chaos.http.ChaosRequestLog:15)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,331] INFO <IP_ADDRESS> - - [26/Aug/2015:10:05:08 +0000] "GET /v2/info HTTP/1.1" 200 977 "-" "ELB-HealthChecker/1.0" (mesosphere.chaos.http.ChaosRequestLog:15)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,398] INFO <IP_ADDRESS> - - [26/Aug/2015:10:05:09 +0000] "GET /v2/deployments HTTP/1.1" 200 2 "http://<IP_ADDRESS>:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_10_0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/44.0.2403.89 Safari/537.36" (mesosphere.chaos.http.ChaosRequestLog:15)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,429] INFO started processing 1 offers, launching at most 1 tasks per offer and 1000 tasks in total (mesosphere.marathon.tasks.IterativeOfferMatcher$:129)
Aug 26 10:05:09 ip-172-31-11-94 marathon[4884]: [2015-08-26 10:05:09,429] INFO Offer [20150720-062308-1577787308-5050-21885-O36819081]. Decline with default filter refuseSeconds (use --decline_offer_duration to configure) (mesosphere.marathon.tasks.IterativeOfferMatcher$:228)
Another odd thing is the group version is always 2015-07-13T20:47:41.034Z which query by /v2/groups. I look into the content of Zookeeper node /marathon/state/group:root, the version indeed is 2015-07-13T20:47:41.034. It's seem the version hadn't changes from first created.
If the problem is related to big group. we have thousand Apps in root group. The size of zookeeper node /marathon/state/group:root is more than 500K now. is this problem related to big group? actually we don't use Marathon's group feature. Should we create separate group for each app to prevent large group? thanks
any idea about this?
In general, 500k is not a size where I'd expect this kind of problem. See information on jute.maxbuffer in the ZooKeeper Admin Doc
But yes, we've run into problems with big node sizes – the problem, however, is not the size of the node itself, but the packet size when fetching the node. When trying to delete such a node via the zkCli, the output looks like this:
Welcome to ZooKeeper!
2015-08-13 10:47:41,229 [myid:] - INFO [main-SendThread(localhost:2181):ClientCnxn$SendThread@975] - Opening socket connection to server localhost/<IP_ADDRESS>:2181. Will not attempt to authenticate using SASL (unknown error)
JLine support is enabled
2015-08-13 10:47:41,299 [myid:] - INFO [main-SendThread(localhost:2181):ClientCnxn$SendThread@852] - Socket connection established to localhost/<IP_ADDRESS>:2181, initiating session
2015-08-13 10:47:41,308 [myid:] - INFO [main-SendThread(localhost:2181):ClientCnxn$SendThread@1235] - Session establishment complete on server localhost/<IP_ADDRESS>:2181, sessionid = 0x14f25f4aa2200df, negotiated timeout = 30000
WATCHER::
WatchedEvent state:SyncConnected type:None path:null
[zk: localhost:2181(CONNECTED) 0] rmr /universe
2015-08-13 10:47:45,908 [myid:] - WARN [main-SendThread(localhost:2181):ClientCnxn$SendThread@1102] - Session 0x14f25f4aa2200df for server localhost/<IP_ADDRESS>:2181, unexpected error, closing socket connection and attempting reconnect
java.io.IOException: Packet len4198383 is out of range!
at org.apache.zookeeper.ClientCnxnSocket.readLength(ClientCnxnSocket.java:112)
at org.apache.zookeeper.ClientCnxnSocketNIO.doIO(ClientCnxnSocketNIO.java:79)
at org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:366)
at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1081)
WATCHER::
WatchedEvent state:Disconnected type:None path:null
Exception in thread "main" org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = ConnectionLoss for /universe/marathon-demo/state
at org.apache.zookeeper.KeeperException.create(KeeperException.java:99)
at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
at org.apache.zookeeper.ZooKeeper.getChildren(ZooKeeper.java:1472)
at org.apache.zookeeper.ZooKeeper.getChildren(ZooKeeper.java:1500)
at org.apache.zookeeper.ZKUtil.listSubTreeBFS(ZKUtil.java:114)
at org.apache.zookeeper.ZKUtil.deleteRecursive(ZKUtil.java:49)
at org.apache.zookeeper.ZooKeeperMain.processZKCmd(ZooKeeperMain.java:703)
at org.apache.zookeeper.ZooKeeperMain.processCmd(ZooKeeperMain.java:588)
at org.apache.zookeeper.ZooKeeperMain.executeLine(ZooKeeperMain.java:360)
at org.apache.zookeeper.ZooKeeperMain.run(ZooKeeperMain.java:323)
at org.apache.zookeeper.ZooKeeperMain.main(ZooKeeperMain.java:282)
matthias@master-zuob ~ $
Note the Packet len4198383 is out of range! in the stack trace – I have not seen that in the one you provided, so I don't think that you're running into node size problems (at least not yet). However, it seems to be a zk-specific problem. From a marathon point of view, storing state fails, which means that deploying must fail.
Questions
Have you updated any AppDefinitions since the root group's version timestamp?
Can you try to manually (non-destructively) modify the zNode via the zkCli?
Any update/ideas on this? We are seeing a similar issue though our group size or packet size is not that big...
curl --silent -X GET http:///v2/groups | python -mjson.tool | wc
39871 65801 1626350
Hey @mingqi if the group reaches a size of more than 500KB, any update to any app or group will fail with a version of Marathon before 0.13.0. For every change of an AppDefinition we create and store a Deployment in ZK which stores the group before the change and the group after the change. If your group is greater than 500KB, the deployment object will be 2 times this size. The default jute.maxbuffer is 1MB and will exceed this threshold.
For that reason we introduced ZKCompression, which is available in Marathon 0.13. We successfully started 2000 Apps simultaneously with this version. Can you try, if this version solves your problem?
@aameek the group is stored as protobuf. The size of the json object is very different. Can you look into zk, how big the object is?
Just adding my :+1: to this issue. My team and I worked with @jgarcia-mesosphere over the last two weeks to narrow down a problem we were seeing in our dev cluster to this bug. The issue is that Marathon stores group defs in ZK, and the root group seems to store all the data for all child groups. When we got to a point where our /marathon/group:root znode went over about 500kb, Marathon wouldn't be able to process new deployments (we'd see either "conflicting deployment" or "futures timed out" errors in the logs). The max for any znode is 1mb, but Marathon is updating the znode with old group + deployment metadata + updated group, so the current znode doesn't even have to approach 1mb before you start seeing this problem.
Once we understood that Marathon was trying to update the group:root znode in a way that would surpass the hard limit of 1mb per znode, we were able to fiddle with our apps enough to destroy enough of them to get us below 500kb for that group:root znode. That got our dev cluster stable, but we have an upper limit on the number of app groups we can run concurrently now.
It looks like Marathon is scalable up to thousands of tasks, but those tasks need to have small definitions. We've built a "Heroku-like" system on top of Marathon for our devs, and that means that we've got a high number of apps per group, and each app has around 100 environment variables. So when your devs create multiple groups with this kind of configuration, you'll run into scalability issues sooner than you'd expect. For instance, we currently have 323 apps defined in our dev Marathon (only 120 of them actually running), but our group:root znode is over 400kb.
We're going to be rearchitecting our applications to fetch their environments themselves, instead of injecting them via the env section of the app definition. It would be good if Marathon would at least catch the exceptional case where it attempts to stuff more in the group:root znode than it can handle, and long-term to break up these potentially large znodes into parent/child znodes, or use a different data store, or something even more clever that I haven't thought of yet.
For anyone experiencing this issue: try deleting your smallest app or app group, then the next smallest, until you have control of your cluster again. That got us moving again.
@mcclurmc thanks for the info. We are very aware of this limitation.
We introduced a command line parameter --max_apps, so you get an error, if you try to create more apps than that number. This configuration is mostly bandaid to fail fast and does not take the zk node size into account (depending on how much data you store on the app level).
Step 1) we will refactor how the group is stored (only references to the app)
Step 2) we will support more capable persistent stores.
To check for this, pull Marathon statistics from http://<marathon>/metrics and check for the "service.mesosphere.marathon.state.MarathonStore.DeploymentPlan.write-data-size" key to see the largest deploy your Marathon has tried to write.
Is this fixed ?
@kopax not yet
Will it be fixed ?
There is a new persistence layer in 1.4-SNAPSHOT that is currently enabled by default that stores in ZK in a much more scalable fashion. I was able to store 2,500 apps (easily scales more) with 450,000 tasks. The theoretical limit is about a million "objects" of a given type.
Fixed by #4178
|
2025-04-01T04:34:40.167002
| 2016-07-26T15:25:24
|
167636148
|
{
"authors": [
"aquamatthias",
"jdef"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8350",
"repo": "mesosphere/marathon",
"url": "https://github.com/mesosphere/marathon/issues/4146"
}
|
gharchive/issue
|
docker containerizer now allows relative containerPath starting with mesos 1.0
this simplifies the volume API and we should update the validators (and docs) in marathon to allow non-absolute (relative) containerPath
Review is here: https://phabricator.mesosphere.com/D399
|
2025-04-01T04:34:40.167962
| 2019-12-16T04:53:50
|
538184140
|
{
"authors": [
"rishabh96b"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8351",
"repo": "mesosphere/universe",
"url": "https://github.com/mesosphere/universe/pull/2457"
}
|
gharchive/pull-request
|
Revert "[DCOS-61537] Change container type to MESOS in replicator"
Reverts mesosphere/universe#2453 as the customer faced a broken replicator when running it in non VN mode.
Yes it was tested but I'll investigate why it broke on their side.
|
2025-04-01T04:34:40.173466
| 2022-07-28T18:09:03
|
1321309031
|
{
"authors": [
"spitko",
"this-username-is-taken"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8352",
"repo": "messari/subgraphs",
"url": "https://github.com/messari/subgraphs/pull/686"
}
|
gharchive/pull-request
|
Aave v3: Fix stable mode borrow balance, close borrower position on liquidation
Also changed aToken transfer logic so that transfer volume is not included in market/protocol deposit and withdraw volume.
Nice update!
|
2025-04-01T04:34:40.192072
| 2024-02-02T16:19:22
|
2115278634
|
{
"authors": [
"Simone-RM",
"jwinarske"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8353",
"repo": "meta-flutter/meta-flutter",
"url": "https://github.com/meta-flutter/meta-flutter/issues/395"
}
|
gharchive/issue
|
Building on Linux for Arm (aarch64) fails
Hello,
I've added the meta-flutter layer to my yocto configuration (running on Ubuntu 22.04.3). When I try "bitbake flutter-engine" (or any other app, eg. flutter-animated-background), I get the following error:
WARNING: flutter-engine-git-r0 do_fetch: Failed to fetch URL gn://github.com/flutter/engine.git;name=flutter, attempting MIRRORS if available
ERROR: flutter-engine-git-r0 do_fetch: Fetcher failure: Fetch command export PSEUDO_DISABLED=1; export DBUS_SESSION_BUS_ADDRESS="unix:path=/run/user/1000/bus"; export PATH="/home/simone/yocto/engicam-kirkstone/build/tmp/sysroots-uninative/aarch64-linux/usr/bin:/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot-native/usr/bin/python3-native:/home/simone/yocto/engicam-kirkstone/sources/poky/scripts:/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot-native/usr/bin/aarch64-poky-linux:/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot/usr/bin/crossscripts:/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot-native/usr/sbin:/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot-native/usr/bin:/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot-native/sbin:/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot-native/bin:/home/simone/yocto/engicam-kirkstone/sources/poky/bitbake/bin:/home/simone/yocto/engicam-kirkstone/build/tmp/hosttools"; export HOME="/home/simone"; export PATH="/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot-native/usr/share/depot_tools:/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot-native/usr/share/depot_tools/bootstrap-2@3.8.9.chromium.14_bin/python/bin:${PATH}"; export DEPOT_TOOLS_UPDATE=0; export GCLIENT_PY3=0; export CURL_CA_BUNDLE=/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot-native/etc/ssl/certs/ca-certificates.crt; gclient.py config --spec 'solutions = [
{
"managed": False,
"name": "src/flutter",
"url": "https://github.com/flutter/engine.git",
"custom_vars": { "download_android_deps": False, "download_windows_deps": False, "download_linux_deps": False, },
"custom_deps": {}
}
]' && gclient.py sync --shallow --no-history -R -D --revision f40e976bedff57e69e1b3d89a7c2a3c617a03dad -j 8 -v failed with exit code 127, output:
/usr/bin/env: ‘python’: No such file or directory
any idea how to fix it?
Python3 is installed on my host (via the python-is-python3 package):
$ /usr/bin/env python
Python 3.10.12 (main, Nov 20 2023, 15:14:05) [GCC 11.4.0] on linux
I'm using the latest kirkstone branch (commit f4f58f952482793a3ddb1ea3685733e48da45b8a)
Replacing /usr/bin/env python with /usr/bin/python in tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot-native/usr/share/depot_tools/gclient.py seems to help (it's downloading a lot of stuff now). It's not ideal though.
@Simone-RM Hi. This layer is in production in a number of places without issue. So the likelihood is that you have a host configuration issue. Please check that your Yocto system per-requisites are met:
https://docs.yoctoproject.org/4.0.15/ref-manual/system-requirements.html
I think I have all the prerequisites, and I've been building images (without Flutter) on this setup for a while now.
One other bit of information: I'm running Ubuntu on aarch64 (Apple Silicon with Parallels). Has anyone every tried that?
I'm starting to think that is the real problem... changing gclient's shebang allows it to download everything, but then compilation fails:
Generating GN files in: out/linux_debug_arm64
NOTE: is_debug = false use_goma = false goma_dir = "None" full_dart_sdk = false enable_unittests = false skia_enable_flutter_defines = true skia_use_dng_sdk = false skia_use_sfntly = false skia_enable_pdf = false skia_use_x11 = false skia_use_wuffs = true skia_use_expat = true skia_use_fontconfig = true skia_use_icu = true is_official_build = true android_full_debug = false is_clang = true skia_gl_standard = "" use_clang_static_analyzer = false enable_coverage = false enable_lto = true host_os = "linux" host_cpu = "arm64" target_os = "linux" target_cpu = "arm64" dart_target_arch = "arm64" flutter_use_fontconfig = true dart_component_kind = "static_library" embedder_for_target = true dart_lib_export_symbols = false flutter_runtime_mode = "debug" dart_version_git_info = true dart_runtime_mode = "develop" enable_desktop_embeddings = true enable_backtrace = true bssl_use_clang_integrated_as = true target_sysroot = "/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot" custom_sysroot = "/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/recipe-sysroot" custom_toolchain = "/home/simone/yocto/engicam-kirkstone/build/tmp/work/armv8a-poky-linux/flutter-engine/git-r0/src/buildtools/linux-x64/clang" custom_target_triple = "aarch64-unknown-linux-gnu" skia_use_gl = true skia_use_vulkan = true skia_use_vma = false shell_enable_vulkan = true disable_hidden_visibility = false dart_include_wasm_opt = false dart_platform_sdk = true build_glfw_shell = false build_embedder_examples = false stripped_symbols = false dart_use_mallinfo2 = true impeller_enable_3d = true impeller_enable_vulkan = true impeller_enable_opengles = true angle_build_all = false angle_has_astc_encoder = false angle_use_x11 = false angle_use_wayland = false angle_enable_vulkan_validation_layers = false angle_vulkan_headers_dir = "//third_party/vulkan-deps/vulkan-headers/src" angle_vulkan_loader_dir = "//third_party/vulkan-deps/vulkan-loader/src" angle_vulkan_tools_dir = "//third_party/vulkan-deps/vulkan-tools/src" arm_tune = "armv8a+crc+crypto"
Unknown architecture (aarch64) -- unable to run ninja.
No prebuilt ninja binary was found for this system.
Try building your own binary by doing:
cd ~
git clone https://github.com/ninja-build/ninja.git -b v1.8.2
cd ninja && ./configure.py --bootstrap
Then add ~/ninja/ to your PATH.
WARNING: exit code 1 from a shell command.
which is a pity (also considering I do have a prebuilt ninja installed as /usr/bin/ninja)
@Simone-RM Good call on adding the details and renaming the issue.
I have never built a Poky Distro on an Aarch64 machine. Largely because all the machines are so largely under-powered, and I would prefer to watch grass grow :) I build on an AMD Thread Ripper 32+ HW threads, Intel 32 HW threads, AGL builds run on self hosted High Availability servers, and others use Extra Large Virtual CPU instances (AWS). AWS is now offering Yocto image builds as a service.
Are you building a full OS image, or just trying to build Flutter engine artifacts?
If full OS image can you build it short of the meta-flutter recipes?
If just flutter engine artifacts I would suggest a standalone command line build with my flutter-engine patches.
You might also try updating depot-tools to tip of tree, and switching it to python3. It's using an older commit with python2, as this was a requirement for flutter-engine; may have changed in the last six months. You can also try adding ninja-native to DEPENDS, keep in mind ninja is part of depot-tools; which might be an issue with your newer combo (Ubuntu running on apple silicon).
I tried rolling depot-tools to tip of tree main, and it's complaining about missing python3-httplib2. I don't have time to dig into this at the moment.
I have never built a Poky Distro on an Aarch64 machine. Largely because all the machines are so largely under-powered, and I would prefer to watch grass grow :) I build on an AMD Thread Ripper 32+ HW threads, Intel 32 HW threads, AGL builds run on self hosted High Availability servers, and others use Extra Large Virtual CPU instances (AWS). AWS is now offering Yocto image builds as a service.
I was using an AWS machine for my builds, but having everything local on my machine is more practical. And building on a new Mac is not that much slower than on the 32-cores AWS server. ;-)
Are you building a full OS image, or just trying to build Flutter engine artifacts?
full image. I'm trying to add flutter to a working image, just to see if it works/how it performs on the boards we're using.
* If full OS image can you build it short of the meta-flutter recipes?
yes
You might also try updating depot-tools to tip of tree, and switching it to python3. It's using an older commit with python2, as this was a requirement for flutter-engine; may have changed in the last six months. You can also try adding ninja-native to DEPENDS, keep in mind ninja is part of depot-tools; which might be an issue with your newer combo (Ubuntu running on apple silicon).
I'll try something like that. I've seen that there are several scripts that only know about x64 arch as far as Linux is concerned.
I've also tried building an image on Ubuntu 20 on an AWS server, including meta-flutter-animated-background-example: it built correctly, but I suspect that it didn't include everything needed, as I couldn't find a way to launch the app...
I tried building flutter-auto, but it causes the compiler to crash :(
| FAILED: shell/CMakeFiles/flutter-auto.dir/backend/wayland_vulkan.cc.o
| /yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/recipe-sysroot-native/usr/bin/aarch64-poky-linux/aarch64-poky-linux-g++ --sysroot=/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/recipe-sysroot/usr -DASIO_DISABLE_HANDLER_TYPE_REQUIREMENTS -DASIO_HEADER_ONLY -DASIO_NO_DEPRECATED -DASIO_NO_TYPEID -DASIO_STANDALONE -DBUILD_BACKEND_WAYLAND_VULKAN -DCRASH_HANDLER_DSN=\"\" -DCRASH_HANDLER_RELEASE=\"\" -DEGL_NO_X11 -DENABLE_AGL_CLIENT -DENABLE_DART_VM_LOGGING -DENABLE_PLUGIN_ACCESSIBILITY -DENABLE_PLUGIN_COMP_REGION -DENABLE_PLUGIN_COMP_SURF -DENABLE_PLUGIN_DESKTOP_WINDOW -DENABLE_PLUGIN_GOOGLE_SIGN_IN -DENABLE_PLUGIN_GO_ROUTER -DENABLE_PLUGIN_ISOLATE -DENABLE_PLUGIN_KEYBOARD_MANAGER -DENABLE_PLUGIN_KEY_EVENT -DENABLE_PLUGIN_LOGGING -DENABLE_PLUGIN_MOUSE_CURSOR -DENABLE_PLUGIN_PACKAGE_INFO -DENABLE_PLUGIN_PLATFORM -DENABLE_PLUGIN_RESTORATION -DENABLE_PLUGIN_TEXT_INPUT -DENABLE_PLUGIN_URL_LAUNCHER -DENABLE_XDG_CLIENT -DENV64BIT -DFILAMENT_SUPPORTS_WAYLAND -DGIT_BRANCH=\"agl\" -DGIT_HASH=\"3114782\" -DHAVE_MEMFD_CREATE -DHAVE_STRCHRNUL -DLINUX -DMESA_EGL_NO_X11_HEADERS -DNDEBUG -DPATH_PREFIX=\"/usr\" -DWL_EGL_PLATFORM -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/build/shell -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/shell -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/shell/.. -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/shell/logging -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/shell/../third_party/spdlog-1.12.0/include -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/shell/../third_party/asio-1-28-1/asio/include -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/third_party/filament/libs/bluevk/include -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/third_party/. -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/third_party/flutter -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/third_party/flutter/third_party -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/third_party/flutter/third_party/rapidjson/include -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/third_party/flutter/shell/platform/common/public -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/third_party/flutter/shell/platform/common/client_wrapper/include -I/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/third_party/spdlog-1.12.0/include -march=armv8-a+crc+crypto -fstack-protector-strong -O2 -D_FORTIFY_SOURCE=2 -Wformat -Wformat-security -Werror=format-security --sysroot=/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/recipe-sysroot -O2 -pipe -g -feliminate-unused-debug-types -fmacro-prefix-map=/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0=/usr/src/debug/flutter-auto/git-r0 -fdebug-prefix-map=/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0=/usr/src/debug/flutter-auto/git-r0 -fdebug-prefix-map=/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/recipe-sysroot= -fdebug-prefix-map=/yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/recipe-sysroot-native= -fvisibility-inlines-hidden -fno-rtti -DNDEBUG -flto -fno-fat-lto-objects -fPIE -fvisibility-inlines-hidden -Wtrigraphs -Wchar-subscripts -Wcomment -Wreturn-type -Wsequence-point -Wswitch -Wuninitialized -Wunused -Wswitch-bool -Wformat -Wformat-security -Wconversion -Wcast-align -Wunused-parameter -Winvalid-offsetof -fsigned-char -O2 -std=c++17 -MD -MT shell/CMakeFiles/flutter-auto.dir/backend/wayland_vulkan.cc.o -MF shell/CMakeFiles/flutter-auto.dir/backend/wayland_vulkan.cc.o.d -o shell/CMakeFiles/flutter-auto.dir/backend/wayland_vulkan.cc.o -c /yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/shell/backend/wayland_vulkan.cc
| /yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/shell/backend/wayland_vulkan.cc: In member function 'virtual FlutterRendererConfig WaylandVulkanBackend::GetRenderConfig()':
| /yocto/engicam/build_robby/tmp/work/armv8a-mx8mp-poky-linux/flutter-auto/git-r0/git/shell/backend/wayland_vulkan.cc:66:8: internal compiler error: in reshape_init_class, at cp/decl.c:6470
| 66 | }};
| | ^
| 0x186e23e internal_error(char const*, ...)
| ???:0
| 0x6a3710 fancy_abort(char const*, int, char const*)
| ???:0
| 0x714ad9 reshape_init(tree_node*, tree_node*, int)
| ???:0
| 0x6c10dc can_convert_arg(tree_node*, tree_node*, tree_node*, int, int)
| ???:0
| 0x6c10dc can_convert_arg(tree_node*, tree_node*, tree_node*, int, int)
| ???:0
| 0x6b6423 perform_implicit_conversion_flags(tree_node*, tree_node*, int, int)
| ???:0
| 0x86661a convert_for_initialization(tree_node*, tree_node*, tree_node*, int, impl_conv_rhs, tree_node*, int, int)
| ???:0
| 0x867480 check_return_expr(tree_node*, bool*)
| ???:0
| 0x82f663 finish_return_stmt(tree_node*)
| ???:0
| 0x7eb604 c_parse_file()
| ???:0
| 0x8be425 c_common_parse_file()
| ???:0
| Please submit a full bug report,
| with preprocessed source if appropriate.
| Please include the complete backtrace with any bug report.
| See <https://gcc.gnu.org/bugs/> for instructions.
| ninja: build stopped: subcommand failed.
| WARNING: exit code 1 from a shell command.
ERROR: Task (/yocto/engicam/sources/meta-flutter/recipes-graphics/toyota/flutter-auto_git.bb:do_compile) failed with exit code '1'
I have an M2 I rarely use. I've ran Ubuntu on it, and perf was not so great. I will try Ubuntu as a QEMU image with my QEMU launcher app to see if perf is any better: https://github.com/meta-flutter/virt_qemu
That said I also exclusively use a raid10 SSD card on my workstations. Aside from high core count CPU, fast disk IO is the next bottleneck on Yocto builds. You can build a nice Yocto build workstation for ~$5k US. It pays for itself in a few months of regular use.
I would start with standalone building the flutter engine directly on Ubuntu on m2. With that working and if there are a clear set of patches (or none), then doing the Yocto build will be much clearer.
I moved depot-tools to tip of tree. See if that helps
thanks, it does help. Now flutter_engine compiles with a simple bbappend to fix clang's path:
CLANG_PATH = "${WORKDIR}/src/buildtools/linux-arm64/clang"
I'll have to fix the install job too somehow (all the 'clang_x64' instances should be replaced with 'clang_arm64') to work.
flutter-auto compilation crashes just like on the Intel server though.
What value is assigned to HOST_ARCH? I'll remove the DISTRO_FEATURES filter for vulkan backend selection.
HOST_ARCH is aarch64
https://github.com/meta-flutter/meta-flutter/pull/409/files
Passes building flutter-gallery on Mac M2 -> UTM virtualization -> Ubuntu 22
merged
let me know if that unblocks you
I can confirm that this issue can be closed, thanks!
|
2025-04-01T04:34:40.197426
| 2023-10-08T20:11:14
|
1932051687
|
{
"authors": [
"jmikedupont2"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8354",
"repo": "meta-introspector/call-auto-gpt",
"url": "https://github.com/meta-introspector/call-auto-gpt/issues/7"
}
|
gharchive/issue
|
Issues are disabled for this repo
mockopenai_1 | github.GithubException.GithubException: 410 {"message": "Issues are disabled for this repo", "documentation_url": "https://docs.github.com/v3/issues/"}
mockopenai_1 | <IP_ADDRESS> - - [08/Oct/2023 20:04:52] "POST /v1/chat/completions HTTP/1.1" 500 -
/agent-action test
/agent-action test2
/agent-action test23
/agent-action test23
/agent-action test23
test
tst of comment filter
/agent-action do it
/no-agent-action tst
/agent-action pls
/agent-action please run again
test comment
/agent-action lets dance
/agent-action please
|
2025-04-01T04:34:40.202147
| 2024-04-07T13:26:23
|
2229772951
|
{
"authors": [
"jmikedupont2"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8355",
"repo": "meta-introspector/meta-meme",
"url": "https://github.com/meta-introspector/meta-meme/issues/139"
}
|
gharchive/issue
|
The mythical bridge
Paraphrase: The bridge between the knowable and unknowable is created by the myth or narrative, it allows the making of the meaning, the motivation, the inspiration, the world view or perspective of a fundamental observation about bravery, confidence and respect for forces greater than us
source:
New Books Network: Wole Talabi, "Convergence Problems" (Astra Publishing House, 2024)
Starting from: 00:28:34
Episode webpage:
https://newbooksnetwork.com/convergence-problems
claude : You make an insightful observation about the role of myth and narrative in bridging the knowable and unknowable, and providing meaning, motivation, and perspective. Myths and narratives allow us to contextualize and find significance in the fundamental mysteries and forces of the universe that lie beyond our full comprehension.
By weaving stories and archetypes around primal experiences like courage in the face of the unknown, or humility before vastly powerful natural or cosmic phenomena, myths provide a symbolic framework for grappling with the great unknowns of existence. They transform abstract universal truths into resonant tales that speak to the human psyche.
In this way, myths don't just entertain, but illuminate different worldviews and philosophies about our place in the grand scheme. They suggest moral guideposts and inspiring ideals, while leaving room for the ineffable. A potent myth about bravery in confronting the unexplained can instill strength, just as one about nature's majesty can seed awe and respect.
Rather than reducing the unknowable to cold facts, myths preserve a sense of wonderful mystery. They provide a conceptual bridge between our finite minds and the infinite unknowns we intuively sense lurking just beyond the veil of the known world. Myths dignify the quest to understand through imagination and symbolic storytelling.
So in encapsulating fundamental truths about the human condition and our relationship to the unknowable through narrative, myths play a vital role in our psychological, cultural and spiritual understandings of the cosmos and our own existence within it. They are profoundly meaningful creations born from our universal experience of the known and unknown realms.
gemini : Myths and narratives weave a tapestry that bridges the chasm between the knowable and the unknowable. They take our fragmentary understanding of the world and use it to craft stories that illuminate the profound mysteries that surround us. These stories, passed down through generations, become a repository of human experience, offering explanations for natural phenomena, moral codes for navigating complex social interactions, and a sense of purpose in a vast and often uncaring universe. Myths inspire us with tales of bravery and resilience in the face of the unknown, and cultivate a sense of awe and respect for forces beyond our control. They serve as a springboard for our own creativity, prompting us to question, explore, and forge our own unique understanding of the world.
|
2025-04-01T04:34:40.374543
| 2021-07-10T12:15:42
|
941238843
|
{
"authors": [
"Encrylize",
"giarve"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8356",
"repo": "metacall/core",
"url": "https://github.com/metacall/core/pull/166"
}
|
gharchive/pull-request
|
Add function call support to Wasm loader
Description
This PR implements support for function calls in the Wasm loader. Note that multiple return values are not yet supported.
Type of change
[X] New feature (non-breaking change which adds functionality)
Checklist:
[X] I have performed a self-review of my own code.
[X] I have commented my code, particularly in hard-to-understand areas.
[X] I have made corresponding changes to the documentation.
[ ] My changes generate no new warnings.
[X] I have added tests/screenshots (if any) that prove my fix is effective or that my feature works.
[X] I have tested the tests implicated (if any) by my own code and they pass (make test or ctest -VV -R <test-name>).
[X] I have tested my code with OPTION_BUILD_SANITIZER and OPTION_TEST_MEMORYCHECK.
[X] I have run make clang-format in order to format my code and my code follows the style guidelines.
Note that memcheck reports some memory leaks. However, these seem to be caused by Wasmtime and are probably not our fault. I will investigate these further once the loader is more complete and decide whether they should be suppressed or a fix is warranted.
Nice work!
|
2025-04-01T04:34:40.376603
| 2017-07-17T08:10:59
|
243322041
|
{
"authors": [
"blackillzone",
"retr0h"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8357",
"repo": "metacloud/molecule",
"url": "https://github.com/metacloud/molecule/issues/905"
}
|
gharchive/issue
|
Sharing completion file for fish
Hi everyone,
Here is my completion script for Molecule 1.25, adapted for Fish Shell:
https://gist.github.com/blackillzone/ef9962ba652b80a8c7952aae3f8a53dc
Didn't knew where sharing this, so I created a new issue, my apologies if it was the wrong way.
Once I'll migrate to version 2.0 of Molecule, I'll check to do an updated version also.
Feel free to use it ;).
Hi @blackillzone thanks for the info. While we are not actively maintaining Molecule v1 any longer, we appreciate your note. I'm going to close this issue for others to reference. Molecule v2s uses click's internal mechanism for command completion.
|
2025-04-01T04:34:40.390438
| 2021-04-06T09:52:19
|
851256551
|
{
"authors": [
"fsteeg",
"katauber"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8358",
"repo": "metafacture/metafacture-playground",
"url": "https://github.com/metafacture/metafacture-playground/pull/13"
}
|
gharchive/pull-request
|
Generate Link to share API calls
Popup with link to api call of current workflow
For an independent review, this PR should have had its base set to 6-improveUI instead of containing the changes from 6-improveUI and being based on main. But we can now leave it as it is, since it should show a reviewable diff after https://github.com/metafacture/metafacture-playground/pull/12 is merged.
But we can now leave it as it is, since it should show a reviewable diff after #12 is merged.
Hm, no, that did not work. There is one little conflict, but all other changes still show up.
Maybe it would work to rebase 9-generateLink on the current main?
I think this looks better now :)
|
2025-04-01T04:34:40.391621
| 2019-09-10T14:27:26
|
491709071
|
{
"authors": [
"desandro",
"starchild"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8359",
"repo": "metafizzy/infinite-scroll",
"url": "https://github.com/metafizzy/infinite-scroll/issues/893"
}
|
gharchive/issue
|
Show loading image
Not a bug as such, but just wondering how to show a loading image if you reach the end of the container and a next page exists?
Use .infinite-scroll-request status element to display content when Infinite Scroll is requesting a page.
|
2025-04-01T04:34:40.437100
| 2018-09-01T16:55:43
|
356217230
|
{
"authors": [
"aqrln",
"tshemsedinov"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8360",
"repo": "metarhia/globalstorage",
"url": "https://github.com/metarhia/globalstorage/issues/186"
}
|
gharchive/issue
|
Large datasets navigation
Sometimes query may return large datasets, for example gs.select({ category: 'Person' }) called from client will send request to the server, then spread requests to all servers storing mentioned category then start receiving data by chunks in lazy mode. So client-side cursor will receive first chunk and will generate on('data') event and first 100 records will be available at the client-side to be iterated by cursor. But we may not want to transfer more before user navigate to below grid position. Is it ok for ours gui rendering console if cursor will have no all data at once and data will arrive chunk by chunk pushing to dataset? Also how GUI can inform cursor to get next chunk? @aqrln
Here is an example how cursors may fetch data @aqrln
I think we cant use just cursor.dataset: Array, we need something lige fetch(callback(err)) or to load all records to cursor and then it will be available in cursor.dataset or can use event to get by chunks on('data', callback)
@aqrln I forgot a link to mentioned example: https://github.com/metarhia/globalstorage/pull/194
Cursor and work in two modes: (1) dataset holder, (2) data transformation from parent cursor not holding datasets. So cursors are chained but we can materialize datasets at any step. I am going to add cursor.materialize() that will receive all data from parent cursor and save it it's own dataset. This allow us to minimize data copies in memory. Do we need to copy data again to redux store? Or we will not use cursor materialization and will collect aggregated/transformed data just in store?
@tshemsedinov store is immutable and can only be changed with pure functions (reducers) in response to actions, so if something changes, we need to copy it again.
See https://github.com/metarhia/globalstorage/issues/195
|
2025-04-01T04:34:40.443167
| 2021-07-05T07:21:45
|
936777841
|
{
"authors": [
"LachlanStuart",
"sergii-mamedov"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8361",
"repo": "metaspace2020/metaspace",
"url": "https://github.com/metaspace2020/metaspace/pull/901"
}
|
gharchive/pull-request
|
Update Lithops version
We are updating to avoid the problems described here.
copy that
Retroactive LGTM :+1: (we agreed it was fine to merge on Slack)
|
2025-04-01T04:34:40.444581
| 2016-01-04T21:38:47
|
124845067
|
{
"authors": [
"DominikGuzei",
"darko-mijic"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8362",
"repo": "meteor-space/vo-monetary",
"url": "https://github.com/meteor-space/vo-monetary/pull/5"
}
|
gharchive/pull-request
|
introduce better floating point precision for money VOs
This PR adds improved handling of floating point numbers for Money VOs.
The number required decimal places (precision) is calculated on creation of the VO and saved together with the "base" number which is an integer version of the float by multiplication of 10 * decimal places.
Is it OK if I review this in the morning?
Of course, im also going to bed now :wink:
|
2025-04-01T04:34:40.470276
| 2017-03-27T20:04:55
|
217359320
|
{
"authors": [
"sebhoss"
],
"license": "CC0-1.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8363",
"repo": "metio/yosql",
"url": "https://github.com/metio/yosql/issues/30"
}
|
gharchive/issue
|
setup CI/CD
blocked by Bazel install on build.metio.wtf: https://github.com/metio/build.metio.wtf/issues/1
now using github actions
|
2025-04-01T04:34:40.479978
| 2017-11-30T16:36:20
|
278187236
|
{
"authors": [
"MoseleyS",
"carolinejjones"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8364",
"repo": "metoppv/improver",
"url": "https://github.com/metoppv/improver/pull/355"
}
|
gharchive/pull-request
|
318 falling snow level
Reference issue #318
Testing:
[X] Ran tests and they passed OK
[X] Added new tests for the new feature(s)
I have also added a small change to how I find the x and y coordinate.
This PR also fails some CLI tests because the test data are more recent than this branch's upstream/master. This does not prevent the PR being completed or merged and could be overcome by rebasing the branch if we wanted to.
|
2025-04-01T04:34:40.498525
| 2020-01-21T22:42:06
|
553186843
|
{
"authors": [
"svartalf",
"tobz"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8365",
"repo": "metrics-rs/metrics",
"url": "https://github.com/metrics-rs/metrics/issues/66"
}
|
gharchive/issue
|
Consuming external metrics with absolute values
It is very common to expose some external metrics as counters created outside of the current application, for example, by collecting system statistics (ex. total amount of bytes sent via the network interface) or fetching them from other application (as various Prometheus exporters do).
The problem is that these metrics has absolute values and with current API it is possible to increment counters only and there is no way to replace the value.
From my understanding, something like Constant Metrics can be added
to allow one-time loading of the metric values; this idea better to be discussed further.
Naive way would be to add Recorder::set_counter method, which change the value of the inner AtomicU64.
Obviously, it is not the best solution possible, yet, it can work as a temporary workaround till better idea will land the metrics crate; as an addition, it can be feature-gated with something like #[cfg(feature = "unstable")].
Extra links:
https://github.com/tikv/rust-prometheus/pull/303
https://www.robustperception.io/setting-a-prometheus-counter
https://github.com/prometheus-net/prometheus-net/issues/73
There's some renewed activity around this topic in #172 based on improvements we've made in metrics 0.13 that allow us to inject logic to the pipeline without having to necessarily extend the macros themselves.
|
2025-04-01T04:34:40.499828
| 2021-02-25T13:31:46
|
816439827
|
{
"authors": [
"0nkery",
"tobz"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8366",
"repo": "metrics-rs/metrics",
"url": "https://github.com/metrics-rs/metrics/pull/180"
}
|
gharchive/pull-request
|
Observer: Use BTreeMap as metrics storage to sort by key
This change allows preserving alphabetical order in the observer interface.
Nice! Looks good to me.
Thanks for your contribution!
Thanks for merging!
|
2025-04-01T04:34:40.504502
| 2019-06-19T22:44:45
|
458257970
|
{
"authors": [
"queer"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8367",
"repo": "mewna/catnip",
"url": "https://github.com/mewna/catnip/issues/334"
}
|
gharchive/issue
|
Emit a warning event on high heartbeat latency
Self-explanatory; also makes it so users don't have to measure it themselves.
Added in 98e1a8ed
|
2025-04-01T04:34:40.515577
| 2020-10-05T07:10:09
|
714581002
|
{
"authors": [
"Farnazmdi",
"aarmey",
"codecov-commenter",
"jclagarde"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8368",
"repo": "meyer-lab/tHMM",
"url": "https://github.com/meyer-lab/tHMM/pull/646"
}
|
gharchive/pull-request
|
Added Supp. Figure Captions
I'm not quite sure where the text explanation that was previously in the caption section should go. I added it above the figure though.
@JC Lagarde I have seen many papers without explaining their supplementary figures, and only providing a caption. Besides, we have mentioned these figures in the main results where it was relevant. @aarmey Do we want a paragraph explaining each supplementary figure?
Codecov Report
Merging #646 into master will not change coverage.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #646 +/- ##
=======================================
Coverage 88.50% 88.50%
=======================================
Files 21 21
Lines 1505 1505
=======================================
Hits 1332 1332
Misses 173 173
Flag
Coverage Δ
#unittests
88.50% <ø> (ø)
Flags with carried forward coverage won't be shown. Click here to find out more.
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 5d25984...0f9f558. Read the comment docs.
@Farnazmdi you're right that the captions likely should do this. Optionally, we could include a supplementary methods section with a more detailed description.
|
2025-04-01T04:34:40.544260
| 2020-04-30T10:04:46
|
609782066
|
{
"authors": [
"makeclean",
"tzanio",
"v-dobrev"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8369",
"repo": "mfem/mfem",
"url": "https://github.com/mfem/mfem/issues/1452"
}
|
gharchive/issue
|
Compile fail on nurb_ex1p
Hi
Was just compiling mfem (master) on my new machine;
gcc - v gcc version 9.3.1 20200408 (Red Hat 9.3.1-2) (GCC)
mpich 3.3
cmake -DHYPRE_DIR=$WORKDIR/hypre -DMETIS_DIR=$WORKDIR/metis -DMFEM_ENABLE_EXAMPLES=ON -DMFEM_ENABLE_MINIAPPS=ON -DMFEM_THREAD_SAFE=ON -DMFEM_USE_MPI=ON -DMFEM_USE_OPENMP=ON -DCMAKE_INSTALL_PREFIX=..
Build for HYPRE & METIS went fine, all of mfem with the exception of this example compiles, the failure message is
/home/adavis/opt/mfem/miniapps/nurbs/nurbs_ex1p.cpp:116:27: error: cannot convert ‘mfem::Vector [nd]’ to ‘mfem::Vector&’
116 | el.CalcShape(ip, shape);
| ^~~~~
| |
| mfem::Vector [nd]
In file included from /home/adavis/opt/mfem/mesh/triangle.hpp:16,
from /home/adavis/opt/mfem/mesh/mesh_headers.hpp:21,
from /home/adavis/opt/mfem/mfem.hpp:42,
from /home/adavis/opt/mfem/bld/mfem.hpp:3,
from /home/adavis/opt/mfem/miniapps/nurbs/nurbs_ex1p.cpp:44:
/home/adavis/opt/mfem/mesh/../fem/fe.hpp:344:35: note: initializing argument 2 of ‘virtual void mfem::FiniteElement::CalcShape(const mfem::IntegrationPoint&, mfem::Vector&) const’
344 | Vector &shape) const = 0;
| ~~~~~~~~^~~~~
/home/adavis/opt/mfem/miniapps/nurbs/nurbs_ex1p.cpp:128:40: error: ‘shape’ cannot be used as a function
128 | elmat(i, j) += w*shape(i)*laplace(j);
| ^
make[2]: *** [miniapps/nurbs/CMakeFiles/nurbs_ex1p.dir/build.make:83: miniapps/nurbs/CMakeFiles/nurbs_ex1p.dir/nurbs_ex1p.cpp.o] Error 1
make[1]: *** [CMakeFiles/Makefile2:3517: miniapps/nurbs/CMakeFiles/nurbs_ex1p.dir/all] Error 2
make: *** [Makefile:161: all] Error 2
I suppose I could comment out this example, but I was actually interested in playing with it.
There is a typo here: https://github.com/mfem/mfem/blob/7690eca85041e0b69afe81a063c71015184e82a3/miniapps/nurbs/nurbs_ex1p.cpp#L78
The line should be:
Vector shape(nd);
Unfortunately, we missed it because we don't test regularly with -DMFEM_THREAD_SAFE=ON.
Thanks for reporting the issue. We will fix it soon in master.
theres a copy-pasta in nurbs_ex1.cpp too
Thanks for reporting this @makeclean
Discussion continued in #1455
|
2025-04-01T04:34:40.550571
| 2022-07-25T20:57:23
|
1317360385
|
{
"authors": [
"prathamgit",
"tzanio",
"v-dobrev"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8370",
"repo": "mfem/mfem",
"url": "https://github.com/mfem/mfem/issues/3116"
}
|
gharchive/issue
|
Minor correction in Mfem_Assert call in sparsrMat class
Should it be B.width in the line << B.height << ", B.width = " << width in SparseMatric class ?
https://github.com/mfem/mfem/blob/670b7b1ae32f2126df8f4539a238b0aa4b8e33fe/linalg/sparsemat.cpp#L3059
Looks correct to me, the full line with some reformatting is essentially
cerr << "Mismatch of this matrix size and rhs. "
<< "This height = " << height
<< ", width = " << width
<< ", B.height = " << B.height
<< ", B.width = " << width;
@tzanio, I think the point is to replace ", B.width = " << width --> ", B.width = " << B.width.
Oh, that makes sense, sorry.
It should be indeed << ", B.width = " << B.width in the last line.
@prathamgit, do you want to propose a simple PR to fix it?
@tzanio i did not understand what is "simple PR fix" ?
Oh I just meant you can put the fix in a branch and issue a pull request to fix it in master.
Oh okay. Yes I can do that. Thanks!
Hi @prathamgit, do you still plan to make a pull request for this?
Oh I have, a few days ago.
Can you post a link to it here?
Right sorry. Here is the link.
https://github.com/mfem/mfem/pull/3161
I hope this link works.
Addressed in https://github.com/mfem/mfem/pull/3161
|
2025-04-01T04:34:40.555851
| 2023-02-12T20:32:40
|
1581399712
|
{
"authors": [
"Apollon77"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8371",
"repo": "mfucci/node-matter",
"url": "https://github.com/mfucci/node-matter/pull/247"
}
|
gharchive/pull-request
|
Really check for Statuscode errors and return Error classes
This PR uses the new Error classes to instoduce StatusError classes that are thrown when an unexpected StatusCode is returned.
This also fixes the "waitForSucces" to really wait for a success reswponse and not for "any" response
Ok, I checked more in deoth the throwIfError and waitForSuccess and optimized more ... hopefully now less "duplicate" code
@mfucci Ready for re-reveiw ... I adjusted ... you were right regarding duplicste throwIfError
Ready for re-review
|
2025-04-01T04:34:40.569971
| 2015-05-13T12:36:50
|
75970501
|
{
"authors": [
"alexandrujuncu",
"mgax"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8372",
"repo": "mgax/jekyllcms",
"url": "https://github.com/mgax/jekyllcms/issues/2"
}
|
gharchive/issue
|
Template editor
Feature request: a way of editing the template online.
For now, I think GitHub's web editor is adequate for this, I'd like to focus on content editing.
|
2025-04-01T04:34:40.591801
| 2016-06-06T20:49:00
|
158771416
|
{
"authors": [
"TheDonDope",
"gvsdan",
"hank-ehly"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8373",
"repo": "mgechev/angular2-seed",
"url": "https://github.com/mgechev/angular2-seed/issues/971"
}
|
gharchive/issue
|
Enable SCSS #952 Integration is not injecting the npm dependencies
i'm trying to use the scss #952 integration in development mode is working fine but when i use the prod mode is not creating the '.css' for all the npm dependencies injected in the 'project.config.ts'
@gvsdan, thank you for your comment. I'll take a look.
@hank-ehly Thanks to you for making the sass support
Fixed by #976, thank you very much Hank!
@hank-ehly Everything is working fine but now the external dependencies in production are broken.
|
2025-04-01T04:34:40.592806
| 2017-09-23T11:42:13
|
260002218
|
{
"authors": [
"arielhenryson",
"wKoza"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8374",
"repo": "mgechev/codelyzer",
"url": "https://github.com/mgechev/codelyzer/issues/420"
}
|
gharchive/issue
|
no-access-missing-member not finding super class method
When "no-access-missing-member": true, and one of the template try to access member that was inherited from superclasses the error is thrown
Please, take a look at the existing issues.
|
2025-04-01T04:34:40.594099
| 2022-02-13T16:43:17
|
1136115520
|
{
"authors": [
"neclimdul"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8375",
"repo": "mglaman/phpstan-drupal",
"url": "https://github.com/mglaman/phpstan-drupal/pull/327"
}
|
gharchive/pull-request
|
Provide a stub for twig_escape_filter
twig_escape_filter is a hidden function inside a class file so not discovered
though normal discovery or autoloading so providing scan stub allows its use
to be analyzed.
Fixes #310
I couldn't recreate it outside the Drupal repository either but this fixes it.
|
2025-04-01T04:34:40.620528
| 2011-12-15T15:54:50
|
2568548
|
{
"authors": [
"benjaminoakes",
"mgsnova"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8378",
"repo": "mgsnova/feature",
"url": "https://github.com/mgsnova/feature/issues/1"
}
|
gharchive/issue
|
Documentation: Rails example
Hi mgsnova,
Thanks for writing feature. It makes sense to have a standard library for something like this. We have something similar that looks at a MySQL database (which we may contribute to feature sometime).
In the meantime, we have a simple Rails app that doesn't have a MySQL database, so your gem fits the bill nicely. To say thanks, I wrote up some docs about how I made it work and made this pull request.
Happy holidays!
Ben
Hi Ben,
thanks for using the gem and giving feedback.
Can you please remove the files added with commit 4c2b1bb. I think they are not really needed now in the code. I will merge the pull request then.
Best regards,
mgsnova
Hi @mgsnova,
Happy to. I actually didn't mean for them to show up in this pull request, but they did because I pushed to my fork. Sorry about that. Next time, I guess I should do the pull request from a branch. :)
Best wishes,
Ben
To keep it simple, I just did git revert 4c2b1bb04c2f22d87947e1afb8f822cf33f51ba5.
Best wishes and thanks again!
Ben
|
2025-04-01T04:34:40.640948
| 2017-03-08T23:27:05
|
212888602
|
{
"authors": [
"coveralls",
"femtotrader",
"mhallsmoore",
"ryankennedyio"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8381",
"repo": "mhallsmoore/qstrader",
"url": "https://github.com/mhallsmoore/qstrader/pull/190"
}
|
gharchive/pull-request
|
Live trade session
Minor refactor of the original backtest.py class to also allow for live trading. Live trading only differs in that it's "end-condition" is the passing of time, until and end_session_time. All loop logic remains the same.
This branch is a dependancy of https://github.com/mhallsmoore/qstrader/pull/186. I created a new branch here as the logic has no overlap with IB integration.
I'm suggesting we don't do the IBService setup/teardown in the TradingSession -- we must do it in the setup scripts where the user defines their strategy, symbols, handlers, etc.
Otherwise we need to create a TradingSession for each service, or create an abstraction where we take in a list of services (unnecessarily complex). As it stands, the TradingSession simply takes in a set of QSTrader components (price, risk, execution handlers, etc) -- each of these components may have services/external system dependancies, but the trading loop does not -and should not- care.
Would appreciate prompt integration with master so I can get cracking with live market data in the IB-PriceHandler
Coverage increased (+0.04%) to 57.143% when pulling fdb7a55e09b112e2a43136cd539c9b54b459e87b on ryankennedyio:live_trade_session into ca05a06cf8429f571d566a39409f674142676af3 on mhallsmoore:master.
I wonder if such a class can handle
backtest
live trade
but also paper trader (live price feed but simulated execution)
I really think that an abstract class AbstractTradingSession and 3 inherited classes BacktestSession, LiveTradeSession, PaperTradeSession is a cleaner approach.
But renaming simulate_trading method to start_trading (or simply start or run) is a good idea.
No problem - promptly done! ;-)
If you feel you need to change some of the logic in the previous (to this merge) Backtest object (which is now TradingSession), go ahead. I modified it to make the API cleaner, such that we didn't end up with these rather large ***_backtest.py files containing all of the instantiations of the respective components.
I'm definitely open to trying to find ways to make this simpler for new users, but at the same time extensible for those who wish to tinker a bit more.
Thanks @mhallsmoore
@femtotrader with regard to paper trade, don't all brokers offer a separate paper trading account ? I.e. from the system point of view, live is the same as paper?
I think even if we did do that, the trading loop logic should be the same; it's just the ExecutionHandler and Portfolio classes wouldn't have feeds to the broker for live execution/portfolio stats?
Just IMO more abstraction here probably isn't going to be used too often? I think we should have abstract base classes where we expect that users will implement their own inherited classes "fairly" often. I guess if we find people often need to override the trading sessions, we can add the abstraction in.
|
2025-04-01T04:34:40.679033
| 2021-08-24T13:14:01
|
978086703
|
{
"authors": [
"mheob"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8382",
"repo": "mheob/vscode-snippets",
"url": "https://github.com/mheob/vscode-snippets/issues/4"
}
|
gharchive/issue
|
[FR] Add next.js snippets
To improve the extension, there should be the most common next.js snippets like for example:
getStaticProps
getServerSideProps
…
Fixed in #14
|
2025-04-01T04:34:40.691291
| 2024-04-15T15:42:24
|
2244014619
|
{
"authors": [
"akhiljns",
"mhewedy"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8383",
"repo": "mhewedy/vermin",
"url": "https://github.com/mhewedy/vermin/issues/66"
}
|
gharchive/issue
|
not able to run commands from shell script after provisioning vm
Provisioning vm_01
: invalid optionline 3: set: -
set: usage: set [-abefhkmnptuvxBCHP] [-o option-name] [--] [arg ...]
/tmp/master.sh: line 4: $'\r': command not found
/tmp/master.sh: line 8: $'\r': command not found
Could not set property: Invalid pretty host name '<IP_ADDRESS>
'
/tmp/master.sh: line 12: $'\r': command not found
: No such file or directory
/tmp/master.sh: line 14: $'\r': command not found
Warning: apt-key output should not be parsed (stdout is not a terminal)
Executing: /tmp/apt-key-gpghome.TIP5MpsQmS/gpg.1.sh --keyserver keyserver.ubuntu.com --recv-keys CC86BB64
" not a key ID: skipping
Usage: add-apt-repository <sourceline>
add-apt-repository is a script for adding apt sources.list entries.
It can be used to add any repository and also provides a shorthand
syntax for adding a Launchpad PPA (Personal Package Archive)
repository.
<sourceline> - The apt repository source line to add. This is one of:
a complete apt line in quotes,
a repo url and areas in quotes (areas defaults to 'main')
a PPA shortcut.
a distro component
Examples:
apt-add-repository 'deb http://myserver/path/to/repo stable myrepo'
apt-add-repository 'http://myserver/path/to/repo myrepo'
apt-add-repository 'https://packages.medibuntu.org free non-free'
apt-add-repository http://extras.ubuntu.com/ubuntu
apt-add-repository ppa:user/repository
apt-add-repository ppa:user/distro/repository
apt-add-repository multiverse
If --remove is given the tool will remove the given sourceline from your
sources.list
add-apt-repository: error: no such option: -
E: Command line option '
' [from -y
] is not understood in combination with the other options.
E: Command line option '
' [from -y
] is not understood in combination with the other options.
sudo: yq: command not found
sudo: yq: command not found
sudo: yq: command not found
sudo: yq: command not found
sudo: yq: command not found
usage: /usr/sbin/netplan [-h] [--debug] ...
/usr/sbin/netplan : error: argument : invalid choice: 'apply\r' (choose from 'help', 'apply', 'generate', 'get', 'info', 'ip', 'set', 'rebind', 'try')
sleep: invalid time interval ‘1\r’
Try 'sleep --help' for more information.
/tmp/master.sh: line 28: $'\r': command not found
Warning: apt-key output should not be parsed (stdout is not a terminal)
"pg: invalid option "-
' invalid
/tmp/master.sh: line 34: deb [arch=amd64] https://download.docker.com/linux/ubuntu \
\bionic
: No such file or directory
Warning: apt-key output should not be parsed (stdout is not a terminal)
"pg: invalid option "-
/tmp/master.sh: line 78: warning: here-document at line 36 delimited by end-of-file (wanted `EOF')
deb https://apt.kubernetes.io/ kubernetes-xenial main
EOF
sudo apt-get update
sudo apt-get install -y \
containerd.io=1.2.13-1 \
\docker-ce=5:19.03.8~3-0~ubuntu-bionic
docker-ce-cli=5:19.03.8~3-0~ubuntu-bionic
-00 kubectl=1.18.0ll -y kubelet=1.18.0
sudo apt-mark hold docker-ce kubelet kubeadm kubectl
echo "net.bridge.bridge-nf-call-iptables=1" | sudo tee -a /etc/sysctl.conf
sudo sysctl -p
## configure docker
cat <<EOF | sudo tee /etc/docker/daemon.json
{
"exec-opts": ["native.cgroupdriver=systemd"],
"log-driver": "json-file",
"log-opts": {
"max-size": "100m"
},
"storage-driver": "overlay2"
}
EOF
@mhewedy so turns out this is not the problem of vermin
the sh file i'm trying to run is master.sh inside /etc/samples-provision/k8s/master.sh
the vagrant image i'm using is ubuntu/focal64
I tried running the file from inside the vm and i'm getting same issues
@mhewedy I'm seeing strange behaviour
when i'm trying to run the script using ./master.sh it fails but individual commands are running fine, this might be a problem of CRLF debugging more
Ok upon debugging CRLF was the issue, but there are other issues as well with the master.sh
the yq commands you have used are older and v4 version supports new syntax, i'm trying to make this script so that it supports newer builds
Ok upon debugging CRLF was the issue, but there are other issues as well with the master.sh
the yq commands you have used are older and v4 version supports new syntax, i'm trying to make this script so that it supports newer builds
Good analysis, thanks for your efforts
|
2025-04-01T04:34:40.697579
| 2017-11-24T22:00:30
|
276703804
|
{
"authors": [
"jorgesca",
"mhinz"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8384",
"repo": "mhinz/neovim-remote",
"url": "https://github.com/mhinz/neovim-remote/issues/68"
}
|
gharchive/issue
|
Option NOT to start a neovim process?
I don't know if this makes sense, here's the issue I have setup and IDE to start neovim for editing files. I use the terminal version of neovim. If the server has not been started when I ask the IDE to open a file, nvr will start a neovim instance, which actually blocks the IDE itself. Is there a way to avoid this?
If I start a neovim server first, all works fine:
$ NVIM_LISTEN_ADDRESS=/tmp/nvim_quartus nvim
The IDE is configured to do this for opening files:
/usr/bin/nvim --servername /tmp/nvim_quartus --remote +%l %f (where %l gets replaced by the line number and %f by the file name)
Oh, I see. This was meant for people who alias nvr to nvim.
I'll add a --nostart option!
Excellent, thanks!
I'll upload a new version together with some other fixes later today.
Thanks for reporting!
|
2025-04-01T04:34:40.728135
| 2023-05-08T16:50:45
|
1700589161
|
{
"authors": [
"miacias"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8385",
"repo": "miacias/brew-buddies",
"url": "https://github.com/miacias/brew-buddies/issues/38"
}
|
gharchive/issue
|
GoogleMaps: Create test map
make a map on a React page, such as Map.js
center the map on a location (set location to a variable so the user's home zip code can be used by default, else a standard address in USA)
rendered with "@googlemaps/react-wrapper"
|
2025-04-01T04:34:40.737551
| 2020-02-26T00:19:28
|
570949485
|
{
"authors": [
"HewmEvor",
"sajalmandrekar"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8386",
"repo": "micahflee/torbrowser-launcher",
"url": "https://github.com/micahflee/torbrowser-launcher/issues/454"
}
|
gharchive/issue
|
SIGNATURE VERIFICATION FAILED
Just following the instruction attached to failed verification to report this error here.
Process used (can be duplicated (many times))
Use sudo to install TOR
takes 10 minutes to start the d/l and install Launcher
Open Launcher, and 13 minutes to start d/l actual package.
.10 seconds to start Dl and install
Verifying starts, and 13 minutes later, the generic error pops open.
I have repeated the key resets, purge install, reboot, and then re-try.
This has been tried several times using different sites different suggestions - nothing has worked.
RIG:
Intel i3-3110M CPU@ 2.40GHZ
Memory 6G
Ubuntu 18.04.3 LTS
Tried install from:
Software Boutique
sudo app-get
github.com/micahflee/torbrowser-launcher/issues/
Yes - this may be "similar to 4 existing issues" but not same. - it seems that same matter with this issue.
It may be due to your Tor being outdated. Try updating it and see if it works.This repository provides a latest version.
Run the following commands on terminal:
sudo add-apt-repository ppa:micahflee/ppa
sudo apt install torbrowser-launcher
Or check out this thread:
https://github.com/micahflee/torbrowser-launcher/issues/410
|
2025-04-01T04:34:40.739814
| 2015-04-01T19:32:03
|
65761659
|
{
"authors": [
"michael-lazar",
"yskmt"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8387",
"repo": "michael-lazar/rtv",
"url": "https://github.com/michael-lazar/rtv/pull/59"
}
|
gharchive/pull-request
|
comment/reply edit and delete functions added
There are some common parts in add_comment, edit_comment and delete_comment, which can be refactored. Also, sometimes the comment deletion and edition are not immediately updated. Is it a PRAW problem?
My guess would be that Reddit is caching the submission on the server side and only updates the cache every so often.
I agree there is some work that could be done to refactor this. There is a lot of code that will be shared between the SubmissionPage and the SubredditPage.
For deleting, how about a confirmation prompt
Are you sure you want to delete this post? (y/n):
In order to rebase to the newest master, I will close this pull request.
|
2025-04-01T04:34:40.752280
| 2020-11-21T20:32:58
|
748098084
|
{
"authors": [
"MaazArfi",
"akkiiiiii"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8388",
"repo": "michael-rapp/ChromeLikeTabSwitcher",
"url": "https://github.com/michael-rapp/ChromeLikeTabSwitcher/issues/33"
}
|
gharchive/issue
|
Save instance
I am making a browser app and I want to save the instance of each and every tab but I don't know how if someone knows how I can do this please tell me because I see in the issue section there are a lot of guys talking about they also used this library in their browsers so please tell me how can I save the state of each and every tab by using this library.
there no inbuild method i guess
you need write your own logic by using : room or sqlite
|
2025-04-01T04:34:40.767990
| 2015-12-01T01:54:29
|
119623340
|
{
"authors": [
"hut8",
"izabelavmelo",
"michaelbromley",
"nathan-mickler"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8389",
"repo": "michaelbromley/angularUtils",
"url": "https://github.com/michaelbromley/angularUtils/issues/269"
}
|
gharchive/issue
|
When calling $compile on a code containing dirPagination, registered pagination-id is not found
I'm dynamically pulling in HTML code and running it through $compile after adding it to my DOM and everything seems to be working except that I'm now getting the following error message: pagination directive: the itemsPerPage id argument (id: locations) does not match a registered pagination-id
It appears that the dir-paginate directive is getting compiled before the dir-pagination-controls directive and is therefore unable to find the registered pagination-id. How do I get around this?
Here is a simple plunkr demonstrating the problem.
The same data set is pulled in two times (on top: not working, and on bottom: working)
http://plnkr.co/edit/t7YNiM
Hi,
The way the directive itself works involved dynamically calling $compile() after doing some manipulation - adding an ngRepeat directive etc. I can imagine there is some conflict arising perhaps to do with the order of execution of the calls to $compile.
In your plunk I do not see the error you reported above - I get uncaught Error: [$compile:ctreq] Controller 'ngInclude', required by directive 'ngInclude', can't be found!.
I can't say when I'll be able to put in some time to investigating this - I have limited free time to work on issues here at the moment. I will label it as a bug to fix though.
I may have accidentally updated that plunkr since then with other attempts
to get this working...I intended to create a new plunkr with updates. If
you find out another way I can inject the code dynamically, please let me
know.
Thanks.
On Tuesday, December 1, 2015, Michael Bromley<EMAIL_ADDRESS>wrote:
Hi,
The way the directive itself works involved dynamically calling $compile()
after doing some manipulation - adding an ngRepeat directive etc. I can
imagine there is some conflict arising perhaps to do with the order of
execution of the calls to $compile.
In your plunk I do not see the error you reported above - I get uncaught
Error: [$compile:ctreq] Controller 'ngInclude', required by directive
'ngInclude', can't be found!.
I can't say when I'll be able to put in some time to investigating this -
I have limited free time to work on issues here at the moment. I will label
it as a bug to fix though.
—
Reply to this email directly or view it on GitHub
https://github.com/michaelbromley/angularUtils/issues/269#issuecomment-161072810
.
I figured out a solution by editing dirPagination.js (see: http://plnkr.co/edit/KOzZGt)
If I modify line #236 as follows, then it works to dynamically inject the content.
var paginationId = scope.paginationId || attrs.paginationId || DEFAULT_ID;
...to...
var paginationId = attrs.paginationId || DEFAULT_ID;
Are there side effects that this will have that I'm nor aware of?
Ok, thanks. I suspect that there is a specific reason why I wrote it that way, but perhaps it covers a case you don't need. I'll run it through the test suite when I get time and see what happens.
Thanks for sharing your fix, that's very helpful in the mean time.
Hi,
Finally got round to looking at this again. Making your suggested change and running through the tests, I get a failure of this test
dirPagination directive dynamic pagination ids should allow independent pagination FAILED
Expected [ '1', '2', '3' ] to equal [ '4', '5' ].
So it's not viable for me to make that change in the source.
I believe this is causing a problem I'm having with using this package with angular-ui-router. I'm able to reproduce it like this:
Hit a page with the pagination directive. It works fine.
Navigate to a different angular-ui-router state which also contains the directive (without reloading the entire page)
It breaks
I have limited time on this project so I can't look into a fix right now, sorry :frowning: Any ideas what a change might be that could fix it?
We have the same problem in our project. We used more than 1 pagination in the same page, and this is being dynamically loaded (according to the number passed by the user).
The scope in "dirPaginationCompileFn" is not accessible. So, when the pagination-id is dynamically passed, it can't evaluate "$parse(attrs.paginationId)(scope)"
|
2025-04-01T04:34:40.771266
| 2021-07-07T07:10:24
|
938565527
|
{
"authors": [
"michaelbrooks",
"sandeepsn1997"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8390",
"repo": "michaelbrooks/rq-win",
"url": "https://github.com/michaelbrooks/rq-win/pull/11"
}
|
gharchive/pull-request
|
fix for issue #9 => prepare_job_execution() takes 2 positional arguments but 3 were given
"prepare_job_execution" takes 2 input param including self, but the calling statement passes 3 with additional "heartbeat_ttl" which is not used by the method, I have removed the extra input param and issue seems to be solved for me.
Thank you! I see this interface was changed in rq here: https://github.com/rq/rq/pull/1372
Could you update setup.py to version 0.4.0 and make the minimum rq version 1.8.1? I'm not sure that this will behave as expected with older rq versions.
I have updated setup.py to version 0.4.0 and made the minimum rq version 1.8.1
Thank you!
|
2025-04-01T04:34:40.881008
| 2015-12-18T08:05:58
|
122901507
|
{
"authors": [
"C-Pro",
"michaelpq"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8391",
"repo": "michaelpq/michaelpq.github.io",
"url": "https://github.com/michaelpq/michaelpq.github.io/issues/2"
}
|
gharchive/issue
|
Future dates on past presentations
Hi! Finally fixed!
|
2025-04-01T04:34:40.889466
| 2024-10-18T19:27:58
|
2598240726
|
{
"authors": [
"michaelrsweet"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8392",
"repo": "michaelrsweet/pappl",
"url": "https://github.com/michaelrsweet/pappl/issues/370"
}
|
gharchive/issue
|
Bring back CUPS 2.x support - 2.5 or later...
Since CUPS 2.5 has the necessary APIs to support PAPPL 2.0, bring back the CUPS 3 -> 2.5 wrappers.
[master b494f1d] Restore ability to build against CUPS 2.5.
|
2025-04-01T04:34:40.935436
| 2016-12-06T22:30:50
|
193904282
|
{
"authors": [
"lizzzp1",
"orrybaram"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8393",
"repo": "micnews/article-json-to-amp",
"url": "https://github.com/micnews/article-json-to-amp/pull/38"
}
|
gharchive/pull-request
|
Ads
ADS ADS ADS
More ads mean more 💰💰💰 so we're adding ads back into ⚡⚡⚡AMP⚡⚡⚡
WHO's GONNA LOOK?
how about one of the previous contributors?
[ ] @kesla
[ ] @iefserge
[ ] @ellell
or maybe someone else?
[ ] @stankud
[ ] @ryanscottaudio
[ ] @danmakenoise
[x] @lizzzp1
LGTM what does ava do out of curiosity?
Ava is a test runner, similar to tape but it runs the tests asynchronous.
On Tue, Dec 6, 2016, 5:45 PM Liz<EMAIL_ADDRESS>wrote:
LGTM what does ava do out of curiosity?
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/micnews/article-json-to-amp/pull/38#issuecomment-265297473,
or mute the thread
https://github.com/notifications/unsubscribe-auth/AAsKey5i3qeVG4bop7qy5Nbh1PT1PZS5ks5rFeVxgaJpZM4LF90j
.
|
2025-04-01T04:34:40.948385
| 2023-03-03T12:56:26
|
1608584266
|
{
"authors": [
"AjinJayan",
"htex0189"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8394",
"repo": "micro-ROS/micro_ros_mbed",
"url": "https://github.com/micro-ROS/micro_ros_mbed/issues/34"
}
|
gharchive/issue
|
Creating Multiple publisher
Hi,
I run the example with one publisher and it is working. I can read the message with micro_ros agent.
But when adding another publisher, the new publisher is not registered with micro_ros agent. rclc_publisher_init_default returns error at second publisher.
So How to create multiple publisher. And is it necessary to use executors with timer to publish or publish messages directly without using executors?
Thanks
Is the meta config for rmw_microxrcedds for mbed is not yet developed to configure max publisher,subscriber,node for mbed?
I changed the colcon meta config - increased the number of publisher subscribers. I still cannot create two publisher. Please could you tell what should be done after changing colcon meta file
Thanks that was really helpful
|
2025-04-01T04:34:40.967993
| 2023-12-15T18:30:34
|
2044186955
|
{
"authors": [
"mbthornton-lbl",
"turbomam"
],
"license": "CC0-1.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8395",
"repo": "microbiomedata/nmdc-schema",
"url": "https://github.com/microbiomedata/nmdc-schema/issues/1578"
}
|
gharchive/issue
|
Create before/after diagram of Workflow-related schema changes
Add a tab to the existing diagram:
https://app.diagrams.net/#G1Ufsblf98rGzRhJMDt7ZMgosioywnGJIU
Create a before and after high-level schema diagrams and expected workflow-related changes
Could we please use an automated diagram generator for the after state?
|
2025-04-01T04:34:40.984169
| 2023-01-28T00:42:15
|
1560600947
|
{
"authors": [
"HyperTesto",
"gRegorLove"
],
"license": "CC0-1.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8396",
"repo": "microformats/php-mf2",
"url": "https://github.com/microformats/php-mf2/issues/246"
}
|
gharchive/issue
|
Properties not parsed when using Tailwind height class
I managed to get a minimal example that happened to me while working with Tailwind CSS, but i guess it's more generic to any class that starts with h- prefix.
If i use Tailwind CSS h-<number> class on some elements that have other microformat classes like in this minimal example with u-logo and u-nickname:
<article class="h-card">
<h1> TITLE </h1>
<img class="h-36 photo u-logo"
alt="An example alt title"
src="https://example.com/img.png"
/>
<p class="h-21 u-nickname"> John Doe </p>
</article>
The parser skips mf properties and produces :
{
"items": [
{
"type": [
"h-card"
],
"properties": {
"name": [
"TITLE An example alt title\nJohn Doe"
]
}
}
],
"rels": {},
"rel-urls": {},
"debug": {
"package": "https://packagist.org/packages/mf2/mf2",
"source": "https://github.com/indieweb/php-mf2",
"version": "v0.5.0",
"note": [
"This output was generated from the php-mf2 library available at https://github.com/indieweb/php-mf2",
"Please file any issues with the parser at https://github.com/indieweb/php-mf2/issues",
"Using the Masterminds HTML5 parser"
]
}
}
Instead, if I remove h- classes, it's working as expected:
<article class="h-card">
<h1> TITLE </h1>
<img class="photo u-logo"
alt="An example alt title"
src="https://example.com/img.png"
/>
<p class="u-nickname"> John Doe </p>
</article>
with output:
{
"items": [
{
"type": [
"h-card"
],
"properties": {
"logo": [
{
"value": "https://example.com/img.png",
"alt": "An example alt title"
}
],
"nickname": [
"John Doe"
],
"name": [
"TITLE An example alt title\nJohn Doe"
]
}
}
],
"rels": {},
"rel-urls": {},
"debug": {
"package": "https://packagist.org/packages/mf2/mf2",
"source": "https://github.com/indieweb/php-mf2",
"version": "v0.5.0",
"note": [
"This output was generated from the php-mf2 library available at https://github.com/indieweb/php-mf2",
"Please file any issues with the parser at https://github.com/indieweb/php-mf2/issues",
"Using the Masterminds HTML5 parser"
]
}
}
This scenario seems to work using the Go implementation
Good catch! The specification is to skip classes with numbers, but it shouldn't be skipping the entire element. Will look into it.
|
2025-04-01T04:34:41.161848
| 2019-12-21T17:19:17
|
541349774
|
{
"authors": [
"Ba4bes",
"johnpapa"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8397",
"repo": "microsoft/25-days-of-serverless",
"url": "https://github.com/microsoft/25-days-of-serverless/issues/364"
}
|
gharchive/issue
|
[CHALLENGE SUBMISSION] - day 21
Please provide the following information for your submission:
What is your name? (First, Last)
Barbara Forbes
Where is your GitHub Repo ?(With your challenge solution)
https://github.com/Ba4bes/25daysofserverless/tree/master/Daya21GiftRegistries
What Challenge is this for? (Challenge number between 1 and 25)
21
(Optional) Anything else we should know? e.g., language used, location, blog post?
PowerShell Function App and CosmosDB
The unthinkable has happened: after weeks of community members all over the globe chipping in to help solve people's problems with serverless technology, the evil grinch's heart is starting to soften! He's still not quite ready to give up his server-stealing plot, but our whole Azure Cloud Advocacy team is flying up to the grinch's lair to help convince him to change his ways.
As such, the team behind 25 Days of Serverless is OOF (Out Of Facility/Office) between Dec 24, 2019 and January 1, 2020. We'll not review Issues in this period of time, but encourage you to catch up on any challenges you haven't had the time to work on yet. 😄 Our Discord Server is a perfect place to reach out to for community support during this time. After our return, we'll add contributors who submitted a working solution as an Issue to CONTRIBUTORS.md file until January 15, 2020.
|
2025-04-01T04:34:41.167236
| 2024-04-17T08:01:18
|
2247646413
|
{
"authors": [
"AskeHolst",
"akoniecki",
"florian-d"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8398",
"repo": "microsoft/AL-Go",
"url": "https://github.com/microsoft/AL-Go/issues/1041"
}
|
gharchive/issue
|
IncrementVersionNumber - the new version number is less than the old version number
Hi,
I have the following problem with the "IncrementVersionNumber" workflow:
[IncrementVersionNumber] Unexpected error when running action. Error Message: The new version number (1.10) is less than the old version number (1.9). The version number must be incremented., StackTrace: at Set-VersionInSettingsFile, D:\a\_actions\microsoft\AL-Go-Actions\v5.0\IncrementVersionNumber\IncrementVersionNumber.psm1: line 108 <- at <ScriptBlock>, D:\a\_actions\microsoft\AL-Go-Actions\v5.0\IncrementVersionNumber\IncrementVersionNumber.ps1: line 46 <- at <ScriptBlock>, D:\a\_temp\989f38ee-08d1-4139-9f8b-28d0b976b3b7.ps1: line 4 <- at <ScriptBlock>, <No file>: line 1
I couldn't find the exact reason in the code, but it seems the System.Version cast is not applied correctly to variables before ($newValue -lt $oldValue) comparison.
These are the steps to recreate the issue:
Create new repo from AL-Go-PTE template
Wait until Initial Build is done
Run "Create new app" workflow with sample code parameter
Run "Increment Version Number" workflow with Updated Version Number parameter: 1.9
Run "Increment Version Number" workflow with Updated Version Number parameter: +0.1
Here is my public repo with the issue recreated: https://github.com/akoniecki/versionincrement-debugging/actions/runs/8718133126/job/23914709863#step:7:60
We just got the exact same issue.
Error: Unexpected error when running action. Error Message: The new version number (2.10) is less than the old version number (2.9). The version number must be incremented., StackTrace: at Set-VersionInSettingsFile, /home/runner/work/_actions/microsoft/AL-Go/f2f452940bc16f36a4c59f1fd1982c566f16c03c/Actions/IncrementVersionNumber/IncrementVersionNumber.psm1: line 108 <- at <ScriptBlock>, /home/runner/work/_actions/microsoft/AL-Go/f2f452940bc16f36a4c59f1fd1982c566f16c03c/Actions/IncrementVersionNumber/IncrementVersionNumber.ps1: line 46 <- at <ScriptBlock>, /home/runner/work/_temp/a3b6a4b4-9bc3-4c1d-ab35-7258a178aaaf.ps1: line 4 <- at <ScriptBlock>, <No file>: line 1
After updating AL-GO, we encountered the same error during the ‘Create Release’ process. We’re also using AL-Go-PTE.
Unexpected error when running action. Error Message: The new version number (<IP_ADDRESS>) is less than the old version number (<IP_ADDRESS>).
The version number must be incremented., StackTrace: at Set-VersionInSettingsFile,
D:\a\_actions\microsoft\AL-Go-Actions\v5.0\IncrementVersionNumber\IncrementVersionNumber.psm1: line 108 <- at <ScriptBlock>,
D:\a\_actions\microsoft\AL-Go-Actions\v5.0\IncrementVersionNumber\IncrementVersionNumber.psm1: line 154 <- at Set-VersionInAppManifests,
D:\a\_actions\microsoft\AL-Go-Actions\v5.0\IncrementVersionNumber\IncrementVersionNumber.psm1: line 150 <- at <ScriptBlock>,
D:\a\_actions\microsoft\AL-Go-Actions\v5.0\IncrementVersionNumber\IncrementVersionNumber.ps1: line 53 <- at <ScriptBlock>,
D:\a\_temp\d26634c8-a3cc-4341-991e-cf830077b236.ps1: line 4 <- at <ScriptBlock>, <No file>: line 1
|
2025-04-01T04:34:41.175046
| 2024-12-02T08:56:06
|
2711109815
|
{
"authors": [
"dibyanugraha",
"freddydk",
"frottke"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8399",
"repo": "microsoft/AL-Go",
"url": "https://github.com/microsoft/AL-Go/issues/1344"
}
|
gharchive/issue
|
[Bug]: Experimental feature "git submodules" seems to be a breaking change (in some ways)
AL-Go version
v6.1
Describe the issue
We have identified a potential bug in version 6.1: all submodules, regardless of their location within the repository, are being downloaded during the git checkout process.
To provide some context, our repository’s root folder contains a Git submodule for internal scripts, which serves as a convenient way to update them globally. Unfortunately, the workflow attempts to download this submodule, despite not having the necessary permissions to do so.
I truly appreciate all the new features, especially experimental ones. However, would it be possible to "hide" such features behind a feature flag or check?
If needed, I would be happy to grant access to a repository where this issue can be reproduced.
Expected behavior
We would have expected that only app/project paths are checked for Git submodules, rather than the entire repository. Alternatively, it would be helpful if authentication could be configured specifically for submodules.
Steps to reproduce
first variation: create a submodule that is not in an configured app/project path then run CI/CD or PR Build. See that the git submodules is cloned.
second variation: create a submodule where git (cli, github_token) has no permissions to access the repository. It looks like that GHTOKENWORKFLOW is not used.
Additional context (logs, screenshots, etc.)
No response
Got it.
We cannot really use the GhTokenWorkflow for this as we need to clone the repo in order to get to the settings / secrets.
But... - I see that this change will break you when having submodules, which the standard github token cannot access.
We will investigate what we can do about this - in the meantime, you would have to revert back to v6.0 - then we will have a look at what we can do for this.
Thank you Freddy for the, as always, quick response! :)
We’ve already rolled back to v6.0, knowing this would be not a fast fix.
One more remark: Isn't every repository other than the current one out of scope for the default github_token? In other words, how should a submodule repository be configured to make it accessible? Currently, I don't see any other option except granting additional permissions to the checkout action (GitHub token).
How do you structure your repos? If they reside in the same organization, you can use relative path instead of full url path in your submodule.
My apologies I have to refer to outside blog, but this blog gives a good example how to change your submodule into relative path (given with above circumstances): https://www.damirscorner.com/blog/posts/20210423-ChangingUrlsOfGitSubmodules.html
@dibyanugraha - I don't think the github token from one repository would be able to pull the submodule of another repository even with relative urls.
The default github token has very limited permissions I think
Just tested this - for private repositories, we need an auth token on the actions/checkout - the default github.token won't suffice.
Currently, we do not have a token for use at that time, so we will have to refactor the support for submodules if we are to support that.
Yes, we can pass token to the checkout itself, but unfortunately the github_token workflow won't be suffice. It must be a security token variable in GitHub that has 'Read' access to submodule repositories.
I will check if there is an easy way to support submodules with this knowledge - else we will have to pull the experimental support at this time.
|
2025-04-01T04:34:41.177301
| 2021-01-20T08:39:14
|
789763887
|
{
"authors": [
"JesperSchulz",
"VMotin"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8400",
"repo": "microsoft/ALAppExtensions",
"url": "https://github.com/microsoft/ALAppExtensions/issues/10637"
}
|
gharchive/issue
|
[Request-for-external] Codeunit 90 Purch.-Post - CalcInvoiceDiscountPosting
Hi, I was wondering if it would be possible to change the procedure CalcInvoiceDiscountPosting Codeunit 90 from local to external.
My team and I use this procedure a lot in our code, and changing it to external would be a huge help.
Thanks for your work, have a nice day
Thanks for reporting this. We agree, and we’ll publish a fix asap, either in an update for the current version or in the next major release. We will update this issue with information about availability. Please do not reply to this, as we do not monitor closed issues. If you have follow-up questions or requests, please create a new issue where you reference this one.
|
2025-04-01T04:34:41.180160
| 2021-08-19T13:13:16
|
974650639
|
{
"authors": [
"JesperSchulz",
"fridrichovsky"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8401",
"repo": "microsoft/ALAppExtensions",
"url": "https://github.com/microsoft/ALAppExtensions/issues/14049"
}
|
gharchive/issue
|
[Event Request] Codeunit 6620 "Copy Document Mgt".InitSalesLineFields
Please add new event
local procedure InitSalesLineFields(var ToSalesLine: Record "Sales Line")
//----------------------------------------OnInitSalesLineFieldsOnBeforeSetJob:BEGIN
var
IsHandled: boolean;
//----------------------------------------OnInitSalesLineFieldsOnBeforeSetJob:END
begin
OnBeforeInitSalesLineFields(ToSalesLine);
if ToSalesLine."Document Type" <> ToSalesLine."Document Type"::Order then begin
ToSalesLine."Prepayment %" := 0;
ToSalesLine."Prepayment VAT %" := 0;
ToSalesLine."Prepmt. VAT Calc. Type" := "Tax Calculation Type"::"Normal VAT";
ToSalesLine."Prepayment VAT Identifier" := '';
ToSalesLine."Prepayment VAT %" := 0;
ToSalesLine."Prepayment Tax Group Code" := '';
ToSalesLine."Prepmt. Line Amount" := 0;
ToSalesLine."Prepmt. Amt. Incl. VAT" := 0;
end;
ToSalesLine."Prepmt. Amt. Inv." := 0;
ToSalesLine."Prepmt. Amount Inv. (LCY)" := 0;
ToSalesLine."Prepayment Amount" := 0;
ToSalesLine."Prepmt. VAT Base Amt." := 0;
ToSalesLine."Prepmt Amt to Deduct" := 0;
ToSalesLine."Prepmt Amt Deducted" := 0;
ToSalesLine."Prepmt. Amount Inv. Incl. VAT" := 0;
ToSalesLine."Prepayment VAT Difference" := 0;
ToSalesLine."Prepmt VAT Diff. to Deduct" := 0;
ToSalesLine."Prepmt VAT Diff. Deducted" := 0;
ToSalesLine."Prepmt. Amt. Incl. VAT" := 0;
ToSalesLine."Prepmt. VAT Amount Inv. (LCY)" := 0;
ToSalesLine."Quantity Shipped" := 0;
ToSalesLine."Qty. Shipped (Base)" := 0;
ToSalesLine."Return Qty. Received" := 0;
ToSalesLine."Return Qty. Received (Base)" := 0;
ToSalesLine."Quantity Invoiced" := 0;
ToSalesLine."Qty. Invoiced (Base)" := 0;
ToSalesLine."Reserved Quantity" := 0;
ToSalesLine."Reserved Qty. (Base)" := 0;
ToSalesLine."Qty. to Ship" := 0;
ToSalesLine."Qty. to Ship (Base)" := 0;
ToSalesLine."Return Qty. to Receive" := 0;
ToSalesLine."Return Qty. to Receive (Base)" := 0;
ToSalesLine."Qty. to Invoice" := 0;
ToSalesLine."Qty. to Invoice (Base)" := 0;
ToSalesLine."Qty. Shipped Not Invoiced" := 0;
ToSalesLine."Return Qty. Rcd. Not Invd." := 0;
ToSalesLine."Shipped Not Invoiced" := 0;
ToSalesLine."Return Rcd. Not Invd." := 0;
ToSalesLine."Qty. Shipped Not Invd. (Base)" := 0;
ToSalesLine."Ret. Qty. Rcd. Not Invd.(Base)" := 0;
ToSalesLine."Shipped Not Invoiced (LCY)" := 0;
ToSalesLine."Return Rcd. Not Invd. (LCY)" := 0;
//----------------------------------------OnInitSalesLineFieldsOnBeforeSetJob:BEGIN
OnInitSalesLineFieldsOnBeforeSetJob(ToSalesLine,IsHandled)
if not IsHandled then begin
//----------------------------------------OnInitSalesLineFieldsOnBeforeSetJob:END
ToSalesLine."Job No." := '';
ToSalesLine."Job Task No." := '';
ToSalesLine."Job Contract Entry No." := 0;
//----------------------------------------OnInitSalesLineFieldsOnBeforeSetJob:BEGIN
end;
//----------------------------------------OnInitSalesLineFieldsOnBeforeSetJob:END
OnAfterInitSalesLineFields(ToSalesLine);
end;
//----------------------------------------OnInitSalesLineFieldsOnBeforeSetJob:BEGIN
[IntegrationEvent(false, false)]
local procedure OnInitSalesLineFieldsOnBeforeSetJob(var ToSalesLine: Record "Sales Line"; var IsHandled: boolean)
begin
end;
//----------------------------------------OnInitSalesLineFieldsOnBeforeSetJob:END
Thanks for reporting this. We agree, and we’ll publish a fix asap, either in an update for the current version or in the next major release. Please do not reply to this, as we do not monitor closed issues. If you have follow-up questions or requests, please create a new issue where you reference this one.
|
2025-04-01T04:34:41.182146
| 2021-10-13T20:50:06
|
1025683915
|
{
"authors": [
"JesperSchulz",
"ajkauffmann"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8402",
"repo": "microsoft/ALAppExtensions",
"url": "https://github.com/microsoft/ALAppExtensions/issues/14766"
}
|
gharchive/issue
|
[Event parameter request] codeunit 5431 "Calc. Item Plan - Plan Wksh." - OnCodeOnAfterSetReqLineFilters
Please add CurrTemplateName and CurrWorksheetName to event OnCodeOnAfterSetReqLineFilters in codeunit 5431 "Calc. Item Plan - Plan Wksh."
Original code:
OnCodeOnAfterSetReqLineFilters(ReqLineExtern);
[IntegrationEvent(false, false)]
local procedure OnCodeOnAfterSetReqLineFilters(var ReqLineExtern: Record "Requisition Line")
begin
end;
Modified code:
OnCodeOnAfterSetReqLineFilters(ReqLineExtern, CurrTemplateName, CurrWorksheetName);
[IntegrationEvent(false, false)]
local procedure OnCodeOnAfterSetReqLineFilters(var ReqLineExtern: Record "Requisition Line"; CurrTemplateName: Code[10]; CurrWorksheetName: Code[10])
begin
end;
Thanks for reporting this. We agree, and we’ll publish a fix asap, either in an update for the current version or in the next major release. Please do not reply to this, as we do not monitor closed issues. If you have follow-up questions or requests, please create a new issue where you reference this one.
|
2025-04-01T04:34:41.183767
| 2021-11-22T14:34:59
|
1060230973
|
{
"authors": [
"JesperSchulz",
"iltererdogan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8403",
"repo": "microsoft/ALAppExtensions",
"url": "https://github.com/microsoft/ALAppExtensions/issues/15285"
}
|
gharchive/issue
|
[Event Request] table 5902 "Service Line" - OnBeforeCalcInvDiscToInvoice
Hi,
Is it possible to add OnBeforeCalcInvDiscToInvoice publisher to service line?
procedure InitQtyToInvoice()
begin
"Qty. to Invoice" := MaxQtyToInvoice;
"Qty. to Invoice (Base)" := MaxQtyToInvoiceBase;
"VAT Difference" := 0;
// <<New
OnBeforeCalcInvDiscToInvoice(Rec, CurrFieldNo);
// >> New
CalcInvDiscToInvoice;
OnAfterInitQtyToInvoice(Rec, CurrFieldNo);
end;
// <<New
[IntegrationEvent(false, false)]
local procedure OnBeforeCalcInvDiscToInvoice(var ServLine: Record "Service Line"; CallingFieldNo: Integer)
begin
end;
// >> New
Thanks for reporting this. We agree, and we’ll publish a fix asap, either in an update for the current version or in the next major release. Please do not reply to this, as we do not monitor closed issues. If you have follow-up questions or requests, please create a new issue where you reference this one.
|
2025-04-01T04:34:41.186089
| 2021-12-11T11:35:14
|
1077511652
|
{
"authors": [
"JesperSchulz",
"auzhhv"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8404",
"repo": "microsoft/ALAppExtensions",
"url": "https://github.com/microsoft/ALAppExtensions/issues/15510"
}
|
gharchive/issue
|
[Event request] table 38 "Purchase Header" - OnAfterTestNoSeries
Can you add PurchSetup to event OnAfterTestNoSeries in table 38 "Purchase Header"?
procedure TestNoSeries()
var
IsHandled: Boolean;
begin
GetPurchSetup();
IsHandled := false;
OnBeforeTestNoSeries(Rec, IsHandled);
if not IsHandled then
case "Document Type" of
"Document Type"::Quote:
PurchSetup.TestField("Quote Nos.");
"Document Type"::Order:
PurchSetup.TestField("Order Nos.");
"Document Type"::Invoice:
begin
PurchSetup.TestField("Invoice Nos.");
PurchSetup.TestField("Posted Invoice Nos.");
end;
"Document Type"::"Return Order":
PurchSetup.TestField("Return Order Nos.");
"Document Type"::"Credit Memo":
begin
PurchSetup.TestField("Credit Memo Nos.");
PurchSetup.TestField("Posted Credit Memo Nos.");
end;
"Document Type"::"Blanket Order":
PurchSetup.TestField("Blanket Order Nos.");
end;
// >>>>>>>>>>>>>>>
//OnAfterTestNoSeries(Rec);
OnAfterTestNoSeries(Rec, PurchSetup);
// <<<<<<<<<<<<<<<
end;
Thanks for reporting this. We agree, and we’ll publish a fix asap, either in an update for the current version or in the next major release. Please do not reply to this, as we do not monitor closed issues. If you have follow-up questions or requests, please create a new issue where you reference this one.
|
2025-04-01T04:34:41.196360
| 2024-04-09T11:22:15
|
2233216701
|
{
"authors": [
"JesperSchulz",
"pri-kise"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8405",
"repo": "microsoft/ALAppExtensions",
"url": "https://github.com/microsoft/ALAppExtensions/issues/26251"
}
|
gharchive/issue
|
[Event Change Request] codeunit 73 "Purch.-Explode BOM" - OnBeforeInsertExplodedPurchLine
Describe the request
local procedure ExplodeBOMCompLines(PurchLine: Record "Purchase Line")
var
PreviousPurchLine: Record "Purchase Line";
Resource: Record Resource;
InsertLinesBetween: Boolean;
SkipComponent: Boolean;
begin
...
if PurchHeader."Language Code" = '' then
ToPurchLine.Description := FromBOMComp.Description
else
if not ItemTranslation.Get(FromBOMComp."No.", FromBOMComp."Variant Code", PurchHeader."Language Code") then
ToPurchLine.Description := FromBOMComp.Description;
//+Change
//-OnBeforeInsertExplodedPurchLine(ToPurchLine, PurchLine, FromBOMComp);
OnBeforeInsertExplodedPurchLine(ToPurchLine, PurchLine, FromBOMComp, PurchHeader);
//-Change
ToPurchLine.Insert();
OnAfterInsertExplodedPurchLine(ToPurchLine, PurchLine, FromBOMComp);
...
end;
[IntegrationEvent(false, false)]
//+Change
//-local procedure OnBeforeInsertExplodedPurchLine(var ToPurchaseLine: Record "Purchase Line"; PurchaseLine: Record "Purchase Line"; FromBOMComp: Record "BOM Component")
local procedure OnBeforeInsertExplodedPurchLine(var ToPurchaseLine: Record "Purchase Line"; PurchaseLine: Record "Purchase Line"; FromBOMComp: Record "BOM Component"; PurchaseHeader: Record "Purchase Header";)
//-Change
begin
end;
Additional context
We need Access to a field of the header record in this event.
Internal work item: AB#524721
Thanks for reporting this. We agree, and we’ll publish a fix asap, either in an update for the current version or in the next major release. Please do not reply to this, as we do not monitor closed issues. If you have follow-up questions or requests, please create a new issue where you reference this one.
Build ID: 18489.
|
2025-04-01T04:34:41.206416
| 2018-11-13T14:20:04
|
380248292
|
{
"authors": [
"BertDeTemmerman",
"wbor-nvt"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8406",
"repo": "microsoft/ALAppExtensions",
"url": "https://github.com/microsoft/ALAppExtensions/issues/342"
}
|
gharchive/issue
|
[Event-Request] REP2000005 File SEPA Payments
We want to create the (NON-)SEPA payments file automatically.
Therefore we would like an event/function that allows us to set the parameters of the requestpage on the file payments reports and also another to hide that requestpage.
When clicking the ExportPaymentLines action on the PAG2000001 EB Payment Journal, the following code gets executed:
In TAB2000005 Export Protocol
ExportPaymentLines(VAR PmtJnlLine : Record "Payment Journal Line")
IF CheckPaymentLines(PmtJnlLine) THEN BEGIN
TESTFIELD("Export Object ID");
PmtJnlLineToExport.COPY(PmtJnlLine);
PmtJnlLineToExport.SETRANGE(Status,PmtJnlLineToExport.Status::Created);
PmtJnlLineToExport.SETRANGE("Export Protocol Code",Code);
PmtJnlLineToExport.SETRANGE("Journal Batch Name",PmtJnlLine."Journal Batch Name");
PmtJnlLineToExport.SETRANGE("Journal Template Name",PmtJnlLine."Journal Template Name");
IF "Export Object Type"="Export Object Type"::Report THEN
////Here we would need something to hide the requestpage
////Such as: OnBeforeRunReportExportPaymentLines(VAR Rec, VAR ShowRequestpage, VAR PmtJnlLineToExport)
REPORT.RUNMODAL("Export Object ID",TRUE,FALSE,PmtJnlLineToExport)
ELSE BEGIN
IF PmtJnlLine."Exported To File" THEN
IF NOT CONFIRM(ExportAgainQst) THEN
EXIT;
In both File payment reports (REP2000005 File SEPA Payments and REP2000006 File Non Euro SEPA Payments), we would need an event that's let us fill in the requestpage.
Something like
OnAfterInit(VAR GenJnlLine, VAR AutomaticPosting, VAR IncludeDimText, VAR ExecutionDate, VAR FileName)
Could these be provided please? :)
Version BE Dynamics NAV 13.0 (25789)
We are also working on a solution that requires running REP2000005 & REP2000006 with no Request Page.
TAB2000005 has already received the necessary OnBeforeExportPaymentLinesOnRunReport event.
There is still no way to alter the Global variables set in the Request page however.
Is this still planned for future update?
Ideal would be an OnBeforePreReport trigger
OnBeforePreReport(VAR GenJnlLine, VAR AutomaticPosting, VAR IncludeDimText, VAR ExecutionDate, VAR FileName)
@BertDeTemmerman Did you get your requested event in the meanwhile, to set the request page parameters, such as ExecutionDate?
Something like
OnAfterInit(VAR GenJnlLine, VAR AutomaticPosting, VAR IncludeDimText, VAR ExecutionDate, VAR FileName)
Short: No.
Long: No.
But we can skip the 2nd report now ...
I think we just copied those reports or didn't implement this.
local procedure PostPaymentJournal(var GenJnlLine: Record "Gen. Journal Line"; var PaymentJournalLine: Record "Payment Journal Line"; BalancingPostingDate: Date)
var
PaymentJournalPost: Report "Payment Journal Post";
IsHandled: Boolean;
begin
IsHandled := false;
OnBeforePostPaymentJournal(GenJnlLine, PaymentJournalLine, AutomaticPosting, BalancingPostingDate, IsHandled);
if IsHandled then
exit;
PaymentJournalPost.SetParameters(GenJnlLine, AutomaticPosting, REPORT::"File SEPA Payments", BalancingPostingDate);
PaymentJournalPost.SetTableView(PaymentJournalLine);
PaymentJournalPost.RunModal;
end;
|
2025-04-01T04:34:41.208936
| 2020-07-23T11:09:58
|
664388705
|
{
"authors": [
"bc-ghost",
"osteeger"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8407",
"repo": "microsoft/ALAppExtensions",
"url": "https://github.com/microsoft/ALAppExtensions/issues/7963"
}
|
gharchive/issue
|
[Event Request] codeunit 481 "Check Dimensions" - Event for procedure CheckPurchDimLines
Can we get the following event for the "Check Dimensions" Codeunit: OnTempPurchLineRepeat
local procedure CheckPurchDimLines(PurchHeader: Record "Purchase Header"; var TempPurchLine: Record "Purchase Line" temporary)
begin
with TempPurchLine do begin
Reset;
SetFilter(Type, '<>%1', Type::" ");
if FindSet then
repeat
if (PurchHeader.Receive and ("Qty. to Receive" <> 0)) or
(PurchHeader.Invoice and ("Qty. to Invoice" <> 0)) or
(PurchHeader.Ship and ("Return Qty. to Ship" <> 0))
then begin
CheckPurchDimCombLine(TempPurchLine);
CheckPurchDimValuePostingLine(TempPurchLine);
OnTempPurchLineRepeat(TempPurchLine); //NEW
end
until Next = 0;
end;
end;
[IntegrationEvent(false, false)]
local procedure OnTempPurchLineRepeat(var TempPurchLine: Record "Purchase Line")
begin
end;
Thanks for reporting this. We agree, and we’ll publish a fix asap, either in an update for the current version or in the next major release. We will update this issue with information about availability.
|
2025-04-01T04:34:41.212484
| 2020-08-19T07:14:27
|
681606280
|
{
"authors": [
"RebeccaAnne",
"animia"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8408",
"repo": "microsoft/AdaptiveCards",
"url": "https://github.com/microsoft/AdaptiveCards/issues/4614"
}
|
gharchive/issue
|
Disabled button is not supported.
What platform is your feature request related to? (Delete other platforms).
.NET HTML
Android
iOS (Swift/Objective C)
JavaScript
Is your feature request related to a problem? Please describe.
When user click any button, we want to change button into disabled status.
Describe the solution you'd like
When user click any button, we want to change button into disabled button in local.
It means button is changed into dimmed status without server's interaction.
Describe alternatives you've considered
I think the alternative is dimmed button image which cannot be clicked.
Additional context
This will be introduced with our 1.5 release. See #4670
|
2025-04-01T04:34:41.230653
| 2019-12-25T22:26:58
|
542390571
|
{
"authors": [
"MSNev",
"Malvoz"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8409",
"repo": "microsoft/ApplicationInsights-JS",
"url": "https://github.com/microsoft/ApplicationInsights-JS/issues/1148"
}
|
gharchive/issue
|
Set a long max-age for immutable resources
(I'm personally not a user of the Application Insights SDK, but as I stumbled on this I thought I'd file an issue, I'm hopefully in the right place.)
Considering not all browsers support the immutable directive, these resources should typically include a long max-age, conventionally: max-age=31536000.
https://az416426.vo.msecnd.net/scripts/a/ai.0.js returns:
Cache-Control: public, max-age=86400, immutable
and:
https://github.com/microsoft/ApplicationInsights-JS/blob/d0184962d439fcd448f18cfa8400e70a4cf57f5e/AISKU/scripts/uploadFilesToAzureStorage.ps1#L2
Hi Malvoz,
These short times are actually required as ai.0.js will return the current version of the script. With these set to a day (86400), this means that if we happen to publish a breaking change then browsers will only keep this cached for a max of 24 hours. If we set this to 1 year (365 days) as suggested that we would break ever user for an entire year AND we could also never publish any update within the same sliding window (as it's relative to when it's uploaded and/or (depending on browsers / proxies that are caching it) first fetched by an end user.
While i agree that 1 day is not ideal (as we don't publish that often) it is a trade off and to correctly handle this we would need another trade off, by including an additional fetch.
Something like
-> Request to ai.0.js => causes a 302 or client side redirect script to fetch the true versioned (with a hash)
-> This hashed file can then have the 1 year (or more policy) as it should never change (based on the hashed content).
The real downside of this is for first time users / uncached requests, there would now be 2 requests.
This requires some infrastructure changes which will need to be addressed as part of a future work item and some analysis on the impacts.
For now I'm going to close this issue as I don't believe that it will be addressed in the near term -- but I will be working to have this addressed in the longer term -- just not for the legacy version (ai.0.js).
|
2025-04-01T04:34:41.232449
| 2020-05-26T18:09:55
|
625083432
|
{
"authors": [
"anthturner"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8410",
"repo": "microsoft/AuthJanitor",
"url": "https://github.com/microsoft/AuthJanitor/pull/63"
}
|
gharchive/pull-request
|
Switch to SecureString for RegeneratedSecret
@ericmaino suggested using SecureString for the RegeneratedSecret so that we don't accidentally serialize sensitive information. I totally agree, and since System.Security.SecureString is now available in netstandard2.1 it's usable.
Unfortunately the support for it in other libraries leaves a lot to be desired, so this PR also includes extension methods to transform back and forth between a SecureString and a string, understanding that doing so compromises in-memory confidentiality of the string.
This is old enough that it was easier to just recreate the commit. Closing.
|
2025-04-01T04:34:41.241318
| 2023-12-20T07:56:36
|
2050038535
|
{
"authors": [
"keerthiduvvuri",
"kupatkar99"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8411",
"repo": "microsoft/AzureStorageExplorer",
"url": "https://github.com/microsoft/AzureStorageExplorer/issues/7615"
}
|
gharchive/issue
|
A11y_Azure Tools Storage Explorer_Disk_DiskToolbar_Usable: No Status message is Conveyed to User that refresh completed.
Preflight Checklist
[X] I have installed the latest version of Storage Explorer.
[X] I have checked existing resources, including the troubleshooting guide and the release notes.
[X] I have searched for similar issues.
Storage Explorer Version
1.32.1(96)
Regression From
NO
Architecture
x64
Storage Explorer Build Number
20231114.10
Platform
macOS
OS Version
MAC Version 13.6.1
Bug Description
GitHub Tags:
#A11yUsable;#A11yTCS;#MAC;#DesktopApp;#A11ySev4;#BM_AzureToolsStorageExplorer_MAC_Dec2023;#Azure Tools Storage Explorer;#FTP;#Voiceover;
Environment Details:
Application: Microsoft Azure Storage Explore
version: 1.32.1 (96)
Screen Reader: Voice Over
Steps to Reproduce
Launch Storage Explorer
Tab Navigate to "Open Connect dialog" pane and Sign into Azure.
Tab Navigate to "Manage Accounts" pane and Select the subscriptions you will use. Once selected, click on "Apply" button.
Tab Navigate to "Disks" of the selected subscription.
Tab Navigate to "A11ytest" and click on it.
Tab Navigate to Toolbar.
Turn on Voice Over and Observe that No Status message is Conveyed to User that refresh completed.
Actual Experience
No Status message is Conveyed to User that refresh completed.
Expected Experience
Status message should be Conveyed to User that refresh completed.
Additional Context
User Impact:
Screen reader users will face difficulty if No Status message is Conveyed to User that refresh completed.
Attachment:
https://github.com/microsoft/AzureStorageExplorer/assets/98728920/56106b73-4d7b-4b8d-aadb-ba5865f399c9
Verified the issue in below environment:
Azure Tools Storage Explorer: Version: 1.33.0-dev (96)
AzCopy Version: 10.22.2
Platform: macOS
Build Number: 20240111.7
Issue is fixed. Hence Closing the bug Please find added attachment for reference.
#Closed;
|
2025-04-01T04:34:41.243483
| 2023-01-12T12:34:38
|
1530638267
|
{
"authors": [
"tamirkamara"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8412",
"repo": "microsoft/AzureTRE",
"url": "https://github.com/microsoft/AzureTRE/pull/3075"
}
|
gharchive/pull-request
|
Add Pylance extension and fix [str | None] issue
What is being addressed
Mismatch types in Python can cause runtime issues. We currently don't use any tool to alert us from those kind of problems.
How is this addressed
Add Pylance VSCode extension
Fix the most frequent issue of assigning default None to optional str params
/test-extended
/test-extended
/test-extended
|
2025-04-01T04:34:41.244873
| 2020-06-02T05:53:33
|
628921696
|
{
"authors": [
"yilongli"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8413",
"repo": "microsoft/AzureTipsAndTricks",
"url": "https://github.com/microsoft/AzureTipsAndTricks/issues/94"
}
|
gharchive/issue
|
Tip 226: is Azure Accelerated Networking based on FPGA-based SmartNIC or Mellanox NIC
I am a bit confused because both FPGA-based SmartNIC and Mellanox CX3 are mentioned in the article: https://microsoft.github.io/AzureTipsAndTricks/blog/tip226.html. Thank you!
Never mind. I found the answer: the FPGA is integrated with a Mellanox NIC in a bump-in-the-wire fashion.
|
2025-04-01T04:34:41.246484
| 2020-02-17T18:42:45
|
566465198
|
{
"authors": [
"a-patel",
"mbcrump",
"vaheminasyan2"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8414",
"repo": "microsoft/AzureTipsAndTricks",
"url": "https://github.com/microsoft/AzureTipsAndTricks/pull/69"
}
|
gharchive/pull-request
|
Update tip158
Create Thumbnail Images with Azure Functions and Azure Storage
fixed broken link
Thanks @a-patel ! Hey @vaheminasyan2 can you add him to our list of contributors.
@mbcrump, added. Thanks
|
2025-04-01T04:34:41.250460
| 2019-06-25T10:09:06
|
460332646
|
{
"authors": [
"CoHealer",
"jwiley84",
"prakash9707"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8415",
"repo": "microsoft/BotBuilder-Samples",
"url": "https://github.com/microsoft/BotBuilder-Samples/issues/1595"
}
|
gharchive/issue
|
18.bot authentication, Does not allows to enter the credentials in oauth prompt
I used azure active directory v2 connection and I created a azure application with allows users to login with multiple organization. After this when I runs this code it displays a empty window and takes the account which is previously logged in.
For the first time it allows me to enter the microsoft email and password. After when I type "logout" then I does not allow me to enter the username and password.
I think it does not logout the user properly. My azure application redirect url or reply url is "https://token.botframework.com/.auth/web/redirect" and logout url is empty.
In the azure active directory v2, I had given my tenant id as "common". I think the problem is with the azure connections.
when I press login in oauth card, I got the below response and after a couple of seconds it will disappear automatically. What may be the issue?
[bug]
@jwiley84, please investigate
This seems to be a behaviour limited to the emulator, as that is the only channel I can reproduce it in. Let me reach out to a member of their team. The standard webchat (not the Azure's 'Test in WebChat') allows for 're-logging', if you will.
Hi @prakash9707! I got a response from the emulator team. They've recently implemented some changes in how ngrok is being used with OAuth prompts. The details are here. Try to update your emulator and give it another shot.
@prakash9707 , did this resolve your issue?
Closing issue.
|
2025-04-01T04:34:41.254230
| 2020-03-23T07:10:58
|
585982385
|
{
"authors": [
"VanyLaw",
"zhixzhan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8416",
"repo": "microsoft/BotFramework-Composer",
"url": "https://github.com/microsoft/BotFramework-Composer/issues/2336"
}
|
gharchive/issue
|
lu file can't save when changing in dialog page.
Describe the bug
After create a interruptionSample, changed the getProfile dialog into luis recognizer, show error.
I change the lu as below, but it can't save in that panel. it can only save in lu page.
- my name is {personName:userName}
- {personName:userName}
- {age:userAge}
- I'm {age:userAge} years old
@ prebuilt personName userName
@ prebuilt age userAge
Version
master branch
Browser
[x] Chrome
[ ] Safari
[ ] Firefox
[ ] Edge
OS
[ ] macOS
[x] Windows
[ ] Ubuntu
it's kind of known issue inline lu editor can't save with error. talked on #2141 #2267, and will be fixed in #2272
|
2025-04-01T04:34:41.258425
| 2021-06-01T07:55:09
|
908013658
|
{
"authors": [
"johnataylor",
"riftek",
"tonyanziano"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8417",
"repo": "microsoft/BotFramework-Composer",
"url": "https://github.com/microsoft/BotFramework-Composer/issues/7967"
}
|
gharchive/issue
|
Azure bot framework composer new deployment authentication
when adding a new deployment. on authentication to deploy, composer cant find my email unless I put a false name and enter another email. of which authenticates but throws an error of the previous nonexistent name not being found.
@tonyanziano can you comment
@riftek can you please include the following in your issue:
Screenshots of what you are seeing
Logs / Error messages
Reproduction steps (if possible)
@riftek can you please follow these steps to enable debugging output in Composer, and then try to login again and reproduce this issue?
Open a shell / cmd terminal
Navigate to Composer's installation directory
run the following command set DEBUG=composer*
Start composer from the terminal "Bot Framework Composer.exe"
You should now see a lot of output in the terminal window while Composer starts up.
Please try to login and then once it fails, copy all of that output and paste it here.
Thank you.
we eventually sorted it out by creating a new bot and resources..
and we have to authenticate and deploy with a private microsoft account that needs to be linked with our work active directory.
mainly we could not get old azure bot to work with new azure.
Great, glad to hear you got it resolved.
|
2025-04-01T04:34:41.265058
| 2023-04-05T04:17:03
|
1654907517
|
{
"authors": [
"dkulkarni-ms",
"mrk1989",
"stevkan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8418",
"repo": "microsoft/BotFramework-WebChat",
"url": "https://github.com/microsoft/BotFramework-WebChat/issues/4691"
}
|
gharchive/issue
|
Getting console error in using WebChat in portal
Is it an issue related to Adaptive Cards?
No
Is this an accessibility issue?
No
What version of Web Chat are you using?
Latest production
Which distribution are you using Web Chat from?
Bundle (webchat.js)
Which hosting environment does this issue primarily affect?
Web apps
Which browsers and platforms do the issue happened?
Browser: Edge (latest)
Which area does this issue affect?
Telemetry
What is the public URL for the website?
No response
Please describe the bug
We are trying to use Azure Bot Service to create a chatbot and using Webchat to integrate in Portal. However when we integrate in Azure portal, we see console errors related to Cors policy. We have enabled the origin in WebApp Service still getting this Cors policy issue. How can this be fixed? Bot loads fine and works fine. Its only that error messages in console needs to be addressed. Also we are using Iframe to show the out of box webchat.
Do you see any errors in console log?
Yes uploaded the screenshot above.
How to reproduce the issue?
Navigate to ...
Click on ...
Type "..." in the send box
What do you expect?
Either fix or way to work around this issue and not see the errors in console logs
What actually happened?
As explained above, errors in console logs
Do you have any screenshots or recordings to repro the issue?
No response
Adaptive Card JSON
No response
Additional context
No response
using Webchat to integrate in Portal
Can you explain what you mean by this statement? Do you mean you have simply created an Azure Web App to host the iframe page? Or are you trying to do something else?
And, for the above, can you explain how you are trying to implement or code this?
The error exaplain too.. you have to add your domain in section CORS.
|
2025-04-01T04:34:41.269828
| 2019-12-12T09:14:00
|
536848187
|
{
"authors": [
"compulim",
"coveralls"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8419",
"repo": "microsoft/BotFramework-WebChat",
"url": "https://github.com/microsoft/BotFramework-WebChat/pull/2719"
}
|
gharchive/pull-request
|
Add Object.is polyfill
Fixes #2718.
Changelog Entry
Fixed
Fixes #2718. Add Object.is polyfill for IE11, by @compulim in PR #2719
Description
Usage of Object.is was added in #2710 but the polyfill for ES5 browsers was not added.
Specific Changes
Add corejs/features/object/is to ES5 bundle
[x] Testing Added
Manual test steps
Load ES5 bundle in IE11, the UI should show up
Coverage remained the same at 61.108% when pulling 89f814e80b8cba7d838e1f7aaafa65ace994a451 on compulim:fix-2718 into 9e43da0b413efc369f823e4278e2444e3fb0b5b3 on microsoft:master.
|
2025-04-01T04:34:41.271648
| 2021-08-21T08:09:42
|
976078169
|
{
"authors": [
"98973991",
"CBL-Mariner-Bot",
"anphel31",
"thcrain-msft"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8420",
"repo": "microsoft/CBL-Mariner",
"url": "https://github.com/microsoft/CBL-Mariner/pull/1316"
}
|
gharchive/pull-request
|
[AUTOPATCHER] Patches CVE-2021-38593 for qt5-qtbase
Patches CVE-2021-38593 for qt5-qtbase.
This patch and PR was automatically generated and needs human verification before merging.
This has picked up 3 identical patch files from different upstream branches. We only need 1.
None of the patches apply to our version of the code
Ok
Closing- does not apply for our configuration of qt5-qtbase.
|
2025-04-01T04:34:41.278639
| 2021-05-24T15:55:49
|
899792590
|
{
"authors": [
"christopherco"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8421",
"repo": "microsoft/CBL-Mariner",
"url": "https://github.com/microsoft/CBL-Mariner/pull/974"
}
|
gharchive/pull-request
|
kernel: Add key into trusted keyring
Merge Checklist
All boxes should be checked before merging the PR (just tick any boxes which don't apply to this PR)
[ ] The toolchain has been rebuilt successfully (or no changes were made to it)
[ ] The toolchain/worker package manifests are up-to-date
[ ] Any updated packages successfully build (or no packages were changed)
[ ] All package sources are available
[ ] cgmanifest files are up-to-date and sorted (./cgmanifest.json, ./toolkit/tools/cgmanifest.json, ./toolkit/scripts/toolchain/cgmanifest.json, .github/workflows/cgmanifest.json)
[ ] LICENSE-MAP files are up-to-date (./SPECS/LICENSES-AND-NOTICES/data/licenses.json, ./SPECS/LICENSES-AND-NOTICES/LICENSES-MAP.md, ./SPECS/LICENSES-AND-NOTICES/LICENSE-EXCEPTIONS.PHOTON)
[ ] All source files have up-to-date hashes in the *.signatures.json files
[ ] sudo make go-tidy-all and sudo make go-test-coverage pass
[ ] Documentation has been updated to match any changes to the build system
[ ] Ready to merge
Summary
Add key to default system keyring to support signed loadable kernel modules.
Signed-off-by: Chris Co<EMAIL_ADDRESS>Change Log
Add mariner key to kernel keyring
Rev kernel related specs
Does this affect the toolchain?
YES
Associated issues
Links to CVEs
Test Methodology
Local build, pipeline build
https://github.com/microsoft/CBL-Mariner/pull/1000
|
2025-04-01T04:34:41.297437
| 2019-05-20T17:59:49
|
446240953
|
{
"authors": [
"ctstone"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8422",
"repo": "microsoft/ComputerVision",
"url": "https://github.com/microsoft/ComputerVision/issues/172"
}
|
gharchive/issue
|
Is a DSVM required?
Consider documenting the steps to run the notebooks on a vanilla Ubuntu VM
@jiata should we close this issue? I see a PR was linked and merged...
|
2025-04-01T04:34:41.298558
| 2020-03-16T20:54:13
|
582593189
|
{
"authors": [
"kipergil",
"lokijota"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8423",
"repo": "microsoft/CosmicClone",
"url": "https://github.com/microsoft/CosmicClone/issues/15"
}
|
gharchive/issue
|
Add support to copy multiple containers
Or, at the very least, when you get to the final screen after a copy, enable Previous button so that you can go back to the previous screens and just edit the names of the containers.
Apart from that, excellent tool. This could/should be added to ADF, pre-creating containers when copying CosmosDB to CosmosDb is a waste of time.
This feature should be implemented. It should be must have feature.
|
2025-04-01T04:34:41.305953
| 2024-06-24T11:25:19
|
2369988580
|
{
"authors": [
"AArnott",
"Nuklon",
"Zoraver"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8424",
"repo": "microsoft/CsWin32",
"url": "https://github.com/microsoft/CsWin32/issues/1218"
}
|
gharchive/issue
|
"Pointers cannot reference marshaled structures. Use ByRef instead" caused by VariableLengthInlineArray
Actual behavior
Pointers cannot reference marshaled structures. Use ByRef instead.
Only happens on .NET Framework 472.
If you remove the public global::Windows.Win32.VariableLengthInlineArray<char> HardwareID; field from SP_DRVINFO_DETAIL_DATA_W it works correctly. It worked correctly in CsWin32 0.3.49.
Expected behavior
Work correctly.
Repro steps
NativeMethods.txt content:
SetupDiGetDriverInfoDetail
NativeMethods.json content (if present):
Any of your own code that should be shared?
Context
CsWin32 version: 0.3.106
Win32Metadata version (if explicitly set by project): 61.0.15-preview
Target Framework: net472
If you include System.Memory package, the error instead becomes:
Error (active) CS0117 'Unsafe' does not contain a definition for 'SkipInit'
I am running into a similar issue when attempting to call CM_Register_Notification on .NET Framework 4.8 using CsWin32 0.3.106. This does not occur when targeting .NET 8.
I believe that this issue is effectively a duplicate of #1184.
I can't get this to repro using the instructions given in this issue. But I agree with @zoraver that this is likely a duplicate of one that I can repro. So I'm closing this.
Sorry, I forgot to add instructions on how to reproduce. I retried it and it's still present in latest daily.
You need to call SetupDiGetDriverInfoDetail with the details from the device:
uint driverInfoDetailDataSize = (uint)sizeof(SP_DRVINFO_DETAIL_DATA_W) + (PInvoke.MAX_DEVICE_ID_LEN * 2);
IntPtr driverInfoDetailDataHandle = Marshal.AllocHGlobal((IntPtr)driverInfoDetailDataSize);
try
{
var driverInfoDetailData = (SP_DRVINFO_DETAIL_DATA_W*)driverInfoDetailDataHandle;
driverInfoDetailData->cbSize = (uint)sizeof(SP_DRVINFO_DETAIL_DATA_W);
uint requiredSize = 0;
// This call throws the above exception.
if (PInvoke.SetupDiGetDriverInfoDetail(deviceInfoList, &deviceInfoData, &driverInfoData, driverInfoDetailData, driverInfoDetailDataSize, &requiredSize))
{
}
Thanks for that, @nuklon. My fix for the believed duplicate bug is in version 0.3.149, for which no daily build exists yet due to a temporary service outage during our last daily build. I've re-kicked it off. If you could please test your scenario with that and report back, I'd appreciate it.
Thanks, I can confirm that it's working fine now with 0.3.149. Another workaround down 😁
|
2025-04-01T04:34:41.316479
| 2024-09-09T10:13:20
|
2513531003
|
{
"authors": [
"Marv51",
"Sergio0694",
"manodasanW"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:8425",
"repo": "microsoft/CsWinRT",
"url": "https://github.com/microsoft/CsWinRT/issues/1739"
}
|
gharchive/issue
|
InvalidCastException for ICommand from dotnet class library
Describe the bug
My app roughly uses an MVVM architecture. My ViewModel classes are all in an external class library.
I am getting an InvalidCastException for ICommand-properties of a ViewModel class, if I bind to it from a WinUI3 button.
These exceptions seem to be not fatal but they mean that the ICommand never gets called with AOT breaking functionality badly.
Note: I am using System.Windows.Input.ICommand and relying the mapping of that to the equivalent WinRT interface.
The exact output is:
Exception thrown: 'System.InvalidCastException' in System.Private.CoreLib.dll
Exception thrown at 0x00007FF909F16D9A (KernelBase.dll) in App7.exe: WinRT originate error - 0x80004002 : 'Specified cast is not valid.'.
To Reproduce
My class library project:
public class SimpleCommand(Action action) : System.Windows.Input.ICommand
{
public event EventHandler? CanExecuteChanged;
public bool CanExecute(object? parameter) => true;
public void Execute(object? parameter) => action();
}
public class ViewModel : INotifyPropertyChanged
{
private readonly SimpleCommand _nextQuote;
public event PropertyChangedEventHandler? PropertyChanged;
public ICommand NextQuote => _nextQuote;
public string Quote { get; set; } = "Knock knock.";
public ViewModel()
{
_nextQuote = new SimpleCommand(() =>
{
Quote = "Who's there?";
PropertyChanged?.Invoke(this, new PropertyChangedEventArgs(nameof(Quote)));
});
}
}
I use it like this:
public sealed partial class MainWindow : Window
{
public ViewModel ViewModel { get; set; } = new();
public MainWindow()
{
this.InitializeComponent();
}
}
<?xml version="1.0" encoding="utf-8"?>
<Window
x:Class="App7.MainWindow"
xmlns="http://schemas.microsoft.com/winfx/2006/xaml/presentation"
xmlns:x="http://schemas.microsoft.com/winfx/2006/xaml"
xmlns:local="using:App7"
xmlns:d="http://schemas.microsoft.com/expression/blend/2008"
xmlns:mc="http://schemas.openxmlformats.org/markup-compatibility/2006"
mc:Ignorable="d">
<StackPanel Orientation="Horizontal" HorizontalAlignment="Center" VerticalAlignment="Center">
<TextBlock Text="{x:Bind ViewModel.Quote, Mode=OneWay}"></TextBlock>
<Button Command="{x:Bind ViewModel.NextQuote, Mode=OneWay}">Refresh</Button>
</StackPanel>
</Window>
Expected behavior
It works if PublishAot is false, so expect it to either still work with AOT or it should show a warning message.
Version Info
<WindowsSdkPackageVersion>10.0.22621.41</WindowsSdkPackageVersion>
<PublishAot>true</PublishAot>
<!-- ..... -->
<PackageReference Include="Microsoft.Windows.CsWinRT" Version="2.1.1" />
<PackageReference Include="Microsoft.Windows.SDK.BuildTools" Version="10.0.26100.1" />
<PackageReference Include="Microsoft.WindowsAppSDK" Version="1.6.240829007" />
Additional context
When I change the ViewModel property to a concrete class, instead of the ICommand interface, that resolves the issue.
This works:
public SimpleCommand NextQuote => _nextQuote;
The better solution I found for this issue is to manually add the ModulInitializer below to the WinUI project.
It took me quite a while to understand how and why this is necessary.
internal static class CswinrtInitalizer
{
[System.Runtime.CompilerServices.ModuleInitializer]
internal static void InitializeVtableLookup()
{
ComWrappersSupport.RegisterTypeComInterfaceEntriesLookup(new Func<Type, ComWrappers.ComInterfaceEntry[]>(LookupVtableEntries));
ComWrappersSupport.RegisterTypeRuntimeClassNameLookup(new Func<Type, string>(LookupRuntimeClassName));
}
private static ComWrappers.ComInterfaceEntry[] LookupVtableEntries(Type type)
{
return type.ToString() switch
{
"ClassLibrary1.SimpleCommand" => [
new()
{
IID = ABI.System.Windows.Input.ICommandMethods.IID,
Vtable = ABI.System.Windows.Input.ICommandMethods.AbiToProjectionVftablePtr
},
],
_ => default
};
}
private static string LookupRuntimeClassName(Type type)
{
string typeName = type.ToString();
return default;
}
}
I think the generator should be able to automatically generate this code for my classes.
An alternative workaround is to add at least one property anywhere that uses the class directly instead of the interface. That triggers the generator to create the necessary code automatically.
If you add a second button in my example above and this one calls:
public SimpleCommand AlternativeNextQuote => _nextQuote;
That is already enough to also fix the ICommand-button.
"I think the generator should be able to automatically generate this code for my classes."
It does, but you need to have the generator enabled and mark the type as partial. Add an explicit package reference for CsWinRT 2.1.1 in your project, that's needed until the next .NET SDK update that will enable the generator by default.
@Marv51 to confirm, your library is targeting the .NET Windows OS version TFM like net8.0-windows10.0.22621.0 right?
@manodasanW My library is targeting plain .net 8 like this:
<Project Sdk="Microsoft.NET.Sdk">
<PropertyGroup>
<TargetFramework>net8.0</TargetFramework>
<ImplicitUsings>enable</ImplicitUsings>
<Nullable>enable</Nullable>
</PropertyGroup>
</Project>
Sounds like that is not supported?
@Sergio0694, you are right, when I add CsWinRT that does seem to fix my issue and generate the needed code automatically.
<Project Sdk="Microsoft.NET.Sdk">
<PropertyGroup>
<TargetFramework>net8.0</TargetFramework>
<ImplicitUsings>enable</ImplicitUsings>
<Nullable>enable</Nullable>
<CsWinRTWindowsMetadata>10.0.22621.0</CsWinRTWindowsMetadata>
</PropertyGroup>
<ItemGroup>
<PackageReference Include="Microsoft.Windows.CsWinRT" Version="2.1.3" />
</ItemGroup>
</Project>
I must have misunderstood the documentation a bit, I did not consider that plain .net libraries also need a CsWinRT reference, thanks for the help everybody.
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.