Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
84,275
24,267,200,999
IssuesEvent
2022-09-28 07:25:08
appsmithorg/appsmith
https://api.github.com/repos/appsmithorg/appsmith
closed
[Bug]: Max height tooltip appears without hover - doesn't disappear
Bug Low Needs Triaging UI Builders Pod Dynamic Height
### Is there an existing issue for this? - [X] I have searched the existing issues ### Description When we D&D a widget with Auto height, and choose Auto height with limits, and click on the widget, Min height tooltip appears only when we hover over the min height signifier, but in case of Max height it appears whenever we select or click on the widget (Without hovering) and doesn't disappear. ![image](https://user-images.githubusercontent.com/104895602/192709047-6360a0a4-1cfb-4e07-a4a0-f4bc1d85fb7c.png) ### Steps To Reproduce 1. D&D a widget with Auto height enabled 2. Select Auto height with limits in the height property 3. Click on the widget 4. Observe the Min height signifies tooltip disappears once we move the mouse to somewhere but the Max height signifier tooltip doesn't disappear in its own, and it stays until we unselect the widget. https://appsmith-git-fix-auto-height-limits-scroll-issue-get-appsmith.vercel.app/ ### Public Sample App _No response_ ### Version Deploy preview
1.0
[Bug]: Max height tooltip appears without hover - doesn't disappear - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Description When we D&D a widget with Auto height, and choose Auto height with limits, and click on the widget, Min height tooltip appears only when we hover over the min height signifier, but in case of Max height it appears whenever we select or click on the widget (Without hovering) and doesn't disappear. ![image](https://user-images.githubusercontent.com/104895602/192709047-6360a0a4-1cfb-4e07-a4a0-f4bc1d85fb7c.png) ### Steps To Reproduce 1. D&D a widget with Auto height enabled 2. Select Auto height with limits in the height property 3. Click on the widget 4. Observe the Min height signifies tooltip disappears once we move the mouse to somewhere but the Max height signifier tooltip doesn't disappear in its own, and it stays until we unselect the widget. https://appsmith-git-fix-auto-height-limits-scroll-issue-get-appsmith.vercel.app/ ### Public Sample App _No response_ ### Version Deploy preview
non_process
max height tooltip appears without hover doesn t disappear is there an existing issue for this i have searched the existing issues description when we d d a widget with auto height and choose auto height with limits and click on the widget min height tooltip appears only when we hover over the min height signifier but in case of max height it appears whenever we select or click on the widget without hovering and doesn t disappear steps to reproduce d d a widget with auto height enabled select auto height with limits in the height property click on the widget observe the min height signifies tooltip disappears once we move the mouse to somewhere but the max height signifier tooltip doesn t disappear in its own and it stays until we unselect the widget public sample app no response version deploy preview
0
8,151
11,354,735,928
IssuesEvent
2020-01-24 18:20:13
googleapis/nodejs-logging-winston
https://api.github.com/repos/googleapis/nodejs-logging-winston
closed
lw.express.makeMiddleware is not a function
type: process
Following the instructions in the ReadMe for the express middleware: ```javascript module.exports = async function(){ const {mylogger, mw} = await lw.express.makeMiddleware(logger); return mw; } ``` But it says makeMiddlware doesn't exist? `UnhandledPromiseRejectionWarning: TypeError: lw.express.makeMiddleware is not a function`
1.0
lw.express.makeMiddleware is not a function - Following the instructions in the ReadMe for the express middleware: ```javascript module.exports = async function(){ const {mylogger, mw} = await lw.express.makeMiddleware(logger); return mw; } ``` But it says makeMiddlware doesn't exist? `UnhandledPromiseRejectionWarning: TypeError: lw.express.makeMiddleware is not a function`
process
lw express makemiddleware is not a function following the instructions in the readme for the express middleware javascript module exports async function const mylogger mw await lw express makemiddleware logger return mw but it says makemiddlware doesn t exist unhandledpromiserejectionwarning typeerror lw express makemiddleware is not a function
1
56,614
8,101,913,171
IssuesEvent
2018-08-12 19:07:19
FreeUKGen/MyopicVicar
https://api.github.com/repos/FreeUKGen/MyopicVicar
closed
Changing my Profile email address has not changed the my Transcriber email value in some batches (was #1319)
documentation ready
Documentation needs to explain that the email address in a batch header is a feature that was needed in FR1 (when we did not have user IDs). Now we have user IDs, it does not matter what is in this field.
1.0
Changing my Profile email address has not changed the my Transcriber email value in some batches (was #1319) - Documentation needs to explain that the email address in a batch header is a feature that was needed in FR1 (when we did not have user IDs). Now we have user IDs, it does not matter what is in this field.
non_process
changing my profile email address has not changed the my transcriber email value in some batches was documentation needs to explain that the email address in a batch header is a feature that was needed in when we did not have user ids now we have user ids it does not matter what is in this field
0
54,715
11,296,038,714
IssuesEvent
2020-01-17 00:17:47
aws-amplify/amplify-cli
https://api.github.com/repos/aws-amplify/amplify-cli
reopened
How do I update src/graphql/schema.json
code-gen pending-triage
**Which Category is your question related to?** Updating `src/graphql/schema.json`. None of the commands I've seen such as `amplify push` do that, and older commands like `amplify code generate --download` no longer work. **Amplify CLI Version** 4.11.0. **What AWS Services are you utilizing?** AppSync + Aurora. My `.graphconfig` (regenerated with latest): ``` projects: TestAPI: schemaPath: amplify/backend/api/TestAPI/build/schema.graphql includes: - src/graphql/**/*.js excludes: - ./amplify/** extensions: amplify: codeGenTarget: javascript generatedFileName: '' docsFilePath: src/graphql region: eu-west-1 apiId: null maxDepth: 2 extensions: amplify: version: 3 ```
1.0
How do I update src/graphql/schema.json - **Which Category is your question related to?** Updating `src/graphql/schema.json`. None of the commands I've seen such as `amplify push` do that, and older commands like `amplify code generate --download` no longer work. **Amplify CLI Version** 4.11.0. **What AWS Services are you utilizing?** AppSync + Aurora. My `.graphconfig` (regenerated with latest): ``` projects: TestAPI: schemaPath: amplify/backend/api/TestAPI/build/schema.graphql includes: - src/graphql/**/*.js excludes: - ./amplify/** extensions: amplify: codeGenTarget: javascript generatedFileName: '' docsFilePath: src/graphql region: eu-west-1 apiId: null maxDepth: 2 extensions: amplify: version: 3 ```
non_process
how do i update src graphql schema json which category is your question related to updating src graphql schema json none of the commands i ve seen such as amplify push do that and older commands like amplify code generate download no longer work amplify cli version what aws services are you utilizing appsync aurora my graphconfig regenerated with latest projects testapi schemapath amplify backend api testapi build schema graphql includes src graphql js excludes amplify extensions amplify codegentarget javascript generatedfilename docsfilepath src graphql region eu west apiid null maxdepth extensions amplify version
0
17,448
23,268,927,961
IssuesEvent
2022-08-04 20:25:53
ppy/osu-web
https://api.github.com/repos/ppy/osu-web
closed
Allow for selectable difficulty icons
proposal area:beatmap-discussions area:beatmap-processing
Potentially, this could take the form of a box on the beatmap listing where you can choose one of the icons which would appear when hovering a difficulty icon. The ones that aren't selected would be darker/grayed out and colored on hover/selection. ![image](https://user-images.githubusercontent.com/33286824/48960485-fad1df80-ef3a-11e8-816a-9ec343243648.png) This would be good because icons are often inaccurate to the intended difficulty level, especially on edge cases like 3.6* insanes or 2.3* normals. Users can already set the name of the difficulty, so being able to set the icon to what difficulty it is meant to be would make sense. With this it'd be easier to see at a glance on the beatmap listing what the spread of a map is actually like, such as ENHI instead of seeing NNHH - which makes it look like there's two normals and two hards when that's not the case at all. This would be changeable by: - the mapper, when a mapset is in pending/wip/graveyard - QAT and higher when a mapset is in any state. By default, icons would be set automatically by star rating the same way they currently are.
1.0
Allow for selectable difficulty icons - Potentially, this could take the form of a box on the beatmap listing where you can choose one of the icons which would appear when hovering a difficulty icon. The ones that aren't selected would be darker/grayed out and colored on hover/selection. ![image](https://user-images.githubusercontent.com/33286824/48960485-fad1df80-ef3a-11e8-816a-9ec343243648.png) This would be good because icons are often inaccurate to the intended difficulty level, especially on edge cases like 3.6* insanes or 2.3* normals. Users can already set the name of the difficulty, so being able to set the icon to what difficulty it is meant to be would make sense. With this it'd be easier to see at a glance on the beatmap listing what the spread of a map is actually like, such as ENHI instead of seeing NNHH - which makes it look like there's two normals and two hards when that's not the case at all. This would be changeable by: - the mapper, when a mapset is in pending/wip/graveyard - QAT and higher when a mapset is in any state. By default, icons would be set automatically by star rating the same way they currently are.
process
allow for selectable difficulty icons potentially this could take the form of a box on the beatmap listing where you can choose one of the icons which would appear when hovering a difficulty icon the ones that aren t selected would be darker grayed out and colored on hover selection this would be good because icons are often inaccurate to the intended difficulty level especially on edge cases like insanes or normals users can already set the name of the difficulty so being able to set the icon to what difficulty it is meant to be would make sense with this it d be easier to see at a glance on the beatmap listing what the spread of a map is actually like such as enhi instead of seeing nnhh which makes it look like there s two normals and two hards when that s not the case at all this would be changeable by the mapper when a mapset is in pending wip graveyard qat and higher when a mapset is in any state by default icons would be set automatically by star rating the same way they currently are
1
124,590
10,317,157,271
IssuesEvent
2019-08-30 11:59:12
alan-turing-institute/MLJLinearModels.jl
https://api.github.com/repos/alan-turing-institute/MLJLinearModels.jl
closed
Either remove or test (F)ADMM on simpler problem
tests
Could not get it to work well on LAD but maybe can make it work on something else like LASSO and can test the rest of the implementation on the go. In the mean time maybe isolate the code in another branch as it makes code cov look bad.
1.0
Either remove or test (F)ADMM on simpler problem - Could not get it to work well on LAD but maybe can make it work on something else like LASSO and can test the rest of the implementation on the go. In the mean time maybe isolate the code in another branch as it makes code cov look bad.
non_process
either remove or test f admm on simpler problem could not get it to work well on lad but maybe can make it work on something else like lasso and can test the rest of the implementation on the go in the mean time maybe isolate the code in another branch as it makes code cov look bad
0
15,849
20,029,314,308
IssuesEvent
2022-02-02 02:18:40
fluent/fluent-bit
https://api.github.com/repos/fluent/fluent-bit
closed
Benchmark
work-in-process Stale
In order for users to assess Fluent Bit, it would be desirable to have a benchmark that covers performance and resource usage. This benchmark could include throughput for various payloads and protocols (HTTP/TCP) as well as resource usage (CPU/MEM, compare with what we did in [Centralized Container Logging with Fluent Bit](https://aws.amazon.com/blogs/opensource/centralized-container-logging-fluent-bit/)). In addition, the benchmark could be used to compare different Fluent Bit releases.
1.0
Benchmark - In order for users to assess Fluent Bit, it would be desirable to have a benchmark that covers performance and resource usage. This benchmark could include throughput for various payloads and protocols (HTTP/TCP) as well as resource usage (CPU/MEM, compare with what we did in [Centralized Container Logging with Fluent Bit](https://aws.amazon.com/blogs/opensource/centralized-container-logging-fluent-bit/)). In addition, the benchmark could be used to compare different Fluent Bit releases.
process
benchmark in order for users to assess fluent bit it would be desirable to have a benchmark that covers performance and resource usage this benchmark could include throughput for various payloads and protocols http tcp as well as resource usage cpu mem compare with what we did in in addition the benchmark could be used to compare different fluent bit releases
1
4,565
7,393,767,531
IssuesEvent
2018-03-17 01:33:12
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Useless Link to Windows PowerShell cmdlets for Azure API Management API
api-management cxp doc-bug in-process triaged
The link given to Windows PowerShell cmdlets for Azure API Management API goes to a useless page. It would be better to send readers to the page for the actual module where the commands are here: https://docs.microsoft.com/en-us/powershell/module/azurerm.apimanagement/?view=azurermps-5.5.0#api_management. Otherwise the link is a dead-end. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 9329324a-cc58-0824-6ad7-a0e5323f68f8 * Version Independent ID: 82200310-9caa-6d03-e02c-7edcdb2bee61 * Content: [Manage Azure API Management using Azure Automation | Microsoft Docs](https://docs.microsoft.com/en-us/azure/api-management/automation-manage-api-management) * Content Source: [articles/api-management/automation-manage-api-management.md](https://github.com/Microsoft/azure-docs/blob/master/articles/api-management/automation-manage-api-management.md) * Service: **api-management** * GitHub Login: @vladvino * Microsoft Alias: **apimpm**
1.0
Useless Link to Windows PowerShell cmdlets for Azure API Management API - The link given to Windows PowerShell cmdlets for Azure API Management API goes to a useless page. It would be better to send readers to the page for the actual module where the commands are here: https://docs.microsoft.com/en-us/powershell/module/azurerm.apimanagement/?view=azurermps-5.5.0#api_management. Otherwise the link is a dead-end. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 9329324a-cc58-0824-6ad7-a0e5323f68f8 * Version Independent ID: 82200310-9caa-6d03-e02c-7edcdb2bee61 * Content: [Manage Azure API Management using Azure Automation | Microsoft Docs](https://docs.microsoft.com/en-us/azure/api-management/automation-manage-api-management) * Content Source: [articles/api-management/automation-manage-api-management.md](https://github.com/Microsoft/azure-docs/blob/master/articles/api-management/automation-manage-api-management.md) * Service: **api-management** * GitHub Login: @vladvino * Microsoft Alias: **apimpm**
process
useless link to windows powershell cmdlets for azure api management api the link given to windows powershell cmdlets for azure api management api goes to a useless page it would be better to send readers to the page for the actual module where the commands are here otherwise the link is a dead end document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service api management github login vladvino microsoft alias apimpm
1
316,220
27,146,178,756
IssuesEvent
2023-02-16 20:08:50
hashicorp/terraform-provider-google
https://api.github.com/repos/hashicorp/terraform-provider-google
closed
Failing test(s): TestAccFirebaseAppleApp_*
size/s priority/1 test failure
### Affected Resource(s) <!--- Please list the affected resources and data sources. ---> * google_firebase_apple_app <!--- This is a template for reporting test failures on nightly builds. It should only be used by core contributors who have access to our CI/CD results. ---> <!-- i.e. "Consistently since X date" or "X% failure in MONTH" --> Failure rate: 100% since 2022-11-22 <!-- List all impacted tests for searchability. The title of the issue can instead list one or more groups of tests, or describe the overall root cause. --> Impacted tests: - TestAccFirebaseAppleApp_firebaseAppleAppBasicExample - TestAccFirebaseAppleApp_update - TestAccFirebaseAppleApp_firebaseAppleAppFullExample <!-- Link to the nightly build(s), ideally with one impacted test opened --> Nightly builds: - https://ci-oss.hashicorp.engineering/buildConfiguration/GoogleCloudBeta_ProviderGoogleCloudBetaGoogleProject/360401?buildTab=tests&expandedTest=-7026834308340857772 <!-- The error message that displays in the tests tab, for reference --> Message: ``` Skip deleting App "projects/ci-test-project-188019/iosApps/1:1067888929963:ios:68211e93cd713a4644dbee" due to deletion_policy: "" testing_new.go:84: Error running post-test destroy, there may be dangling resources: FirebaseAppleApp still exists ``` ``` Error: Error creating AppleApp: googleapi: Error 409: Requested entity already exists ```
1.0
Failing test(s): TestAccFirebaseAppleApp_* - ### Affected Resource(s) <!--- Please list the affected resources and data sources. ---> * google_firebase_apple_app <!--- This is a template for reporting test failures on nightly builds. It should only be used by core contributors who have access to our CI/CD results. ---> <!-- i.e. "Consistently since X date" or "X% failure in MONTH" --> Failure rate: 100% since 2022-11-22 <!-- List all impacted tests for searchability. The title of the issue can instead list one or more groups of tests, or describe the overall root cause. --> Impacted tests: - TestAccFirebaseAppleApp_firebaseAppleAppBasicExample - TestAccFirebaseAppleApp_update - TestAccFirebaseAppleApp_firebaseAppleAppFullExample <!-- Link to the nightly build(s), ideally with one impacted test opened --> Nightly builds: - https://ci-oss.hashicorp.engineering/buildConfiguration/GoogleCloudBeta_ProviderGoogleCloudBetaGoogleProject/360401?buildTab=tests&expandedTest=-7026834308340857772 <!-- The error message that displays in the tests tab, for reference --> Message: ``` Skip deleting App "projects/ci-test-project-188019/iosApps/1:1067888929963:ios:68211e93cd713a4644dbee" due to deletion_policy: "" testing_new.go:84: Error running post-test destroy, there may be dangling resources: FirebaseAppleApp still exists ``` ``` Error: Error creating AppleApp: googleapi: Error 409: Requested entity already exists ```
non_process
failing test s testaccfirebaseappleapp affected resource s google firebase apple app failure rate since impacted tests testaccfirebaseappleapp firebaseappleappbasicexample testaccfirebaseappleapp update testaccfirebaseappleapp firebaseappleappfullexample nightly builds message skip deleting app projects ci test project iosapps ios due to deletion policy testing new go error running post test destroy there may be dangling resources firebaseappleapp still exists error error creating appleapp googleapi error requested entity already exists
0
8,042
11,217,766,486
IssuesEvent
2020-01-07 09:58:42
prisma/prisma2
https://api.github.com/repos/prisma/prisma2
closed
CLI tries to download npm packages using `npm add`
bug/2-confirmed kind/bug process/candidate
https://github.com/prisma/prisma2/blob/34f5e82d201fc5ceaad6041bf670838161188f1b/cli/sdk/src/predefinedGeneratorResolvers.ts#L88-L90 It tries to download packages using `npm add` even though npm packages are installed using `npm install`
1.0
CLI tries to download npm packages using `npm add` - https://github.com/prisma/prisma2/blob/34f5e82d201fc5ceaad6041bf670838161188f1b/cli/sdk/src/predefinedGeneratorResolvers.ts#L88-L90 It tries to download packages using `npm add` even though npm packages are installed using `npm install`
process
cli tries to download npm packages using npm add it tries to download packages using npm add even though npm packages are installed using npm install
1
7,665
10,756,857,286
IssuesEvent
2019-10-31 12:08:08
kubeflow/kubeflow
https://api.github.com/repos/kubeflow/kubeflow
opened
Qualify 0.7.0 with kfctl_k8s_istio.yaml
area/kfctl kind/process priority/p0
We are now on 0.7.0RC7 for kfctl https://github.com/kubeflow/kubeflow/releases There are currently no known P0 issues. https://github.com/orgs/kubeflow/projects/22?card_filter_query=label%3Apriority%2Fp0 Opening this bug to track qualifying the `kfctl_k8s_istio.yaml` config. Ideally we'd like to aim to finalize 0.7.0 today. So it would be great to do run through the deployment and identify and fix any issues that come up. There was a bug #4415 with the registration flow in the case with no identity. We'd like to verify that issue has been fixed. Related to: #4249 /assign @krishnadurai
1.0
Qualify 0.7.0 with kfctl_k8s_istio.yaml - We are now on 0.7.0RC7 for kfctl https://github.com/kubeflow/kubeflow/releases There are currently no known P0 issues. https://github.com/orgs/kubeflow/projects/22?card_filter_query=label%3Apriority%2Fp0 Opening this bug to track qualifying the `kfctl_k8s_istio.yaml` config. Ideally we'd like to aim to finalize 0.7.0 today. So it would be great to do run through the deployment and identify and fix any issues that come up. There was a bug #4415 with the registration flow in the case with no identity. We'd like to verify that issue has been fixed. Related to: #4249 /assign @krishnadurai
process
qualify with kfctl istio yaml we are now on for kfctl there are currently no known issues opening this bug to track qualifying the kfctl istio yaml config ideally we d like to aim to finalize today so it would be great to do run through the deployment and identify and fix any issues that come up there was a bug with the registration flow in the case with no identity we d like to verify that issue has been fixed related to assign krishnadurai
1
16,687
21,791,068,442
IssuesEvent
2022-05-14 22:53:25
lynnandtonic/nestflix.fun
https://api.github.com/repos/lynnandtonic/nestflix.fun
closed
Ask Mr. Lizard from Jim Henson's Dinosaurs
suggested title in process
Please add as much of the following info as you can: Title: Ask Mr. Lizard Type (film/tv show): TV show Film or show in which it appears: Jim Henson's Dinosaurs Is the parent film/show streaming anywhere? DISNEY+ About when in the parent film/show does it appear? Episode 201, 215 and 413. Actual footage of the film/show can be seen (yes/no)? Yes. The running joke is that the little dinosaur child Timmy always dies and Mr Lizard tells the crew "We're gonna need another Timmy"
1.0
Ask Mr. Lizard from Jim Henson's Dinosaurs - Please add as much of the following info as you can: Title: Ask Mr. Lizard Type (film/tv show): TV show Film or show in which it appears: Jim Henson's Dinosaurs Is the parent film/show streaming anywhere? DISNEY+ About when in the parent film/show does it appear? Episode 201, 215 and 413. Actual footage of the film/show can be seen (yes/no)? Yes. The running joke is that the little dinosaur child Timmy always dies and Mr Lizard tells the crew "We're gonna need another Timmy"
process
ask mr lizard from jim henson s dinosaurs please add as much of the following info as you can title ask mr lizard type film tv show tv show film or show in which it appears jim henson s dinosaurs is the parent film show streaming anywhere disney about when in the parent film show does it appear episode and actual footage of the film show can be seen yes no yes the running joke is that the little dinosaur child timmy always dies and mr lizard tells the crew we re gonna need another timmy
1
11,434
9,367,510,782
IssuesEvent
2019-04-03 05:55:02
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
There is a missing line in the program list.
assigned-to-author cognitive-services/svc doc-bug triaged
@linyixian commented on [Sun Mar 17 2019](https://github.com/MicrosoftDocs/azure-docs.ja-jp/issues/2223) モデルとラベルを読み込む際にファイル名を代入する部分が抜けています。 filename = "model.pb" labels_filename = "labels.txt" --- #### ドキュメントの詳細 ⚠ *このセクションを編集しないでください。 docs.microsoft.com で必須です ➟ GitHub の問題のリンク。* * ID: 754dd7d5-dd8b-4a75-5d79-0ba621a847cd * Version Independent ID: 3cca3535-1c16-2da1-46c6-421cbf808bc3 * Content: [チュートリアル:Python での TensorFlow モデルの実行 - Custom Vision Service](https://docs.microsoft.com/ja-jp/azure/cognitive-services/custom-vision-service/export-model-python#feedback) * Content Source: [articles/cognitive-services/Custom-Vision-Service/export-model-python.md](https://github.com/MicrosoftDocs/azure-docs.ja-jp/blob/live/articles/cognitive-services/Custom-Vision-Service/export-model-python.md) * Service: **cognitive-services** * GitHub Login: @areddish * Microsoft Alias: **areddish**
1.0
There is a missing line in the program list. - @linyixian commented on [Sun Mar 17 2019](https://github.com/MicrosoftDocs/azure-docs.ja-jp/issues/2223) モデルとラベルを読み込む際にファイル名を代入する部分が抜けています。 filename = "model.pb" labels_filename = "labels.txt" --- #### ドキュメントの詳細 ⚠ *このセクションを編集しないでください。 docs.microsoft.com で必須です ➟ GitHub の問題のリンク。* * ID: 754dd7d5-dd8b-4a75-5d79-0ba621a847cd * Version Independent ID: 3cca3535-1c16-2da1-46c6-421cbf808bc3 * Content: [チュートリアル:Python での TensorFlow モデルの実行 - Custom Vision Service](https://docs.microsoft.com/ja-jp/azure/cognitive-services/custom-vision-service/export-model-python#feedback) * Content Source: [articles/cognitive-services/Custom-Vision-Service/export-model-python.md](https://github.com/MicrosoftDocs/azure-docs.ja-jp/blob/live/articles/cognitive-services/Custom-Vision-Service/export-model-python.md) * Service: **cognitive-services** * GitHub Login: @areddish * Microsoft Alias: **areddish**
non_process
there is a missing line in the program list linyixian commented on モデルとラベルを読み込む際にファイル名を代入する部分が抜けています。 filename model pb labels filename labels txt ドキュメントの詳細 ⚠ このセクションを編集しないでください。 docs microsoft com で必須です ➟ github の問題のリンク。 id version independent id content content source service cognitive services github login areddish microsoft alias areddish
0
727,681
25,044,031,678
IssuesEvent
2022-11-05 02:39:47
PMEAL/OpenPNM
https://api.github.com/repos/PMEAL/OpenPNM
closed
Remove the py module
high priority maintenance
This is causing our nightly builds to fail, no sign of what went wrong but while digging I see that the devs [discourage](https://github.com/pytest-dev/py) its use
1.0
Remove the py module - This is causing our nightly builds to fail, no sign of what went wrong but while digging I see that the devs [discourage](https://github.com/pytest-dev/py) its use
non_process
remove the py module this is causing our nightly builds to fail no sign of what went wrong but while digging i see that the devs its use
0
4,242
7,005,164,884
IssuesEvent
2017-12-19 00:23:26
websharks/comet-cache
https://api.github.com/repos/websharks/comet-cache
opened
create_function() is deprecated in PHP 7.2
compatibility
> Seen with WP4.9.1 running on a PHP7.2.0 server: PHP Deprecated: Function create_function() is deprecated in …\wp-content\plugins\comet-cache\src\includes\traits\Shared\FsUtils.php on line 33 Reported here: https://wordpress.org/support/topic/version-170220-function-create_function-is-deprecated-in-php-7-2/
True
create_function() is deprecated in PHP 7.2 - > Seen with WP4.9.1 running on a PHP7.2.0 server: PHP Deprecated: Function create_function() is deprecated in …\wp-content\plugins\comet-cache\src\includes\traits\Shared\FsUtils.php on line 33 Reported here: https://wordpress.org/support/topic/version-170220-function-create_function-is-deprecated-in-php-7-2/
non_process
create function is deprecated in php seen with running on a server php deprecated function create function is deprecated in … wp content plugins comet cache src includes traits shared fsutils php on line reported here
0
595
3,070,997,438
IssuesEvent
2015-08-19 09:13:08
Graylog2/graylog2-server
https://api.github.com/repos/Graylog2/graylog2-server
closed
Add JSON converter
feature processing
For extracting information from log messages which contain JSON and might appear in logged query URLs and similar data, a JSON extractor would be useful.
1.0
Add JSON converter - For extracting information from log messages which contain JSON and might appear in logged query URLs and similar data, a JSON extractor would be useful.
process
add json converter for extracting information from log messages which contain json and might appear in logged query urls and similar data a json extractor would be useful
1
1,843
4,647,114,221
IssuesEvent
2016-10-01 09:09:48
AllenFang/react-bootstrap-table
https://api.github.com/repos/AllenFang/react-bootstrap-table
closed
TableHeaderColumn does not accept function as className attribute
enhancement inprocess
According to the [docs](http://allenfang.github.io/react-bootstrap-table/docs.html), the TableHeaderColumn should accept passing a function to the className attribute. However, the current implementation restricts the PropType to String: ` TableHeaderColumn.propTypes = { ... className: PropTypes.string, ... `
1.0
TableHeaderColumn does not accept function as className attribute - According to the [docs](http://allenfang.github.io/react-bootstrap-table/docs.html), the TableHeaderColumn should accept passing a function to the className attribute. However, the current implementation restricts the PropType to String: ` TableHeaderColumn.propTypes = { ... className: PropTypes.string, ... `
process
tableheadercolumn does not accept function as classname attribute according to the the tableheadercolumn should accept passing a function to the classname attribute however the current implementation restricts the proptype to string tableheadercolumn proptypes classname proptypes string
1
19,503
25,812,450,721
IssuesEvent
2022-12-12 00:08:13
nkdAgility/azure-devops-migration-tools
https://api.github.com/repos/nkdAgility/azure-devops-migration-tools
closed
Issue Migrating Pipeline with Service Connection
no-issue-activity Pipeline Processor
## Describe your issue: When trying to migrate a pipeline that references a service connection from one AzDO Project/Org to another AzDO Project/Org, I get the following error on the service connection migration: ``` [08:57:34 ERR] Error migrating ServiceConnection: Artifactory. Please migrate it manually. Url: POST https://dev.azure.com/TargetOrg/Test/_apis/serviceendpoint/endpoints/ {"$id":"1","innerException":null,"message":"At least one project reference required to create an endpoint.","typeName":"System.ArgumentException, mscorlib","typeKey":"ArgumentException","errorCode":0,"eventId":0} ``` Then when the migration tool tries to migrate the pipeline that uses the service connection I receive this error: ``` [08:57:40 ERR] Error migrating BuildDefinition: Test-CI. Please migrate it manually. Url: POST https://dev.azure.com/TargetOrg/Test/_apis/build/definitions/ {"$id":"1","innerException":null,"message":"The pipeline is not valid. Job Job_1: Step input artifactoryService references service connection 6696c25f-ac9a-4745-bd11-949dfba9c2cc which could not be found. The service connection does not exist or has not been authorized for use. For authorization details, refer to https://aka.ms/yamlauthz.","typeName":"Microsoft.TeamFoundation.DistributedTask.Pipelines.PipelineValidationException, Microsoft.TeamFoundation.DistributedTask.WebApi","typeKey":"PipelineValidationException","errorCode":0,"eventId":3000} ``` Even if I create the service connection in the target project I get the above error message. Please see the below config json: ``` { "ChangeSetMappingFile":null, "Endpoints":{ "AzureDevOpsEndpoints":[ { "Name":"Source", "$type":"AzureDevOpsEndpointOptions", "Organisation":"https://dev.azure.com/SourceOrg", "Project":"Test", "ReflectedWorkItemIDFieldName":"Custom.ReflectedWorkItemId", "AuthenticationMode":"AccessToken", "AccessToken":"token", "EndpointEnrichers":null }, { "Name":"Target", "$type":"AzureDevOpsEndpointOptions", "Organisation":"https://dev.azure.com/TargetOrg", "Project":"Test", "ReflectedWorkItemIDFieldName":"TfsMigrationTool.ReflectedWorkItemId", "AuthenticationMode":"AccessToken", "AccessToken":"token", "EndpointEnrichers":null } ] }, "LogLevel":"Debug", "Processors":[ { "$type":"AzureDevOpsPipelineProcessorOptions", "Enabled":true, "MigrateBuildPipelines":true, "MigrateReleasePipelines":true, "MigrateTaskGroups":true, "MigrateVariableGroups":true, "MigrateServiceConnections":true, "BuildPipelines":null, "ReleasePipelines":null, "RefName":null, "SourceName":"Source", "TargetName":"Target", "RepositoryNameMaps":{ "Test":"Test" } } ], "Version":"12.0", "Endpoints":{ "InMemoryWorkItemEndpoints":[ { "Name":"Source", "EndpointEnrichers":null }, { "Name":"Target", "EndpointEnrichers":null } ] } } ``` ## Source Details Azure DevOps Services ## Target Details Azure DevOps Services [test.txt](https://github.com/nkdAgility/azure-devops-migration-tools/files/9769048/test.txt)
1.0
Issue Migrating Pipeline with Service Connection - ## Describe your issue: When trying to migrate a pipeline that references a service connection from one AzDO Project/Org to another AzDO Project/Org, I get the following error on the service connection migration: ``` [08:57:34 ERR] Error migrating ServiceConnection: Artifactory. Please migrate it manually. Url: POST https://dev.azure.com/TargetOrg/Test/_apis/serviceendpoint/endpoints/ {"$id":"1","innerException":null,"message":"At least one project reference required to create an endpoint.","typeName":"System.ArgumentException, mscorlib","typeKey":"ArgumentException","errorCode":0,"eventId":0} ``` Then when the migration tool tries to migrate the pipeline that uses the service connection I receive this error: ``` [08:57:40 ERR] Error migrating BuildDefinition: Test-CI. Please migrate it manually. Url: POST https://dev.azure.com/TargetOrg/Test/_apis/build/definitions/ {"$id":"1","innerException":null,"message":"The pipeline is not valid. Job Job_1: Step input artifactoryService references service connection 6696c25f-ac9a-4745-bd11-949dfba9c2cc which could not be found. The service connection does not exist or has not been authorized for use. For authorization details, refer to https://aka.ms/yamlauthz.","typeName":"Microsoft.TeamFoundation.DistributedTask.Pipelines.PipelineValidationException, Microsoft.TeamFoundation.DistributedTask.WebApi","typeKey":"PipelineValidationException","errorCode":0,"eventId":3000} ``` Even if I create the service connection in the target project I get the above error message. Please see the below config json: ``` { "ChangeSetMappingFile":null, "Endpoints":{ "AzureDevOpsEndpoints":[ { "Name":"Source", "$type":"AzureDevOpsEndpointOptions", "Organisation":"https://dev.azure.com/SourceOrg", "Project":"Test", "ReflectedWorkItemIDFieldName":"Custom.ReflectedWorkItemId", "AuthenticationMode":"AccessToken", "AccessToken":"token", "EndpointEnrichers":null }, { "Name":"Target", "$type":"AzureDevOpsEndpointOptions", "Organisation":"https://dev.azure.com/TargetOrg", "Project":"Test", "ReflectedWorkItemIDFieldName":"TfsMigrationTool.ReflectedWorkItemId", "AuthenticationMode":"AccessToken", "AccessToken":"token", "EndpointEnrichers":null } ] }, "LogLevel":"Debug", "Processors":[ { "$type":"AzureDevOpsPipelineProcessorOptions", "Enabled":true, "MigrateBuildPipelines":true, "MigrateReleasePipelines":true, "MigrateTaskGroups":true, "MigrateVariableGroups":true, "MigrateServiceConnections":true, "BuildPipelines":null, "ReleasePipelines":null, "RefName":null, "SourceName":"Source", "TargetName":"Target", "RepositoryNameMaps":{ "Test":"Test" } } ], "Version":"12.0", "Endpoints":{ "InMemoryWorkItemEndpoints":[ { "Name":"Source", "EndpointEnrichers":null }, { "Name":"Target", "EndpointEnrichers":null } ] } } ``` ## Source Details Azure DevOps Services ## Target Details Azure DevOps Services [test.txt](https://github.com/nkdAgility/azure-devops-migration-tools/files/9769048/test.txt)
process
issue migrating pipeline with service connection describe your issue when trying to migrate a pipeline that references a service connection from one azdo project org to another azdo project org i get the following error on the service connection migration error migrating serviceconnection artifactory please migrate it manually url post id innerexception null message at least one project reference required to create an endpoint typename system argumentexception mscorlib typekey argumentexception errorcode eventid then when the migration tool tries to migrate the pipeline that uses the service connection i receive this error error migrating builddefinition test ci please migrate it manually url post id innerexception null message the pipeline is not valid job job step input artifactoryservice references service connection which could not be found the service connection does not exist or has not been authorized for use for authorization details refer to microsoft teamfoundation distributedtask webapi typekey pipelinevalidationexception errorcode eventid even if i create the service connection in the target project i get the above error message please see the below config json changesetmappingfile null endpoints azuredevopsendpoints name source type azuredevopsendpointoptions organisation project test reflectedworkitemidfieldname custom reflectedworkitemid authenticationmode accesstoken accesstoken token endpointenrichers null name target type azuredevopsendpointoptions organisation project test reflectedworkitemidfieldname tfsmigrationtool reflectedworkitemid authenticationmode accesstoken accesstoken token endpointenrichers null loglevel debug processors type azuredevopspipelineprocessoroptions enabled true migratebuildpipelines true migratereleasepipelines true migratetaskgroups true migratevariablegroups true migrateserviceconnections true buildpipelines null releasepipelines null refname null sourcename source targetname target repositorynamemaps test test version endpoints inmemoryworkitemendpoints name source endpointenrichers null name target endpointenrichers null source details azure devops services target details azure devops services
1
2,948
5,930,207,519
IssuesEvent
2017-05-24 00:24:43
ncbo/bioportal-project
https://api.github.com/repos/ncbo/bioportal-project
opened
CHEAR: latest submission failed to parse
ontology processing problem
The latest submission of the [CHEAR ontology](http://bioportal.bioontology.org/ontologies/CHEAR) failed to parse (status shows "Error Rdf") in the BioPortal UI. Parsing logs indicate that the ontology was successfully parsed by the OWL API, but failed the secondary rapper parsing step: ``` I, [2017-05-23T07:57:20.760933 #3345] INFO -- : ["OWLAPI Java command: parsing finished successfully."] I, [2017-05-23T07:57:20.761117 #3345] INFO -- : ["Output size 11890550 in `/srv/ncbo/repository/CHEAR/3/owlapi.xrdf`"] E, [2017-05-23T07:57:21.540084 #3345] ERROR -- : ["Exception: Rapper cannot parse rdfxml file at /srv/ncbo/repository/CHEAR/3/owlapi.xrdf: rapper: Parsing URI file:///srv/ncbo/repository/CHEAR/3/owlapi.xrdf with parser rdfxml rapper: Serializing with serializer ntriples rapper: Error - URI file:///srv/ncbo/repository/CHEAR/3/owlapi.xrdf:169752 - Illegal rdf:nodeID value '_:genid8154' ``` Manually running rapper command confirms that rapper is unable to serialize the content in the owlapi.xrdf file: ``` [ncbo-deployer@ncbo-prd-app-25 3]$ rapper -i rdfxml -o ntriples owlapi.xrdf > data.triples rapper: Parsing URI file:///srv/ncbo/share/env/production/repository/CHEAR/3/owlapi.xrdf with parser rdfxml rapper: Serializing with serializer ntriples rapper: Error - URI file:///srv/ncbo/share/env/production/repository/CHEAR/3/owlapi.xrdf:169752 - Illegal rdf:nodeID value '_:genid8154' rapper: Failed to parse file owlapi.xrdf rdfxml content rapper: Parsing returned 104863 triples ```
1.0
CHEAR: latest submission failed to parse - The latest submission of the [CHEAR ontology](http://bioportal.bioontology.org/ontologies/CHEAR) failed to parse (status shows "Error Rdf") in the BioPortal UI. Parsing logs indicate that the ontology was successfully parsed by the OWL API, but failed the secondary rapper parsing step: ``` I, [2017-05-23T07:57:20.760933 #3345] INFO -- : ["OWLAPI Java command: parsing finished successfully."] I, [2017-05-23T07:57:20.761117 #3345] INFO -- : ["Output size 11890550 in `/srv/ncbo/repository/CHEAR/3/owlapi.xrdf`"] E, [2017-05-23T07:57:21.540084 #3345] ERROR -- : ["Exception: Rapper cannot parse rdfxml file at /srv/ncbo/repository/CHEAR/3/owlapi.xrdf: rapper: Parsing URI file:///srv/ncbo/repository/CHEAR/3/owlapi.xrdf with parser rdfxml rapper: Serializing with serializer ntriples rapper: Error - URI file:///srv/ncbo/repository/CHEAR/3/owlapi.xrdf:169752 - Illegal rdf:nodeID value '_:genid8154' ``` Manually running rapper command confirms that rapper is unable to serialize the content in the owlapi.xrdf file: ``` [ncbo-deployer@ncbo-prd-app-25 3]$ rapper -i rdfxml -o ntriples owlapi.xrdf > data.triples rapper: Parsing URI file:///srv/ncbo/share/env/production/repository/CHEAR/3/owlapi.xrdf with parser rdfxml rapper: Serializing with serializer ntriples rapper: Error - URI file:///srv/ncbo/share/env/production/repository/CHEAR/3/owlapi.xrdf:169752 - Illegal rdf:nodeID value '_:genid8154' rapper: Failed to parse file owlapi.xrdf rdfxml content rapper: Parsing returned 104863 triples ```
process
chear latest submission failed to parse the latest submission of the failed to parse status shows error rdf in the bioportal ui parsing logs indicate that the ontology was successfully parsed by the owl api but failed the secondary rapper parsing step i info i info e error exception rapper cannot parse rdfxml file at srv ncbo repository chear owlapi xrdf rapper parsing uri file srv ncbo repository chear owlapi xrdf with parser rdfxml rapper serializing with serializer ntriples rapper error uri file srv ncbo repository chear owlapi xrdf illegal rdf nodeid value manually running rapper command confirms that rapper is unable to serialize the content in the owlapi xrdf file rapper i rdfxml o ntriples owlapi xrdf data triples rapper parsing uri file srv ncbo share env production repository chear owlapi xrdf with parser rdfxml rapper serializing with serializer ntriples rapper error uri file srv ncbo share env production repository chear owlapi xrdf illegal rdf nodeid value rapper failed to parse file owlapi xrdf rdfxml content rapper parsing returned triples
1
100,028
4,075,711,373
IssuesEvent
2016-05-29 12:02:46
HabitRPG/habitrpg
https://api.github.com/repos/HabitRPG/habitrpg
closed
"Error request entity too large" - can't join / leave / edit / end challenge or guild
API bug Memorable priority - critical
``` 2013-09-07T23:05:25.959760+00:00 app[web.1]: Error: Request Entity Too Large 2013-09-07T23:05:25.959760+00:00 app[web.1]: at Object.exports.error (/app/node_modules/express/node_modules/connect/lib/utils.js:62:13) ``` ---- _edit by admin_: # What to do if you see "request entity too large" from any action involving a challenge or a guild: **This problem occurs when guilds and challenges have many players in them.** **It can be a temporary error, so try again a few hours later.** **If you are trying to join a challenge, use [this "Join Challenge" form](https://jsfiddle.net/robcthegeek/qnc4g5g6/embedded/result/). It is likely to work in many cases but sometimes might also produce an error (due to the challenge being too large or popular, not due to a bug in the form).** **If you're still unable to join the challenge, please comment here and include your User ID (see step 6 of the Help -> Report a Bug instructions), the URL of the challenge or guild (the URL will have a long string of random characters in it, such as https://habitica.com/#/options/groups/challenges/3b4f33e9-727b-4b9e-b41c-babfbcdb7ca4), and the action you were trying to take (e.g., join, leave, etc).** **Especially comment if you are the challenge or guild owner and are trying to edit it but can't. Tell us what changes you need to have made.** **If you are trying to close a challenge and cannot because of this error, make a note of the challenge's name, its URL, and the User ID and name of the person who is the winner. Add that information here and an admin will help you award the challenge achievement and prize.** This is due to #5830 <!--- @huboard:{"order":85.21591186523438} --> <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/965314-error-request-entity-too-large-can-t-join-leave-edit-end-challenge-or-guild?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github). </bountysource-plugin>
1.0
"Error request entity too large" - can't join / leave / edit / end challenge or guild - ``` 2013-09-07T23:05:25.959760+00:00 app[web.1]: Error: Request Entity Too Large 2013-09-07T23:05:25.959760+00:00 app[web.1]: at Object.exports.error (/app/node_modules/express/node_modules/connect/lib/utils.js:62:13) ``` ---- _edit by admin_: # What to do if you see "request entity too large" from any action involving a challenge or a guild: **This problem occurs when guilds and challenges have many players in them.** **It can be a temporary error, so try again a few hours later.** **If you are trying to join a challenge, use [this "Join Challenge" form](https://jsfiddle.net/robcthegeek/qnc4g5g6/embedded/result/). It is likely to work in many cases but sometimes might also produce an error (due to the challenge being too large or popular, not due to a bug in the form).** **If you're still unable to join the challenge, please comment here and include your User ID (see step 6 of the Help -> Report a Bug instructions), the URL of the challenge or guild (the URL will have a long string of random characters in it, such as https://habitica.com/#/options/groups/challenges/3b4f33e9-727b-4b9e-b41c-babfbcdb7ca4), and the action you were trying to take (e.g., join, leave, etc).** **Especially comment if you are the challenge or guild owner and are trying to edit it but can't. Tell us what changes you need to have made.** **If you are trying to close a challenge and cannot because of this error, make a note of the challenge's name, its URL, and the User ID and name of the person who is the winner. Add that information here and an admin will help you award the challenge achievement and prize.** This is due to #5830 <!--- @huboard:{"order":85.21591186523438} --> <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/965314-error-request-entity-too-large-can-t-join-leave-edit-end-challenge-or-guild?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github). </bountysource-plugin>
non_process
error request entity too large can t join leave edit end challenge or guild app error request entity too large app at object exports error app node modules express node modules connect lib utils js edit by admin what to do if you see request entity too large from any action involving a challenge or a guild this problem occurs when guilds and challenges have many players in them it can be a temporary error so try again a few hours later if you are trying to join a challenge use it is likely to work in many cases but sometimes might also produce an error due to the challenge being too large or popular not due to a bug in the form if you re still unable to join the challenge please comment here and include your user id see step of the help report a bug instructions the url of the challenge or guild the url will have a long string of random characters in it such as and the action you were trying to take e g join leave etc especially comment if you are the challenge or guild owner and are trying to edit it but can t tell us what changes you need to have made if you are trying to close a challenge and cannot because of this error make a note of the challenge s name its url and the user id and name of the person who is the winner add that information here and an admin will help you award the challenge achievement and prize this is due to huboard order want to back this issue we accept bounties via
0
7,822
10,987,928,466
IssuesEvent
2019-12-02 10:11:58
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
Add new parameter types for documenting algorithms (file, matrix and layer)
Guidelines Processing Alg backport release_3.4 enhancement
## Description <!-- Cleaning the queue is a process done by project maintainers, mostly on a volunteer basis. We try to keep the overhead as small as possible and appreciate if you help us to do so by completing the following items. Include sentences with details describing the issue you have encountered (e.g., actual behavior, expected behavior, steps to reproduce). --> Goal: Add new parameter types for the documentation of algorithms. * ``file`` (for ``QgsProcessingParameterFileDestination`` - used for instance in the QGIS "Download file" algorithm) * ``matrix`` (for ``QgsProcessingParameterMatrix`` - used for instance in the QGIS "Reclassify by table" algorithm) Page URL: https://docs.qgis.org/testing/en/docs/documentation_guidelines/writing.html
1.0
Add new parameter types for documenting algorithms (file, matrix and layer) - ## Description <!-- Cleaning the queue is a process done by project maintainers, mostly on a volunteer basis. We try to keep the overhead as small as possible and appreciate if you help us to do so by completing the following items. Include sentences with details describing the issue you have encountered (e.g., actual behavior, expected behavior, steps to reproduce). --> Goal: Add new parameter types for the documentation of algorithms. * ``file`` (for ``QgsProcessingParameterFileDestination`` - used for instance in the QGIS "Download file" algorithm) * ``matrix`` (for ``QgsProcessingParameterMatrix`` - used for instance in the QGIS "Reclassify by table" algorithm) Page URL: https://docs.qgis.org/testing/en/docs/documentation_guidelines/writing.html
process
add new parameter types for documenting algorithms file matrix and layer description cleaning the queue is a process done by project maintainers mostly on a volunteer basis we try to keep the overhead as small as possible and appreciate if you help us to do so by completing the following items include sentences with details describing the issue you have encountered e g actual behavior expected behavior steps to reproduce goal add new parameter types for the documentation of algorithms file for qgsprocessingparameterfiledestination used for instance in the qgis download file algorithm matrix for qgsprocessingparametermatrix used for instance in the qgis reclassify by table algorithm page url
1
11,507
14,383,016,813
IssuesEvent
2020-12-02 08:30:16
carp-lang/Carp
https://api.github.com/repos/carp-lang/Carp
closed
Style guide
haskell process under discussion
I'm opening a separate issue (from PR #874) for the discussion of a style guide. Here are some existing style guides for Haskell, as references (thanks @hellerve) https://kowainik.github.io/posts/2019-02-06-style-guide https://wiki.haskell.org/Programming_guidelines They both look pretty good as a basis for ours! Please add more if you know of some great ones. I'm fine with just adopting the first one (or any similar one) wholesale, with just two clarifications: (1) Aligning things like imports and tokens like `=`, `->`, etc, should be avoided or left up to the programmer. Even though we can have this "for free" (through automation) I think it slows you down unnecessarily when writing the code and it sometimes looks really weird (if the code on the left hand side varies wildly in length). I think it makes sense to let this be up to the programmer in each instance (the first of the style guides above even say so.) If we want a hard and fast rule I'd err on not having it. (2) I'd like to stick to a "lispier" style when it comes to the use of `$`. This is personal preference, but I think it has some merit (and a lot of the code is already using this style.) The simple rule is that `$` should *only* be used to avoid parenthesis across multiple lines, like here: ```haskell foo = blah $ do a b c ``` which admittedly reads much nicer than ```haskell foo = blah (do a b c) ``` In all other instances I'd much prefer the lisp-inspired ```haskell foo x y z = f x (g y (h z)) ``` rather than ```haskell foo x y z = f x $ g y $ h z ``` and the like. This does *not* mean to always avoid useful operators like `<*>` when those make sense, it just makes grouping of code more clear, and easy to manipulate with structural editing tools (like Paredit).
1.0
Style guide - I'm opening a separate issue (from PR #874) for the discussion of a style guide. Here are some existing style guides for Haskell, as references (thanks @hellerve) https://kowainik.github.io/posts/2019-02-06-style-guide https://wiki.haskell.org/Programming_guidelines They both look pretty good as a basis for ours! Please add more if you know of some great ones. I'm fine with just adopting the first one (or any similar one) wholesale, with just two clarifications: (1) Aligning things like imports and tokens like `=`, `->`, etc, should be avoided or left up to the programmer. Even though we can have this "for free" (through automation) I think it slows you down unnecessarily when writing the code and it sometimes looks really weird (if the code on the left hand side varies wildly in length). I think it makes sense to let this be up to the programmer in each instance (the first of the style guides above even say so.) If we want a hard and fast rule I'd err on not having it. (2) I'd like to stick to a "lispier" style when it comes to the use of `$`. This is personal preference, but I think it has some merit (and a lot of the code is already using this style.) The simple rule is that `$` should *only* be used to avoid parenthesis across multiple lines, like here: ```haskell foo = blah $ do a b c ``` which admittedly reads much nicer than ```haskell foo = blah (do a b c) ``` In all other instances I'd much prefer the lisp-inspired ```haskell foo x y z = f x (g y (h z)) ``` rather than ```haskell foo x y z = f x $ g y $ h z ``` and the like. This does *not* mean to always avoid useful operators like `<*>` when those make sense, it just makes grouping of code more clear, and easy to manipulate with structural editing tools (like Paredit).
process
style guide i m opening a separate issue from pr for the discussion of a style guide here are some existing style guides for haskell as references thanks hellerve they both look pretty good as a basis for ours please add more if you know of some great ones i m fine with just adopting the first one or any similar one wholesale with just two clarifications aligning things like imports and tokens like etc should be avoided or left up to the programmer even though we can have this for free through automation i think it slows you down unnecessarily when writing the code and it sometimes looks really weird if the code on the left hand side varies wildly in length i think it makes sense to let this be up to the programmer in each instance the first of the style guides above even say so if we want a hard and fast rule i d err on not having it i d like to stick to a lispier style when it comes to the use of this is personal preference but i think it has some merit and a lot of the code is already using this style the simple rule is that should only be used to avoid parenthesis across multiple lines like here haskell foo blah do a b c which admittedly reads much nicer than haskell foo blah do a b c in all other instances i d much prefer the lisp inspired haskell foo x y z f x g y h z rather than haskell foo x y z f x g y h z and the like this does not mean to always avoid useful operators like when those make sense it just makes grouping of code more clear and easy to manipulate with structural editing tools like paredit
1
228,227
17,439,446,773
IssuesEvent
2021-08-05 01:17:53
amzn/selling-partner-api-docs
https://api.github.com/repos/amzn/selling-partner-api-docs
opened
sortOrder not working for purchaseOrders
bug documentation
endpoint: /vendor/orders/v1/purchaseOrders&sortOrder=DESC It does not seem to be sorting( ASC or DESC ) the purchase orders by creation date
1.0
sortOrder not working for purchaseOrders - endpoint: /vendor/orders/v1/purchaseOrders&sortOrder=DESC It does not seem to be sorting( ASC or DESC ) the purchase orders by creation date
non_process
sortorder not working for purchaseorders endpoint vendor orders purchaseorders sortorder desc it does not seem to be sorting asc or desc the purchase orders by creation date
0
21,806
30,316,364,218
IssuesEvent
2023-07-10 15:50:15
tdwg/dwc
https://api.github.com/repos/tdwg/dwc
closed
Change term - county
Term - change Class - Location non-normative Process - complete
Submitter: John Wieczorek (following issue raised by Ian Engelbrecht @ianengelbrecht Issue #221 and tdwg/dwc-qa#141) Justification (why is this change necessary?): Clarity Proponents (who needs this change): Everyone Current Term definition: https://dwc.tdwg.org/list/#dwc_county Proposed new attributes of the term: Usage comments (recommendations regarding content, etc.): Recommended best practice is to use a controlled vocabulary such as the Getty Thesaurus of Geographic Names. Recommended best practice is to leave this field blank if the Location spans multiple entities at this administrative level or if the Location might be in one or another of multiple possible entities at this level. Multiplicity and uncertainty of the geographic entity can be captured either in the term higherGeography or in the term locality, or both. Replaces (identifier of the existing term that would be deprecated and replaced by this term, if applicable): http://rs.tdwg.org/dwc/terms/version/county-2017-10-06
1.0
Change term - county - Submitter: John Wieczorek (following issue raised by Ian Engelbrecht @ianengelbrecht Issue #221 and tdwg/dwc-qa#141) Justification (why is this change necessary?): Clarity Proponents (who needs this change): Everyone Current Term definition: https://dwc.tdwg.org/list/#dwc_county Proposed new attributes of the term: Usage comments (recommendations regarding content, etc.): Recommended best practice is to use a controlled vocabulary such as the Getty Thesaurus of Geographic Names. Recommended best practice is to leave this field blank if the Location spans multiple entities at this administrative level or if the Location might be in one or another of multiple possible entities at this level. Multiplicity and uncertainty of the geographic entity can be captured either in the term higherGeography or in the term locality, or both. Replaces (identifier of the existing term that would be deprecated and replaced by this term, if applicable): http://rs.tdwg.org/dwc/terms/version/county-2017-10-06
process
change term county submitter john wieczorek following issue raised by ian engelbrecht ianengelbrecht issue and tdwg dwc qa justification why is this change necessary clarity proponents who needs this change everyone current term definition proposed new attributes of the term usage comments recommendations regarding content etc recommended best practice is to use a controlled vocabulary such as the getty thesaurus of geographic names recommended best practice is to leave this field blank if the location spans multiple entities at this administrative level or if the location might be in one or another of multiple possible entities at this level multiplicity and uncertainty of the geographic entity can be captured either in the term highergeography or in the term locality or both replaces identifier of the existing term that would be deprecated and replaced by this term if applicable
1
1,638
4,188,118,621
IssuesEvent
2016-06-23 19:40:02
ForestryMC/ForestryMC
https://api.github.com/repos/ForestryMC/ForestryMC
closed
[1.10] - Caught exception from forestry java.lang.NoSuchFieldError: field_180413_ao
Incompatibility
Using Forge 1981 for Minecraft 1.10 Forestry: *forestry_1.9.4-5.0.23.170* Crashlog: http://paste.ee/p/tddVJ And before anyone who doesn't have the correct information comments on this with "You're using 1.10 stupid, forestry is for 1.9.4". 1.9.4 mods in MOST cases will work on 1.10, this has been confirmed yesterday/today by a tweeted picture from Cpw aswell as internal testing within the modding community. For example of a working modpack: http://i.imgur.com/8rFNiOY.png
True
[1.10] - Caught exception from forestry java.lang.NoSuchFieldError: field_180413_ao - Using Forge 1981 for Minecraft 1.10 Forestry: *forestry_1.9.4-5.0.23.170* Crashlog: http://paste.ee/p/tddVJ And before anyone who doesn't have the correct information comments on this with "You're using 1.10 stupid, forestry is for 1.9.4". 1.9.4 mods in MOST cases will work on 1.10, this has been confirmed yesterday/today by a tweeted picture from Cpw aswell as internal testing within the modding community. For example of a working modpack: http://i.imgur.com/8rFNiOY.png
non_process
caught exception from forestry java lang nosuchfielderror field ao using forge for minecraft forestry forestry crashlog and before anyone who doesn t have the correct information comments on this with you re using stupid forestry is for mods in most cases will work on this has been confirmed yesterday today by a tweeted picture from cpw aswell as internal testing within the modding community for example of a working modpack
0
125,416
10,341,790,530
IssuesEvent
2019-09-04 03:45:22
rancher/rke
https://api.github.com/repos/rancher/rke
closed
selinux-enabled nodes have hyperkube components receive permission denied for files from service-sidekick
[zube]: To Test internal kind/bug team/ca
**RKE version:** 0.2.6 **Docker version: (`docker version`,`docker info` preferred)** 18.09.8 **Operating system and kernel: (`cat /etc/os-release`, `uname -r` preferred)** CentOS Linux release 7.6.1810 (Core) **Type/provider of hosts: (VirtualBox/Bare-metal/AWS/GCE/DO)** N/A, doesn't matter **cluster.yml file:** ``` nodes: - address: 172.16.134.48 user: centos role: [etcd, controlplane, worker] ``` **Steps to Reproduce:** Start CentOS 7 or RHEL 7 with ``` [root@ck-selinux-test _data]# cat /etc/docker/daemon.json { "selinux-enabled": true } [root@ck-selinux-test ~]# sestatus SELinux status: enabled SELinuxfs mount: /sys/fs/selinux SELinux root directory: /etc/selinux Loaded policy name: targeted Current mode: enforcing Mode from config file: enforcing Policy MLS status: enabled Policy deny_unknown status: allowed Max kernel policy version: 31 [root@ck-selinux-test ~]# ``` and try to run `rke up` from another machine that has the capability to ssh **Results:** ``` Endeavor:ck-selinux-test chriskim$ rke up INFO[0000] Initiating Kubernetes cluster INFO[0000] [dialer] Setup tunnel for host [172.16.134.48] INFO[0003] [state] Pulling image [rancher/rke-tools:v0.1.34] on host [172.16.134.48] INFO[0015] [state] Successfully pulled image [rancher/rke-tools:v0.1.34] on host [172.16.134.48] INFO[0016] [state] Successfully started [cluster-state-deployer] container on host [172.16.134.48] INFO[0017] [certificates] Generating CA kubernetes certificates INFO[0017] [certificates] Generating Kubernetes API server aggregation layer requestheader client CA certificates INFO[0017] [certificates] Generating Kubernetes API server proxy client certificates INFO[0017] [certificates] Generating Kubernetes API server certificates INFO[0017] [certificates] Generating Service account token key INFO[0017] [certificates] Generating Kube Controller certificates INFO[0017] [certificates] Generating Kube Scheduler certificates INFO[0018] [certificates] Generating Kube Proxy certificates INFO[0018] [certificates] Generating Node certificate INFO[0018] [certificates] Generating admin certificates and kubeconfig INFO[0018] [certificates] Generating etcd-172.16.134.48 certificate and key INFO[0018] Successfully Deployed state file at [./cluster.rkestate] INFO[0018] Building Kubernetes cluster INFO[0018] [dialer] Setup tunnel for host [172.16.134.48] INFO[0020] [network] Deploying port listener containers INFO[0022] [network] Successfully started [rke-etcd-port-listener] container on host [172.16.134.48] INFO[0023] [network] Successfully started [rke-cp-port-listener] container on host [172.16.134.48] INFO[0025] [network] Successfully started [rke-worker-port-listener] container on host [172.16.134.48] INFO[0025] [network] Port listener containers deployed successfully INFO[0025] [network] Running control plane -> etcd port checks INFO[0026] [network] Successfully started [rke-port-checker] container on host [172.16.134.48] INFO[0027] [network] Running control plane -> worker port checks INFO[0028] [network] Successfully started [rke-port-checker] container on host [172.16.134.48] INFO[0028] [network] Running workers -> control plane port checks INFO[0030] [network] Successfully started [rke-port-checker] container on host [172.16.134.48] INFO[0030] [network] Checking KubeAPI port Control Plane hosts INFO[0031] [network] Removing port listener containers INFO[0031] [remove/rke-etcd-port-listener] Successfully removed container on host [172.16.134.48] INFO[0032] [remove/rke-cp-port-listener] Successfully removed container on host [172.16.134.48] INFO[0033] [remove/rke-worker-port-listener] Successfully removed container on host [172.16.134.48] INFO[0033] [network] Port listener containers removed successfully INFO[0033] [certificates] Deploying kubernetes certificates to Cluster nodes INFO[0040] [reconcile] Rebuilding and updating local kube config INFO[0040] Successfully Deployed local admin kubeconfig at [./kube_config_cluster.yml] INFO[0040] [certificates] Successfully deployed kubernetes certificates to Cluster nodes INFO[0040] [reconcile] Reconciling cluster state INFO[0040] [reconcile] This is newly generated cluster INFO[0040] Pre-pulling kubernetes images INFO[0040] [pre-deploy] Pulling image [rancher/hyperkube:v1.14.3-rancher1] on host [172.16.134.48] INFO[0101] [pre-deploy] Successfully pulled image [rancher/hyperkube:v1.14.3-rancher1] on host [172.16.134.48] INFO[0101] Kubernetes images pulled successfully INFO[0101] [etcd] Building up etcd plane.. INFO[0101] [etcd] Pulling image [rancher/coreos-etcd:v3.3.10-rancher1] on host [172.16.134.48] INFO[0106] [etcd] Successfully pulled image [rancher/coreos-etcd:v3.3.10-rancher1] on host [172.16.134.48] INFO[0107] [etcd] Successfully started [etcd] container on host [172.16.134.48] INFO[0107] [etcd] Saving snapshot [etcd-rolling-snapshots] on host [172.16.134.48] INFO[0109] [etcd] Successfully started [etcd-rolling-snapshots] container on host [172.16.134.48] INFO[0116] [certificates] Successfully started [rke-bundle-cert] container on host [172.16.134.48] INFO[0116] Waiting for [rke-bundle-cert] container to exit on host [172.16.134.48] INFO[0116] [certificates] successfully saved certificate bundle [/opt/rke/etcd-snapshots//pki.bundle.tar.gz] on host [172.16.134.48] INFO[0118] [etcd] Successfully started [rke-log-linker] container on host [172.16.134.48] INFO[0119] [remove/rke-log-linker] Successfully removed container on host [172.16.134.48] INFO[0119] [etcd] Successfully started etcd plane.. Checking etcd cluster health INFO[0121] [controlplane] Building up Controller Plane.. INFO[0123] [controlplane] Successfully started [kube-apiserver] container on host [172.16.134.48] INFO[0123] [healthcheck] Start Healthcheck on service [kube-apiserver] on host [172.16.134.48] FATA[0193] [controlPlane] Failed to bring up Control Plane: [Failed to verify healthcheck: Failed to check https://localhost:6443/healthz for service [kube-apiserver] on host [172.16.134.48]: Get https://localhost:6443/healthz: Unable to access the service on localhost:6443. The service might be still starting up. Error: ssh: rejected: connect failed (Connection refused), log: /bin/bash: /opt/rke-tools/entrypoint.sh: Permission denied] Endeavor:ck-selinux-test chriskim$ ``` When you inspect the `service-sidekick` container you see that there is a shared volume in `/var/lib/docker/volumes` that is then shared into the various kubernetes component containers. On my system for example, it is at `/var/lib/docker/volumes/46f993a6a0adee06a8ea07397ddd8bcf963d4f0213022410a513c371b7eb1690/_data`. Changing into this directory and then running an `ls -Z` shows the following contexts: ``` [root@ck-selinux-test _data]# ls -Z drwxr-xr-x. root root system_u:object_r:container_file_t:s0 bin -rwxr-xr-x. root root system_u:object_r:container_file_t:s0:c311,c924 cloud-provider.sh -rwxr-xr-x. root root system_u:object_r:container_file_t:s0:c311,c924 entrypoint.sh -rwxr-xr-x. root root system_u:object_r:container_file_t:s0:c311,c924 rke-etcd-backup -rwxr-xr-x. root root system_u:object_r:container_file_t:s0:c311,c924 weave-plugins-cni.sh [root@ck-selinux-test _data]# ``` Notably, the context is `container_file_t` and the categories on the files are set to `c311,c924` in this case. Running a `restorecon -nrv` on this directory shows that no files are to be changed, even though ``` [root@ck-selinux-test _data]# semanage fcontext -l | grep "var/lib/docker" /var/lib/docker(/.*)? all files system_u:object_r:container_var_lib_t:s0 /var/lib/docker/.*/config\.env all files system_u:object_r:container_share_t:s0 /var/lib/docker/vfs(/.*)? all files system_u:object_r:container_file_t:s0 /var/lib/docker/init(/.*)? all files system_u:object_r:container_share_t:s0 /var/lib/docker-latest(/.*)? all files system_u:object_r:container_var_lib_t:s0 /var/lib/docker/overlay(/.*)? all files system_u:object_r:container_share_t:s0 /var/lib/docker-latest/.*/config\.env all files system_u:object_r:container_share_t:s0 /var/lib/docker/overlay2(/.*)? all files system_u:object_r:container_share_t:s0 /var/lib/docker-latest/init(/.*)? all files system_u:object_r:container_share_t:s0 /var/lib/docker/containers/.*/hosts all files system_u:object_r:container_share_t:s0 /var/lib/docker/containers/.*/.*\.log all files system_u:object_r:container_log_t:s0 /var/lib/docker/containers/.*/hostname all files system_u:object_r:container_share_t:s0 /var/lib/docker-latest/overlay(/.*)? all files system_u:object_r:container_share_t:s0 /var/lib/docker-latest/overlay2(/.*)? all files system_u:object_r:container_share_t:s0 /var/lib/docker-latest/containers/.*/hosts all files system_u:object_r:container_share_t:s0 /var/lib/docker-latest/containers/.*/.*\.log all files system_u:object_r:container_log_t:s0 /var/lib/docker-latest/containers/.*/hostname all files system_u:object_r:container_share_t:s0 ``` shows that the files in `/var/lib/docker/volumes` should end up being `container_var_lib_t`, I'm assuming this is because of Docker and the way it's creating the volume. Looking in `audit.log` we can see the permission denied is due to ``` type=SYSCALL msg=audit(1564494369.647:636): arch=c000003e syscall=2 success=no exit=-13 a0=102bec8 a1=0 a2=68732e746e696f a3=2 items=0 ppid=9468 pid=9485 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="entrypoint.sh" exe="/bin/bash" subj=system_u:system_r:container_t:s0:c443,c824 key=(null) type=AVC msg=audit(1564494430.270:637): avc: denied { read } for pid=9545 comm="entrypoint.sh" name="entrypoint.sh" dev="dm-0" ino=67645448 scontext=system_u:system_r:container_t:s0:c443,c824 tcontext=system_u:object_r:container_file_t:s0:c311,c924 tclass=file permissive=0 type=SYSCALL msg=audit(1564494430.270:637): arch=c000003e syscall=2 success=no exit=-13 a0=a3bec8 a1=0 a2=68732e746e696f a3=2 items=0 ppid=9528 pid=9545 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="entrypoint.sh" exe="/bin/bash" subj=system_u:system_r:container_t:s0:c443,c824 key=(null) type=AVC msg=audit(1564494490.923:638): avc: denied { read } for pid=9607 comm="entrypoint.sh" name="entrypoint.sh" dev="dm-0" ino=67645448 scontext=system_u:system_r:container_t:s0:c443,c824 tcontext=system_u:object_r:container_file_t:s0:c311,c924 tclass=file permissive=0 type=SYSCALL msg=audit(1564494490.923:638): arch=c000003e syscall=2 success=no exit=-13 a0=1320ec8 a1=0 a2=68732e746e696f a3=2 items=0 ppid=9590 pid=9607 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="entrypoint.sh" exe="/bin/bash" subj=system_u:system_r:container_t:s0:c443,c824 key=(null) type=AVC msg=audit(1564494551.607:641): avc: denied { read } for pid=9670 comm="entrypoint.sh" name="entrypoint.sh" dev="dm-0" ino=67645448 scontext=system_u:system_r:container_t:s0:c443,c824 tcontext=system_u:object_r:container_file_t:s0:c311,c924 tclass=file permissive=0 type=SYSCALL msg=audit(1564494551.607:641): arch=c000003e syscall=2 success=no exit=-13 a0=24eaec8 a1=0 a2=68732e746e696f a3=2 items=0 ppid=9654 pid=9670 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="entrypoint.sh" exe="/bin/bash" subj=system_u:system_r:container_t:s0:c443,c824 key=(null) [root@ck-selinux-test audit]# ``` Running ``` chcat -d bin/docker chcat -d cloud-provider.sh chcat -d entrypoint.sh chcat -d rke-etcd-backup chcat -d weave-plugins-cni.sh ``` from the _data directory of the shared volume from service-sidekick and removing the categories from the files in the shared volume will allow the kubernetes components to properly start, and re-running `rke up` will end up having the cluster come up. I have not fully debugged this yet, but am assuming that the categories on that volume are coming from Docker and we aren't specifying `:z` for the `--volumes-from` equivalent option. Still need to investigate this.
1.0
selinux-enabled nodes have hyperkube components receive permission denied for files from service-sidekick - **RKE version:** 0.2.6 **Docker version: (`docker version`,`docker info` preferred)** 18.09.8 **Operating system and kernel: (`cat /etc/os-release`, `uname -r` preferred)** CentOS Linux release 7.6.1810 (Core) **Type/provider of hosts: (VirtualBox/Bare-metal/AWS/GCE/DO)** N/A, doesn't matter **cluster.yml file:** ``` nodes: - address: 172.16.134.48 user: centos role: [etcd, controlplane, worker] ``` **Steps to Reproduce:** Start CentOS 7 or RHEL 7 with ``` [root@ck-selinux-test _data]# cat /etc/docker/daemon.json { "selinux-enabled": true } [root@ck-selinux-test ~]# sestatus SELinux status: enabled SELinuxfs mount: /sys/fs/selinux SELinux root directory: /etc/selinux Loaded policy name: targeted Current mode: enforcing Mode from config file: enforcing Policy MLS status: enabled Policy deny_unknown status: allowed Max kernel policy version: 31 [root@ck-selinux-test ~]# ``` and try to run `rke up` from another machine that has the capability to ssh **Results:** ``` Endeavor:ck-selinux-test chriskim$ rke up INFO[0000] Initiating Kubernetes cluster INFO[0000] [dialer] Setup tunnel for host [172.16.134.48] INFO[0003] [state] Pulling image [rancher/rke-tools:v0.1.34] on host [172.16.134.48] INFO[0015] [state] Successfully pulled image [rancher/rke-tools:v0.1.34] on host [172.16.134.48] INFO[0016] [state] Successfully started [cluster-state-deployer] container on host [172.16.134.48] INFO[0017] [certificates] Generating CA kubernetes certificates INFO[0017] [certificates] Generating Kubernetes API server aggregation layer requestheader client CA certificates INFO[0017] [certificates] Generating Kubernetes API server proxy client certificates INFO[0017] [certificates] Generating Kubernetes API server certificates INFO[0017] [certificates] Generating Service account token key INFO[0017] [certificates] Generating Kube Controller certificates INFO[0017] [certificates] Generating Kube Scheduler certificates INFO[0018] [certificates] Generating Kube Proxy certificates INFO[0018] [certificates] Generating Node certificate INFO[0018] [certificates] Generating admin certificates and kubeconfig INFO[0018] [certificates] Generating etcd-172.16.134.48 certificate and key INFO[0018] Successfully Deployed state file at [./cluster.rkestate] INFO[0018] Building Kubernetes cluster INFO[0018] [dialer] Setup tunnel for host [172.16.134.48] INFO[0020] [network] Deploying port listener containers INFO[0022] [network] Successfully started [rke-etcd-port-listener] container on host [172.16.134.48] INFO[0023] [network] Successfully started [rke-cp-port-listener] container on host [172.16.134.48] INFO[0025] [network] Successfully started [rke-worker-port-listener] container on host [172.16.134.48] INFO[0025] [network] Port listener containers deployed successfully INFO[0025] [network] Running control plane -> etcd port checks INFO[0026] [network] Successfully started [rke-port-checker] container on host [172.16.134.48] INFO[0027] [network] Running control plane -> worker port checks INFO[0028] [network] Successfully started [rke-port-checker] container on host [172.16.134.48] INFO[0028] [network] Running workers -> control plane port checks INFO[0030] [network] Successfully started [rke-port-checker] container on host [172.16.134.48] INFO[0030] [network] Checking KubeAPI port Control Plane hosts INFO[0031] [network] Removing port listener containers INFO[0031] [remove/rke-etcd-port-listener] Successfully removed container on host [172.16.134.48] INFO[0032] [remove/rke-cp-port-listener] Successfully removed container on host [172.16.134.48] INFO[0033] [remove/rke-worker-port-listener] Successfully removed container on host [172.16.134.48] INFO[0033] [network] Port listener containers removed successfully INFO[0033] [certificates] Deploying kubernetes certificates to Cluster nodes INFO[0040] [reconcile] Rebuilding and updating local kube config INFO[0040] Successfully Deployed local admin kubeconfig at [./kube_config_cluster.yml] INFO[0040] [certificates] Successfully deployed kubernetes certificates to Cluster nodes INFO[0040] [reconcile] Reconciling cluster state INFO[0040] [reconcile] This is newly generated cluster INFO[0040] Pre-pulling kubernetes images INFO[0040] [pre-deploy] Pulling image [rancher/hyperkube:v1.14.3-rancher1] on host [172.16.134.48] INFO[0101] [pre-deploy] Successfully pulled image [rancher/hyperkube:v1.14.3-rancher1] on host [172.16.134.48] INFO[0101] Kubernetes images pulled successfully INFO[0101] [etcd] Building up etcd plane.. INFO[0101] [etcd] Pulling image [rancher/coreos-etcd:v3.3.10-rancher1] on host [172.16.134.48] INFO[0106] [etcd] Successfully pulled image [rancher/coreos-etcd:v3.3.10-rancher1] on host [172.16.134.48] INFO[0107] [etcd] Successfully started [etcd] container on host [172.16.134.48] INFO[0107] [etcd] Saving snapshot [etcd-rolling-snapshots] on host [172.16.134.48] INFO[0109] [etcd] Successfully started [etcd-rolling-snapshots] container on host [172.16.134.48] INFO[0116] [certificates] Successfully started [rke-bundle-cert] container on host [172.16.134.48] INFO[0116] Waiting for [rke-bundle-cert] container to exit on host [172.16.134.48] INFO[0116] [certificates] successfully saved certificate bundle [/opt/rke/etcd-snapshots//pki.bundle.tar.gz] on host [172.16.134.48] INFO[0118] [etcd] Successfully started [rke-log-linker] container on host [172.16.134.48] INFO[0119] [remove/rke-log-linker] Successfully removed container on host [172.16.134.48] INFO[0119] [etcd] Successfully started etcd plane.. Checking etcd cluster health INFO[0121] [controlplane] Building up Controller Plane.. INFO[0123] [controlplane] Successfully started [kube-apiserver] container on host [172.16.134.48] INFO[0123] [healthcheck] Start Healthcheck on service [kube-apiserver] on host [172.16.134.48] FATA[0193] [controlPlane] Failed to bring up Control Plane: [Failed to verify healthcheck: Failed to check https://localhost:6443/healthz for service [kube-apiserver] on host [172.16.134.48]: Get https://localhost:6443/healthz: Unable to access the service on localhost:6443. The service might be still starting up. Error: ssh: rejected: connect failed (Connection refused), log: /bin/bash: /opt/rke-tools/entrypoint.sh: Permission denied] Endeavor:ck-selinux-test chriskim$ ``` When you inspect the `service-sidekick` container you see that there is a shared volume in `/var/lib/docker/volumes` that is then shared into the various kubernetes component containers. On my system for example, it is at `/var/lib/docker/volumes/46f993a6a0adee06a8ea07397ddd8bcf963d4f0213022410a513c371b7eb1690/_data`. Changing into this directory and then running an `ls -Z` shows the following contexts: ``` [root@ck-selinux-test _data]# ls -Z drwxr-xr-x. root root system_u:object_r:container_file_t:s0 bin -rwxr-xr-x. root root system_u:object_r:container_file_t:s0:c311,c924 cloud-provider.sh -rwxr-xr-x. root root system_u:object_r:container_file_t:s0:c311,c924 entrypoint.sh -rwxr-xr-x. root root system_u:object_r:container_file_t:s0:c311,c924 rke-etcd-backup -rwxr-xr-x. root root system_u:object_r:container_file_t:s0:c311,c924 weave-plugins-cni.sh [root@ck-selinux-test _data]# ``` Notably, the context is `container_file_t` and the categories on the files are set to `c311,c924` in this case. Running a `restorecon -nrv` on this directory shows that no files are to be changed, even though ``` [root@ck-selinux-test _data]# semanage fcontext -l | grep "var/lib/docker" /var/lib/docker(/.*)? all files system_u:object_r:container_var_lib_t:s0 /var/lib/docker/.*/config\.env all files system_u:object_r:container_share_t:s0 /var/lib/docker/vfs(/.*)? all files system_u:object_r:container_file_t:s0 /var/lib/docker/init(/.*)? all files system_u:object_r:container_share_t:s0 /var/lib/docker-latest(/.*)? all files system_u:object_r:container_var_lib_t:s0 /var/lib/docker/overlay(/.*)? all files system_u:object_r:container_share_t:s0 /var/lib/docker-latest/.*/config\.env all files system_u:object_r:container_share_t:s0 /var/lib/docker/overlay2(/.*)? all files system_u:object_r:container_share_t:s0 /var/lib/docker-latest/init(/.*)? all files system_u:object_r:container_share_t:s0 /var/lib/docker/containers/.*/hosts all files system_u:object_r:container_share_t:s0 /var/lib/docker/containers/.*/.*\.log all files system_u:object_r:container_log_t:s0 /var/lib/docker/containers/.*/hostname all files system_u:object_r:container_share_t:s0 /var/lib/docker-latest/overlay(/.*)? all files system_u:object_r:container_share_t:s0 /var/lib/docker-latest/overlay2(/.*)? all files system_u:object_r:container_share_t:s0 /var/lib/docker-latest/containers/.*/hosts all files system_u:object_r:container_share_t:s0 /var/lib/docker-latest/containers/.*/.*\.log all files system_u:object_r:container_log_t:s0 /var/lib/docker-latest/containers/.*/hostname all files system_u:object_r:container_share_t:s0 ``` shows that the files in `/var/lib/docker/volumes` should end up being `container_var_lib_t`, I'm assuming this is because of Docker and the way it's creating the volume. Looking in `audit.log` we can see the permission denied is due to ``` type=SYSCALL msg=audit(1564494369.647:636): arch=c000003e syscall=2 success=no exit=-13 a0=102bec8 a1=0 a2=68732e746e696f a3=2 items=0 ppid=9468 pid=9485 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="entrypoint.sh" exe="/bin/bash" subj=system_u:system_r:container_t:s0:c443,c824 key=(null) type=AVC msg=audit(1564494430.270:637): avc: denied { read } for pid=9545 comm="entrypoint.sh" name="entrypoint.sh" dev="dm-0" ino=67645448 scontext=system_u:system_r:container_t:s0:c443,c824 tcontext=system_u:object_r:container_file_t:s0:c311,c924 tclass=file permissive=0 type=SYSCALL msg=audit(1564494430.270:637): arch=c000003e syscall=2 success=no exit=-13 a0=a3bec8 a1=0 a2=68732e746e696f a3=2 items=0 ppid=9528 pid=9545 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="entrypoint.sh" exe="/bin/bash" subj=system_u:system_r:container_t:s0:c443,c824 key=(null) type=AVC msg=audit(1564494490.923:638): avc: denied { read } for pid=9607 comm="entrypoint.sh" name="entrypoint.sh" dev="dm-0" ino=67645448 scontext=system_u:system_r:container_t:s0:c443,c824 tcontext=system_u:object_r:container_file_t:s0:c311,c924 tclass=file permissive=0 type=SYSCALL msg=audit(1564494490.923:638): arch=c000003e syscall=2 success=no exit=-13 a0=1320ec8 a1=0 a2=68732e746e696f a3=2 items=0 ppid=9590 pid=9607 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="entrypoint.sh" exe="/bin/bash" subj=system_u:system_r:container_t:s0:c443,c824 key=(null) type=AVC msg=audit(1564494551.607:641): avc: denied { read } for pid=9670 comm="entrypoint.sh" name="entrypoint.sh" dev="dm-0" ino=67645448 scontext=system_u:system_r:container_t:s0:c443,c824 tcontext=system_u:object_r:container_file_t:s0:c311,c924 tclass=file permissive=0 type=SYSCALL msg=audit(1564494551.607:641): arch=c000003e syscall=2 success=no exit=-13 a0=24eaec8 a1=0 a2=68732e746e696f a3=2 items=0 ppid=9654 pid=9670 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="entrypoint.sh" exe="/bin/bash" subj=system_u:system_r:container_t:s0:c443,c824 key=(null) [root@ck-selinux-test audit]# ``` Running ``` chcat -d bin/docker chcat -d cloud-provider.sh chcat -d entrypoint.sh chcat -d rke-etcd-backup chcat -d weave-plugins-cni.sh ``` from the _data directory of the shared volume from service-sidekick and removing the categories from the files in the shared volume will allow the kubernetes components to properly start, and re-running `rke up` will end up having the cluster come up. I have not fully debugged this yet, but am assuming that the categories on that volume are coming from Docker and we aren't specifying `:z` for the `--volumes-from` equivalent option. Still need to investigate this.
non_process
selinux enabled nodes have hyperkube components receive permission denied for files from service sidekick rke version docker version docker version docker info preferred operating system and kernel cat etc os release uname r preferred centos linux release core type provider of hosts virtualbox bare metal aws gce do n a doesn t matter cluster yml file nodes address user centos role steps to reproduce start centos or rhel with cat etc docker daemon json selinux enabled true sestatus selinux status enabled selinuxfs mount sys fs selinux selinux root directory etc selinux loaded policy name targeted current mode enforcing mode from config file enforcing policy mls status enabled policy deny unknown status allowed max kernel policy version and try to run rke up from another machine that has the capability to ssh results endeavor ck selinux test chriskim rke up info initiating kubernetes cluster info setup tunnel for host info pulling image on host info successfully pulled image on host info successfully started container on host info generating ca kubernetes certificates info generating kubernetes api server aggregation layer requestheader client ca certificates info generating kubernetes api server proxy client certificates info generating kubernetes api server certificates info generating service account token key info generating kube controller certificates info generating kube scheduler certificates info generating kube proxy certificates info generating node certificate info generating admin certificates and kubeconfig info generating etcd certificate and key info successfully deployed state file at info building kubernetes cluster info setup tunnel for host info deploying port listener containers info successfully started container on host info successfully started container on host info successfully started container on host info port listener containers deployed successfully info running control plane etcd port checks info successfully started container on host info running control plane worker port checks info successfully started container on host info running workers control plane port checks info successfully started container on host info checking kubeapi port control plane hosts info removing port listener containers info successfully removed container on host info successfully removed container on host info successfully removed container on host info port listener containers removed successfully info deploying kubernetes certificates to cluster nodes info rebuilding and updating local kube config info successfully deployed local admin kubeconfig at info successfully deployed kubernetes certificates to cluster nodes info reconciling cluster state info this is newly generated cluster info pre pulling kubernetes images info pulling image on host info successfully pulled image on host info kubernetes images pulled successfully info building up etcd plane info pulling image on host info successfully pulled image on host info successfully started container on host info saving snapshot on host info successfully started container on host info successfully started container on host info waiting for container to exit on host info successfully saved certificate bundle on host info successfully started container on host info successfully removed container on host info successfully started etcd plane checking etcd cluster health info building up controller plane info successfully started container on host info start healthcheck on service on host fata failed to bring up control plane on host get unable to access the service on localhost the service might be still starting up error ssh rejected connect failed connection refused log bin bash opt rke tools entrypoint sh permission denied endeavor ck selinux test chriskim when you inspect the service sidekick container you see that there is a shared volume in var lib docker volumes that is then shared into the various kubernetes component containers on my system for example it is at var lib docker volumes data changing into this directory and then running an ls z shows the following contexts ls z drwxr xr x root root system u object r container file t bin rwxr xr x root root system u object r container file t cloud provider sh rwxr xr x root root system u object r container file t entrypoint sh rwxr xr x root root system u object r container file t rke etcd backup rwxr xr x root root system u object r container file t weave plugins cni sh notably the context is container file t and the categories on the files are set to in this case running a restorecon nrv on this directory shows that no files are to be changed even though semanage fcontext l grep var lib docker var lib docker all files system u object r container var lib t var lib docker config env all files system u object r container share t var lib docker vfs all files system u object r container file t var lib docker init all files system u object r container share t var lib docker latest all files system u object r container var lib t var lib docker overlay all files system u object r container share t var lib docker latest config env all files system u object r container share t var lib docker all files system u object r container share t var lib docker latest init all files system u object r container share t var lib docker containers hosts all files system u object r container share t var lib docker containers log all files system u object r container log t var lib docker containers hostname all files system u object r container share t var lib docker latest overlay all files system u object r container share t var lib docker latest all files system u object r container share t var lib docker latest containers hosts all files system u object r container share t var lib docker latest containers log all files system u object r container log t var lib docker latest containers hostname all files system u object r container share t shows that the files in var lib docker volumes should end up being container var lib t i m assuming this is because of docker and the way it s creating the volume looking in audit log we can see the permission denied is due to type syscall msg audit arch syscall success no exit items ppid pid auid uid gid euid suid fsuid egid sgid fsgid tty none ses comm entrypoint sh exe bin bash subj system u system r container t key null type avc msg audit avc denied read for pid comm entrypoint sh name entrypoint sh dev dm ino scontext system u system r container t tcontext system u object r container file t tclass file permissive type syscall msg audit arch syscall success no exit items ppid pid auid uid gid euid suid fsuid egid sgid fsgid tty none ses comm entrypoint sh exe bin bash subj system u system r container t key null type avc msg audit avc denied read for pid comm entrypoint sh name entrypoint sh dev dm ino scontext system u system r container t tcontext system u object r container file t tclass file permissive type syscall msg audit arch syscall success no exit items ppid pid auid uid gid euid suid fsuid egid sgid fsgid tty none ses comm entrypoint sh exe bin bash subj system u system r container t key null type avc msg audit avc denied read for pid comm entrypoint sh name entrypoint sh dev dm ino scontext system u system r container t tcontext system u object r container file t tclass file permissive type syscall msg audit arch syscall success no exit items ppid pid auid uid gid euid suid fsuid egid sgid fsgid tty none ses comm entrypoint sh exe bin bash subj system u system r container t key null running chcat d bin docker chcat d cloud provider sh chcat d entrypoint sh chcat d rke etcd backup chcat d weave plugins cni sh from the data directory of the shared volume from service sidekick and removing the categories from the files in the shared volume will allow the kubernetes components to properly start and re running rke up will end up having the cluster come up i have not fully debugged this yet but am assuming that the categories on that volume are coming from docker and we aren t specifying z for the volumes from equivalent option still need to investigate this
0
12,648
15,022,075,359
IssuesEvent
2021-02-01 16:31:44
pystatgen/sgkit
https://api.github.com/repos/pystatgen/sgkit
closed
Add upstream CI build
process + tools
Dask has an "upstream" CI build, which is an interesting take on this problem, it essentially runs dask CI with main/master versions of some of the core dependencies. They run this on schedule as well as on master changes. This could be valuable for us to detect breaking changes before releases. See the upstream CI definition [here](https://github.com/dask/dask/blob/master/.github/workflows/ci-upstream.yml). Edit: but I'm not sure if it leads to "actions", see the state of that [build](https://github.com/dask/dask/actions?query=workflow%3AUpstream). _Originally posted by @ravwojdyla in https://github.com/pystatgen/sgkit/issues/422#issuecomment-757365235_
1.0
Add upstream CI build - Dask has an "upstream" CI build, which is an interesting take on this problem, it essentially runs dask CI with main/master versions of some of the core dependencies. They run this on schedule as well as on master changes. This could be valuable for us to detect breaking changes before releases. See the upstream CI definition [here](https://github.com/dask/dask/blob/master/.github/workflows/ci-upstream.yml). Edit: but I'm not sure if it leads to "actions", see the state of that [build](https://github.com/dask/dask/actions?query=workflow%3AUpstream). _Originally posted by @ravwojdyla in https://github.com/pystatgen/sgkit/issues/422#issuecomment-757365235_
process
add upstream ci build dask has an upstream ci build which is an interesting take on this problem it essentially runs dask ci with main master versions of some of the core dependencies they run this on schedule as well as on master changes this could be valuable for us to detect breaking changes before releases see the upstream ci definition edit but i m not sure if it leads to actions see the state of that originally posted by ravwojdyla in
1
450,141
31,884,153,233
IssuesEvent
2023-09-16 18:41:50
Proyectos-ISPC/proyecto-app-mobile
https://api.github.com/repos/Proyectos-ISPC/proyecto-app-mobile
closed
US12 Plantear Historias de Usuarios (derivadas de los requerimientos) y Tareas dependientes de las US para incorporarlas en el repositorio remoto gitHub. (Issues y Milestones) - Tener en cuenta la redacción adecuada para las US y nomenclatura, ej “#US01 Como usuario quiero ingresar al carrito para poder comprar”
documentation 1ST
- [x] #TK00 Definir tareas dentro de la Historia de Usuario. - [x] #TK1 Una vez realizada la tarea #11 pasar los requerimientos a historias de usuario en el Project en Github.
1.0
US12 Plantear Historias de Usuarios (derivadas de los requerimientos) y Tareas dependientes de las US para incorporarlas en el repositorio remoto gitHub. (Issues y Milestones) - Tener en cuenta la redacción adecuada para las US y nomenclatura, ej “#US01 Como usuario quiero ingresar al carrito para poder comprar” - - [x] #TK00 Definir tareas dentro de la Historia de Usuario. - [x] #TK1 Una vez realizada la tarea #11 pasar los requerimientos a historias de usuario en el Project en Github.
non_process
plantear historias de usuarios derivadas de los requerimientos y tareas dependientes de las us para incorporarlas en el repositorio remoto github issues y milestones tener en cuenta la redacción adecuada para las us y nomenclatura ej “ como usuario quiero ingresar al carrito para poder comprar” definir tareas dentro de la historia de usuario una vez realizada la tarea pasar los requerimientos a historias de usuario en el project en github
0
51,636
3,013,525,936
IssuesEvent
2015-07-29 09:31:44
yawlfoundation/yawl
https://api.github.com/repos/yawlfoundation/yawl
closed
Introduce optional task input parameters
auto-migrated Priority-Medium Type-Enhancement
``` I'm trying to develop a custom yawl service which should have optional parameters. E.g. my param A has the type xs:NCName and should be optional. If I don't set a value in the mapping, yawl complains. If I type an empty string (actually 2 white-spaces because yawl-editor doesn't let me do an empty-string mapping) then I get validation errors because an empty string is not a valid value for type NCName, of course. The only alternative at the moment would be to not include the optional parameters in the describeRequiredParams() method and give them a variable name not starting with "Yawl" because it seems that users can't create task variables starting with "Yawl", probably protected space. I don't like this workaround and think it needs a better solution. ``` Original issue reported on code.google.com by `neothema...@googlemail.com` on 30 Aug 2010 at 12:33
1.0
Introduce optional task input parameters - ``` I'm trying to develop a custom yawl service which should have optional parameters. E.g. my param A has the type xs:NCName and should be optional. If I don't set a value in the mapping, yawl complains. If I type an empty string (actually 2 white-spaces because yawl-editor doesn't let me do an empty-string mapping) then I get validation errors because an empty string is not a valid value for type NCName, of course. The only alternative at the moment would be to not include the optional parameters in the describeRequiredParams() method and give them a variable name not starting with "Yawl" because it seems that users can't create task variables starting with "Yawl", probably protected space. I don't like this workaround and think it needs a better solution. ``` Original issue reported on code.google.com by `neothema...@googlemail.com` on 30 Aug 2010 at 12:33
non_process
introduce optional task input parameters i m trying to develop a custom yawl service which should have optional parameters e g my param a has the type xs ncname and should be optional if i don t set a value in the mapping yawl complains if i type an empty string actually white spaces because yawl editor doesn t let me do an empty string mapping then i get validation errors because an empty string is not a valid value for type ncname of course the only alternative at the moment would be to not include the optional parameters in the describerequiredparams method and give them a variable name not starting with yawl because it seems that users can t create task variables starting with yawl probably protected space i don t like this workaround and think it needs a better solution original issue reported on code google com by neothema googlemail com on aug at
0
515,465
14,963,369,253
IssuesEvent
2021-01-27 10:29:01
celo-org/celo-monorepo
https://api.github.com/repos/celo-org/celo-monorepo
closed
[Suggestion]Add sub-section which shows the payment Request history for each user.
Component: Feed & Notifications Priority: P2 enhancement qa wallet
**Frequency:** 100% **App version:** Staging build v1.5.2 **Repro on:** Samsung Galaxy Note 5 (7.0), Samsung Galaxy s5 (6.0.1) **Pre-condition**: Make sure user has proceeded for Payment request. **Repro Steps:** 1) Launch the app from device A. 2) Launch the app from device B. 3) User A sends the request to user B from Device A. 4) Make sure the Payment you’ve requested section is shown. 5) Tap on the Review button. 6) Tap on the “Cancel” button. 7) Observed Requested payment is removed for User A and User B as well. **Impact:** The user will not be able to view the transaction requests history made by him/her if the request is canceled or declined. **Current Behavior:** Requested payment is removed for User A and User B as well **Expected Behavior:** Need to add Add sub-section which shows the payment Request history for each user **Attachment:** [Android_RequestSuggesion.mp4 ](https://drive.google.com/file/d/11cFwAruLhjzsg3uL4uehoTWDejPIrC7t/view?usp=sharing)
1.0
[Suggestion]Add sub-section which shows the payment Request history for each user. - **Frequency:** 100% **App version:** Staging build v1.5.2 **Repro on:** Samsung Galaxy Note 5 (7.0), Samsung Galaxy s5 (6.0.1) **Pre-condition**: Make sure user has proceeded for Payment request. **Repro Steps:** 1) Launch the app from device A. 2) Launch the app from device B. 3) User A sends the request to user B from Device A. 4) Make sure the Payment you’ve requested section is shown. 5) Tap on the Review button. 6) Tap on the “Cancel” button. 7) Observed Requested payment is removed for User A and User B as well. **Impact:** The user will not be able to view the transaction requests history made by him/her if the request is canceled or declined. **Current Behavior:** Requested payment is removed for User A and User B as well **Expected Behavior:** Need to add Add sub-section which shows the payment Request history for each user **Attachment:** [Android_RequestSuggesion.mp4 ](https://drive.google.com/file/d/11cFwAruLhjzsg3uL4uehoTWDejPIrC7t/view?usp=sharing)
non_process
add sub section which shows the payment request history for each user frequency app version staging build repro on samsung galaxy note samsung galaxy pre condition make sure user has proceeded for payment request repro steps launch the app from device a launch the app from device b user a sends the request to user b from device a make sure the payment you’ve requested section is shown tap on the review button tap on the “cancel” button observed requested payment is removed for user a and user b as well impact the user will not be able to view the transaction requests history made by him her if the request is canceled or declined current behavior requested payment is removed for user a and user b as well expected behavior need to add add sub section which shows the payment request history for each user attachment
0
8,052
11,220,889,776
IssuesEvent
2020-01-07 16:40:27
openenclave/openenclave
https://api.github.com/repos/openenclave/openenclave
closed
Release process is unclear
process triaged
Releasing.md says: > We want to release `v0.7.0` in about a week, so we choose a release manager from the committers. Who is the "we"? Is it the Community Maintenance Committee? How is the choice made? Why is "release" not one of the areas of responsibility on the Maintainers.md page?
1.0
Release process is unclear - Releasing.md says: > We want to release `v0.7.0` in about a week, so we choose a release manager from the committers. Who is the "we"? Is it the Community Maintenance Committee? How is the choice made? Why is "release" not one of the areas of responsibility on the Maintainers.md page?
process
release process is unclear releasing md says we want to release in about a week so we choose a release manager from the committers who is the we is it the community maintenance committee how is the choice made why is release not one of the areas of responsibility on the maintainers md page
1
18,726
24,613,072,730
IssuesEvent
2022-10-15 01:31:57
GehanYangFS/CSCI599
https://api.github.com/repos/GehanYangFS/CSCI599
closed
CSCI-599 Milestone-1: Basic Infra
process
Write a script to generate a `settings.json` file Write a script to handle PointCloud Write a script as primary control of multiple multirotor drones
1.0
CSCI-599 Milestone-1: Basic Infra - Write a script to generate a `settings.json` file Write a script to handle PointCloud Write a script as primary control of multiple multirotor drones
process
csci milestone basic infra write a script to generate a settings json file write a script to handle pointcloud write a script as primary control of multiple multirotor drones
1
20,394
27,051,806,014
IssuesEvent
2023-02-13 13:46:02
brucemiller/LaTeXML
https://api.github.com/repos/brucemiller/LaTeXML
opened
[MathML] remove reliance on mathvariant
enhancement postprocessing
With MathML Core [issue#182](https://github.com/w3c/mathml-core/issues/182), we no longer have mathvariant available, and newly prefer a combination of Unicode and CSS for this kind of styling. An example of the mixed state of affairs is available in [ar5iv#184](https://github.com/dginev/ar5iv/issues/184). I suspect we should add a selection of `math_variant` (or similarly named) CSS classes for the cases where Unicode does not have available codepoints.
1.0
[MathML] remove reliance on mathvariant - With MathML Core [issue#182](https://github.com/w3c/mathml-core/issues/182), we no longer have mathvariant available, and newly prefer a combination of Unicode and CSS for this kind of styling. An example of the mixed state of affairs is available in [ar5iv#184](https://github.com/dginev/ar5iv/issues/184). I suspect we should add a selection of `math_variant` (or similarly named) CSS classes for the cases where Unicode does not have available codepoints.
process
remove reliance on mathvariant with mathml core we no longer have mathvariant available and newly prefer a combination of unicode and css for this kind of styling an example of the mixed state of affairs is available in i suspect we should add a selection of math variant or similarly named css classes for the cases where unicode does not have available codepoints
1
288,278
31,861,238,991
IssuesEvent
2023-09-15 11:04:34
nidhi7598/linux-v4.19.72_CVE-2022-3564
https://api.github.com/repos/nidhi7598/linux-v4.19.72_CVE-2022-3564
opened
CVE-2021-4093 (High) detected in linuxlinux-4.19.294, linuxlinux-4.19.294
Mend: dependency security vulnerability
## CVE-2021-4093 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxlinux-4.19.294</b>, <b>linuxlinux-4.19.294</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in the KVM's AMD code for supporting the Secure Encrypted Virtualization-Encrypted State (SEV-ES). A KVM guest using SEV-ES can trigger out-of-bounds reads and writes in the host kernel via a malicious VMGEXIT for a string I/O instruction (for example, outs or ins) using the exit reason SVM_EXIT_IOIO. This issue results in a crash of the entire system or a potential guest-to-host escape scenario. <p>Publish Date: 2022-02-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-4093>CVE-2021-4093</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-4093">https://www.linuxkernelcves.com/cves/CVE-2021-4093</a></p> <p>Release Date: 2022-02-18</p> <p>Fix Resolution: v5.14.15,v5.15-rc7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-4093 (High) detected in linuxlinux-4.19.294, linuxlinux-4.19.294 - ## CVE-2021-4093 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxlinux-4.19.294</b>, <b>linuxlinux-4.19.294</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in the KVM's AMD code for supporting the Secure Encrypted Virtualization-Encrypted State (SEV-ES). A KVM guest using SEV-ES can trigger out-of-bounds reads and writes in the host kernel via a malicious VMGEXIT for a string I/O instruction (for example, outs or ins) using the exit reason SVM_EXIT_IOIO. This issue results in a crash of the entire system or a potential guest-to-host escape scenario. <p>Publish Date: 2022-02-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-4093>CVE-2021-4093</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-4093">https://www.linuxkernelcves.com/cves/CVE-2021-4093</a></p> <p>Release Date: 2022-02-18</p> <p>Fix Resolution: v5.14.15,v5.15-rc7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in linuxlinux linuxlinux cve high severity vulnerability vulnerable libraries linuxlinux linuxlinux vulnerability details a flaw was found in the kvm s amd code for supporting the secure encrypted virtualization encrypted state sev es a kvm guest using sev es can trigger out of bounds reads and writes in the host kernel via a malicious vmgexit for a string i o instruction for example outs or ins using the exit reason svm exit ioio this issue results in a crash of the entire system or a potential guest to host escape scenario publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope changed impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
19,268
25,461,599,394
IssuesEvent
2022-11-24 20:01:06
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
Processor to enforce an allowlist schema for redaction and compliance
enhancement processor/redaction
**Is your feature request related to a problem? Please describe.** We need to be able to redact span attributes or tags based on key names, key values, and value size. Some attributes are sensitive (e.g. `date_of_birth`, `credit_card`) and for legal compliance reasons cannot be allowed in a trace. The best practice for preventing accidental leaks of sensitive values is to explicitly allow known good values rather than banning known bad values. **Describe the solution you'd like** A processor that can process spans to delete or mask sensitive values. The actions of the processor should be auditable so developers can debug missing fields. Example configuration: ```yaml processors: redaction: allowed_keys: ["id, "name"] blocked_values: - "^4[0-9]{12}(?:[0-9]{3})?$" # Visa credit card number - "^(5[1-5][0-9]{14})$" # Mastercard summary: debug # vs info or silent ``` Example input: ```yaml "id": 5 "date_of_birth": "2021-07-06" "4111111111111111": "such a nice credit card number" "5555555555554444": "another sample number" "name": "Jane Doe 4111111111111111" "unknown": 4 ``` Example non-production output: ```yaml "id": 5 "name": "Jane Doe ****" "masked_keys": "4111111111111111,5555555555554444,date_of_birth,unknown" "masked_key_count": 4 "masked_values": "name" "masked_value_count": 1 ``` Example production output: ```yaml "id": 5 "name": "Jane Doe ****" "masked_key_count": 4 "masked_value_count": 1 ``` **Describe alternatives you've considered** [AttributesProcessor](https://github.com/open-telemetry/opentelemetry-collector/tree/main/processor/attributesprocessor) could be enhanced to include these features. Right now its regex support is limited to splitting an existing attribute into multiple attributes or matching a span by one of its values, but it cannot delete all attributes that do not match a list or mask a value that matches a regex. Because this enhancement is not on the roadmap, it is unlikely to be implemented in AttributesProcessor. It also makes more sense to have all redaction or compliance features in one processor. **Additional context** There is a request for an attribute allowlist at the opentelemetry specification level here: <https://github.com/open-telemetry/opentelemetry-specification/issues/823>. That request is solving a related technical problem (unbounded schema growth) rather than the more narrow business problem (data compliance).
1.0
Processor to enforce an allowlist schema for redaction and compliance - **Is your feature request related to a problem? Please describe.** We need to be able to redact span attributes or tags based on key names, key values, and value size. Some attributes are sensitive (e.g. `date_of_birth`, `credit_card`) and for legal compliance reasons cannot be allowed in a trace. The best practice for preventing accidental leaks of sensitive values is to explicitly allow known good values rather than banning known bad values. **Describe the solution you'd like** A processor that can process spans to delete or mask sensitive values. The actions of the processor should be auditable so developers can debug missing fields. Example configuration: ```yaml processors: redaction: allowed_keys: ["id, "name"] blocked_values: - "^4[0-9]{12}(?:[0-9]{3})?$" # Visa credit card number - "^(5[1-5][0-9]{14})$" # Mastercard summary: debug # vs info or silent ``` Example input: ```yaml "id": 5 "date_of_birth": "2021-07-06" "4111111111111111": "such a nice credit card number" "5555555555554444": "another sample number" "name": "Jane Doe 4111111111111111" "unknown": 4 ``` Example non-production output: ```yaml "id": 5 "name": "Jane Doe ****" "masked_keys": "4111111111111111,5555555555554444,date_of_birth,unknown" "masked_key_count": 4 "masked_values": "name" "masked_value_count": 1 ``` Example production output: ```yaml "id": 5 "name": "Jane Doe ****" "masked_key_count": 4 "masked_value_count": 1 ``` **Describe alternatives you've considered** [AttributesProcessor](https://github.com/open-telemetry/opentelemetry-collector/tree/main/processor/attributesprocessor) could be enhanced to include these features. Right now its regex support is limited to splitting an existing attribute into multiple attributes or matching a span by one of its values, but it cannot delete all attributes that do not match a list or mask a value that matches a regex. Because this enhancement is not on the roadmap, it is unlikely to be implemented in AttributesProcessor. It also makes more sense to have all redaction or compliance features in one processor. **Additional context** There is a request for an attribute allowlist at the opentelemetry specification level here: <https://github.com/open-telemetry/opentelemetry-specification/issues/823>. That request is solving a related technical problem (unbounded schema growth) rather than the more narrow business problem (data compliance).
process
processor to enforce an allowlist schema for redaction and compliance is your feature request related to a problem please describe we need to be able to redact span attributes or tags based on key names key values and value size some attributes are sensitive e g date of birth credit card and for legal compliance reasons cannot be allowed in a trace the best practice for preventing accidental leaks of sensitive values is to explicitly allow known good values rather than banning known bad values describe the solution you d like a processor that can process spans to delete or mask sensitive values the actions of the processor should be auditable so developers can debug missing fields example configuration yaml processors redaction allowed keys blocked values visa credit card number mastercard summary debug vs info or silent example input yaml id date of birth such a nice credit card number another sample number name jane doe unknown example non production output yaml id name jane doe masked keys date of birth unknown masked key count masked values name masked value count example production output yaml id name jane doe masked key count masked value count describe alternatives you ve considered could be enhanced to include these features right now its regex support is limited to splitting an existing attribute into multiple attributes or matching a span by one of its values but it cannot delete all attributes that do not match a list or mask a value that matches a regex because this enhancement is not on the roadmap it is unlikely to be implemented in attributesprocessor it also makes more sense to have all redaction or compliance features in one processor additional context there is a request for an attribute allowlist at the opentelemetry specification level here that request is solving a related technical problem unbounded schema growth rather than the more narrow business problem data compliance
1
9,322
12,338,261,030
IssuesEvent
2020-05-14 16:10:46
DiSSCo/user-stories
https://api.github.com/repos/DiSSCo/user-stories
opened
Digital images and an annotation system
2. Collection Management 2. University/Research institute 4. Data processing ICEDIG-SURVEY Specimen level
As a Scientist I want to correct an identification and add an annotation so that I can file the specimen under the correct taxonomic name for this I need Digital images and an annotation system
1.0
Digital images and an annotation system - As a Scientist I want to correct an identification and add an annotation so that I can file the specimen under the correct taxonomic name for this I need Digital images and an annotation system
process
digital images and an annotation system as a scientist i want to correct an identification and add an annotation so that i can file the specimen under the correct taxonomic name for this i need digital images and an annotation system
1
6,941
10,110,544,478
IssuesEvent
2019-07-30 10:31:03
bisq-network/bisq
https://api.github.com/repos/bisq-network/bisq
closed
Set an automatic on off schedule for 1hr altcoin trades
in:gui in:trade-process was:dropped
Trades that have 24hr completion time can be taken care of in the morning. But everyone sleeps, and might forget to turn off, and it can be time consuming to sort through multiple trades to turn certain trades off. The 1hr alt coin trades need a schedule that can be set by the user, so they can turn off, especially for the time when the might be sleeping.
1.0
Set an automatic on off schedule for 1hr altcoin trades - Trades that have 24hr completion time can be taken care of in the morning. But everyone sleeps, and might forget to turn off, and it can be time consuming to sort through multiple trades to turn certain trades off. The 1hr alt coin trades need a schedule that can be set by the user, so they can turn off, especially for the time when the might be sleeping.
process
set an automatic on off schedule for altcoin trades trades that have completion time can be taken care of in the morning but everyone sleeps and might forget to turn off and it can be time consuming to sort through multiple trades to turn certain trades off the alt coin trades need a schedule that can be set by the user so they can turn off especially for the time when the might be sleeping
1
136,410
30,543,215,856
IssuesEvent
2023-07-20 00:06:58
grpc/grpc-java
https://api.github.com/repos/grpc/grpc-java
opened
AsyncServletOutputStreamWriterConcurrencyTest is too slow
code health
On my laptop, the single `linCheck` test takes over four minutes. I believe this is adding roughly that amount of time to the total build. This test class should either not be part of the normal test execution, or be configured in such a way to take considerably less time (max of 30s?).
1.0
AsyncServletOutputStreamWriterConcurrencyTest is too slow - On my laptop, the single `linCheck` test takes over four minutes. I believe this is adding roughly that amount of time to the total build. This test class should either not be part of the normal test execution, or be configured in such a way to take considerably less time (max of 30s?).
non_process
asyncservletoutputstreamwriterconcurrencytest is too slow on my laptop the single lincheck test takes over four minutes i believe this is adding roughly that amount of time to the total build this test class should either not be part of the normal test execution or be configured in such a way to take considerably less time max of
0
430,269
12,450,430,480
IssuesEvent
2020-05-27 08:47:48
riptano/cassandra-quarkus
https://api.github.com/repos/riptano/cassandra-quarkus
opened
cassandra-quarkus fixes once we get a driver with JAVA-2663
priority:major quarkus status:open type:bug
<p>Imported from <a href="https://datastax-oss.atlassian.net/browse/JAVA-2789">JAVA-2789</a>, originally reported by @absurdfarce on Tue, 19 May 2020 12:01:03 -0700.</p> <hr /> <p>A few small things that need to happen once we upgrade cassandra-quarkus to use a Java driver version which includes JAVA-2663:</p> <ul> <li>The NativeSubstitutions class should be removed entirely</li> <li>The explicit force of DefaultDriverOption.TIMESTAMP_GENERATOR_FORCE_JAVA_CLOCK to true can be removed as well</li> <li>The UnsignedBytesSubstitution can be removed as well <ul> <li>Assuming the same substitution has been added to the driver core as part of <span> <a href="https://datastax-oss.atlassian.net/browse/JAVA-2782" rel="nofollow"> JAVA-2782 </a> <span>In Progress</span> </span> </li> </ul> </li> </ul> <p>See the code review for the cassandra-quarkus side of <span> <a href="https://datastax-oss.atlassian.net/browse/JAVA-2782" rel="nofollow"> JAVA-2782 </a> <span>In Progress</span> </span> (<a href="https://github.com/datastax/cassandra-quarkus/pull/14/files" rel="nofollow">https://github.com/datastax/cassandra-quarkus/pull/14/files</a>) for more details</p>
1.0
cassandra-quarkus fixes once we get a driver with JAVA-2663 - <p>Imported from <a href="https://datastax-oss.atlassian.net/browse/JAVA-2789">JAVA-2789</a>, originally reported by @absurdfarce on Tue, 19 May 2020 12:01:03 -0700.</p> <hr /> <p>A few small things that need to happen once we upgrade cassandra-quarkus to use a Java driver version which includes JAVA-2663:</p> <ul> <li>The NativeSubstitutions class should be removed entirely</li> <li>The explicit force of DefaultDriverOption.TIMESTAMP_GENERATOR_FORCE_JAVA_CLOCK to true can be removed as well</li> <li>The UnsignedBytesSubstitution can be removed as well <ul> <li>Assuming the same substitution has been added to the driver core as part of <span> <a href="https://datastax-oss.atlassian.net/browse/JAVA-2782" rel="nofollow"> JAVA-2782 </a> <span>In Progress</span> </span> </li> </ul> </li> </ul> <p>See the code review for the cassandra-quarkus side of <span> <a href="https://datastax-oss.atlassian.net/browse/JAVA-2782" rel="nofollow"> JAVA-2782 </a> <span>In Progress</span> </span> (<a href="https://github.com/datastax/cassandra-quarkus/pull/14/files" rel="nofollow">https://github.com/datastax/cassandra-quarkus/pull/14/files</a>) for more details</p>
non_process
cassandra quarkus fixes once we get a driver with java imported from a few small things that need to happen once we upgrade cassandra quarkus to use a java driver version which includes java the nativesubstitutions class should be removed entirely the explicit force of defaultdriveroption timestamp generator force java clock to true can be removed as well the unsignedbytessubstitution can be removed as well assuming the same substitution has been added to the driver core as part of java in progress see the code review for the cassandra quarkus side of java in progress for more details
0
741
3,214,326,624
IssuesEvent
2015-10-07 00:51:32
broadinstitute/hellbender
https://api.github.com/repos/broadinstitute/hellbender
closed
Refactor Dataflow transforms by "top level" transform
Dataflow DataflowPreprocessingPipeline
There should be a package for each top level transform. Right now, transforms are organized haphazardly roughly by simple vs composite.
1.0
Refactor Dataflow transforms by "top level" transform - There should be a package for each top level transform. Right now, transforms are organized haphazardly roughly by simple vs composite.
process
refactor dataflow transforms by top level transform there should be a package for each top level transform right now transforms are organized haphazardly roughly by simple vs composite
1
19,303
25,466,620,758
IssuesEvent
2022-11-25 05:30:13
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[IDP] [PM] Admin is able to sign in with invalid password
Bug Blocker P0 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
In Sign in screen , Enter valid email and invalid password and Verify **AR:** Admin is able to sign in with invalid password **ER:** Admin should be able to sign in with valid credentials only.
3.0
[IDP] [PM] Admin is able to sign in with invalid password - In Sign in screen , Enter valid email and invalid password and Verify **AR:** Admin is able to sign in with invalid password **ER:** Admin should be able to sign in with valid credentials only.
process
admin is able to sign in with invalid password in sign in screen enter valid email and invalid password and verify ar admin is able to sign in with invalid password er admin should be able to sign in with valid credentials only
1
603
10,606,799,827
IssuesEvent
2019-10-11 00:57:48
AY1920S1-CS2103T-T13-3/main
https://api.github.com/repos/AY1920S1-CS2103T-T13-3/main
closed
As a user, I want to have information of the person I need to reimburse
Integration People UseCase status.Ongoing type.Story
so that I can easily find information to contact the person for reimbursement
1.0
As a user, I want to have information of the person I need to reimburse - so that I can easily find information to contact the person for reimbursement
non_process
as a user i want to have information of the person i need to reimburse so that i can easily find information to contact the person for reimbursement
0
13,266
15,731,386,020
IssuesEvent
2021-03-29 17:00:39
googleapis/elixir-google-api
https://api.github.com/repos/googleapis/elixir-google-api
opened
Warning: a recent release failed
type: process
The following release PRs may have failed: * #7710 * #7709 * #7708 * #7707 * #7706 * #7710 * #7709 * #7708 * #7707 * #7706 * #7710 * #7709 * #7708 * #7707 * #7706
1.0
Warning: a recent release failed - The following release PRs may have failed: * #7710 * #7709 * #7708 * #7707 * #7706 * #7710 * #7709 * #7708 * #7707 * #7706 * #7710 * #7709 * #7708 * #7707 * #7706
process
warning a recent release failed the following release prs may have failed
1
142,355
19,090,463,950
IssuesEvent
2021-11-29 11:29:42
sultanabubaker/NuGet_Project_SDK_NonSDK
https://api.github.com/repos/sultanabubaker/NuGet_Project_SDK_NonSDK
opened
CVE-2019-0567 (High) detected in multiple libraries
security vulnerability
## CVE-2019-0567 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ChakraCore-1.8.1.0.dll</b>, <b>ChakraCore-1.8.1.0.dll</b>, <b>ChakraCore-1.8.1.0.dll</b>, <b>microsoft.chakracore.1.8.1.nupkg</b></p></summary> <p> <details><summary><b>ChakraCore-1.8.1.0.dll</b></p></summary> <p>Microsoft ? Chakra Core</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p> <p>Path to vulnerable library: /SDK/bin/Debug/net5.0/runtimes/win8-arm/native/ChakraCore.dll,/SDK/bin/Debug/netcoreapp3.1/runtimes/win8-arm/native/ChakraCore.dll</p> <p> Dependency Hierarchy: - :x: **ChakraCore-1.8.1.0.dll** (Vulnerable Library) </details> <details><summary><b>ChakraCore-1.8.1.0.dll</b></p></summary> <p>Microsoft ? Chakra Core</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p> <p>Path to vulnerable library: /SDK/bin/Debug/netcoreapp3.1/runtimes/win7-x86/native/ChakraCore.dll,/SDK/bin/Debug/net5.0/runtimes/win7-x86/native/ChakraCore.dll</p> <p> Dependency Hierarchy: - :x: **ChakraCore-1.8.1.0.dll** (Vulnerable Library) </details> <details><summary><b>ChakraCore-1.8.1.0.dll</b></p></summary> <p>Microsoft ? Chakra Core</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p> <p>Path to vulnerable library: /SDK/bin/Debug/netcoreapp3.1/runtimes/win7-x64/native/ChakraCore.dll,/SDK/bin/Debug/net5.0/runtimes/win7-x64/native/ChakraCore.dll</p> <p> Dependency Hierarchy: - :x: **ChakraCore-1.8.1.0.dll** (Vulnerable Library) </details> <details><summary><b>microsoft.chakracore.1.8.1.nupkg</b></p></summary> <p>ChakraCore is the core part of the Chakra Javascript engine that powers Microsoft Edge.</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p> <p>Path to dependency file: NuGet_Project_SDK_NonSDK/SDK/SDK.csproj</p> <p>Path to vulnerable library: crosoft.chakracore/1.8.1/microsoft.chakracore.1.8.1.nupkg</p> <p> Dependency Hierarchy: - :x: **microsoft.chakracore.1.8.1.nupkg** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/sultanabubaker/NuGet_Project_SDK_NonSDK/commit/2cdcbe42d2efe636b5e9b1d4c29c9da6e2c9b927">2cdcbe42d2efe636b5e9b1d4c29c9da6e2c9b927</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A remote code execution vulnerability exists in the way that the Chakra scripting engine handles objects in memory in Microsoft Edge, aka "Chakra Scripting Engine Memory Corruption Vulnerability." This affects Microsoft Edge, ChakraCore. This CVE ID is unique from CVE-2019-0539, CVE-2019-0568. <p>Publish Date: 2019-01-08 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0567>CVE-2019-0567</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/Microsoft/ChakraCore/pull/5899">https://github.com/Microsoft/ChakraCore/pull/5899</a></p> <p>Release Date: 2020-08-24</p> <p>Fix Resolution: v1.11.5</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":".NET","packageName":"ChakraCore","packageVersion":"1.8.1.0","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"ChakraCore:1.8.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v1.11.5","isBinary":true},{"packageType":".NET","packageName":"ChakraCore","packageVersion":"1.8.1.0","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"ChakraCore:1.8.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v1.11.5","isBinary":true},{"packageType":".NET","packageName":"ChakraCore","packageVersion":"1.8.1.0","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"ChakraCore:1.8.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v1.11.5","isBinary":true},{"packageType":"Nuget","packageName":"Microsoft.ChakraCore","packageVersion":"1.8.1","packageFilePaths":["/SDK/SDK.csproj"],"isTransitiveDependency":false,"dependencyTree":"Microsoft.ChakraCore:1.8.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v1.11.5","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-0567","vulnerabilityDetails":"A remote code execution vulnerability exists in the way that the Chakra scripting engine handles objects in memory in Microsoft Edge, aka \"Chakra Scripting Engine Memory Corruption Vulnerability.\" This affects Microsoft Edge, ChakraCore. This CVE ID is unique from CVE-2019-0539, CVE-2019-0568.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0567","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-0567 (High) detected in multiple libraries - ## CVE-2019-0567 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ChakraCore-1.8.1.0.dll</b>, <b>ChakraCore-1.8.1.0.dll</b>, <b>ChakraCore-1.8.1.0.dll</b>, <b>microsoft.chakracore.1.8.1.nupkg</b></p></summary> <p> <details><summary><b>ChakraCore-1.8.1.0.dll</b></p></summary> <p>Microsoft ? Chakra Core</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p> <p>Path to vulnerable library: /SDK/bin/Debug/net5.0/runtimes/win8-arm/native/ChakraCore.dll,/SDK/bin/Debug/netcoreapp3.1/runtimes/win8-arm/native/ChakraCore.dll</p> <p> Dependency Hierarchy: - :x: **ChakraCore-1.8.1.0.dll** (Vulnerable Library) </details> <details><summary><b>ChakraCore-1.8.1.0.dll</b></p></summary> <p>Microsoft ? Chakra Core</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p> <p>Path to vulnerable library: /SDK/bin/Debug/netcoreapp3.1/runtimes/win7-x86/native/ChakraCore.dll,/SDK/bin/Debug/net5.0/runtimes/win7-x86/native/ChakraCore.dll</p> <p> Dependency Hierarchy: - :x: **ChakraCore-1.8.1.0.dll** (Vulnerable Library) </details> <details><summary><b>ChakraCore-1.8.1.0.dll</b></p></summary> <p>Microsoft ? Chakra Core</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p> <p>Path to vulnerable library: /SDK/bin/Debug/netcoreapp3.1/runtimes/win7-x64/native/ChakraCore.dll,/SDK/bin/Debug/net5.0/runtimes/win7-x64/native/ChakraCore.dll</p> <p> Dependency Hierarchy: - :x: **ChakraCore-1.8.1.0.dll** (Vulnerable Library) </details> <details><summary><b>microsoft.chakracore.1.8.1.nupkg</b></p></summary> <p>ChakraCore is the core part of the Chakra Javascript engine that powers Microsoft Edge.</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p> <p>Path to dependency file: NuGet_Project_SDK_NonSDK/SDK/SDK.csproj</p> <p>Path to vulnerable library: crosoft.chakracore/1.8.1/microsoft.chakracore.1.8.1.nupkg</p> <p> Dependency Hierarchy: - :x: **microsoft.chakracore.1.8.1.nupkg** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/sultanabubaker/NuGet_Project_SDK_NonSDK/commit/2cdcbe42d2efe636b5e9b1d4c29c9da6e2c9b927">2cdcbe42d2efe636b5e9b1d4c29c9da6e2c9b927</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A remote code execution vulnerability exists in the way that the Chakra scripting engine handles objects in memory in Microsoft Edge, aka "Chakra Scripting Engine Memory Corruption Vulnerability." This affects Microsoft Edge, ChakraCore. This CVE ID is unique from CVE-2019-0539, CVE-2019-0568. <p>Publish Date: 2019-01-08 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0567>CVE-2019-0567</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/Microsoft/ChakraCore/pull/5899">https://github.com/Microsoft/ChakraCore/pull/5899</a></p> <p>Release Date: 2020-08-24</p> <p>Fix Resolution: v1.11.5</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":".NET","packageName":"ChakraCore","packageVersion":"1.8.1.0","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"ChakraCore:1.8.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v1.11.5","isBinary":true},{"packageType":".NET","packageName":"ChakraCore","packageVersion":"1.8.1.0","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"ChakraCore:1.8.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v1.11.5","isBinary":true},{"packageType":".NET","packageName":"ChakraCore","packageVersion":"1.8.1.0","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"ChakraCore:1.8.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v1.11.5","isBinary":true},{"packageType":"Nuget","packageName":"Microsoft.ChakraCore","packageVersion":"1.8.1","packageFilePaths":["/SDK/SDK.csproj"],"isTransitiveDependency":false,"dependencyTree":"Microsoft.ChakraCore:1.8.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v1.11.5","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-0567","vulnerabilityDetails":"A remote code execution vulnerability exists in the way that the Chakra scripting engine handles objects in memory in Microsoft Edge, aka \"Chakra Scripting Engine Memory Corruption Vulnerability.\" This affects Microsoft Edge, ChakraCore. This CVE ID is unique from CVE-2019-0539, CVE-2019-0568.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0567","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries chakracore dll chakracore dll chakracore dll microsoft chakracore nupkg chakracore dll microsoft chakra core library home page a href path to vulnerable library sdk bin debug runtimes arm native chakracore dll sdk bin debug runtimes arm native chakracore dll dependency hierarchy x chakracore dll vulnerable library chakracore dll microsoft chakra core library home page a href path to vulnerable library sdk bin debug runtimes native chakracore dll sdk bin debug runtimes native chakracore dll dependency hierarchy x chakracore dll vulnerable library chakracore dll microsoft chakra core library home page a href path to vulnerable library sdk bin debug runtimes native chakracore dll sdk bin debug runtimes native chakracore dll dependency hierarchy x chakracore dll vulnerable library microsoft chakracore nupkg chakracore is the core part of the chakra javascript engine that powers microsoft edge library home page a href path to dependency file nuget project sdk nonsdk sdk sdk csproj path to vulnerable library crosoft chakracore microsoft chakracore nupkg dependency hierarchy x microsoft chakracore nupkg vulnerable library found in head commit a href found in base branch master vulnerability details a remote code execution vulnerability exists in the way that the chakra scripting engine handles objects in memory in microsoft edge aka chakra scripting engine memory corruption vulnerability this affects microsoft edge chakracore this cve id is unique from cve cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree chakracore isminimumfixversionavailable true minimumfixversion isbinary true packagetype net packagename chakracore packageversion packagefilepaths istransitivedependency false dependencytree chakracore isminimumfixversionavailable true minimumfixversion isbinary true packagetype net packagename chakracore packageversion packagefilepaths istransitivedependency false dependencytree chakracore isminimumfixversionavailable true minimumfixversion isbinary true packagetype nuget packagename microsoft chakracore packageversion packagefilepaths istransitivedependency false dependencytree microsoft chakracore isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails a remote code execution vulnerability exists in the way that the chakra scripting engine handles objects in memory in microsoft edge aka chakra scripting engine memory corruption vulnerability this affects microsoft edge chakracore this cve id is unique from cve cve vulnerabilityurl
0
278,439
30,702,322,476
IssuesEvent
2023-07-27 01:20:27
AlexRogalskiy/javascript-patterns
https://api.github.com/repos/AlexRogalskiy/javascript-patterns
closed
CVE-2022-2900 (Critical) detected in parse-url-6.0.0.tgz - autoclosed
Mend: dependency security vulnerability
## CVE-2022-2900 - Critical Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>parse-url-6.0.0.tgz</b></p></summary> <p>An advanced url parser supporting git urls too.</p> <p>Library home page: <a href="https://registry.npmjs.org/parse-url/-/parse-url-6.0.0.tgz">https://registry.npmjs.org/parse-url/-/parse-url-6.0.0.tgz</a></p> <p> Dependency Hierarchy: - release-it-14.11.8.tgz (Root Library) - git-url-parse-11.6.0.tgz - git-up-4.0.5.tgz - :x: **parse-url-6.0.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Server-Side Request Forgery (SSRF) in GitHub repository ionicabizau/parse-url prior to 8.1.0. <p>Publish Date: 2022-09-14 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-2900>CVE-2022-2900</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-09-14</p> <p>Fix Resolution (parse-url): 8.0.0</p> <p>Direct dependency fix Resolution (release-it): 15.4.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-2900 (Critical) detected in parse-url-6.0.0.tgz - autoclosed - ## CVE-2022-2900 - Critical Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>parse-url-6.0.0.tgz</b></p></summary> <p>An advanced url parser supporting git urls too.</p> <p>Library home page: <a href="https://registry.npmjs.org/parse-url/-/parse-url-6.0.0.tgz">https://registry.npmjs.org/parse-url/-/parse-url-6.0.0.tgz</a></p> <p> Dependency Hierarchy: - release-it-14.11.8.tgz (Root Library) - git-url-parse-11.6.0.tgz - git-up-4.0.5.tgz - :x: **parse-url-6.0.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Server-Side Request Forgery (SSRF) in GitHub repository ionicabizau/parse-url prior to 8.1.0. <p>Publish Date: 2022-09-14 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-2900>CVE-2022-2900</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-09-14</p> <p>Fix Resolution (parse-url): 8.0.0</p> <p>Direct dependency fix Resolution (release-it): 15.4.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve critical detected in parse url tgz autoclosed cve critical severity vulnerability vulnerable library parse url tgz an advanced url parser supporting git urls too library home page a href dependency hierarchy release it tgz root library git url parse tgz git up tgz x parse url tgz vulnerable library found in base branch master vulnerability details server side request forgery ssrf in github repository ionicabizau parse url prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution parse url direct dependency fix resolution release it step up your open source security game with mend
0
2,710
2,623,018,670
IssuesEvent
2015-03-04 09:34:25
laterpay/laterpay-wordpress-plugin
https://api.github.com/repos/laterpay/laterpay-wordpress-plugin
closed
Remove VAT settings
New Feature To be Tested
Because of VATMOSS, LaterPay has to get the correct VAT **from the user**. The merchant does not have to provide a VAT with the purchase link anymore. All things related to passing the VAT to the purchase dialogs have to be removed.
1.0
Remove VAT settings - Because of VATMOSS, LaterPay has to get the correct VAT **from the user**. The merchant does not have to provide a VAT with the purchase link anymore. All things related to passing the VAT to the purchase dialogs have to be removed.
non_process
remove vat settings because of vatmoss laterpay has to get the correct vat from the user the merchant does not have to provide a vat with the purchase link anymore all things related to passing the vat to the purchase dialogs have to be removed
0
21,079
28,026,074,930
IssuesEvent
2023-03-28 09:12:47
fluent/fluent-bit
https://api.github.com/repos/fluent/fluent-bit
closed
input forward: plugin doesn't close TCP port when it caught SIGTERM
enhancement work-in-process
## Bug Report **Describe the bug** The input forward plugin doesn't close TCP port when it caught SIGTERM. This causes losing some logs on shutdown. **To Reproduce** 1. Run a fluent-bit container with Docker: ``` $ docker run -it --rm --name flb -p 24224:24224 fluent/fluent-bit:1.5.6 /fluent-bit/bin/fluent-bit -v -i forward -o stdout -f 1 ``` 2. Run a `nc` command (port scan) container with Docker fluent log driver and forward its log to the fluent-bit container: ``` $ docker run -it --rm --log-driver=fluentd --log-opt fluentd-address=127.0.0.1:24224 --link flb busybox sh -c 'while true; do date; nc -vz flb 24224; sleep 1; done' ``` 3. Stop the fluent-bit container: ``` $ date -u; time docker stop -t 10 flb; date -u ``` The outputs of command are as follows: ``` $ docker run -it --rm --name flb -p 24224:24224 fluent/fluent-bit:1.5.6 /fluent-bit/bin/fluent-bit -v -i forward -o stdout -f 1 Fluent Bit v1.5.6 * Copyright (C) 2019-2020 The Fluent Bit Authors * Copyright (C) 2015-2018 Treasure Data * Fluent Bit is a CNCF sub-project under the umbrella of Fluentd * https://fluentbit.io [2020/09/29 00:09:55] [ info] Configuration: [2020/09/29 00:09:55] [ info] flush time | 1.000000 seconds [2020/09/29 00:09:55] [ info] grace | 5 seconds [2020/09/29 00:09:55] [ info] daemon | 0 [2020/09/29 00:09:55] [ info] ___________ [2020/09/29 00:09:55] [ info] inputs: [2020/09/29 00:09:55] [ info] forward [2020/09/29 00:09:55] [ info] ___________ [2020/09/29 00:09:55] [ info] filters: [2020/09/29 00:09:55] [ info] ___________ [2020/09/29 00:09:55] [ info] outputs: [2020/09/29 00:09:55] [ info] stdout.0 [2020/09/29 00:09:55] [ info] ___________ [2020/09/29 00:09:55] [ info] collectors: [2020/09/29 00:09:55] [ info] [engine] started (pid=1) [2020/09/29 00:09:55] [debug] [engine] coroutine stack size: 24576 bytes (24.0K) [2020/09/29 00:09:55] [debug] [storage] [cio stream] new stream registered: forward.0 [2020/09/29 00:09:55] [ info] [storage] version=1.0.5, initializing... [2020/09/29 00:09:55] [ info] [storage] in-memory [2020/09/29 00:09:55] [ info] [storage] normal synchronization mode, checksum disabled, max_chunks_up=128 [2020/09/29 00:09:55] [debug] [in_fw] Listen='0.0.0.0' TCP_Port=24224 [2020/09/29 00:09:55] [ info] [input:forward:forward.0] listening on 0.0.0.0:24224 [2020/09/29 00:09:55] [debug] [router] default match rule forward.0:stdout.0 [2020/09/29 00:09:55] [ info] [sp] stream processor started [0] e9ed8c030c0a: [1601338201.000000000, {"container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b"}]73", "container_name"=>"/laughing_feistel", "source"=>"stdout", "log"=>"Tue Sep 29 00:10:01 UTC 2020 [1] e9ed8c030c0a: [1601338201.000000000, {"container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b"}]73", "container_name"=>"/laughing_feistel", "source"=>"stdout", "log"=>"flb (172.17.0.2:24224) open [2020/09/29 00:10:02] [debug] [task] created task=0x7f4257a2f3a0 id=0 OK [2020/09/29 00:10:02] [debug] [task] destroy task=0x7f4257a2f3a0 (task_id=0) [0] e9ed8c030c0a: [1601338202.000000000, {"container_name"=>"/laughing_feistel", "source"=>"stdout", "log"=>"Tue Sep 2", "container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b3e973"}] [1] e9ed8c030c0a: [1601338202.000000000, {"container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b"}]73", "container_name"=>"/laughing_feistel", "source"=>"stdout", "log"=>"flb (172.17.0.2:24224) open [2020/09/29 00:10:03] [debug] [task] created task=0x7f4257a2f3a0 id=0 OK [2020/09/29 00:10:03] [debug] [task] destroy task=0x7f4257a2f3a0 (task_id=0) ", "container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b3e973", "container_name"=>"/laughing_feistel", "source"=>"stdout"}] ", "container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b3e973", "container_name"=>"/laughing_feistel", "source"=>"stdout"}] [2020/09/29 00:10:04] [debug] [task] created task=0x7f4257a2f3a0 id=0 OK [2020/09/29 00:10:04] [debug] [task] destroy task=0x7f4257a2f3a0 (task_id=0) [engine] caught signal (SIGTERM) ", "container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b3e973", "container_name"=>"/laughing_feistel", "source"=>"stdout"}] [1] e9ed8c030c0a: [1601338204.000000000, {"container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b"}]73", "container_name"=>"/laughing_feistel", "source"=>"stdout", "log"=>"flb (172.17.0.2:24224) open [2020/09/29 00:10:04] [debug] [task] created task=0x7f4257a2f3a0 id=0 OK [2020/09/29 00:10:04] [ warn] [engine] service will stop in 5 seconds [2020/09/29 00:10:04] [debug] [task] destroy task=0x7f4257a2f3a0 (task_id=0) [0] e9ed8c030c0a: [1601338205.000000000, {"container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b"}]73", "container_name"=>"/laughing_feistel", "source"=>"stdout", "log"=>"Tue Sep 29 00:10:05 UTC 2020 ", "container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b3e973", "container_name"=>"/laughing_feistel"}] [2020/09/29 00:10:06] [debug] [task] created task=0x7f4257a2f3a0 id=0 OK [2020/09/29 00:10:06] [debug] [task] destroy task=0x7f4257a2f3a0 (task_id=0) [0] e9ed8c030c0a: [1601338206.000000000, {"container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b"}]73", "container_name"=>"/laughing_feistel", "source"=>"stdout", "log"=>"Tue Sep 29 00:10:06 UTC 2020 [1] e9ed8c030c0a: [1601338206.000000000, {"container_name"=>"/laughing_feistel", "source"=>"stdout", "log"=>"flb (172.", "container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b3e973"}] [2020/09/29 00:10:07] [debug] [task] created task=0x7f4257a2f3a0 id=0 OK [2020/09/29 00:10:07] [debug] [task] destroy task=0x7f4257a2f3a0 (task_id=0) [0] e9ed8c030c0a: [1601338207.000000000, {"container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b"}]73", "container_name"=>"/laughing_feistel", "source"=>"stdout", "log"=>"Tue Sep 29 00:10:07 UTC 2020 [1] e9ed8c030c0a: [1601338207.000000000, {"container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b"}]73", "container_name"=>"/laughing_feistel", "source"=>"stdout", "log"=>"flb (172.17.0.2:24224) open [2020/09/29 00:10:08] [debug] [task] created task=0x7f4257a2f3a0 id=0 OK [2020/09/29 00:10:08] [debug] [task] destroy task=0x7f4257a2f3a0 (task_id=0) [2020/09/29 00:10:09] [ info] [engine] service stopped ``` ``` $ docker run -it --rm --log-driver=fluentd --log-opt fluentd-address=127.0.0.1:24224 --link flb busybox sh -c 'while true; do date; nc -vz flb 24224; sleep 1; done' Tue Sep 29 00:10:01 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:10:02 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:10:03 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:10:04 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:10:05 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:10:06 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:10:07 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:10:08 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:10:09 UTC 2020 ``` ``` $ date -u; time docker stop -t 10 flb; date -u Tue Sep 29 00:10:04 UTC 2020 flb docker stop -t 10 flb 0.04s user 0.02s system 1% cpu 4.409 total Tue Sep 29 00:10:09 UTC 2020 ``` According to the above output, the Docker tried to stop the fluent-bit container at 00:10:04. The engine caught signal (SIGTERM) at 00:10:04, but some logs are still ingested and the timestamp of the last log entry is 1601338207(=2020/09/29 00:10:07) and the result of nc command shows the TCP port was still open. The engine completed shutdown at 00:10:09 after grace period (5 sec) I tested multiple times and I found its behavior seems to be inconsistent. The input was sometimes paused (but TCP port was still open). ``` [0] cfedccd0310b: [1601337750.000000000, {"container_id"=>"cfedccd0310b0b4b747d4bd6b8362d8d009ff26b24cdd59d3b58faa3ea3"}]42", "container_name"=>"/vigilant_aryabhata", "source"=>"stdout", "log"=>"Tue Sep 29 00:02:30 UTC 2020 ", "container_id"=>"cfedccd0310b0b4b747d4bd6b8362d8d009ff26b24cdd59d3b58faa3ea300b42", "container_name"=>"/vigilant_aryabhata", "source"=>"stdout"}] [2020/09/29 00:02:30] [debug] [task] created task=0x7f8ee0e2f3a0 id=0 OK [2020/09/29 00:02:30] [debug] [task] destroy task=0x7f8ee0e2f3a0 (task_id=0) [engine] caught signal (SIGTERM) [2020/09/29 00:02:30] [ warn] [engine] service will stop in 5 seconds [2020/09/29 00:02:31] [debug] [input chunk] forward.0 is paused, cannot append records [2020/09/29 00:02:31] [debug] [input chunk] forward.0 is paused, cannot append records [2020/09/29 00:02:32] [debug] [input chunk] forward.0 is paused, cannot append records [2020/09/29 00:02:32] [debug] [input chunk] forward.0 is paused, cannot append records [2020/09/29 00:02:33] [debug] [input chunk] forward.0 is paused, cannot append records [2020/09/29 00:02:33] [debug] [input chunk] forward.0 is paused, cannot append records [2020/09/29 00:02:34] [debug] [input chunk] forward.0 is paused, cannot append records [2020/09/29 00:02:34] [debug] [input chunk] forward.0 is paused, cannot append records [2020/09/29 00:02:35] [debug] [input chunk] forward.0 is paused, cannot append records [2020/09/29 00:02:35] [debug] [input chunk] forward.0 is paused, cannot append records [2020/09/29 00:02:35] [ info] [engine] service stopped ``` ``` Tue Sep 29 00:02:28 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:02:29 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:02:30 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:02:31 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:02:32 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:02:33 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:02:34 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:02:35 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:02:36 UTC 2020 nc: flb (172.17.0.2:24224): Connection timed out Tue Sep 29 00:04:46 UTC 2020 ``` ``` $ date -u; time docker stop -t 10 flb; date -u Tue Sep 29 00:02:30 UTC 2020 flb docker stop -t 10 flb 0.04s user 0.02s system 1% cpu 5.190 total Tue Sep 29 00:02:35 UTC 2020 ``` **Expected behavior** The input forward plugin should close TCP port whent it caught SIGTERM not to lose any logs on shutdown. **Your Environment** * Version used: v1.5.6 * Configuration: (passed with arguments described above the reproduction case) * Environment name and version (e.g. Kubernetes? What version?): AWS ECS Fargate (platform version: 1.4.0) * Server type and version: AWS * Operating System and version: Docker (fluent/fluent-bit:1.5.6) * Filters and plugins: input forward **Additional context** I'm running fluent-bit containers on AWS ECS Fargate behind AWS NLB and forwarding other container's log with the Docker fluentd log driver. When I need to update my configuration for fluent-bit, I'm using the Blue-Green deployment strategy not to lose any logs during the deployment. However, AWS NLB doesn't reset the existing TCP connection even if NLB's deregistration delay passed. The deregistration delay block a new connection, but it doesn't reset the existing one. If my understanding is correct, the grace period doesn't ensure to flush all buffers. This means that if fluent-bit ingest new log entries during shutdown they will be lost. I don't think it only affects fluent-bit behind AWS NLB. I can reproduce a simple docker container setup described above. Blocking input on shutdown consistently may be a minimum solution, but it would be great if the fluent-bit close TCP port on shutdown because it allows sender to detect quickly receiver no longer available. **Related issue** https://github.com/fluent/fluent-bit/issues/1496
1.0
input forward: plugin doesn't close TCP port when it caught SIGTERM - ## Bug Report **Describe the bug** The input forward plugin doesn't close TCP port when it caught SIGTERM. This causes losing some logs on shutdown. **To Reproduce** 1. Run a fluent-bit container with Docker: ``` $ docker run -it --rm --name flb -p 24224:24224 fluent/fluent-bit:1.5.6 /fluent-bit/bin/fluent-bit -v -i forward -o stdout -f 1 ``` 2. Run a `nc` command (port scan) container with Docker fluent log driver and forward its log to the fluent-bit container: ``` $ docker run -it --rm --log-driver=fluentd --log-opt fluentd-address=127.0.0.1:24224 --link flb busybox sh -c 'while true; do date; nc -vz flb 24224; sleep 1; done' ``` 3. Stop the fluent-bit container: ``` $ date -u; time docker stop -t 10 flb; date -u ``` The outputs of command are as follows: ``` $ docker run -it --rm --name flb -p 24224:24224 fluent/fluent-bit:1.5.6 /fluent-bit/bin/fluent-bit -v -i forward -o stdout -f 1 Fluent Bit v1.5.6 * Copyright (C) 2019-2020 The Fluent Bit Authors * Copyright (C) 2015-2018 Treasure Data * Fluent Bit is a CNCF sub-project under the umbrella of Fluentd * https://fluentbit.io [2020/09/29 00:09:55] [ info] Configuration: [2020/09/29 00:09:55] [ info] flush time | 1.000000 seconds [2020/09/29 00:09:55] [ info] grace | 5 seconds [2020/09/29 00:09:55] [ info] daemon | 0 [2020/09/29 00:09:55] [ info] ___________ [2020/09/29 00:09:55] [ info] inputs: [2020/09/29 00:09:55] [ info] forward [2020/09/29 00:09:55] [ info] ___________ [2020/09/29 00:09:55] [ info] filters: [2020/09/29 00:09:55] [ info] ___________ [2020/09/29 00:09:55] [ info] outputs: [2020/09/29 00:09:55] [ info] stdout.0 [2020/09/29 00:09:55] [ info] ___________ [2020/09/29 00:09:55] [ info] collectors: [2020/09/29 00:09:55] [ info] [engine] started (pid=1) [2020/09/29 00:09:55] [debug] [engine] coroutine stack size: 24576 bytes (24.0K) [2020/09/29 00:09:55] [debug] [storage] [cio stream] new stream registered: forward.0 [2020/09/29 00:09:55] [ info] [storage] version=1.0.5, initializing... [2020/09/29 00:09:55] [ info] [storage] in-memory [2020/09/29 00:09:55] [ info] [storage] normal synchronization mode, checksum disabled, max_chunks_up=128 [2020/09/29 00:09:55] [debug] [in_fw] Listen='0.0.0.0' TCP_Port=24224 [2020/09/29 00:09:55] [ info] [input:forward:forward.0] listening on 0.0.0.0:24224 [2020/09/29 00:09:55] [debug] [router] default match rule forward.0:stdout.0 [2020/09/29 00:09:55] [ info] [sp] stream processor started [0] e9ed8c030c0a: [1601338201.000000000, {"container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b"}]73", "container_name"=>"/laughing_feistel", "source"=>"stdout", "log"=>"Tue Sep 29 00:10:01 UTC 2020 [1] e9ed8c030c0a: [1601338201.000000000, {"container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b"}]73", "container_name"=>"/laughing_feistel", "source"=>"stdout", "log"=>"flb (172.17.0.2:24224) open [2020/09/29 00:10:02] [debug] [task] created task=0x7f4257a2f3a0 id=0 OK [2020/09/29 00:10:02] [debug] [task] destroy task=0x7f4257a2f3a0 (task_id=0) [0] e9ed8c030c0a: [1601338202.000000000, {"container_name"=>"/laughing_feistel", "source"=>"stdout", "log"=>"Tue Sep 2", "container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b3e973"}] [1] e9ed8c030c0a: [1601338202.000000000, {"container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b"}]73", "container_name"=>"/laughing_feistel", "source"=>"stdout", "log"=>"flb (172.17.0.2:24224) open [2020/09/29 00:10:03] [debug] [task] created task=0x7f4257a2f3a0 id=0 OK [2020/09/29 00:10:03] [debug] [task] destroy task=0x7f4257a2f3a0 (task_id=0) ", "container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b3e973", "container_name"=>"/laughing_feistel", "source"=>"stdout"}] ", "container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b3e973", "container_name"=>"/laughing_feistel", "source"=>"stdout"}] [2020/09/29 00:10:04] [debug] [task] created task=0x7f4257a2f3a0 id=0 OK [2020/09/29 00:10:04] [debug] [task] destroy task=0x7f4257a2f3a0 (task_id=0) [engine] caught signal (SIGTERM) ", "container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b3e973", "container_name"=>"/laughing_feistel", "source"=>"stdout"}] [1] e9ed8c030c0a: [1601338204.000000000, {"container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b"}]73", "container_name"=>"/laughing_feistel", "source"=>"stdout", "log"=>"flb (172.17.0.2:24224) open [2020/09/29 00:10:04] [debug] [task] created task=0x7f4257a2f3a0 id=0 OK [2020/09/29 00:10:04] [ warn] [engine] service will stop in 5 seconds [2020/09/29 00:10:04] [debug] [task] destroy task=0x7f4257a2f3a0 (task_id=0) [0] e9ed8c030c0a: [1601338205.000000000, {"container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b"}]73", "container_name"=>"/laughing_feistel", "source"=>"stdout", "log"=>"Tue Sep 29 00:10:05 UTC 2020 ", "container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b3e973", "container_name"=>"/laughing_feistel"}] [2020/09/29 00:10:06] [debug] [task] created task=0x7f4257a2f3a0 id=0 OK [2020/09/29 00:10:06] [debug] [task] destroy task=0x7f4257a2f3a0 (task_id=0) [0] e9ed8c030c0a: [1601338206.000000000, {"container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b"}]73", "container_name"=>"/laughing_feistel", "source"=>"stdout", "log"=>"Tue Sep 29 00:10:06 UTC 2020 [1] e9ed8c030c0a: [1601338206.000000000, {"container_name"=>"/laughing_feistel", "source"=>"stdout", "log"=>"flb (172.", "container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b3e973"}] [2020/09/29 00:10:07] [debug] [task] created task=0x7f4257a2f3a0 id=0 OK [2020/09/29 00:10:07] [debug] [task] destroy task=0x7f4257a2f3a0 (task_id=0) [0] e9ed8c030c0a: [1601338207.000000000, {"container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b"}]73", "container_name"=>"/laughing_feistel", "source"=>"stdout", "log"=>"Tue Sep 29 00:10:07 UTC 2020 [1] e9ed8c030c0a: [1601338207.000000000, {"container_id"=>"e9ed8c030c0af0df30f15bad9a5e62c49e78dfc581738c6ccf480457e0b"}]73", "container_name"=>"/laughing_feistel", "source"=>"stdout", "log"=>"flb (172.17.0.2:24224) open [2020/09/29 00:10:08] [debug] [task] created task=0x7f4257a2f3a0 id=0 OK [2020/09/29 00:10:08] [debug] [task] destroy task=0x7f4257a2f3a0 (task_id=0) [2020/09/29 00:10:09] [ info] [engine] service stopped ``` ``` $ docker run -it --rm --log-driver=fluentd --log-opt fluentd-address=127.0.0.1:24224 --link flb busybox sh -c 'while true; do date; nc -vz flb 24224; sleep 1; done' Tue Sep 29 00:10:01 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:10:02 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:10:03 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:10:04 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:10:05 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:10:06 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:10:07 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:10:08 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:10:09 UTC 2020 ``` ``` $ date -u; time docker stop -t 10 flb; date -u Tue Sep 29 00:10:04 UTC 2020 flb docker stop -t 10 flb 0.04s user 0.02s system 1% cpu 4.409 total Tue Sep 29 00:10:09 UTC 2020 ``` According to the above output, the Docker tried to stop the fluent-bit container at 00:10:04. The engine caught signal (SIGTERM) at 00:10:04, but some logs are still ingested and the timestamp of the last log entry is 1601338207(=2020/09/29 00:10:07) and the result of nc command shows the TCP port was still open. The engine completed shutdown at 00:10:09 after grace period (5 sec) I tested multiple times and I found its behavior seems to be inconsistent. The input was sometimes paused (but TCP port was still open). ``` [0] cfedccd0310b: [1601337750.000000000, {"container_id"=>"cfedccd0310b0b4b747d4bd6b8362d8d009ff26b24cdd59d3b58faa3ea3"}]42", "container_name"=>"/vigilant_aryabhata", "source"=>"stdout", "log"=>"Tue Sep 29 00:02:30 UTC 2020 ", "container_id"=>"cfedccd0310b0b4b747d4bd6b8362d8d009ff26b24cdd59d3b58faa3ea300b42", "container_name"=>"/vigilant_aryabhata", "source"=>"stdout"}] [2020/09/29 00:02:30] [debug] [task] created task=0x7f8ee0e2f3a0 id=0 OK [2020/09/29 00:02:30] [debug] [task] destroy task=0x7f8ee0e2f3a0 (task_id=0) [engine] caught signal (SIGTERM) [2020/09/29 00:02:30] [ warn] [engine] service will stop in 5 seconds [2020/09/29 00:02:31] [debug] [input chunk] forward.0 is paused, cannot append records [2020/09/29 00:02:31] [debug] [input chunk] forward.0 is paused, cannot append records [2020/09/29 00:02:32] [debug] [input chunk] forward.0 is paused, cannot append records [2020/09/29 00:02:32] [debug] [input chunk] forward.0 is paused, cannot append records [2020/09/29 00:02:33] [debug] [input chunk] forward.0 is paused, cannot append records [2020/09/29 00:02:33] [debug] [input chunk] forward.0 is paused, cannot append records [2020/09/29 00:02:34] [debug] [input chunk] forward.0 is paused, cannot append records [2020/09/29 00:02:34] [debug] [input chunk] forward.0 is paused, cannot append records [2020/09/29 00:02:35] [debug] [input chunk] forward.0 is paused, cannot append records [2020/09/29 00:02:35] [debug] [input chunk] forward.0 is paused, cannot append records [2020/09/29 00:02:35] [ info] [engine] service stopped ``` ``` Tue Sep 29 00:02:28 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:02:29 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:02:30 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:02:31 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:02:32 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:02:33 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:02:34 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:02:35 UTC 2020 flb (172.17.0.2:24224) open Tue Sep 29 00:02:36 UTC 2020 nc: flb (172.17.0.2:24224): Connection timed out Tue Sep 29 00:04:46 UTC 2020 ``` ``` $ date -u; time docker stop -t 10 flb; date -u Tue Sep 29 00:02:30 UTC 2020 flb docker stop -t 10 flb 0.04s user 0.02s system 1% cpu 5.190 total Tue Sep 29 00:02:35 UTC 2020 ``` **Expected behavior** The input forward plugin should close TCP port whent it caught SIGTERM not to lose any logs on shutdown. **Your Environment** * Version used: v1.5.6 * Configuration: (passed with arguments described above the reproduction case) * Environment name and version (e.g. Kubernetes? What version?): AWS ECS Fargate (platform version: 1.4.0) * Server type and version: AWS * Operating System and version: Docker (fluent/fluent-bit:1.5.6) * Filters and plugins: input forward **Additional context** I'm running fluent-bit containers on AWS ECS Fargate behind AWS NLB and forwarding other container's log with the Docker fluentd log driver. When I need to update my configuration for fluent-bit, I'm using the Blue-Green deployment strategy not to lose any logs during the deployment. However, AWS NLB doesn't reset the existing TCP connection even if NLB's deregistration delay passed. The deregistration delay block a new connection, but it doesn't reset the existing one. If my understanding is correct, the grace period doesn't ensure to flush all buffers. This means that if fluent-bit ingest new log entries during shutdown they will be lost. I don't think it only affects fluent-bit behind AWS NLB. I can reproduce a simple docker container setup described above. Blocking input on shutdown consistently may be a minimum solution, but it would be great if the fluent-bit close TCP port on shutdown because it allows sender to detect quickly receiver no longer available. **Related issue** https://github.com/fluent/fluent-bit/issues/1496
process
input forward plugin doesn t close tcp port when it caught sigterm bug report describe the bug the input forward plugin doesn t close tcp port when it caught sigterm this causes losing some logs on shutdown to reproduce run a fluent bit container with docker docker run it rm name flb p fluent fluent bit fluent bit bin fluent bit v i forward o stdout f run a nc command port scan container with docker fluent log driver and forward its log to the fluent bit container docker run it rm log driver fluentd log opt fluentd address link flb busybox sh c while true do date nc vz flb sleep done stop the fluent bit container date u time docker stop t flb date u the outputs of command are as follows docker run it rm name flb p fluent fluent bit fluent bit bin fluent bit v i forward o stdout f fluent bit copyright c the fluent bit authors copyright c treasure data fluent bit is a cncf sub project under the umbrella of fluentd configuration flush time seconds grace seconds daemon inputs forward filters outputs stdout collectors started pid coroutine stack size bytes new stream registered forward version initializing in memory normal synchronization mode checksum disabled max chunks up listen tcp port listening on default match rule forward stdout stream processor started container name laughing feistel source stdout log tue sep utc container name laughing feistel source stdout log flb open created task id ok destroy task task id container name laughing feistel source stdout log flb open created task id ok destroy task task id container id container name laughing feistel source stdout container id container name laughing feistel source stdout created task id ok destroy task task id caught signal sigterm container id container name laughing feistel source stdout container name laughing feistel source stdout log flb open created task id ok service will stop in seconds destroy task task id container name laughing feistel source stdout log tue sep utc container id container name laughing feistel created task id ok destroy task task id container name laughing feistel source stdout log tue sep utc created task id ok destroy task task id container name laughing feistel source stdout log tue sep utc container name laughing feistel source stdout log flb open created task id ok destroy task task id service stopped docker run it rm log driver fluentd log opt fluentd address link flb busybox sh c while true do date nc vz flb sleep done tue sep utc flb open tue sep utc flb open tue sep utc flb open tue sep utc flb open tue sep utc flb open tue sep utc flb open tue sep utc flb open tue sep utc flb open tue sep utc date u time docker stop t flb date u tue sep utc flb docker stop t flb user system cpu total tue sep utc according to the above output the docker tried to stop the fluent bit container at the engine caught signal sigterm at but some logs are still ingested and the timestamp of the last log entry is and the result of nc command shows the tcp port was still open the engine completed shutdown at after grace period sec i tested multiple times and i found its behavior seems to be inconsistent the input was sometimes paused but tcp port was still open container name vigilant aryabhata source stdout log tue sep utc container id container name vigilant aryabhata source stdout created task id ok destroy task task id caught signal sigterm service will stop in seconds forward is paused cannot append records forward is paused cannot append records forward is paused cannot append records forward is paused cannot append records forward is paused cannot append records forward is paused cannot append records forward is paused cannot append records forward is paused cannot append records forward is paused cannot append records forward is paused cannot append records service stopped tue sep utc flb open tue sep utc flb open tue sep utc flb open tue sep utc flb open tue sep utc flb open tue sep utc flb open tue sep utc flb open tue sep utc flb open tue sep utc nc flb connection timed out tue sep utc date u time docker stop t flb date u tue sep utc flb docker stop t flb user system cpu total tue sep utc expected behavior the input forward plugin should close tcp port whent it caught sigterm not to lose any logs on shutdown your environment version used configuration passed with arguments described above the reproduction case environment name and version e g kubernetes what version aws ecs fargate platform version server type and version aws operating system and version docker fluent fluent bit filters and plugins input forward additional context i m running fluent bit containers on aws ecs fargate behind aws nlb and forwarding other container s log with the docker fluentd log driver when i need to update my configuration for fluent bit i m using the blue green deployment strategy not to lose any logs during the deployment however aws nlb doesn t reset the existing tcp connection even if nlb s deregistration delay passed the deregistration delay block a new connection but it doesn t reset the existing one if my understanding is correct the grace period doesn t ensure to flush all buffers this means that if fluent bit ingest new log entries during shutdown they will be lost i don t think it only affects fluent bit behind aws nlb i can reproduce a simple docker container setup described above blocking input on shutdown consistently may be a minimum solution but it would be great if the fluent bit close tcp port on shutdown because it allows sender to detect quickly receiver no longer available related issue
1
59,317
24,727,453,976
IssuesEvent
2022-10-20 14:58:41
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Kensho web hook lacks info
triaged assigned-to-author doc-enhancement Pri2 metrics-advisor/subsvc applied-ai-services/svc
Kensho2 (Anomaly detector) send very different info in the web hook than stated in the page. the body of the request is very narrow with almost no info and certainly no callBackUrl, example: _"body": { "timestamp": "2022-10-02T00:00:00Z", "alertSettingGuid": "***********************" }_ Please help me getting the right body or fix this as this is needed for our product. --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: ad86463f-baff-a804-0aa2-f87fc4356041 * Version Independent ID: 1a4c5b6d-9f45-83fe-439f-7436b0d669dd * Content: [Configure Metrics Advisor alerts - Azure Cognitive Services](https://learn.microsoft.com/en-us/azure/applied-ai-services/metrics-advisor/how-tos/alerts) * Content Source: [articles/applied-ai-services/metrics-advisor/how-tos/alerts.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/applied-ai-services/metrics-advisor/how-tos/alerts.md) * Service: **applied-ai-services** * Sub-service: **metrics-advisor** * GitHub Login: @mrbullwinkle * Microsoft Alias: **mbullwin**
1.0
Kensho web hook lacks info - Kensho2 (Anomaly detector) send very different info in the web hook than stated in the page. the body of the request is very narrow with almost no info and certainly no callBackUrl, example: _"body": { "timestamp": "2022-10-02T00:00:00Z", "alertSettingGuid": "***********************" }_ Please help me getting the right body or fix this as this is needed for our product. --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: ad86463f-baff-a804-0aa2-f87fc4356041 * Version Independent ID: 1a4c5b6d-9f45-83fe-439f-7436b0d669dd * Content: [Configure Metrics Advisor alerts - Azure Cognitive Services](https://learn.microsoft.com/en-us/azure/applied-ai-services/metrics-advisor/how-tos/alerts) * Content Source: [articles/applied-ai-services/metrics-advisor/how-tos/alerts.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/applied-ai-services/metrics-advisor/how-tos/alerts.md) * Service: **applied-ai-services** * Sub-service: **metrics-advisor** * GitHub Login: @mrbullwinkle * Microsoft Alias: **mbullwin**
non_process
kensho web hook lacks info anomaly detector send very different info in the web hook than stated in the page the body of the request is very narrow with almost no info and certainly no callbackurl example body timestamp alertsettingguid please help me getting the right body or fix this as this is needed for our product document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id baff version independent id content content source service applied ai services sub service metrics advisor github login mrbullwinkle microsoft alias mbullwin
0
3,222
2,824,204,984
IssuesEvent
2015-05-21 13:36:13
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
Default hreflang tag and including full list of translated pages (including own url)
No Code Attached Yet
#### Steps to reproduce the issue Make a site with multiple languages #### Expected result The main/default language for a page should have a hreflang="x-default" tag as explained here: http://googlewebmastercentral.blogspot.nl/2013/04/x-default-hreflang-for-international-pages.html Additionally each page should have all languages with their respective URL's listed, including its own URL. (currently it only lists different ones) #### Actual result Only the other languages have a hreflang tag set, the loaded page language is not included. Additionally the default language is not marked with an "x-default" tag. #### System information (as much as possible) Not relevant #### Additional comments I've discovered this issue trough using [http://seorch.eu/](http://seorch.eu/ "") ![screen shot 2015-04-20 at 14 09 22](http://issues.joomla.org/uploads/1/c92c6251edc039772d13ec25c0603995.png) I also suggest reading: [http://moz.com/learn/seo/hreflang-tag](http://moz.com/learn/seo/hreflang-tag "") For best practises.
1.0
Default hreflang tag and including full list of translated pages (including own url) - #### Steps to reproduce the issue Make a site with multiple languages #### Expected result The main/default language for a page should have a hreflang="x-default" tag as explained here: http://googlewebmastercentral.blogspot.nl/2013/04/x-default-hreflang-for-international-pages.html Additionally each page should have all languages with their respective URL's listed, including its own URL. (currently it only lists different ones) #### Actual result Only the other languages have a hreflang tag set, the loaded page language is not included. Additionally the default language is not marked with an "x-default" tag. #### System information (as much as possible) Not relevant #### Additional comments I've discovered this issue trough using [http://seorch.eu/](http://seorch.eu/ "") ![screen shot 2015-04-20 at 14 09 22](http://issues.joomla.org/uploads/1/c92c6251edc039772d13ec25c0603995.png) I also suggest reading: [http://moz.com/learn/seo/hreflang-tag](http://moz.com/learn/seo/hreflang-tag "") For best practises.
non_process
default hreflang tag and including full list of translated pages including own url steps to reproduce the issue make a site with multiple languages expected result the main default language for a page should have a hreflang x default tag as explained here additionally each page should have all languages with their respective url s listed including its own url currently it only lists different ones actual result only the other languages have a hreflang tag set the loaded page language is not included additionally the default language is not marked with an x default tag system information as much as possible not relevant additional comments i ve discovered this issue trough using i also suggest reading for best practises
0
195,493
14,737,153,572
IssuesEvent
2021-01-07 01:01:23
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
opened
Manual test run on Win x64 for 1.18.x Release #4
OS/Desktop OS/Windows QA/Yes release-notes/exclude tests
### Installer - [ ] Check signature: If OS Run `spctl --assess --verbose /Applications/Brave-Browser-Beta.app/` and make sure it returns `accepted`. If Windows right click on the `brave_installer-x64.exe` and go to Properties, go to the Digital Signatures tab and double click on the signature. Make sure it says "The digital signature is OK" in the popup window ### Widevine - [ ] Verify `Widevine Notification` is shown when you visit Netflix for the first time - [ ] Test that you can stream on Netflix on a fresh profile after installing Widevine ### Rewards - [ ] Verify account balance shows correct BAT and USD value - [ ] Verify you are able to restore a wallet - [ ] Verify actions taken (claiming grant, tipping, auto-contribute) display in wallet panel - [ ] Verify when you click on the BR panel while on a site, the panel displays site specific information (site favicon, domain, attention %) - [ ] Verify you are able to make one-time tip and they display in tips panel - [ ] Verify you are able to make recurring tip and they display in tips panel - [ ] Verify you can tip a verified publisher - [ ] Verify you can tip a verified YouTube creator - [ ] Verify you are able to perform a contribution - [ ] Verify if you disable auto-contribute you are still able to tip regular sites and YouTube creators ## Update tests - [ ] Verify visiting `brave://settings/help` triggers update check - [ ] Verify once update is downloaded, prompts to `Relaunch` to install update #### Components - [ ] Delete Adblock folder from browser profile and restart browser. Visit `brave://components` and verify `Brave Ad Block Updater` downloads and update the component. Repeat for all Brave components ### Upgrade - [ ] Make sure that data from the last version appears in the new version OK - [ ] Ensure that `brave://version` lists the expected Brave & Chromium versions - [ ] With data from the last version, verify that - [ ] Bookmarks on the bookmark toolbar and bookmark folders can be opened - [ ] Cookies are preserved - [ ] Installed extensions are retained and work correctly - [ ] Opened tabs can be reloaded - [ ] Stored passwords are preserved - [ ] Sync chain created in previous version is retained - [ ] Social media blocking buttons changes are retained - [ ] Rewards - [ ] Wallet balance is retained - [ ] Auto-contribute list is retained - [ ] Both Tips and Monthly Contributions are retained - [ ] Wallet panel transactions list is retained - [ ] Changes to rewards settings are retained - [ ] Ads - [ ] Both `Estimated pending rewards` & `Ad notifications received this month` are retained - [ ] Changes to ads settings are retained - [ ] Ensure that ads are not being enabled when upgrading to a new version if they were disabled - [ ] Ensure that ads are not disabled when upgrading to a new version if they were enabled
1.0
Manual test run on Win x64 for 1.18.x Release #4 - ### Installer - [ ] Check signature: If OS Run `spctl --assess --verbose /Applications/Brave-Browser-Beta.app/` and make sure it returns `accepted`. If Windows right click on the `brave_installer-x64.exe` and go to Properties, go to the Digital Signatures tab and double click on the signature. Make sure it says "The digital signature is OK" in the popup window ### Widevine - [ ] Verify `Widevine Notification` is shown when you visit Netflix for the first time - [ ] Test that you can stream on Netflix on a fresh profile after installing Widevine ### Rewards - [ ] Verify account balance shows correct BAT and USD value - [ ] Verify you are able to restore a wallet - [ ] Verify actions taken (claiming grant, tipping, auto-contribute) display in wallet panel - [ ] Verify when you click on the BR panel while on a site, the panel displays site specific information (site favicon, domain, attention %) - [ ] Verify you are able to make one-time tip and they display in tips panel - [ ] Verify you are able to make recurring tip and they display in tips panel - [ ] Verify you can tip a verified publisher - [ ] Verify you can tip a verified YouTube creator - [ ] Verify you are able to perform a contribution - [ ] Verify if you disable auto-contribute you are still able to tip regular sites and YouTube creators ## Update tests - [ ] Verify visiting `brave://settings/help` triggers update check - [ ] Verify once update is downloaded, prompts to `Relaunch` to install update #### Components - [ ] Delete Adblock folder from browser profile and restart browser. Visit `brave://components` and verify `Brave Ad Block Updater` downloads and update the component. Repeat for all Brave components ### Upgrade - [ ] Make sure that data from the last version appears in the new version OK - [ ] Ensure that `brave://version` lists the expected Brave & Chromium versions - [ ] With data from the last version, verify that - [ ] Bookmarks on the bookmark toolbar and bookmark folders can be opened - [ ] Cookies are preserved - [ ] Installed extensions are retained and work correctly - [ ] Opened tabs can be reloaded - [ ] Stored passwords are preserved - [ ] Sync chain created in previous version is retained - [ ] Social media blocking buttons changes are retained - [ ] Rewards - [ ] Wallet balance is retained - [ ] Auto-contribute list is retained - [ ] Both Tips and Monthly Contributions are retained - [ ] Wallet panel transactions list is retained - [ ] Changes to rewards settings are retained - [ ] Ads - [ ] Both `Estimated pending rewards` & `Ad notifications received this month` are retained - [ ] Changes to ads settings are retained - [ ] Ensure that ads are not being enabled when upgrading to a new version if they were disabled - [ ] Ensure that ads are not disabled when upgrading to a new version if they were enabled
non_process
manual test run on win for x release installer check signature if os run spctl assess verbose applications brave browser beta app and make sure it returns accepted if windows right click on the brave installer exe and go to properties go to the digital signatures tab and double click on the signature make sure it says the digital signature is ok in the popup window widevine verify widevine notification is shown when you visit netflix for the first time test that you can stream on netflix on a fresh profile after installing widevine rewards verify account balance shows correct bat and usd value verify you are able to restore a wallet verify actions taken claiming grant tipping auto contribute display in wallet panel verify when you click on the br panel while on a site the panel displays site specific information site favicon domain attention verify you are able to make one time tip and they display in tips panel verify you are able to make recurring tip and they display in tips panel verify you can tip a verified publisher verify you can tip a verified youtube creator verify you are able to perform a contribution verify if you disable auto contribute you are still able to tip regular sites and youtube creators update tests verify visiting brave settings help triggers update check verify once update is downloaded prompts to relaunch to install update components delete adblock folder from browser profile and restart browser visit brave components and verify brave ad block updater downloads and update the component repeat for all brave components upgrade make sure that data from the last version appears in the new version ok ensure that brave version lists the expected brave chromium versions with data from the last version verify that bookmarks on the bookmark toolbar and bookmark folders can be opened cookies are preserved installed extensions are retained and work correctly opened tabs can be reloaded stored passwords are preserved sync chain created in previous version is retained social media blocking buttons changes are retained rewards wallet balance is retained auto contribute list is retained both tips and monthly contributions are retained wallet panel transactions list is retained changes to rewards settings are retained ads both estimated pending rewards ad notifications received this month are retained changes to ads settings are retained ensure that ads are not being enabled when upgrading to a new version if they were disabled ensure that ads are not disabled when upgrading to a new version if they were enabled
0
109,151
16,832,263,173
IssuesEvent
2021-06-18 07:15:08
Vento-Nuenenen/MatSchlacht
https://api.github.com/repos/Vento-Nuenenen/MatSchlacht
closed
CVE-2021-33502 (High) detected in normalize-url-3.3.0.tgz - autoclosed
security vulnerability
## CVE-2021-33502 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>normalize-url-3.3.0.tgz</b></p></summary> <p>Normalize a URL</p> <p>Library home page: <a href="https://registry.npmjs.org/normalize-url/-/normalize-url-3.3.0.tgz">https://registry.npmjs.org/normalize-url/-/normalize-url-3.3.0.tgz</a></p> <p>Path to dependency file: MatSchlacht/package.json</p> <p>Path to vulnerable library: MatSchlacht/node_modules/normalize-url/package.json</p> <p> Dependency Hierarchy: - laravel-mix-6.0.19.tgz (Root Library) - cssnano-4.1.11.tgz - cssnano-preset-default-4.0.8.tgz - postcss-normalize-url-4.0.1.tgz - :x: **normalize-url-3.3.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Vento-Nuenenen/MatSchlacht/commit/81a4d313160355bfe563fdc594c11c58ffcbe11f">81a4d313160355bfe563fdc594c11c58ffcbe11f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The normalize-url package before 4.5.1, 5.x before 5.3.1, and 6.x before 6.0.1 for Node.js has a ReDoS (regular expression denial of service) issue because it has exponential performance for data: URLs. <p>Publish Date: 2021-05-24 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33502>CVE-2021-33502</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33502">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33502</a></p> <p>Release Date: 2021-05-24</p> <p>Fix Resolution: normalize-url - 4.5.1, 5.3.1, 6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-33502 (High) detected in normalize-url-3.3.0.tgz - autoclosed - ## CVE-2021-33502 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>normalize-url-3.3.0.tgz</b></p></summary> <p>Normalize a URL</p> <p>Library home page: <a href="https://registry.npmjs.org/normalize-url/-/normalize-url-3.3.0.tgz">https://registry.npmjs.org/normalize-url/-/normalize-url-3.3.0.tgz</a></p> <p>Path to dependency file: MatSchlacht/package.json</p> <p>Path to vulnerable library: MatSchlacht/node_modules/normalize-url/package.json</p> <p> Dependency Hierarchy: - laravel-mix-6.0.19.tgz (Root Library) - cssnano-4.1.11.tgz - cssnano-preset-default-4.0.8.tgz - postcss-normalize-url-4.0.1.tgz - :x: **normalize-url-3.3.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Vento-Nuenenen/MatSchlacht/commit/81a4d313160355bfe563fdc594c11c58ffcbe11f">81a4d313160355bfe563fdc594c11c58ffcbe11f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The normalize-url package before 4.5.1, 5.x before 5.3.1, and 6.x before 6.0.1 for Node.js has a ReDoS (regular expression denial of service) issue because it has exponential performance for data: URLs. <p>Publish Date: 2021-05-24 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33502>CVE-2021-33502</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33502">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33502</a></p> <p>Release Date: 2021-05-24</p> <p>Fix Resolution: normalize-url - 4.5.1, 5.3.1, 6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in normalize url tgz autoclosed cve high severity vulnerability vulnerable library normalize url tgz normalize a url library home page a href path to dependency file matschlacht package json path to vulnerable library matschlacht node modules normalize url package json dependency hierarchy laravel mix tgz root library cssnano tgz cssnano preset default tgz postcss normalize url tgz x normalize url tgz vulnerable library found in head commit a href found in base branch master vulnerability details the normalize url package before x before and x before for node js has a redos regular expression denial of service issue because it has exponential performance for data urls publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution normalize url step up your open source security game with whitesource
0
320,904
23,831,503,197
IssuesEvent
2022-09-05 21:43:08
Dog-Face-Development/ProgramVer
https://api.github.com/repos/Dog-Face-Development/ProgramVer
closed
Update README
enhancement documentation
From Auto-Anouncements created by [willtheorangeguy](https://github.com/willtheorangeguy): Dog-Face-Development/Auto-Anouncements#22 From PyWorkout created by [willtheorangeguy](https://github.com/willtheorangeguy): willtheorangeguy/PyWorkout#3 Create a proper `README.md`, using bookmarked templates. Should have many areas of key program information. Ideas: - [Electron-Markdownify](https://github.com/amitmerchant1990/electron-markdownify#readme) - [Choo](https://github.com/choojs/choo#readme) - [Redoc](https://github.com/Redocly/redoc#readme)
1.0
Update README - From Auto-Anouncements created by [willtheorangeguy](https://github.com/willtheorangeguy): Dog-Face-Development/Auto-Anouncements#22 From PyWorkout created by [willtheorangeguy](https://github.com/willtheorangeguy): willtheorangeguy/PyWorkout#3 Create a proper `README.md`, using bookmarked templates. Should have many areas of key program information. Ideas: - [Electron-Markdownify](https://github.com/amitmerchant1990/electron-markdownify#readme) - [Choo](https://github.com/choojs/choo#readme) - [Redoc](https://github.com/Redocly/redoc#readme)
non_process
update readme from auto anouncements created by dog face development auto anouncements from pyworkout created by willtheorangeguy pyworkout create a proper readme md using bookmarked templates should have many areas of key program information ideas
0
22,490
31,464,794,586
IssuesEvent
2023-08-30 00:30:35
FasterXML/jackson-core
https://api.github.com/repos/FasterXML/jackson-core
closed
Add configurable limit for the maximum length of Object property names to parse before failing
2.16 processing-limits
(note: part of #637) Similar to token-length limits for Numbers (see #815) and Strings (see #863), we need an option to limit maximum Object Property name length to something reasonable, as otherwise name tokenization can become performance issue for very long names. As with other limits, units would related to underlying input units -- `byte`s or `char`s depending on input source. We need to decide on reasonable defaults to use: my initial suggestion is to limit to 64k but this probably needs to correlated to: 1. Performance-characteristics (can we find a pivot point where length increase has super-linear negative effect?) 2. Longest legit (non-malicious) property names used in the wild. That is: figure out highest limit that works for DoS aspect, balanced with lowest limit that would be unlikely to break existing legit usage.
1.0
Add configurable limit for the maximum length of Object property names to parse before failing - (note: part of #637) Similar to token-length limits for Numbers (see #815) and Strings (see #863), we need an option to limit maximum Object Property name length to something reasonable, as otherwise name tokenization can become performance issue for very long names. As with other limits, units would related to underlying input units -- `byte`s or `char`s depending on input source. We need to decide on reasonable defaults to use: my initial suggestion is to limit to 64k but this probably needs to correlated to: 1. Performance-characteristics (can we find a pivot point where length increase has super-linear negative effect?) 2. Longest legit (non-malicious) property names used in the wild. That is: figure out highest limit that works for DoS aspect, balanced with lowest limit that would be unlikely to break existing legit usage.
process
add configurable limit for the maximum length of object property names to parse before failing note part of similar to token length limits for numbers see and strings see we need an option to limit maximum object property name length to something reasonable as otherwise name tokenization can become performance issue for very long names as with other limits units would related to underlying input units byte s or char s depending on input source we need to decide on reasonable defaults to use my initial suggestion is to limit to but this probably needs to correlated to performance characteristics can we find a pivot point where length increase has super linear negative effect longest legit non malicious property names used in the wild that is figure out highest limit that works for dos aspect balanced with lowest limit that would be unlikely to break existing legit usage
1
75,549
20,870,415,296
IssuesEvent
2022-03-22 11:23:44
harness-software/wp-graphql-gravity-forms
https://api.github.com/repos/harness-software/wp-graphql-gravity-forms
closed
[Bug] PHP Fatal error: Uncaught Error: Failed opening required .../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/autoload_real.php
bug 🐛 build scripts 👷‍♂️
<!-- Thank you for taking the time to report a possible bug. Remember, an issue is not the place to ask questions. You can use Slack for that, or start a topic in Github Discussions. --> ## Describe the bug <!-- Please write a clear and concise description of the bug. --> Installed and activated WPGraphQL. Installed and activated Gravity Forms. Downloaded the repo zip and uploaded to WordPress and activated. Get the following WordPress error (There has been a critical error on this website. Please check your site admin email inbox for instructions.): ``` PHP Fatal error: Uncaught Error: Failed opening required '/nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/../yahnis-elsts/plugin-update-checker/load-v4p11.php' (include_path='.:/usr/share/pear/php:/usr/share/php') in /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/autoload_real.php:78\nStack trace:\n#0 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/autoload_real.php(61): composerRequire0e8f6c0730ce438688b3484326e9667b('49a1299791c25c6...', '/nas/content/li...')\n#1 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/autoload.php(7): ComposerAutoloaderInit0e8f6c0730ce438688b3484326e9667b::getLoader()\n#2 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/src/GF.php(55): require_once('/nas/content/li...')\n#3 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/src/GF.php(34): WPGraphQL\\GF\\GF->includes()\n#4 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/wp-graphql-gravity-forms.php(90): WPGraphQL\\GF\\GF::instance()\n#5 /nas/content/live/../wp-includes/class-wp-hook.php(307): gf_graphql_init(Object(WPGraphQL))\n#6 /nas/content/live/../wp-includes/class-wp-hook.php(331): WP_Hook->apply_filters('', Array)\n#7 /nas/content/live/../wp-includes/plugin.php(474): WP_Hook->do_action(Array)\n#8 /nas/content/live/../wp-content/plugins/wp-graphql/src/WPGraphQL.php(255): do_action('graphql_init', Object(WPGraphQL))\n#9 /nas/content/live/../wp-includes/class-wp-hook.php(307): WPGraphQL->{closure}('')\n#10 /nas/content/live/../wp-includes/class-wp-hook.php(331): WP_Hook->apply_filters(NULL, Array)\n#11 /nas/content/live/../wp-includes/plugin.php(474): WP_Hook->do_action(Array)\n#12 /nas/content/live/../wp-settings.php(565): do_action('after_setup_the...')\n#13 /nas/content/live/../wp-config.php(119): require_once('/nas/content/li...')\n#14 /nas/content/live/../wp-load.php(50): require_once('/nas/content/li...')\n#15 /nas/content/live/../wp-admin/admin.php(34): require_once('/nas/content/li...')\n#16 /nas/content/live/../wp-admin/plugins.php(10): require_once('/nas/content/li...')\n#17 {main}\n thrown in /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/autoload_real.php on line 78, referer: https://...wpengine.com/wp-admin/update.php?action=upload-plugin PHP Fatal error: Uncaught Error: Failed opening required '/nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/../yahnis-elsts/plugin-update-checker/load-v4p11.php' (include_path='.:/usr/share/pear/php:/usr/share/php') in /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/autoload_real.php:78\nStack trace:\n#0 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/autoload_real.php(61): composerRequire0e8f6c0730ce438688b3484326e9667b('49a1299791c25c6...', '/nas/content/li...')\n#1 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/autoload.php(7): ComposerAutoloaderInit0e8f6c0730ce438688b3484326e9667b::getLoader()\n#2 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/src/GF.php(55): require_once('/nas/content/li...')\n#3 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/src/GF.php(34): WPGraphQL\\GF\\GF->includes()\n#4 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/wp-graphql-gravity-forms.php(90): WPGraphQL\\GF\\GF::instance()\n#5 /nas/content/live/../wp-includes/class-wp-hook.php(307): gf_graphql_init(Object(WPGraphQL))\n#6 /nas/content/live/../wp-includes/class-wp-hook.php(331): WP_Hook->apply_filters('', Array)\n#7 /nas/content/live/../wp-includes/plugin.php(474): WP_Hook->do_action(Array)\n#8 /nas/content/live/../wp-content/plugins/wp-graphql/src/WPGraphQL.php(255): do_action('graphql_init', Object(WPGraphQL))\n#9 /nas/content/live/../wp-includes/class-wp-hook.php(307): WPGraphQL->{closure}('')\n#10 /nas/content/live/../wp-includes/class-wp-hook.php(331): WP_Hook->apply_filters(NULL, Array)\n#11 /nas/content/live/../wp-includes/plugin.php(474): WP_Hook->do_action(Array)\n#12 /nas/content/live/../wp-settings.php(565): do_action('after_setup_the...')\n#13 /nas/content/live/../wp-config.php(119): require_once('/nas/content/li...')\n#14 /nas/content/live/../wp-load.php(50): require_once('/nas/content/li...')\n#15 /nas/content/live/../wp-admin/admin.php(34): require_once('/nas/content/li...')\n#16 /nas/content/live/../wp-admin/update.php(16): require_once('/nas/content/li...')\n#17 {main}\n thrown in /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/autoload_real.php on line 78, referer: https://...wpengine.com/wp-admin/plugin-install.php?s=wpgraphql&tab=search&type=term PHP Fatal error: Uncaught Error: Failed opening required '/nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/../yahnis-elsts/plugin-update-checker/load-v4p11.php' (include_path='.:/usr/share/pear/php:/usr/share/php') in /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/autoload_real.php:78\nStack trace:\n#0 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/autoload_real.php(61): composerRequire0e8f6c0730ce438688b3484326e9667b('49a1299791c25c6...', '/nas/content/li...')\n#1 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/autoload.php(7): ComposerAutoloaderInit0e8f6c0730ce438688b3484326e9667b::getLoader()\n#2 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/src/GF.php(55): require_once('/nas/content/li...')\n#3 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/src/GF.php(34): WPGraphQL\\GF\\GF->includes()\n#4 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/wp-graphql-gravity-forms.php(90): WPGraphQL\\GF\\GF::instance()\n#5 /nas/content/live/../wp-includes/class-wp-hook.php(307): gf_graphql_init(Object(WPGraphQL))\n#6 /nas/content/live/../wp-includes/class-wp-hook.php(331): WP_Hook->apply_filters('', Array)\n#7 /nas/content/live/../wp-includes/plugin.php(474): WP_Hook->do_action(Array)\n#8 /nas/content/live/../wp-content/plugins/wp-graphql/src/WPGraphQL.php(255): do_action('graphql_init', Object(WPGraphQL))\n#9 /nas/content/live/../wp-includes/class-wp-hook.php(307): WPGraphQL->{closure}('')\n#10 /nas/content/live/../wp-includes/class-wp-hook.php(331): WP_Hook->apply_filters(NULL, Array)\n#11 /nas/content/live/../wp-includes/plugin.php(474): WP_Hook->do_action(Array)\n#12 /nas/content/live/../wp-settings.php(565): do_action('after_setup_the...')\n#13 /nas/content/live/../wp-config.php(119): require_once('/nas/content/li...')\n#14 /nas/content/live/../wp-load.php(50): require_once('/nas/content/li...')\n#15 /nas/content/live/../wp-admin/admin.php(34): require_once('/nas/content/li...')\n#16 /nas/content/live/../wp-admin/plugin-install.php(16): require_once('/nas/content/li...')\n#17 {main}\n thrown in /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/autoload_real.php on line 78, referer: https://...wpengine.com/wp-admin/plugins.php ``` ## To Reproduce <!-- Please list the all the steps needed to reproduce the bug. Ideally, this should be in the form of a GraphQL snippet that can be used in GraphiQL IDE. --> 1. Install & activate WPGraphQL. 2. Install & activate Gravity Forms and any supported addons. 3. Download the zip of this repository and upload it to your WordPress install, and activate the plugin. ## Expected behavior <!--- Please write a clear and concise description of what you expected to happen. --> Plugin to not crash WordPress. ## Additional context Add any other context about the problem here. Hosted on WPEngine. ### Technical Info * WordPress version: 5.9.2 * WPGraphQL version: 1.7.2 * Gravity Forms version: 2.5.16 * WPGraphQL for Gravity Forms version: v0.10.5 * Other relevant frameworks/plugins: <html> <body> <!--StartFragment--> Plugin Name | Plugin Version -- | -- Advanced Custom Fields PRO | by Delicious Brains - 5.12 Atlas Content Modeler | by WP Engine - 0.14.0 Code Snippets | by Code Snippets Pro - 2.14.3 FaustWP | by WP Engine - 0.7.6 Genesis Custom Blocks | by Genesis Custom Blocks - 1.4.0 WordPress Importer | by wordpressdotorg - 0.7 WP GraphQL | by WPGraphQL - 1.7.2 WPGraphQL for Advanced Custom Fields | by WPGraphQL, Jason Bahl - 0.5.3 WP GraphQL Gutenberg | by pristas-peter - 0.3.11 <!--EndFragment--> </body> </html>
1.0
[Bug] PHP Fatal error: Uncaught Error: Failed opening required .../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/autoload_real.php - <!-- Thank you for taking the time to report a possible bug. Remember, an issue is not the place to ask questions. You can use Slack for that, or start a topic in Github Discussions. --> ## Describe the bug <!-- Please write a clear and concise description of the bug. --> Installed and activated WPGraphQL. Installed and activated Gravity Forms. Downloaded the repo zip and uploaded to WordPress and activated. Get the following WordPress error (There has been a critical error on this website. Please check your site admin email inbox for instructions.): ``` PHP Fatal error: Uncaught Error: Failed opening required '/nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/../yahnis-elsts/plugin-update-checker/load-v4p11.php' (include_path='.:/usr/share/pear/php:/usr/share/php') in /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/autoload_real.php:78\nStack trace:\n#0 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/autoload_real.php(61): composerRequire0e8f6c0730ce438688b3484326e9667b('49a1299791c25c6...', '/nas/content/li...')\n#1 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/autoload.php(7): ComposerAutoloaderInit0e8f6c0730ce438688b3484326e9667b::getLoader()\n#2 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/src/GF.php(55): require_once('/nas/content/li...')\n#3 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/src/GF.php(34): WPGraphQL\\GF\\GF->includes()\n#4 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/wp-graphql-gravity-forms.php(90): WPGraphQL\\GF\\GF::instance()\n#5 /nas/content/live/../wp-includes/class-wp-hook.php(307): gf_graphql_init(Object(WPGraphQL))\n#6 /nas/content/live/../wp-includes/class-wp-hook.php(331): WP_Hook->apply_filters('', Array)\n#7 /nas/content/live/../wp-includes/plugin.php(474): WP_Hook->do_action(Array)\n#8 /nas/content/live/../wp-content/plugins/wp-graphql/src/WPGraphQL.php(255): do_action('graphql_init', Object(WPGraphQL))\n#9 /nas/content/live/../wp-includes/class-wp-hook.php(307): WPGraphQL->{closure}('')\n#10 /nas/content/live/../wp-includes/class-wp-hook.php(331): WP_Hook->apply_filters(NULL, Array)\n#11 /nas/content/live/../wp-includes/plugin.php(474): WP_Hook->do_action(Array)\n#12 /nas/content/live/../wp-settings.php(565): do_action('after_setup_the...')\n#13 /nas/content/live/../wp-config.php(119): require_once('/nas/content/li...')\n#14 /nas/content/live/../wp-load.php(50): require_once('/nas/content/li...')\n#15 /nas/content/live/../wp-admin/admin.php(34): require_once('/nas/content/li...')\n#16 /nas/content/live/../wp-admin/plugins.php(10): require_once('/nas/content/li...')\n#17 {main}\n thrown in /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/autoload_real.php on line 78, referer: https://...wpengine.com/wp-admin/update.php?action=upload-plugin PHP Fatal error: Uncaught Error: Failed opening required '/nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/../yahnis-elsts/plugin-update-checker/load-v4p11.php' (include_path='.:/usr/share/pear/php:/usr/share/php') in /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/autoload_real.php:78\nStack trace:\n#0 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/autoload_real.php(61): composerRequire0e8f6c0730ce438688b3484326e9667b('49a1299791c25c6...', '/nas/content/li...')\n#1 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/autoload.php(7): ComposerAutoloaderInit0e8f6c0730ce438688b3484326e9667b::getLoader()\n#2 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/src/GF.php(55): require_once('/nas/content/li...')\n#3 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/src/GF.php(34): WPGraphQL\\GF\\GF->includes()\n#4 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/wp-graphql-gravity-forms.php(90): WPGraphQL\\GF\\GF::instance()\n#5 /nas/content/live/../wp-includes/class-wp-hook.php(307): gf_graphql_init(Object(WPGraphQL))\n#6 /nas/content/live/../wp-includes/class-wp-hook.php(331): WP_Hook->apply_filters('', Array)\n#7 /nas/content/live/../wp-includes/plugin.php(474): WP_Hook->do_action(Array)\n#8 /nas/content/live/../wp-content/plugins/wp-graphql/src/WPGraphQL.php(255): do_action('graphql_init', Object(WPGraphQL))\n#9 /nas/content/live/../wp-includes/class-wp-hook.php(307): WPGraphQL->{closure}('')\n#10 /nas/content/live/../wp-includes/class-wp-hook.php(331): WP_Hook->apply_filters(NULL, Array)\n#11 /nas/content/live/../wp-includes/plugin.php(474): WP_Hook->do_action(Array)\n#12 /nas/content/live/../wp-settings.php(565): do_action('after_setup_the...')\n#13 /nas/content/live/../wp-config.php(119): require_once('/nas/content/li...')\n#14 /nas/content/live/../wp-load.php(50): require_once('/nas/content/li...')\n#15 /nas/content/live/../wp-admin/admin.php(34): require_once('/nas/content/li...')\n#16 /nas/content/live/../wp-admin/update.php(16): require_once('/nas/content/li...')\n#17 {main}\n thrown in /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/autoload_real.php on line 78, referer: https://...wpengine.com/wp-admin/plugin-install.php?s=wpgraphql&tab=search&type=term PHP Fatal error: Uncaught Error: Failed opening required '/nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/../yahnis-elsts/plugin-update-checker/load-v4p11.php' (include_path='.:/usr/share/pear/php:/usr/share/php') in /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/autoload_real.php:78\nStack trace:\n#0 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/autoload_real.php(61): composerRequire0e8f6c0730ce438688b3484326e9667b('49a1299791c25c6...', '/nas/content/li...')\n#1 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/autoload.php(7): ComposerAutoloaderInit0e8f6c0730ce438688b3484326e9667b::getLoader()\n#2 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/src/GF.php(55): require_once('/nas/content/li...')\n#3 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/src/GF.php(34): WPGraphQL\\GF\\GF->includes()\n#4 /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/wp-graphql-gravity-forms.php(90): WPGraphQL\\GF\\GF::instance()\n#5 /nas/content/live/../wp-includes/class-wp-hook.php(307): gf_graphql_init(Object(WPGraphQL))\n#6 /nas/content/live/../wp-includes/class-wp-hook.php(331): WP_Hook->apply_filters('', Array)\n#7 /nas/content/live/../wp-includes/plugin.php(474): WP_Hook->do_action(Array)\n#8 /nas/content/live/../wp-content/plugins/wp-graphql/src/WPGraphQL.php(255): do_action('graphql_init', Object(WPGraphQL))\n#9 /nas/content/live/../wp-includes/class-wp-hook.php(307): WPGraphQL->{closure}('')\n#10 /nas/content/live/../wp-includes/class-wp-hook.php(331): WP_Hook->apply_filters(NULL, Array)\n#11 /nas/content/live/../wp-includes/plugin.php(474): WP_Hook->do_action(Array)\n#12 /nas/content/live/../wp-settings.php(565): do_action('after_setup_the...')\n#13 /nas/content/live/../wp-config.php(119): require_once('/nas/content/li...')\n#14 /nas/content/live/../wp-load.php(50): require_once('/nas/content/li...')\n#15 /nas/content/live/../wp-admin/admin.php(34): require_once('/nas/content/li...')\n#16 /nas/content/live/../wp-admin/plugin-install.php(16): require_once('/nas/content/li...')\n#17 {main}\n thrown in /nas/content/live/../wp-content/plugins/wp-graphql-gravity-forms-develop/vendor/composer/autoload_real.php on line 78, referer: https://...wpengine.com/wp-admin/plugins.php ``` ## To Reproduce <!-- Please list the all the steps needed to reproduce the bug. Ideally, this should be in the form of a GraphQL snippet that can be used in GraphiQL IDE. --> 1. Install & activate WPGraphQL. 2. Install & activate Gravity Forms and any supported addons. 3. Download the zip of this repository and upload it to your WordPress install, and activate the plugin. ## Expected behavior <!--- Please write a clear and concise description of what you expected to happen. --> Plugin to not crash WordPress. ## Additional context Add any other context about the problem here. Hosted on WPEngine. ### Technical Info * WordPress version: 5.9.2 * WPGraphQL version: 1.7.2 * Gravity Forms version: 2.5.16 * WPGraphQL for Gravity Forms version: v0.10.5 * Other relevant frameworks/plugins: <html> <body> <!--StartFragment--> Plugin Name | Plugin Version -- | -- Advanced Custom Fields PRO | by Delicious Brains - 5.12 Atlas Content Modeler | by WP Engine - 0.14.0 Code Snippets | by Code Snippets Pro - 2.14.3 FaustWP | by WP Engine - 0.7.6 Genesis Custom Blocks | by Genesis Custom Blocks - 1.4.0 WordPress Importer | by wordpressdotorg - 0.7 WP GraphQL | by WPGraphQL - 1.7.2 WPGraphQL for Advanced Custom Fields | by WPGraphQL, Jason Bahl - 0.5.3 WP GraphQL Gutenberg | by pristas-peter - 0.3.11 <!--EndFragment--> </body> </html>
non_process
php fatal error uncaught error failed opening required wp content plugins wp graphql gravity forms develop vendor composer autoload real php thank you for taking the time to report a possible bug remember an issue is not the place to ask questions you can use slack for that or start a topic in github discussions describe the bug installed and activated wpgraphql installed and activated gravity forms downloaded the repo zip and uploaded to wordpress and activated get the following wordpress error there has been a critical error on this website please check your site admin email inbox for instructions php fatal error uncaught error failed opening required nas content live wp content plugins wp graphql gravity forms develop vendor composer yahnis elsts plugin update checker load php include path usr share pear php usr share php in nas content live wp content plugins wp graphql gravity forms develop vendor composer autoload real php nstack trace n nas content live wp content plugins wp graphql gravity forms develop vendor composer autoload real php nas content li n nas content live wp content plugins wp graphql gravity forms develop vendor autoload php getloader n nas content live wp content plugins wp graphql gravity forms develop src gf php require once nas content li n nas content live wp content plugins wp graphql gravity forms develop src gf php wpgraphql gf gf includes n nas content live wp content plugins wp graphql gravity forms develop wp graphql gravity forms php wpgraphql gf gf instance n nas content live wp includes class wp hook php gf graphql init object wpgraphql n nas content live wp includes class wp hook php wp hook apply filters array n nas content live wp includes plugin php wp hook do action array n nas content live wp content plugins wp graphql src wpgraphql php do action graphql init object wpgraphql n nas content live wp includes class wp hook php wpgraphql closure n nas content live wp includes class wp hook php wp hook apply filters null array n nas content live wp includes plugin php wp hook do action array n nas content live wp settings php do action after setup the n nas content live wp config php require once nas content li n nas content live wp load php require once nas content li n nas content live wp admin admin php require once nas content li n nas content live wp admin plugins php require once nas content li n main n thrown in nas content live wp content plugins wp graphql gravity forms develop vendor composer autoload real php on line referer php fatal error uncaught error failed opening required nas content live wp content plugins wp graphql gravity forms develop vendor composer yahnis elsts plugin update checker load php include path usr share pear php usr share php in nas content live wp content plugins wp graphql gravity forms develop vendor composer autoload real php nstack trace n nas content live wp content plugins wp graphql gravity forms develop vendor composer autoload real php nas content li n nas content live wp content plugins wp graphql gravity forms develop vendor autoload php getloader n nas content live wp content plugins wp graphql gravity forms develop src gf php require once nas content li n nas content live wp content plugins wp graphql gravity forms develop src gf php wpgraphql gf gf includes n nas content live wp content plugins wp graphql gravity forms develop wp graphql gravity forms php wpgraphql gf gf instance n nas content live wp includes class wp hook php gf graphql init object wpgraphql n nas content live wp includes class wp hook php wp hook apply filters array n nas content live wp includes plugin php wp hook do action array n nas content live wp content plugins wp graphql src wpgraphql php do action graphql init object wpgraphql n nas content live wp includes class wp hook php wpgraphql closure n nas content live wp includes class wp hook php wp hook apply filters null array n nas content live wp includes plugin php wp hook do action array n nas content live wp settings php do action after setup the n nas content live wp config php require once nas content li n nas content live wp load php require once nas content li n nas content live wp admin admin php require once nas content li n nas content live wp admin update php require once nas content li n main n thrown in nas content live wp content plugins wp graphql gravity forms develop vendor composer autoload real php on line referer php fatal error uncaught error failed opening required nas content live wp content plugins wp graphql gravity forms develop vendor composer yahnis elsts plugin update checker load php include path usr share pear php usr share php in nas content live wp content plugins wp graphql gravity forms develop vendor composer autoload real php nstack trace n nas content live wp content plugins wp graphql gravity forms develop vendor composer autoload real php nas content li n nas content live wp content plugins wp graphql gravity forms develop vendor autoload php getloader n nas content live wp content plugins wp graphql gravity forms develop src gf php require once nas content li n nas content live wp content plugins wp graphql gravity forms develop src gf php wpgraphql gf gf includes n nas content live wp content plugins wp graphql gravity forms develop wp graphql gravity forms php wpgraphql gf gf instance n nas content live wp includes class wp hook php gf graphql init object wpgraphql n nas content live wp includes class wp hook php wp hook apply filters array n nas content live wp includes plugin php wp hook do action array n nas content live wp content plugins wp graphql src wpgraphql php do action graphql init object wpgraphql n nas content live wp includes class wp hook php wpgraphql closure n nas content live wp includes class wp hook php wp hook apply filters null array n nas content live wp includes plugin php wp hook do action array n nas content live wp settings php do action after setup the n nas content live wp config php require once nas content li n nas content live wp load php require once nas content li n nas content live wp admin admin php require once nas content li n nas content live wp admin plugin install php require once nas content li n main n thrown in nas content live wp content plugins wp graphql gravity forms develop vendor composer autoload real php on line referer to reproduce please list the all the steps needed to reproduce the bug ideally this should be in the form of a graphql snippet that can be used in graphiql ide install activate wpgraphql install activate gravity forms and any supported addons download the zip of this repository and upload it to your wordpress install and activate the plugin expected behavior plugin to not crash wordpress additional context add any other context about the problem here hosted on wpengine technical info wordpress version wpgraphql version gravity forms version wpgraphql for gravity forms version other relevant frameworks plugins plugin name plugin version advanced custom fields pro by delicious brains  atlas content modeler by wp engine  code snippets by code snippets pro  faustwp by wp engine  genesis custom blocks by genesis custom blocks  wordpress importer by wordpressdotorg  wp graphql by wpgraphql  wpgraphql for advanced custom fields by wpgraphql jason bahl  wp graphql gutenberg by pristas peter
0
13,252
15,724,249,674
IssuesEvent
2021-03-29 08:34:02
q191201771/lal
https://api.github.com/repos/q191201771/lal
closed
期望老板 rtsp的支持
#Feature *In process
rtsp进入lal; 从lal拉rtsp的流出来 ; 以及rtsp的流进来,rtmp的流出来(包括flv,hls); and 从lal拉rtsp/rtmp/flv/hls; 无论 是从哪个协议过来的; 期望的结果是rtsp得到很好的支持. 谢谢老板 .
1.0
期望老板 rtsp的支持 - rtsp进入lal; 从lal拉rtsp的流出来 ; 以及rtsp的流进来,rtmp的流出来(包括flv,hls); and 从lal拉rtsp/rtmp/flv/hls; 无论 是从哪个协议过来的; 期望的结果是rtsp得到很好的支持. 谢谢老板 .
process
期望老板 rtsp的支持 rtsp进入lal 从lal拉rtsp的流出来 以及rtsp的流进来 rtmp的流出来 包括flv hls and 从lal拉rtsp rtmp flv hls 无论 是从哪个协议过来的 期望的结果是rtsp得到很好的支持 谢谢老板
1
9,403
12,401,307,302
IssuesEvent
2020-05-21 09:36:26
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
closed
Deploy Blockchain ETL
P1 process
**Components:** - Mirror node importer (publishes to pubsub) - Dataflow jobs - Dedupe job **Resources needed:** 1. BigQuery tables : transactions, errors, dedupe_state 2. PubSub topic for transactions 3. GCS bucket : Used for dataflow templates, staging and as temp location 4. ETL Pipeline from PubSub to BigQuery: - PubSub subscription - Service account with following roles: BigQuery Data Editor, Dataflow Worker, Pub/Sub Subscriber, and Storage Admin 5. Deduplication Task - Service account with following roles: BigQuery Data Editor, BigQuery Job User, Monitoring Metric Writer 6. Mirror Importer - Service account with following roles: PubSub Publisher 7. (Optional) ETL Pipeline from PubSub to GCS - GCS Bucket: For output of pipeline - Service account with following roles: Dataflow Worker, Pub/Sub Editor (for creating subscription), and Storage Admin
1.0
Deploy Blockchain ETL - **Components:** - Mirror node importer (publishes to pubsub) - Dataflow jobs - Dedupe job **Resources needed:** 1. BigQuery tables : transactions, errors, dedupe_state 2. PubSub topic for transactions 3. GCS bucket : Used for dataflow templates, staging and as temp location 4. ETL Pipeline from PubSub to BigQuery: - PubSub subscription - Service account with following roles: BigQuery Data Editor, Dataflow Worker, Pub/Sub Subscriber, and Storage Admin 5. Deduplication Task - Service account with following roles: BigQuery Data Editor, BigQuery Job User, Monitoring Metric Writer 6. Mirror Importer - Service account with following roles: PubSub Publisher 7. (Optional) ETL Pipeline from PubSub to GCS - GCS Bucket: For output of pipeline - Service account with following roles: Dataflow Worker, Pub/Sub Editor (for creating subscription), and Storage Admin
process
deploy blockchain etl components mirror node importer publishes to pubsub dataflow jobs dedupe job resources needed bigquery tables transactions errors dedupe state pubsub topic for transactions gcs bucket used for dataflow templates staging and as temp location etl pipeline from pubsub to bigquery pubsub subscription service account with following roles bigquery data editor dataflow worker pub sub subscriber and storage admin deduplication task service account with following roles bigquery data editor bigquery job user monitoring metric writer mirror importer service account with following roles pubsub publisher optional etl pipeline from pubsub to gcs gcs bucket for output of pipeline service account with following roles dataflow worker pub sub editor for creating subscription and storage admin
1
75,993
3,479,425,879
IssuesEvent
2015-12-28 20:12:36
Mobicents/RestComm
https://api.github.com/repos/Mobicents/RestComm
opened
Dialogic JSR309 cannot join call legs to conference
1. Bug High-Priority XMS-1.0.0
The Dialogic JSR309 driver (5.0-alpha) has problems joining all call legs to a conference bridge. The problem was first noticed when performing load tests for conference scenarios. Problem is noticeable even with small call rates (10 CPS)
1.0
Dialogic JSR309 cannot join call legs to conference - The Dialogic JSR309 driver (5.0-alpha) has problems joining all call legs to a conference bridge. The problem was first noticed when performing load tests for conference scenarios. Problem is noticeable even with small call rates (10 CPS)
non_process
dialogic cannot join call legs to conference the dialogic driver alpha has problems joining all call legs to a conference bridge the problem was first noticed when performing load tests for conference scenarios problem is noticeable even with small call rates cps
0
667,490
22,475,128,664
IssuesEvent
2022-06-22 11:36:59
ctm/mb2-doc
https://api.github.com/repos/ctm/mb2-doc
closed
rebuy doesn't get player new lammers
bug high priority easy
Fix so that rebuy lammer tournaments hand out the appropriate number of lammers on a rebuy. Currently, only the WSOP-Style 2022 Short-Deck No-Limit tournament is the only rebuy tournament that uses lammers. However, due to me not thinking about lammers when I partially implemented rebuys a long time ago, players who rebuy only get T20,000 rather than the T20,000 and two lammers that people start with. This should be trivial to implement. It was just an oversight and not a technical limitation.
1.0
rebuy doesn't get player new lammers - Fix so that rebuy lammer tournaments hand out the appropriate number of lammers on a rebuy. Currently, only the WSOP-Style 2022 Short-Deck No-Limit tournament is the only rebuy tournament that uses lammers. However, due to me not thinking about lammers when I partially implemented rebuys a long time ago, players who rebuy only get T20,000 rather than the T20,000 and two lammers that people start with. This should be trivial to implement. It was just an oversight and not a technical limitation.
non_process
rebuy doesn t get player new lammers fix so that rebuy lammer tournaments hand out the appropriate number of lammers on a rebuy currently only the wsop style short deck no limit tournament is the only rebuy tournament that uses lammers however due to me not thinking about lammers when i partially implemented rebuys a long time ago players who rebuy only get rather than the and two lammers that people start with this should be trivial to implement it was just an oversight and not a technical limitation
0
92,261
11,616,544,194
IssuesEvent
2020-02-26 15:52:33
department-of-veterans-affairs/caseflow
https://api.github.com/repos/department-of-veterans-affairs/caseflow
opened
Motions to Vacate - URL instructions copy (for motions attorney and VLJ)
design 💅 foxtrot
## User story <!-- Include a user story, if it exists. Be specific about which users you're talking about, please. --> Motions Attorneys and Judges need a way to understand the value of adding their draft and signed denial/dismissal letters in the Shared Drive to help streamline communication and their workflow. ## Problem statement <!-- Describe the problem the design, writing, or research is intended to solve. --> Litigation Support requested the feature of using URLs to handle denial and dismissal draft letters between motions attorneys and VLJs. Currently, it is unknown where the files will be stored, and how access will be handled, so the team has decided to make this feature **optional**. For now, we will add instructions copy to help users understand the value of using URLs in the handoff process. ## What is out of scope? <!-- This can be particularly helpful for research tickets. Does not need to be an exhaustive list, but should clearly define the boundaries of the work --> Making the URL field required. ## Background/context <!-- Why are we designing/writing this? Who is it for? What research has been done that tells us this needs to be designed, written, or researched? --> This is a key feature requested by Litigation Support, and is an important element within the designs for denied/dismissed Motions to Vacate. We have validated the need for this, however are still unsure how it will work in reality and have thus decided to make it an optional field. We would still like to encourage users to use this. ## What are the unknowns? <!-- If there are key unknowns or assumptions, add them here. If we're accepting the risks associated with the unknowns or assumptions, let us know that too. --> - Do users really pay attention to help-text? - Is help-text enough to communicate this process? - What information would be most useful - explaining the process, the value of it, or both? ## Existing design and content <!--If we're replacing something, please include a screenshot(s) or link to the existing solution, even if it's not in Caseflow. --> The existing designs are attached. ![Address MTV _ Deny](https://user-images.githubusercontent.com/51714306/75361927-03277180-5886-11ea-8ef9-63a72965bf76.jpg) ![Motions Attorney_Review_Deny](https://user-images.githubusercontent.com/51714306/75361929-03277180-5886-11ea-9ed9-7ea43987954b.jpg) ## Success criteria <!-- Include as needed, especially for issues that aren't part of epics. if no measurable success criteria, what does success look like? --> - Designed screen added and shared with engineers.
1.0
Motions to Vacate - URL instructions copy (for motions attorney and VLJ) - ## User story <!-- Include a user story, if it exists. Be specific about which users you're talking about, please. --> Motions Attorneys and Judges need a way to understand the value of adding their draft and signed denial/dismissal letters in the Shared Drive to help streamline communication and their workflow. ## Problem statement <!-- Describe the problem the design, writing, or research is intended to solve. --> Litigation Support requested the feature of using URLs to handle denial and dismissal draft letters between motions attorneys and VLJs. Currently, it is unknown where the files will be stored, and how access will be handled, so the team has decided to make this feature **optional**. For now, we will add instructions copy to help users understand the value of using URLs in the handoff process. ## What is out of scope? <!-- This can be particularly helpful for research tickets. Does not need to be an exhaustive list, but should clearly define the boundaries of the work --> Making the URL field required. ## Background/context <!-- Why are we designing/writing this? Who is it for? What research has been done that tells us this needs to be designed, written, or researched? --> This is a key feature requested by Litigation Support, and is an important element within the designs for denied/dismissed Motions to Vacate. We have validated the need for this, however are still unsure how it will work in reality and have thus decided to make it an optional field. We would still like to encourage users to use this. ## What are the unknowns? <!-- If there are key unknowns or assumptions, add them here. If we're accepting the risks associated with the unknowns or assumptions, let us know that too. --> - Do users really pay attention to help-text? - Is help-text enough to communicate this process? - What information would be most useful - explaining the process, the value of it, or both? ## Existing design and content <!--If we're replacing something, please include a screenshot(s) or link to the existing solution, even if it's not in Caseflow. --> The existing designs are attached. ![Address MTV _ Deny](https://user-images.githubusercontent.com/51714306/75361927-03277180-5886-11ea-8ef9-63a72965bf76.jpg) ![Motions Attorney_Review_Deny](https://user-images.githubusercontent.com/51714306/75361929-03277180-5886-11ea-9ed9-7ea43987954b.jpg) ## Success criteria <!-- Include as needed, especially for issues that aren't part of epics. if no measurable success criteria, what does success look like? --> - Designed screen added and shared with engineers.
non_process
motions to vacate url instructions copy for motions attorney and vlj user story motions attorneys and judges need a way to understand the value of adding their draft and signed denial dismissal letters in the shared drive to help streamline communication and their workflow problem statement litigation support requested the feature of using urls to handle denial and dismissal draft letters between motions attorneys and vljs currently it is unknown where the files will be stored and how access will be handled so the team has decided to make this feature optional for now we will add instructions copy to help users understand the value of using urls in the handoff process what is out of scope making the url field required background context this is a key feature requested by litigation support and is an important element within the designs for denied dismissed motions to vacate we have validated the need for this however are still unsure how it will work in reality and have thus decided to make it an optional field we would still like to encourage users to use this what are the unknowns do users really pay attention to help text is help text enough to communicate this process what information would be most useful explaining the process the value of it or both existing design and content the existing designs are attached success criteria designed screen added and shared with engineers
0
301,168
22,739,127,750
IssuesEvent
2022-07-07 00:42:31
pedroAkiraDanno/C2
https://api.github.com/repos/pedroAkiraDanno/C2
opened
Organize / create folder and file about random
bug documentation duplicate
Organize / create folder and file about random because randon files are in q_sort files
1.0
Organize / create folder and file about random - Organize / create folder and file about random because randon files are in q_sort files
non_process
organize create folder and file about random organize create folder and file about random because randon files are in q sort files
0
32,174
12,093,639,998
IssuesEvent
2020-04-19 20:27:45
AOSC-Dev/aosc-os-abbs
https://api.github.com/repos/AOSC-Dev/aosc-os-abbs
closed
dovecot: security update to 2.3.7.2
aosa-pending security to-stable upgrade
<!-- Please remove items do not apply. --> **CVE IDs:** CVE-2019-11494, CVE-2019-11499, CVE-2019-11500 **Other security advisory IDs:** USN-3961-1, DSA-4510-1, USN-4110-1, GLSA 201908-29, ASA-201908-18 **Descriptions:** https://dovecot.org/list/dovecot-news/2019-April/000408.html * CVE-2019-11494: Submission-login crashed with signal 11 due to null pointer access when authentication was aborted by disconnecting. * CVE-2019-11499: Submission-login crashed when authentication was started over TLS secured channel and invalid authentication message was sent. https://dovecot.org/pipermail/dovecot-news/2019-August/000415.html * CVE-2019-11500: IMAP protocol parser does not properly handle NUL byte when scanning data in quoted strings, leading to out of bounds heap memory writes. Found by Nick Roessler and Rafi Rubin. **Architectural progress:** <!-- Please remove any architecture to which the security vulnerabilities do not apply. --> - [x] AMD64 `amd64` - [x] AArch64 `arm64` - [x] ARMv7 `armel` - [x] PowerPC 64-bit BE `ppc64`
True
dovecot: security update to 2.3.7.2 - <!-- Please remove items do not apply. --> **CVE IDs:** CVE-2019-11494, CVE-2019-11499, CVE-2019-11500 **Other security advisory IDs:** USN-3961-1, DSA-4510-1, USN-4110-1, GLSA 201908-29, ASA-201908-18 **Descriptions:** https://dovecot.org/list/dovecot-news/2019-April/000408.html * CVE-2019-11494: Submission-login crashed with signal 11 due to null pointer access when authentication was aborted by disconnecting. * CVE-2019-11499: Submission-login crashed when authentication was started over TLS secured channel and invalid authentication message was sent. https://dovecot.org/pipermail/dovecot-news/2019-August/000415.html * CVE-2019-11500: IMAP protocol parser does not properly handle NUL byte when scanning data in quoted strings, leading to out of bounds heap memory writes. Found by Nick Roessler and Rafi Rubin. **Architectural progress:** <!-- Please remove any architecture to which the security vulnerabilities do not apply. --> - [x] AMD64 `amd64` - [x] AArch64 `arm64` - [x] ARMv7 `armel` - [x] PowerPC 64-bit BE `ppc64`
non_process
dovecot security update to cve ids cve cve cve other security advisory ids usn dsa usn glsa asa descriptions cve submission login crashed with signal due to null pointer access when authentication was aborted by disconnecting cve submission login crashed when authentication was started over tls secured channel and invalid authentication message was sent cve imap protocol parser does not properly handle nul byte when scanning data in quoted strings leading to out of bounds heap memory writes found by nick roessler and rafi rubin architectural progress armel powerpc bit be
0
412,166
12,036,041,815
IssuesEvent
2020-04-13 19:01:50
kubeflow/kubeflow
https://api.github.com/repos/kubeflow/kubeflow
closed
Istio ingress gateway errors in on-prem environment
kind/bug platform/onprem priority/p2
/kind bug **What steps did you take and what happened:** Installed with kfctl_k8s_istio.yaml on 3 nodes cluster with below specs - CentOS Linux release 7.6.1810 (Core) - kernel version 3.10.0-957.12.2.el7.x86_64 - Calico v3.8 - Kubernetes version: v1.15.7 The istio ingress gateway fails with the below error; ``` 2020-03-12 19:42:14.979][25][warning][config] [bazel-out/k8-opt/bin/external/envoy/source/common/config/_virtual_includes/grpc_stream_lib/common/config/grpc_stream.h:86] gRPC config stream closed: 14, no healthy upstream [2020-03-12 19:42:14.979][25][warning][config] [bazel-out/k8-opt/bin/external/envoy/source/common/config/_virtual_includes/grpc_stream_lib/common/config/grpc_stream.h:49] Unable to establish new stream 2020-03-12T19:42:18.083573Z info Envoy proxy is NOT ready: config not received from Pilot (is Pilot running?): cds updates: 1 successful, 0 rejected; lds updates: 0 successful, 0 rejected 2020-03-12T19:42:19.083333Z info Envoy proxy is NOT ready: config not received from Pilot (is Pilot running?): cds updates: 1 successful, 0 rejected; lds updates: 0 successful, 0 rejected 2020-03-12T19:42:21.781082Z info Envoy proxy is NOT ready: config not received from Pilot (is Pilot running?): cds updates: 1 successful, 0 rejected; lds updates: 0 successful, 0 rejected 2020-03-12T19:42:23.380790Z info Envoy proxy is NOT ready: config not received from Pilot (is Pilot running?): cds updates: 1 successful, 0 rejected; lds updates: 0 successful, 0 rejected 2020-03-12T19:42:25.276445Z info Envoy proxy is NOT ready: config not received from Pilot (is Pilot running?): cds updates: 1 successful, 0 rejected; lds updates: 0 successful, 0 rejected 2020-03-12T19:42:28.948991Z info Envoy proxy is NOT ready: failed retrieving Envoy stats: Get http://127.0.0.1:15000/stats?usedonly: net/http: request canceled (Client.Timeout exceeded while awaiting headers) 2020-03-12T19:42:29.275745Z info Envoy proxy is ready 2020-03-12T19:43:11.074793Z warn Epoch 0 terminated with an error: signal: killed 2020-03-12T19:43:11.074831Z warn Aborted all epochs 2020-03-12T19:43:11.074929Z info Epoch 0: set retry delay to 200ms, budget to 9 2020-03-12T19:43:11.075056Z info Envoy proxy is NOT ready: failed retrieving Envoy stats: Get http://127.0.0.1:15000/stats?usedonly: dial tcp 127.0.0.1:15000: connect: connection refused 2020-03-12T19:43:11.275208Z info Reconciling retry (budget 9) 2020-03-12T19:43:11.275275Z info Epoch 0 starting 2020-03-12T19:43:11.276788Z info Envoy command: [-c /etc/istio/proxy/envoy-rev0.json --restart-epoch 0 --drain-time-s 45 --parent-shutdown-time-s 60 --service-cluster istio-ingressgateway --service-node router~10.41.32.206~istio-ingressgateway-589ff776dd-tdd2d.istio-system~istio-system.svc.cluster.local --max-obj-name-len 189 --allow-unknown-fields -l warning] [2020-03-12 19:43:11.581][97][warning][misc] [external/envoy/source/common/protobuf/utility.cc:174] Using deprecated option 'envoy.api.v2.Cluster.hosts' from file cds.proto. This configuration will be removed from Envoy soon. Please see https://www.envoyproxy.io/docs/envoy/latest/intro/deprecated for details. [2020-03-12 19:43:11.581][97][warning][misc] [external/envoy/source/common/protobuf/utility.cc:174] Using deprecated option 'envoy.api.v2.Cluster.hosts' from file cds.proto. This configuration will be removed from Envoy soon. Please see https://www.envoyproxy.io/docs/envoy/latest/intro/deprecated for details. [2020-03-12 19:43:11.581][97][warning][misc] [external/envoy/source/common/protobuf/utility.cc:174] Using deprecated option 'envoy.api.v2.Cluster.hosts' from file cds.proto. This configuration will be removed from Envoy soon. Please see https://www.envoyproxy.io/docs/envoy/latest/intro/deprecated for details. [2020-03-12 19:43:11.681][97][warning][config] [bazel-out/k8-opt/bin/external/envoy/source/common/config/_virtual_includes/grpc_stream_lib/common/config/grpc_stream.h:86] gRPC config stream closed: 14, no healthy upstream [2020-03-12 19:43:11.681][97][warning][config] [bazel-out/k8-opt/bin/external/envoy/source/common/config/_virtual_includes/grpc_stream_lib/common/config/grpc_stream.h:49] Unable to establish new stream 2020-03-12T19:43:13.575433Z info Envoy proxy is NOT ready: config not received from Pilot (is Pilot running?): cds updates: 1 successful, 0 rejected; lds updates: 0 successful, 0 rejected 2020-03-12T19:43:15.780870Z info Envoy proxy is NOT ready: config not received from Pilot (is Pilot running?): cds updates: 1 successful, 0 rejected; lds updates: 0 successful, 0 rejected 2020-03-12T19:43:17.379880Z info Envoy proxy is NOT ready: config not received from Pilot (is Pilot running?): cds updates: 1 successful, 0 rejected; lds updates: 0 successful, 0 rejected 2020-03-12T19:43:19.378874Z info Envoy proxy is NOT ready: config not received from Pilot (is Pilot running?): cds updates: 1 successful, 0 rejected; lds updates: 0 successful, 0 rejected 2020-03-12T19:43:21.274930Z info Envoy proxy is NOT ready: config not received from Pilot (is Pilot running?): cds updates: 1 successful, 0 rejected; lds updates: 0 successful, 0 rejected 2020-03-12T19:43:22.950029Z info Envoy proxy is NOT ready: config not received from Pilot (is Pilot running?): cds updates: 1 successful, 0 rejected; lds updates: 0 successful, 0 rejected 2020-03-12T19:43:26.974110Z info Envoy proxy is ready 2020-03-12T19:43:28.479465Z warn Epoch 0 terminated with an error: signal: killed 2020-03-12T19:43:28.479503Z warn Aborted all epochs 2020-03-12T19:43:28.479569Z info Epoch 0: set retry delay to 400ms, budget to 8 2020-03-12T19:43:28.479867Z info Envoy proxy is NOT ready: failed retrieving Envoy stats: Get http://127.0.0.1:15000/stats?usedonly: dial tcp 127.0.0.1:15000: connect: connection refused 2020-03-12T19:43:28.879727Z info Reconciling retry (budget 82020-03-12T19:43:28.879845Z info Epoch 0 starting 2020-03-12T19:43:28.881489Z info Envoy command: [-c /etc/istio/proxy/envoy-rev0.json --restart-epoch 0 --drain-time-s 45 --parent-shutdown-time-s 60 --service-cluster istio-ingressgateway --service-node router~10.41.32.206~istio-ingressgateway-589ff776dd-tdd2d.istio-system~istio-system.svc.cluster.local --max-obj-name-len 189 --allow-unknown-fields -l warning] 2020-03-12T19:43:28.948944Z info Envoy proxy is NOT ready: failed retrieving Envoy stats: Get http://127.0.0.1:15000/stats?usedonly: dial tcp 127.0.0.1:15000: connect: connection refused ``` Eventually, the ingress gateway pod will fail with `OOMKilled` **What did you expect to happen:** The ingress gateway to work without issue. **Environment:** - Kubeflow version: (version number can be found at the bottom left corner of the Kubeflow dashboard): - kfctl version: v1.0-rc.1-0-g963c787 - Kubernetes platform: v1.15.7 - Kubernetes version: v1.15.7 - OS (e.g. from `/etc/os-release`): NAME="CentOS Linux" VERSION="7 (Core)"
1.0
Istio ingress gateway errors in on-prem environment - /kind bug **What steps did you take and what happened:** Installed with kfctl_k8s_istio.yaml on 3 nodes cluster with below specs - CentOS Linux release 7.6.1810 (Core) - kernel version 3.10.0-957.12.2.el7.x86_64 - Calico v3.8 - Kubernetes version: v1.15.7 The istio ingress gateway fails with the below error; ``` 2020-03-12 19:42:14.979][25][warning][config] [bazel-out/k8-opt/bin/external/envoy/source/common/config/_virtual_includes/grpc_stream_lib/common/config/grpc_stream.h:86] gRPC config stream closed: 14, no healthy upstream [2020-03-12 19:42:14.979][25][warning][config] [bazel-out/k8-opt/bin/external/envoy/source/common/config/_virtual_includes/grpc_stream_lib/common/config/grpc_stream.h:49] Unable to establish new stream 2020-03-12T19:42:18.083573Z info Envoy proxy is NOT ready: config not received from Pilot (is Pilot running?): cds updates: 1 successful, 0 rejected; lds updates: 0 successful, 0 rejected 2020-03-12T19:42:19.083333Z info Envoy proxy is NOT ready: config not received from Pilot (is Pilot running?): cds updates: 1 successful, 0 rejected; lds updates: 0 successful, 0 rejected 2020-03-12T19:42:21.781082Z info Envoy proxy is NOT ready: config not received from Pilot (is Pilot running?): cds updates: 1 successful, 0 rejected; lds updates: 0 successful, 0 rejected 2020-03-12T19:42:23.380790Z info Envoy proxy is NOT ready: config not received from Pilot (is Pilot running?): cds updates: 1 successful, 0 rejected; lds updates: 0 successful, 0 rejected 2020-03-12T19:42:25.276445Z info Envoy proxy is NOT ready: config not received from Pilot (is Pilot running?): cds updates: 1 successful, 0 rejected; lds updates: 0 successful, 0 rejected 2020-03-12T19:42:28.948991Z info Envoy proxy is NOT ready: failed retrieving Envoy stats: Get http://127.0.0.1:15000/stats?usedonly: net/http: request canceled (Client.Timeout exceeded while awaiting headers) 2020-03-12T19:42:29.275745Z info Envoy proxy is ready 2020-03-12T19:43:11.074793Z warn Epoch 0 terminated with an error: signal: killed 2020-03-12T19:43:11.074831Z warn Aborted all epochs 2020-03-12T19:43:11.074929Z info Epoch 0: set retry delay to 200ms, budget to 9 2020-03-12T19:43:11.075056Z info Envoy proxy is NOT ready: failed retrieving Envoy stats: Get http://127.0.0.1:15000/stats?usedonly: dial tcp 127.0.0.1:15000: connect: connection refused 2020-03-12T19:43:11.275208Z info Reconciling retry (budget 9) 2020-03-12T19:43:11.275275Z info Epoch 0 starting 2020-03-12T19:43:11.276788Z info Envoy command: [-c /etc/istio/proxy/envoy-rev0.json --restart-epoch 0 --drain-time-s 45 --parent-shutdown-time-s 60 --service-cluster istio-ingressgateway --service-node router~10.41.32.206~istio-ingressgateway-589ff776dd-tdd2d.istio-system~istio-system.svc.cluster.local --max-obj-name-len 189 --allow-unknown-fields -l warning] [2020-03-12 19:43:11.581][97][warning][misc] [external/envoy/source/common/protobuf/utility.cc:174] Using deprecated option 'envoy.api.v2.Cluster.hosts' from file cds.proto. This configuration will be removed from Envoy soon. Please see https://www.envoyproxy.io/docs/envoy/latest/intro/deprecated for details. [2020-03-12 19:43:11.581][97][warning][misc] [external/envoy/source/common/protobuf/utility.cc:174] Using deprecated option 'envoy.api.v2.Cluster.hosts' from file cds.proto. This configuration will be removed from Envoy soon. Please see https://www.envoyproxy.io/docs/envoy/latest/intro/deprecated for details. [2020-03-12 19:43:11.581][97][warning][misc] [external/envoy/source/common/protobuf/utility.cc:174] Using deprecated option 'envoy.api.v2.Cluster.hosts' from file cds.proto. This configuration will be removed from Envoy soon. Please see https://www.envoyproxy.io/docs/envoy/latest/intro/deprecated for details. [2020-03-12 19:43:11.681][97][warning][config] [bazel-out/k8-opt/bin/external/envoy/source/common/config/_virtual_includes/grpc_stream_lib/common/config/grpc_stream.h:86] gRPC config stream closed: 14, no healthy upstream [2020-03-12 19:43:11.681][97][warning][config] [bazel-out/k8-opt/bin/external/envoy/source/common/config/_virtual_includes/grpc_stream_lib/common/config/grpc_stream.h:49] Unable to establish new stream 2020-03-12T19:43:13.575433Z info Envoy proxy is NOT ready: config not received from Pilot (is Pilot running?): cds updates: 1 successful, 0 rejected; lds updates: 0 successful, 0 rejected 2020-03-12T19:43:15.780870Z info Envoy proxy is NOT ready: config not received from Pilot (is Pilot running?): cds updates: 1 successful, 0 rejected; lds updates: 0 successful, 0 rejected 2020-03-12T19:43:17.379880Z info Envoy proxy is NOT ready: config not received from Pilot (is Pilot running?): cds updates: 1 successful, 0 rejected; lds updates: 0 successful, 0 rejected 2020-03-12T19:43:19.378874Z info Envoy proxy is NOT ready: config not received from Pilot (is Pilot running?): cds updates: 1 successful, 0 rejected; lds updates: 0 successful, 0 rejected 2020-03-12T19:43:21.274930Z info Envoy proxy is NOT ready: config not received from Pilot (is Pilot running?): cds updates: 1 successful, 0 rejected; lds updates: 0 successful, 0 rejected 2020-03-12T19:43:22.950029Z info Envoy proxy is NOT ready: config not received from Pilot (is Pilot running?): cds updates: 1 successful, 0 rejected; lds updates: 0 successful, 0 rejected 2020-03-12T19:43:26.974110Z info Envoy proxy is ready 2020-03-12T19:43:28.479465Z warn Epoch 0 terminated with an error: signal: killed 2020-03-12T19:43:28.479503Z warn Aborted all epochs 2020-03-12T19:43:28.479569Z info Epoch 0: set retry delay to 400ms, budget to 8 2020-03-12T19:43:28.479867Z info Envoy proxy is NOT ready: failed retrieving Envoy stats: Get http://127.0.0.1:15000/stats?usedonly: dial tcp 127.0.0.1:15000: connect: connection refused 2020-03-12T19:43:28.879727Z info Reconciling retry (budget 82020-03-12T19:43:28.879845Z info Epoch 0 starting 2020-03-12T19:43:28.881489Z info Envoy command: [-c /etc/istio/proxy/envoy-rev0.json --restart-epoch 0 --drain-time-s 45 --parent-shutdown-time-s 60 --service-cluster istio-ingressgateway --service-node router~10.41.32.206~istio-ingressgateway-589ff776dd-tdd2d.istio-system~istio-system.svc.cluster.local --max-obj-name-len 189 --allow-unknown-fields -l warning] 2020-03-12T19:43:28.948944Z info Envoy proxy is NOT ready: failed retrieving Envoy stats: Get http://127.0.0.1:15000/stats?usedonly: dial tcp 127.0.0.1:15000: connect: connection refused ``` Eventually, the ingress gateway pod will fail with `OOMKilled` **What did you expect to happen:** The ingress gateway to work without issue. **Environment:** - Kubeflow version: (version number can be found at the bottom left corner of the Kubeflow dashboard): - kfctl version: v1.0-rc.1-0-g963c787 - Kubernetes platform: v1.15.7 - Kubernetes version: v1.15.7 - OS (e.g. from `/etc/os-release`): NAME="CentOS Linux" VERSION="7 (Core)"
non_process
istio ingress gateway errors in on prem environment kind bug what steps did you take and what happened installed with kfctl istio yaml on nodes cluster with below specs centos linux release core kernel version calico kubernetes version the istio ingress gateway fails with the below error grpc config stream closed no healthy upstream unable to establish new stream info envoy proxy is not ready config not received from pilot is pilot running cds updates successful rejected lds updates successful rejected info envoy proxy is not ready config not received from pilot is pilot running cds updates successful rejected lds updates successful rejected info envoy proxy is not ready config not received from pilot is pilot running cds updates successful rejected lds updates successful rejected info envoy proxy is not ready config not received from pilot is pilot running cds updates successful rejected lds updates successful rejected info envoy proxy is not ready config not received from pilot is pilot running cds updates successful rejected lds updates successful rejected info envoy proxy is not ready failed retrieving envoy stats get net http request canceled client timeout exceeded while awaiting headers info envoy proxy is ready warn epoch terminated with an error signal killed warn aborted all epochs info epoch set retry delay to budget to info envoy proxy is not ready failed retrieving envoy stats get dial tcp connect connection refused info reconciling retry budget info epoch starting info envoy command using deprecated option envoy api cluster hosts from file cds proto this configuration will be removed from envoy soon please see for details using deprecated option envoy api cluster hosts from file cds proto this configuration will be removed from envoy soon please see for details using deprecated option envoy api cluster hosts from file cds proto this configuration will be removed from envoy soon please see for details grpc config stream closed no healthy upstream unable to establish new stream info envoy proxy is not ready config not received from pilot is pilot running cds updates successful rejected lds updates successful rejected info envoy proxy is not ready config not received from pilot is pilot running cds updates successful rejected lds updates successful rejected info envoy proxy is not ready config not received from pilot is pilot running cds updates successful rejected lds updates successful rejected info envoy proxy is not ready config not received from pilot is pilot running cds updates successful rejected lds updates successful rejected info envoy proxy is not ready config not received from pilot is pilot running cds updates successful rejected lds updates successful rejected info envoy proxy is not ready config not received from pilot is pilot running cds updates successful rejected lds updates successful rejected info envoy proxy is ready warn epoch terminated with an error signal killed warn aborted all epochs info epoch set retry delay to budget to info envoy proxy is not ready failed retrieving envoy stats get dial tcp connect connection refused info reconciling retry budget info epoch starting info envoy command info envoy proxy is not ready failed retrieving envoy stats get dial tcp connect connection refused eventually the ingress gateway pod will fail with oomkilled what did you expect to happen the ingress gateway to work without issue environment kubeflow version version number can be found at the bottom left corner of the kubeflow dashboard kfctl version rc kubernetes platform kubernetes version os e g from etc os release name centos linux version core
0
21,670
30,116,734,071
IssuesEvent
2023-06-30 12:10:52
h4sh5/pypi-auto-scanner
https://api.github.com/repos/h4sh5/pypi-auto-scanner
opened
hpcflow-new2 0.2.0a58 has 1 GuardDog issues
guarddog silent-process-execution
https://pypi.org/project/hpcflow-new2 https://inspector.pypi.io/project/hpcflow-new2 ```{ "dependency": "hpcflow-new2", "version": "0.2.0a58", "result": { "issues": 1, "errors": {}, "results": { "silent-process-execution": [ { "location": "hpcflow_new2-0.2.0a58/hpcflow/sdk/helper/helper.py:118", "code": " proc = subprocess.Popen(\n args=args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n **kwargs,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpf1ieqxij/hpcflow-new2" } }```
1.0
hpcflow-new2 0.2.0a58 has 1 GuardDog issues - https://pypi.org/project/hpcflow-new2 https://inspector.pypi.io/project/hpcflow-new2 ```{ "dependency": "hpcflow-new2", "version": "0.2.0a58", "result": { "issues": 1, "errors": {}, "results": { "silent-process-execution": [ { "location": "hpcflow_new2-0.2.0a58/hpcflow/sdk/helper/helper.py:118", "code": " proc = subprocess.Popen(\n args=args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n **kwargs,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpf1ieqxij/hpcflow-new2" } }```
process
hpcflow has guarddog issues dependency hpcflow version result issues errors results silent process execution location hpcflow hpcflow sdk helper helper py code proc subprocess popen n args args n stdin subprocess devnull n stdout subprocess devnull n stderr subprocess devnull n kwargs n message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp hpcflow
1
4,071
3,294,393,827
IssuesEvent
2015-10-31 01:18:14
EricssonResearch/openwebrtc
https://api.github.com/repos/EricssonResearch/openwebrtc
closed
libffi armv7/armv7s crashing
bug iOS JS/Web Non-Build Issue
When using Bowser on armv7/armv7s, the bridge thread segfaults during startup: https://gist.github.com/superdump/996d8550beaf491c99e3 I have tested arm64 and it is only armv7/armv7s that crashes. This basically means that Bowser is currently not usable on devices using CPUs before the A7. That is, devices older than (and including) iPhone 5 / 5C, iPad 4, iPad Mini 2.
1.0
libffi armv7/armv7s crashing - When using Bowser on armv7/armv7s, the bridge thread segfaults during startup: https://gist.github.com/superdump/996d8550beaf491c99e3 I have tested arm64 and it is only armv7/armv7s that crashes. This basically means that Bowser is currently not usable on devices using CPUs before the A7. That is, devices older than (and including) iPhone 5 / 5C, iPad 4, iPad Mini 2.
non_process
libffi crashing when using bowser on the bridge thread segfaults during startup i have tested and it is only that crashes this basically means that bowser is currently not usable on devices using cpus before the that is devices older than and including iphone ipad ipad mini
0
21,673
30,119,992,225
IssuesEvent
2023-06-30 14:27:29
0xPolygonMiden/miden-vm
https://api.github.com/repos/0xPolygonMiden/miden-vm
closed
Range checker improvements
processor air v0.7
In the current implementation, the range checker component (described [here](https://0xpolygonmiden.github.io/miden-vm/design/range.html)) consists of two tables: 8-bit table and 16-bit table. We use a selector column to identify which table we are in. This structure can be greatly simplified if we get rid of the 8-bit table entirely, and instead, use a single constraint to allow "jumps" between two consecutive rows of the 16-bit table to be at most 8. Denoting $\Delta v = v' - v$, the constraint would look as follows: $$ \prod_{i=0}^8 (\Delta v - i) = 0 $$ This constraint would have degree 9, which is fine for us. The advantages of this approach: * The range checker becomes much simpler. * We can get rid of 1 main trace column (the one that is used as the selector) and 1 auxiliary trace column (the one that is used for the 8-bit table). The drawbacks of this approach: * We'd need to increase the minimum trace length from ~1K rows to about ~8K rows. In my mind this is not really a drawback as the vast majority of programs are likely to be 8K rows or more. However, this will affect test runtimes, and we'll probably need to refactor our testing methodology to avoid increasing test times by 8x.
1.0
Range checker improvements - In the current implementation, the range checker component (described [here](https://0xpolygonmiden.github.io/miden-vm/design/range.html)) consists of two tables: 8-bit table and 16-bit table. We use a selector column to identify which table we are in. This structure can be greatly simplified if we get rid of the 8-bit table entirely, and instead, use a single constraint to allow "jumps" between two consecutive rows of the 16-bit table to be at most 8. Denoting $\Delta v = v' - v$, the constraint would look as follows: $$ \prod_{i=0}^8 (\Delta v - i) = 0 $$ This constraint would have degree 9, which is fine for us. The advantages of this approach: * The range checker becomes much simpler. * We can get rid of 1 main trace column (the one that is used as the selector) and 1 auxiliary trace column (the one that is used for the 8-bit table). The drawbacks of this approach: * We'd need to increase the minimum trace length from ~1K rows to about ~8K rows. In my mind this is not really a drawback as the vast majority of programs are likely to be 8K rows or more. However, this will affect test runtimes, and we'll probably need to refactor our testing methodology to avoid increasing test times by 8x.
process
range checker improvements in the current implementation the range checker component described consists of two tables bit table and bit table we use a selector column to identify which table we are in this structure can be greatly simplified if we get rid of the bit table entirely and instead use a single constraint to allow jumps between two consecutive rows of the bit table to be at most denoting delta v v v the constraint would look as follows prod i delta v i this constraint would have degree which is fine for us the advantages of this approach the range checker becomes much simpler we can get rid of main trace column the one that is used as the selector and auxiliary trace column the one that is used for the bit table the drawbacks of this approach we d need to increase the minimum trace length from rows to about rows in my mind this is not really a drawback as the vast majority of programs are likely to be rows or more however this will affect test runtimes and we ll probably need to refactor our testing methodology to avoid increasing test times by
1
14,292
17,265,846,211
IssuesEvent
2021-07-22 13:43:18
GoogleCloudPlatform/cloud-opensource-java
https://api.github.com/repos/GoogleCloudPlatform/cloud-opensource-java
closed
Release script should not need user to specify current or next version of BOM
process
This can be inferred from pom.xml
1.0
Release script should not need user to specify current or next version of BOM - This can be inferred from pom.xml
process
release script should not need user to specify current or next version of bom this can be inferred from pom xml
1
13,499
5,390,114,747
IssuesEvent
2017-02-25 10:37:01
fossasia/open-event-android
https://api.github.com/repos/fossasia/open-event-android
closed
Upgrade gradle version to 2.2.3 for faster build with instant run
Build enhancement
**Actual Behaviour** current gradle version 2.2.2 **Expected Behaviour** Upgrade gradle version to 2.2.3 for faster build with instant run **Would you like to work on the issue?** @mananwason @the-dagger Can i work?
1.0
Upgrade gradle version to 2.2.3 for faster build with instant run - **Actual Behaviour** current gradle version 2.2.2 **Expected Behaviour** Upgrade gradle version to 2.2.3 for faster build with instant run **Would you like to work on the issue?** @mananwason @the-dagger Can i work?
non_process
upgrade gradle version to for faster build with instant run actual behaviour current gradle version expected behaviour upgrade gradle version to for faster build with instant run would you like to work on the issue mananwason the dagger can i work
0
55,678
23,534,324,110
IssuesEvent
2022-08-19 18:45:23
googleapis/java-game-servers
https://api.github.com/repos/googleapis/java-game-servers
closed
The build failed
type: bug priority: p2 api: gameservices flakybot: issue
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: fb81d8abe05ca84f9545521203b095007e36bb24 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/dbfece46-3b89-45fb-b1e9-fb143c89533f), [Sponge](http://sponge2/dbfece46-3b89-45fb-b1e9-fb143c89533f) status: failed <details><summary>Test output</summary><br><pre>java.util.concurrent.ExecutionException: com.google.api.gax.rpc.InvalidArgumentException: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Cannot connect to the provided Kubernetes cluster "projects/java-docs-samples-testing/locations/us-central1-a/clusters/standard-cluster-1", check the name and make sure that the cluster exists at com.google.common.util.concurrent.AbstractFuture.getDoneValue(AbstractFuture.java:588) at com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:463) at com.google.common.util.concurrent.FluentFuture$TrustedFuture.get(FluentFuture.java:100) at com.google.common.util.concurrent.ForwardingFuture.get(ForwardingFuture.java:73) at com.google.api.gax.longrunning.OperationFutureImpl.get(OperationFutureImpl.java:133) at com.example.gameservices.clusters.CreateCluster.createGameServerCluster(CreateCluster.java:71) at com.example.gameservices.ClusterTests.init(ClusterTests.java:82) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunBefores.invokeMethod(RunBefores.java:33) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:364) at org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:272) at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:237) at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:158) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) Caused by: com.google.api.gax.rpc.InvalidArgumentException: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Cannot connect to the provided Kubernetes cluster "projects/java-docs-samples-testing/locations/us-central1-a/clusters/standard-cluster-1", check the name and make sure that the cluster exists at com.google.api.gax.rpc.ApiExceptionFactory.createException(ApiExceptionFactory.java:47) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:72) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:60) at com.google.api.gax.grpc.GrpcExceptionCallable$ExceptionTransformingFuture.onFailure(GrpcExceptionCallable.java:97) at com.google.api.core.ApiFutures$1.onFailure(ApiFutures.java:68) at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1133) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1277) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:1038) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:808) at io.grpc.stub.ClientCalls$GrpcFuture.setException(ClientCalls.java:564) at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:534) at io.grpc.internal.DelayedClientCall$DelayedListener$3.run(DelayedClientCall.java:463) at io.grpc.internal.DelayedClientCall$DelayedListener.delayOrExecute(DelayedClientCall.java:427) at io.grpc.internal.DelayedClientCall$DelayedListener.onClose(DelayedClientCall.java:460) at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:562) at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:70) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:743) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:722) at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37) at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:133) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Cannot connect to the provided Kubernetes cluster "projects/java-docs-samples-testing/locations/us-central1-a/clusters/standard-cluster-1", check the name and make sure that the cluster exists at io.grpc.Status.asRuntimeException(Status.java:535) ... 13 more </pre></details>
1.0
The build failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: fb81d8abe05ca84f9545521203b095007e36bb24 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/dbfece46-3b89-45fb-b1e9-fb143c89533f), [Sponge](http://sponge2/dbfece46-3b89-45fb-b1e9-fb143c89533f) status: failed <details><summary>Test output</summary><br><pre>java.util.concurrent.ExecutionException: com.google.api.gax.rpc.InvalidArgumentException: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Cannot connect to the provided Kubernetes cluster "projects/java-docs-samples-testing/locations/us-central1-a/clusters/standard-cluster-1", check the name and make sure that the cluster exists at com.google.common.util.concurrent.AbstractFuture.getDoneValue(AbstractFuture.java:588) at com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:463) at com.google.common.util.concurrent.FluentFuture$TrustedFuture.get(FluentFuture.java:100) at com.google.common.util.concurrent.ForwardingFuture.get(ForwardingFuture.java:73) at com.google.api.gax.longrunning.OperationFutureImpl.get(OperationFutureImpl.java:133) at com.example.gameservices.clusters.CreateCluster.createGameServerCluster(CreateCluster.java:71) at com.example.gameservices.ClusterTests.init(ClusterTests.java:82) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunBefores.invokeMethod(RunBefores.java:33) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:364) at org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:272) at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:237) at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:158) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) Caused by: com.google.api.gax.rpc.InvalidArgumentException: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Cannot connect to the provided Kubernetes cluster "projects/java-docs-samples-testing/locations/us-central1-a/clusters/standard-cluster-1", check the name and make sure that the cluster exists at com.google.api.gax.rpc.ApiExceptionFactory.createException(ApiExceptionFactory.java:47) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:72) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:60) at com.google.api.gax.grpc.GrpcExceptionCallable$ExceptionTransformingFuture.onFailure(GrpcExceptionCallable.java:97) at com.google.api.core.ApiFutures$1.onFailure(ApiFutures.java:68) at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1133) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1277) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:1038) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:808) at io.grpc.stub.ClientCalls$GrpcFuture.setException(ClientCalls.java:564) at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:534) at io.grpc.internal.DelayedClientCall$DelayedListener$3.run(DelayedClientCall.java:463) at io.grpc.internal.DelayedClientCall$DelayedListener.delayOrExecute(DelayedClientCall.java:427) at io.grpc.internal.DelayedClientCall$DelayedListener.onClose(DelayedClientCall.java:460) at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:562) at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:70) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:743) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:722) at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37) at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:133) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Cannot connect to the provided Kubernetes cluster "projects/java-docs-samples-testing/locations/us-central1-a/clusters/standard-cluster-1", check the name and make sure that the cluster exists at io.grpc.Status.asRuntimeException(Status.java:535) ... 13 more </pre></details>
non_process
the build failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output java util concurrent executionexception com google api gax rpc invalidargumentexception io grpc statusruntimeexception invalid argument cannot connect to the provided kubernetes cluster projects java docs samples testing locations us a clusters standard cluster check the name and make sure that the cluster exists at com google common util concurrent abstractfuture getdonevalue abstractfuture java at com google common util concurrent abstractfuture get abstractfuture java at com google common util concurrent fluentfuture trustedfuture get fluentfuture java at com google common util concurrent forwardingfuture get forwardingfuture java at com google api gax longrunning operationfutureimpl get operationfutureimpl java at com example gameservices clusters createcluster creategameservercluster createcluster java at com example gameservices clustertests init clustertests java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements runbefores invokemethod runbefores java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org apache maven surefire execute java at org apache maven surefire executewithrerun java at org apache maven surefire executetestset java at org apache maven surefire invoke java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter run forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java caused by com google api gax rpc invalidargumentexception io grpc statusruntimeexception invalid argument cannot connect to the provided kubernetes cluster projects java docs samples testing locations us a clusters standard cluster check the name and make sure that the cluster exists at com google api gax rpc apiexceptionfactory createexception apiexceptionfactory java at com google api gax grpc grpcapiexceptionfactory create grpcapiexceptionfactory java at com google api gax grpc grpcapiexceptionfactory create grpcapiexceptionfactory java at com google api gax grpc grpcexceptioncallable exceptiontransformingfuture onfailure grpcexceptioncallable java at com google api core apifutures onfailure apifutures java at com google common util concurrent futures callbacklistener run futures java at com google common util concurrent directexecutor execute directexecutor java at com google common util concurrent abstractfuture executelistener abstractfuture java at com google common util concurrent abstractfuture complete abstractfuture java at com google common util concurrent abstractfuture setexception abstractfuture java at io grpc stub clientcalls grpcfuture setexception clientcalls java at io grpc stub clientcalls unarystreamtofuture onclose clientcalls java at io grpc internal delayedclientcall delayedlistener run delayedclientcall java at io grpc internal delayedclientcall delayedlistener delayorexecute delayedclientcall java at io grpc internal delayedclientcall delayedlistener onclose delayedclientcall java at io grpc internal clientcallimpl closeobserver clientcallimpl java at io grpc internal clientcallimpl access clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runinternal clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runincontext clientcallimpl java at io grpc internal contextrunnable run contextrunnable java at io grpc internal serializingexecutor run serializingexecutor java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java caused by io grpc statusruntimeexception invalid argument cannot connect to the provided kubernetes cluster projects java docs samples testing locations us a clusters standard cluster check the name and make sure that the cluster exists at io grpc status asruntimeexception status java more
0
20,021
26,493,459,621
IssuesEvent
2023-01-18 02:00:06
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Mon, 16 Jan 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events There is no result ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### Reworking geometric morphometrics into a methodology of transformation grids - **Authors:** Fred L. Bookstein - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2301.05623 - **Pdf link:** https://arxiv.org/pdf/2301.05623 - **Abstract** Today's typical application of geometric morphometrics to a quantitative comparison of organismal anatomies begins by standardizing samples of homologously labelled point configurations for location, orientation, and scale, and then renders the ensuing comparisons graphically by thin-plate spline as applied to group averages, principal components, regression predictions, or canonical variates. The scale-standardization step has recently come under criticism as inappropriate, at least for growth studies. This essay argues for a similar rethinking of the centering and rotation, and then the replacement of the thin-plate spline interpolant of the resulting configurations by a different strategy that leaves unexplained residuals at every landmark individually in order to simplify the interpretation of the displayed grid as a whole, the "transformation grid" that has been highlighted as the true underlying topic ever since D'Arcy Thompson's exposition of 1917. For analyses of comparisons involving gradients at large geometric scale, this paper argues for replacement of all the Procrustes conventions by a version of my two-point registration of 1986 (originally Francis Galton's of 1907). The choice of the two points interacts with another non-Procrustes concern, interpretability of the grid lines of a coordinate system deformed according to a fitted polynomial trend rather than an interpolating thin-plate spline. The paper works two examples using previously published cranial data; there result new findings pertinent to the interpretation of both of these classic data sets. A concluding discussion suggests that the current toolkit of geometric morphometrics, centered on Procrustes shape coordinates and thin-plate splines, is too restricted to suit many of the interpretive purposes of evolutionary and developmental biology. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### Neural Image Compression with a Diffusion-Based Decoder - **Authors:** Noor Fathima Goose, Jens Petersen, Auke Wiggers, Tianlin Xu, Guillaume Sautière - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2301.05489 - **Pdf link:** https://arxiv.org/pdf/2301.05489 - **Abstract** Diffusion probabilistic models have recently achieved remarkable success in generating high quality image and video data. In this work, we build on this class of generative models and introduce a method for lossy compression of high resolution images. The resulting codec, which we call DIffuson-based Residual Augmentation Codec (DIRAC),is the first neural codec to allow smooth traversal of the rate-distortion-perception tradeoff at test time, while obtaining competitive performance with GAN-based methods in perceptual quality. Furthermore, while sampling from diffusion probabilistic models is notoriously expensive, we show that in the compression setting the number of steps can be drastically reduced. ## Keyword: RAW There is no result ## Keyword: raw image There is no result
2.0
New submissions for Mon, 16 Jan 23 - ## Keyword: events There is no result ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### Reworking geometric morphometrics into a methodology of transformation grids - **Authors:** Fred L. Bookstein - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2301.05623 - **Pdf link:** https://arxiv.org/pdf/2301.05623 - **Abstract** Today's typical application of geometric morphometrics to a quantitative comparison of organismal anatomies begins by standardizing samples of homologously labelled point configurations for location, orientation, and scale, and then renders the ensuing comparisons graphically by thin-plate spline as applied to group averages, principal components, regression predictions, or canonical variates. The scale-standardization step has recently come under criticism as inappropriate, at least for growth studies. This essay argues for a similar rethinking of the centering and rotation, and then the replacement of the thin-plate spline interpolant of the resulting configurations by a different strategy that leaves unexplained residuals at every landmark individually in order to simplify the interpretation of the displayed grid as a whole, the "transformation grid" that has been highlighted as the true underlying topic ever since D'Arcy Thompson's exposition of 1917. For analyses of comparisons involving gradients at large geometric scale, this paper argues for replacement of all the Procrustes conventions by a version of my two-point registration of 1986 (originally Francis Galton's of 1907). The choice of the two points interacts with another non-Procrustes concern, interpretability of the grid lines of a coordinate system deformed according to a fitted polynomial trend rather than an interpolating thin-plate spline. The paper works two examples using previously published cranial data; there result new findings pertinent to the interpretation of both of these classic data sets. A concluding discussion suggests that the current toolkit of geometric morphometrics, centered on Procrustes shape coordinates and thin-plate splines, is too restricted to suit many of the interpretive purposes of evolutionary and developmental biology. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### Neural Image Compression with a Diffusion-Based Decoder - **Authors:** Noor Fathima Goose, Jens Petersen, Auke Wiggers, Tianlin Xu, Guillaume Sautière - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2301.05489 - **Pdf link:** https://arxiv.org/pdf/2301.05489 - **Abstract** Diffusion probabilistic models have recently achieved remarkable success in generating high quality image and video data. In this work, we build on this class of generative models and introduce a method for lossy compression of high resolution images. The resulting codec, which we call DIffuson-based Residual Augmentation Codec (DIRAC),is the first neural codec to allow smooth traversal of the rate-distortion-perception tradeoff at test time, while obtaining competitive performance with GAN-based methods in perceptual quality. Furthermore, while sampling from diffusion probabilistic models is notoriously expensive, we show that in the compression setting the number of steps can be drastically reduced. ## Keyword: RAW There is no result ## Keyword: raw image There is no result
process
new submissions for mon jan keyword events there is no result keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp reworking geometric morphometrics into a methodology of transformation grids authors fred l bookstein subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract today s typical application of geometric morphometrics to a quantitative comparison of organismal anatomies begins by standardizing samples of homologously labelled point configurations for location orientation and scale and then renders the ensuing comparisons graphically by thin plate spline as applied to group averages principal components regression predictions or canonical variates the scale standardization step has recently come under criticism as inappropriate at least for growth studies this essay argues for a similar rethinking of the centering and rotation and then the replacement of the thin plate spline interpolant of the resulting configurations by a different strategy that leaves unexplained residuals at every landmark individually in order to simplify the interpretation of the displayed grid as a whole the transformation grid that has been highlighted as the true underlying topic ever since d arcy thompson s exposition of for analyses of comparisons involving gradients at large geometric scale this paper argues for replacement of all the procrustes conventions by a version of my two point registration of originally francis galton s of the choice of the two points interacts with another non procrustes concern interpretability of the grid lines of a coordinate system deformed according to a fitted polynomial trend rather than an interpolating thin plate spline the paper works two examples using previously published cranial data there result new findings pertinent to the interpretation of both of these classic data sets a concluding discussion suggests that the current toolkit of geometric morphometrics centered on procrustes shape coordinates and thin plate splines is too restricted to suit many of the interpretive purposes of evolutionary and developmental biology keyword image signal processing there is no result keyword image signal process there is no result keyword compression neural image compression with a diffusion based decoder authors noor fathima goose jens petersen auke wiggers tianlin xu guillaume sautière subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract diffusion probabilistic models have recently achieved remarkable success in generating high quality image and video data in this work we build on this class of generative models and introduce a method for lossy compression of high resolution images the resulting codec which we call diffuson based residual augmentation codec dirac is the first neural codec to allow smooth traversal of the rate distortion perception tradeoff at test time while obtaining competitive performance with gan based methods in perceptual quality furthermore while sampling from diffusion probabilistic models is notoriously expensive we show that in the compression setting the number of steps can be drastically reduced keyword raw there is no result keyword raw image there is no result
1
95,396
27,494,623,022
IssuesEvent
2023-03-05 01:47:24
microsoft/onnxruntime
https://api.github.com/repos/microsoft/onnxruntime
opened
[Build] Error APPX1101 - Payload contains two or more files with the same destination path 'microsoft.ai.machinelearning.dll'
build
### Describe the issue A Visual Studio Customer created a feedback ticket in developer community regarding https://www.nuget.org/packages/Microsoft.AI.MachineLearning NuGet package. Upon clicking the source repository, I was navigated to this repository. https://developercommunity.visualstudio.com/t/Conflict-between-WinUI3-template-and-Mic/10291713 Steps to reproduce. 1. Create New Project > Blank App, Packaged (WinUI 3 in Desktop) 2. Add nuget package Microsoft.AI.MachineLearning 3. Build and get the following linker error I am able to reproduce this error following the steps mentioned above. However the error is raised by the targets file inside package. `$\Microsoft.WindowsAppSDK.1.2.221109.1\build\Microsoft.Build.Msix.Packaging.targets`. Hence I think it is an package authoring error not VS/NuGet error. ### Urgency severity: customer is unable to use this version of the package. ### Target platform win-x64 ### Build script Just click on Build in the Visual Studio will raise this error message. ### Error / output ``` 1>C:\Users\a\source\repos\App5\packages\Microsoft.WindowsAppSDK.1.2.221109.1\build\Microsoft.Build.Msix.Packaging.targets(1504,5): error APPX1101: Payload contains two or more files with the same destination path 'microsoft.ai.machinelearning.dll'. Source files: 1>C:\Users\a\source\repos\App5\packages\Microsoft.WindowsAppSDK.1.2.221109.1\build\Microsoft.Build.Msix.Packaging.targets(1504,5): error APPX1101: C:\Users\a\source\repos\App5\packages\Microsoft.AI.MachineLearning.1.13.1\runtimes\win-x64\_native\microsoft.ai.machinelearning.dll 1>C:\Users\a\source\repos\App5\packages\Microsoft.WindowsAppSDK.1.2.221109.1\build\Microsoft.Build.Msix.Packaging.targets(1504,5): error APPX1101: C:\Users\a\source\repos\App5\packages\Microsoft.AI.MachineLearning.1.13.1\winmds\microsoft.ai.machinelearning.dll Note that there is no dll at C:\Users\a\source\repos\App5\packages\Microsoft.AI.MachineLearning.1.13.1\winmds\microsoft.ai.machinelearning.dll rather just an .winmd file with the same stem. ``` ### Visual Studio Version VS 2022 17.5 or VS 2022 Internal Version 17.6 ### GCC / Compiler Version _No response_
1.0
[Build] Error APPX1101 - Payload contains two or more files with the same destination path 'microsoft.ai.machinelearning.dll' - ### Describe the issue A Visual Studio Customer created a feedback ticket in developer community regarding https://www.nuget.org/packages/Microsoft.AI.MachineLearning NuGet package. Upon clicking the source repository, I was navigated to this repository. https://developercommunity.visualstudio.com/t/Conflict-between-WinUI3-template-and-Mic/10291713 Steps to reproduce. 1. Create New Project > Blank App, Packaged (WinUI 3 in Desktop) 2. Add nuget package Microsoft.AI.MachineLearning 3. Build and get the following linker error I am able to reproduce this error following the steps mentioned above. However the error is raised by the targets file inside package. `$\Microsoft.WindowsAppSDK.1.2.221109.1\build\Microsoft.Build.Msix.Packaging.targets`. Hence I think it is an package authoring error not VS/NuGet error. ### Urgency severity: customer is unable to use this version of the package. ### Target platform win-x64 ### Build script Just click on Build in the Visual Studio will raise this error message. ### Error / output ``` 1>C:\Users\a\source\repos\App5\packages\Microsoft.WindowsAppSDK.1.2.221109.1\build\Microsoft.Build.Msix.Packaging.targets(1504,5): error APPX1101: Payload contains two or more files with the same destination path 'microsoft.ai.machinelearning.dll'. Source files: 1>C:\Users\a\source\repos\App5\packages\Microsoft.WindowsAppSDK.1.2.221109.1\build\Microsoft.Build.Msix.Packaging.targets(1504,5): error APPX1101: C:\Users\a\source\repos\App5\packages\Microsoft.AI.MachineLearning.1.13.1\runtimes\win-x64\_native\microsoft.ai.machinelearning.dll 1>C:\Users\a\source\repos\App5\packages\Microsoft.WindowsAppSDK.1.2.221109.1\build\Microsoft.Build.Msix.Packaging.targets(1504,5): error APPX1101: C:\Users\a\source\repos\App5\packages\Microsoft.AI.MachineLearning.1.13.1\winmds\microsoft.ai.machinelearning.dll Note that there is no dll at C:\Users\a\source\repos\App5\packages\Microsoft.AI.MachineLearning.1.13.1\winmds\microsoft.ai.machinelearning.dll rather just an .winmd file with the same stem. ``` ### Visual Studio Version VS 2022 17.5 or VS 2022 Internal Version 17.6 ### GCC / Compiler Version _No response_
non_process
error payload contains two or more files with the same destination path microsoft ai machinelearning dll describe the issue a visual studio customer created a feedback ticket in developer community regarding nuget package upon clicking the source repository i was navigated to this repository steps to reproduce create new project blank app packaged winui in desktop add nuget package microsoft ai machinelearning build and get the following linker error i am able to reproduce this error following the steps mentioned above however the error is raised by the targets file inside package microsoft windowsappsdk build microsoft build msix packaging targets hence i think it is an package authoring error not vs nuget error urgency severity customer is unable to use this version of the package target platform win build script just click on build in the visual studio will raise this error message error output c users a source repos packages microsoft windowsappsdk build microsoft build msix packaging targets error payload contains two or more files with the same destination path microsoft ai machinelearning dll source files c users a source repos packages microsoft windowsappsdk build microsoft build msix packaging targets error c users a source repos packages microsoft ai machinelearning runtimes win native microsoft ai machinelearning dll c users a source repos packages microsoft windowsappsdk build microsoft build msix packaging targets error c users a source repos packages microsoft ai machinelearning winmds microsoft ai machinelearning dll note that there is no dll at c users a source repos packages microsoft ai machinelearning winmds microsoft ai machinelearning dll rather just an winmd file with the same stem visual studio version vs or vs internal version gcc compiler version no response
0
9,703
12,702,813,493
IssuesEvent
2020-06-22 20:55:45
modi-w/AutoVersionsDB
https://api.github.com/repos/modi-w/AutoVersionsDB
closed
Recfactor The Code by FxCopAnalyzer
area-Core area-Sql Server area-Tests area-UI process-ready-for-implementation type-enhancement
**Action Items:** 1. Append FxCopAnalyzer nuget library to all projects in the solution. 2. Resolve all the warnings and messages from the error windows.
1.0
Recfactor The Code by FxCopAnalyzer - **Action Items:** 1. Append FxCopAnalyzer nuget library to all projects in the solution. 2. Resolve all the warnings and messages from the error windows.
process
recfactor the code by fxcopanalyzer action items append fxcopanalyzer nuget library to all projects in the solution resolve all the warnings and messages from the error windows
1
108,360
16,770,524,541
IssuesEvent
2021-06-14 14:18:48
OpenLiberty/open-liberty
https://api.github.com/repos/OpenLiberty/open-liberty
reopened
Feature Test Summary for OIDC JWE (RP) support
Feature Test Summary team:Security SSO
## Test Strategy **Describe the test strategy & approach for this feature, and describe how the approach verifies the functions delivered by this feature.** _For any feature, be aware that only FAT tests (not unit or BVT) are executed in our cross platform testing. To ensure cross platform testing ensure you have sufficient FAT coverage to verify the feature._ _If delivering tests outside of the standard Liberty FAT framework, do the tests push the results into cognitive testing database (if not, consult with the CSI Team who can provide advice and verify if results are being received)?_ ### List of FAT projects affected * ### Test strategy * What functionality is new or modified by this feature? * What are the positive and negative tests for that functionality? (Tell me the specific scenarios you tested. What kind of tests do you have for when everything ends up working (positive tests)? What about tests that verify we fail gracefully when things go wrong (negative tests)? See the [Positive and negative tests](https://github.ibm.com/websphere/WS-CD-Open/wiki/Feature-Review-(Feature-Test-Summary-Process)#positive-and-negative-tests) section of the Feature Test Summary Process wiki for more detail.) * What manual tests are there (if any)? (Note: Automated testing is expected for all features with manual testing considered an exception to the rule.) ## Confidence Level **Collectively as a team you need to assess your confidence in the testing delivered based on the values below. This should be done as a team and not an individual to ensure more eyes are on it and that pressures to deliver quickly are absorbed by the team as a whole.** Please indicate your confidence in the testing (up to and including FAT) delivered with this feature by selecting one of these values: 0 - No automated testing delivered 1 - We have minimal automated coverage of the feature including golden paths. There is a relatively high risk that defects or issues could be found in this feature. 2 - We have delivered a reasonable automated coverage of the golden paths of this feature but are aware of gaps and extra testing that could be done here. Error/outlying scenarios are not really covered. There are likely risks that issues may exist in the golden paths 3 - We have delivered all automated testing we believe is needed for the golden paths of this feature and minimal coverage of the error/outlying scenarios. There is a risk when the feature is used outside the golden paths however we are confident on the golden path. Note: This may still be a valid end state for a feature... things like Beta features may well suffice at this level. 4 - We have delivered all automated testing we believe is needed for the golden paths of this feature and have good coverage of the error/outlying scenarios. While more testing of the error/outlying scenarios could be added we believe there is minimal risk here and the cost of providing these is considered higher than the benefit they would provide. 5 - We have delivered all automated testing we believe is needed for this feature. The testing covers all golden path cases as well as all the error/outlying scenarios that make sense. We are not aware of any gaps in the testing at this time. No manual testing is required to verify this feature. Based on your answer above, for any answer other than a 4 or 5 please provide details of what drove your answer. Please be aware, it may be perfectly reasonable in some scenarios to deliver with any value above. We may accept no automated testing is needed for some features, we may be happy with low levels of testing on samples for instance so please don't feel the need to drive to a 5. We need your honest assessment as a team and the reasoning for why you believe shipping at that level is valid. What are the gaps, what is the risk etc. Please also provide links to the follow on work that is needed to close the gaps (should you deem it needed)
True
Feature Test Summary for OIDC JWE (RP) support - ## Test Strategy **Describe the test strategy & approach for this feature, and describe how the approach verifies the functions delivered by this feature.** _For any feature, be aware that only FAT tests (not unit or BVT) are executed in our cross platform testing. To ensure cross platform testing ensure you have sufficient FAT coverage to verify the feature._ _If delivering tests outside of the standard Liberty FAT framework, do the tests push the results into cognitive testing database (if not, consult with the CSI Team who can provide advice and verify if results are being received)?_ ### List of FAT projects affected * ### Test strategy * What functionality is new or modified by this feature? * What are the positive and negative tests for that functionality? (Tell me the specific scenarios you tested. What kind of tests do you have for when everything ends up working (positive tests)? What about tests that verify we fail gracefully when things go wrong (negative tests)? See the [Positive and negative tests](https://github.ibm.com/websphere/WS-CD-Open/wiki/Feature-Review-(Feature-Test-Summary-Process)#positive-and-negative-tests) section of the Feature Test Summary Process wiki for more detail.) * What manual tests are there (if any)? (Note: Automated testing is expected for all features with manual testing considered an exception to the rule.) ## Confidence Level **Collectively as a team you need to assess your confidence in the testing delivered based on the values below. This should be done as a team and not an individual to ensure more eyes are on it and that pressures to deliver quickly are absorbed by the team as a whole.** Please indicate your confidence in the testing (up to and including FAT) delivered with this feature by selecting one of these values: 0 - No automated testing delivered 1 - We have minimal automated coverage of the feature including golden paths. There is a relatively high risk that defects or issues could be found in this feature. 2 - We have delivered a reasonable automated coverage of the golden paths of this feature but are aware of gaps and extra testing that could be done here. Error/outlying scenarios are not really covered. There are likely risks that issues may exist in the golden paths 3 - We have delivered all automated testing we believe is needed for the golden paths of this feature and minimal coverage of the error/outlying scenarios. There is a risk when the feature is used outside the golden paths however we are confident on the golden path. Note: This may still be a valid end state for a feature... things like Beta features may well suffice at this level. 4 - We have delivered all automated testing we believe is needed for the golden paths of this feature and have good coverage of the error/outlying scenarios. While more testing of the error/outlying scenarios could be added we believe there is minimal risk here and the cost of providing these is considered higher than the benefit they would provide. 5 - We have delivered all automated testing we believe is needed for this feature. The testing covers all golden path cases as well as all the error/outlying scenarios that make sense. We are not aware of any gaps in the testing at this time. No manual testing is required to verify this feature. Based on your answer above, for any answer other than a 4 or 5 please provide details of what drove your answer. Please be aware, it may be perfectly reasonable in some scenarios to deliver with any value above. We may accept no automated testing is needed for some features, we may be happy with low levels of testing on samples for instance so please don't feel the need to drive to a 5. We need your honest assessment as a team and the reasoning for why you believe shipping at that level is valid. What are the gaps, what is the risk etc. Please also provide links to the follow on work that is needed to close the gaps (should you deem it needed)
non_process
feature test summary for oidc jwe rp support test strategy describe the test strategy approach for this feature and describe how the approach verifies the functions delivered by this feature for any feature be aware that only fat tests not unit or bvt are executed in our cross platform testing to ensure cross platform testing ensure you have sufficient fat coverage to verify the feature if delivering tests outside of the standard liberty fat framework do the tests push the results into cognitive testing database if not consult with the csi team who can provide advice and verify if results are being received list of fat projects affected test strategy what functionality is new or modified by this feature what are the positive and negative tests for that functionality tell me the specific scenarios you tested what kind of tests do you have for when everything ends up working positive tests what about tests that verify we fail gracefully when things go wrong negative tests see the section of the feature test summary process wiki for more detail what manual tests are there if any note automated testing is expected for all features with manual testing considered an exception to the rule confidence level collectively as a team you need to assess your confidence in the testing delivered based on the values below this should be done as a team and not an individual to ensure more eyes are on it and that pressures to deliver quickly are absorbed by the team as a whole please indicate your confidence in the testing up to and including fat delivered with this feature by selecting one of these values no automated testing delivered we have minimal automated coverage of the feature including golden paths there is a relatively high risk that defects or issues could be found in this feature we have delivered a reasonable automated coverage of the golden paths of this feature but are aware of gaps and extra testing that could be done here error outlying scenarios are not really covered there are likely risks that issues may exist in the golden paths we have delivered all automated testing we believe is needed for the golden paths of this feature and minimal coverage of the error outlying scenarios there is a risk when the feature is used outside the golden paths however we are confident on the golden path note this may still be a valid end state for a feature things like beta features may well suffice at this level we have delivered all automated testing we believe is needed for the golden paths of this feature and have good coverage of the error outlying scenarios while more testing of the error outlying scenarios could be added we believe there is minimal risk here and the cost of providing these is considered higher than the benefit they would provide we have delivered all automated testing we believe is needed for this feature the testing covers all golden path cases as well as all the error outlying scenarios that make sense we are not aware of any gaps in the testing at this time no manual testing is required to verify this feature based on your answer above for any answer other than a or please provide details of what drove your answer please be aware it may be perfectly reasonable in some scenarios to deliver with any value above we may accept no automated testing is needed for some features we may be happy with low levels of testing on samples for instance so please don t feel the need to drive to a we need your honest assessment as a team and the reasoning for why you believe shipping at that level is valid what are the gaps what is the risk etc please also provide links to the follow on work that is needed to close the gaps should you deem it needed
0
39,836
5,252,143,632
IssuesEvent
2017-02-02 02:46:18
semperfiwebdesign/all-in-one-seo-pack
https://api.github.com/repos/semperfiwebdesign/all-in-one-seo-pack
closed
Uncaught exception ‘BadMethodCallException’ generate_htaccess_blocklist
Bug Needs Testing Priority - High
Reported here - https://wordpress.org/support/topic/uncaught-exception-badmethodcallexception-generate_htaccess_blocklist/ User states that when updating to WordPress v4.7.2 they get a white screen and this error in the debug log: Method generate_htaccess_blocklist doesn’t exist’ in /path/to/wordpress/wp-content/plugins/all-in-one-seo-pack/admin/aioseop_module_class.php:52 User is running nginx.
1.0
Uncaught exception ‘BadMethodCallException’ generate_htaccess_blocklist - Reported here - https://wordpress.org/support/topic/uncaught-exception-badmethodcallexception-generate_htaccess_blocklist/ User states that when updating to WordPress v4.7.2 they get a white screen and this error in the debug log: Method generate_htaccess_blocklist doesn’t exist’ in /path/to/wordpress/wp-content/plugins/all-in-one-seo-pack/admin/aioseop_module_class.php:52 User is running nginx.
non_process
uncaught exception ‘badmethodcallexception’ generate htaccess blocklist reported here user states that when updating to wordpress they get a white screen and this error in the debug log method generate htaccess blocklist doesn’t exist’ in path to wordpress wp content plugins all in one seo pack admin aioseop module class php user is running nginx
0
127,610
18,013,701,162
IssuesEvent
2021-09-16 11:37:55
artsking/linux-4.1.15_CVE-2021-33034_withPatch
https://api.github.com/repos/artsking/linux-4.1.15_CVE-2021-33034_withPatch
closed
CVE-2015-7566 (Medium) detected in linux-stable-rtv4.1.33 - autoclosed
security vulnerability
## CVE-2015-7566 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/artsking/linux-4.1.15_CVE-2021-33034_withPatch/commit/01646e9f5acca54431418de57befba08e7a6c7c6">01646e9f5acca54431418de57befba08e7a6c7c6</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/serial/visor.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/serial/visor.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The clie_5_attach function in drivers/usb/serial/visor.c in the Linux kernel through 4.4.1 allows physically proximate attackers to cause a denial of service (NULL pointer dereference and system crash) or possibly have unspecified other impact by inserting a USB device that lacks a bulk-out endpoint. <p>Publish Date: 2016-02-08 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-7566>CVE-2015-7566</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Physical - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2015-7566">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2015-7566</a></p> <p>Release Date: 2016-02-08</p> <p>Fix Resolution: v4.5-rc2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2015-7566 (Medium) detected in linux-stable-rtv4.1.33 - autoclosed - ## CVE-2015-7566 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/artsking/linux-4.1.15_CVE-2021-33034_withPatch/commit/01646e9f5acca54431418de57befba08e7a6c7c6">01646e9f5acca54431418de57befba08e7a6c7c6</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/serial/visor.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/serial/visor.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The clie_5_attach function in drivers/usb/serial/visor.c in the Linux kernel through 4.4.1 allows physically proximate attackers to cause a denial of service (NULL pointer dereference and system crash) or possibly have unspecified other impact by inserting a USB device that lacks a bulk-out endpoint. <p>Publish Date: 2016-02-08 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-7566>CVE-2015-7566</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Physical - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2015-7566">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2015-7566</a></p> <p>Release Date: 2016-02-08</p> <p>Fix Resolution: v4.5-rc2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linux stable autoclosed cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers usb serial visor c drivers usb serial visor c vulnerability details the clie attach function in drivers usb serial visor c in the linux kernel through allows physically proximate attackers to cause a denial of service null pointer dereference and system crash or possibly have unspecified other impact by inserting a usb device that lacks a bulk out endpoint publish date url a href cvss score details base score metrics exploitability metrics attack vector physical attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
9,962
12,992,233,444
IssuesEvent
2020-07-23 06:20:17
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
API to check memory usage on UWP (the Process class is not supported)
area-System.Diagnostics.Process
From standard created by [Sergio0694](https://github.com/Sergio0694) : dotnet/standard#641 Hello, this is related to dotnet/runtime#13954 and you can see my original question on StackOverflow [here](https://stackoverflow.com/questions/48622947/how-to-get-process-memory-from-both-net-standard-2-0-and-uwp/48623215?noredirect=1#comment84245854_48623215). Basically, the issue is that using: ```var memory = System.Diagnostics.Process.GetCurrentProcess().PrivateMemorySize64;``` Crashes on UWP. In DEBUG mode, I get a `PlatformNotSupportedException`, while in RELEASE mode (possibly due to the .NET Native compiler) I get a `TypeLoadException` followed by a series of low-level exceptions, and apparently I can't catch any of these exceptions. Even placing the above line into a general `try/catch` block doesn't work, the app just crashes anyways. I understand the `Process` class is locked on UWP for security reasons, but there should be an API to at least be able to check the current memory usage (and CPU usage). Some "read only" APIs from the `Process` class that should be allowed on UWP too I mean. For example, the .NET Standard `System.Diagnostics.Process.GetCurrentProcess().PrivateMemorySize64` call should have an equivalent API that gets mapped to `Windows.System.MemoryManager` APIs on UWP. Is there a workaround, and/or is this something planned in the future? Thanks!
1.0
API to check memory usage on UWP (the Process class is not supported) - From standard created by [Sergio0694](https://github.com/Sergio0694) : dotnet/standard#641 Hello, this is related to dotnet/runtime#13954 and you can see my original question on StackOverflow [here](https://stackoverflow.com/questions/48622947/how-to-get-process-memory-from-both-net-standard-2-0-and-uwp/48623215?noredirect=1#comment84245854_48623215). Basically, the issue is that using: ```var memory = System.Diagnostics.Process.GetCurrentProcess().PrivateMemorySize64;``` Crashes on UWP. In DEBUG mode, I get a `PlatformNotSupportedException`, while in RELEASE mode (possibly due to the .NET Native compiler) I get a `TypeLoadException` followed by a series of low-level exceptions, and apparently I can't catch any of these exceptions. Even placing the above line into a general `try/catch` block doesn't work, the app just crashes anyways. I understand the `Process` class is locked on UWP for security reasons, but there should be an API to at least be able to check the current memory usage (and CPU usage). Some "read only" APIs from the `Process` class that should be allowed on UWP too I mean. For example, the .NET Standard `System.Diagnostics.Process.GetCurrentProcess().PrivateMemorySize64` call should have an equivalent API that gets mapped to `Windows.System.MemoryManager` APIs on UWP. Is there a workaround, and/or is this something planned in the future? Thanks!
process
api to check memory usage on uwp the process class is not supported from standard created by dotnet standard hello this is related to dotnet runtime and you can see my original question on stackoverflow basically the issue is that using var memory system diagnostics process getcurrentprocess crashes on uwp in debug mode i get a platformnotsupportedexception while in release mode possibly due to the net native compiler i get a typeloadexception followed by a series of low level exceptions and apparently i can t catch any of these exceptions even placing the above line into a general try catch block doesn t work the app just crashes anyways i understand the process class is locked on uwp for security reasons but there should be an api to at least be able to check the current memory usage and cpu usage some read only apis from the process class that should be allowed on uwp too i mean for example the net standard system diagnostics process getcurrentprocess call should have an equivalent api that gets mapped to windows system memorymanager apis on uwp is there a workaround and or is this something planned in the future thanks
1
48,418
7,431,759,910
IssuesEvent
2018-03-25 17:50:05
fbpic/fbpic
https://api.github.com/repos/fbpic/fbpic
closed
Document ionization module and add an example script
documentation ionization
We might want to have a simpler API for creating Particles for this... Even a small backward-compatible change would be good. I'll think about it.
1.0
Document ionization module and add an example script - We might want to have a simpler API for creating Particles for this... Even a small backward-compatible change would be good. I'll think about it.
non_process
document ionization module and add an example script we might want to have a simpler api for creating particles for this even a small backward compatible change would be good i ll think about it
0
67,585
12,977,464,482
IssuesEvent
2020-07-21 20:42:22
microsoft/AdaptiveCards
https://api.github.com/repos/microsoft/AdaptiveCards
closed
[JavaScript][Deregistering a property from the parser/render]
AdaptiveCards v20.07 CiscoWebEx-Integration Platform-HTML Request Status-In Code Review javascript
**What platform is your feature request related to?** * JavaScript **Is your feature request related to a problem? Please describe.** We would like to disable the functionality of unsupported properties on parser/render **Describe the solution you'd like** It would be great to have a way to `Deregister a property from the parser/render` **Deregister has to work for the following cases:** * Deregister the property for a particular element( For Example: Deregister the `Height` property of `columnSet`) * Deregister the property for all elements(For Example: Deregister the `verticalContentAlignment` property for all the elements where it supports) ### For Example Unsupported Property : `maxLines` for `TextBlock` element. After deregistering a property, though users try to add `maxLines` property to their JSON then that property shouldn't render/parse. **Current behavior:** <img width="1680" alt="Screenshot 2020-07-17 at 11 37 18 AM" src="https://user-images.githubusercontent.com/55378170/87754154-e9c99680-c821-11ea-93f0-d5e393a428e0.png"> **If possible, validator have to point out that the unsupported properties are not supported as warning message** For Example:(Warning Message) **Unknown/Unsupported property :`maxLines` of `TextBlock` Element**
1.0
[JavaScript][Deregistering a property from the parser/render] - **What platform is your feature request related to?** * JavaScript **Is your feature request related to a problem? Please describe.** We would like to disable the functionality of unsupported properties on parser/render **Describe the solution you'd like** It would be great to have a way to `Deregister a property from the parser/render` **Deregister has to work for the following cases:** * Deregister the property for a particular element( For Example: Deregister the `Height` property of `columnSet`) * Deregister the property for all elements(For Example: Deregister the `verticalContentAlignment` property for all the elements where it supports) ### For Example Unsupported Property : `maxLines` for `TextBlock` element. After deregistering a property, though users try to add `maxLines` property to their JSON then that property shouldn't render/parse. **Current behavior:** <img width="1680" alt="Screenshot 2020-07-17 at 11 37 18 AM" src="https://user-images.githubusercontent.com/55378170/87754154-e9c99680-c821-11ea-93f0-d5e393a428e0.png"> **If possible, validator have to point out that the unsupported properties are not supported as warning message** For Example:(Warning Message) **Unknown/Unsupported property :`maxLines` of `TextBlock` Element**
non_process
what platform is your feature request related to javascript is your feature request related to a problem please describe we would like to disable the functionality of unsupported properties on parser render describe the solution you d like it would be great to have a way to deregister a property from the parser render deregister has to work for the following cases deregister the property for a particular element for example deregister the height property of columnset deregister the property for all elements for example deregister the verticalcontentalignment property for all the elements where it supports for example unsupported property maxlines for textblock element after deregistering a property though users try to add maxlines property to their json then that property shouldn t render parse current behavior img width alt screenshot at am src if possible validator have to point out that the unsupported properties are not supported as warning message for example warning message unknown unsupported property maxlines of textblock element
0
4,019
6,953,489,879
IssuesEvent
2017-12-06 21:13:34
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
SQL Server database table not showing up
Bug Database/SQLServer Help Wanted Priority/P3 Query Processor
I'm running a local instance of Metabase v0.22.2, via the JAR, and connecting to a SQL Server 2012 (Express) instance. When I create a new question, I can see almost all of the tables in my database. The only that I know is missing is the one named *dbo.BATCH*. I'm assuming it's missing because 'batch' is considered reserved or otherwise an invalid table name.
1.0
SQL Server database table not showing up - I'm running a local instance of Metabase v0.22.2, via the JAR, and connecting to a SQL Server 2012 (Express) instance. When I create a new question, I can see almost all of the tables in my database. The only that I know is missing is the one named *dbo.BATCH*. I'm assuming it's missing because 'batch' is considered reserved or otherwise an invalid table name.
process
sql server database table not showing up i m running a local instance of metabase via the jar and connecting to a sql server express instance when i create a new question i can see almost all of the tables in my database the only that i know is missing is the one named dbo batch i m assuming it s missing because batch is considered reserved or otherwise an invalid table name
1
19,579
25,904,525,699
IssuesEvent
2022-12-15 09:11:30
googleapis/python-runtimeconfig
https://api.github.com/repos/googleapis/python-runtimeconfig
opened
Your .repo-metadata.json file has a problem 🤒
type: process repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan 📈: * api_shortname 'runtimeconfig' invalid in .repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file: Result of scan 📈: * api_shortname 'runtimeconfig' invalid in .repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 api shortname runtimeconfig invalid in repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
1
13,523
16,058,259,163
IssuesEvent
2021-04-23 08:49:25
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
Error: [introspection-engine\connectors\sql-introspection-connector\src\introspection_helpers.rs:188:64] called `Option::unwrap()` on a `None` value
bug/1-repro-available kind/bug process/candidate team/migrations
<!-- If required, please update the title to be clear and descriptive --> Command: `prisma introspect` Version: `2.21.2` Binary Version: `e421996c87d5f3c8f7eeadd502d4ad402c89464d` Report: https://prisma-errors.netlify.app/report/13243 OS: `x64 win32 10.0.19042` JS Stacktrace: ``` Error: [introspection-engine\connectors\sql-introspection-connector\src\introspection_helpers.rs:188:64] called `Option::unwrap()` on a `None` value at ChildProcess.<anonymous> (D:\PJ\fym2\node_modules\prisma\build\index.js:39953:28) at ChildProcess.emit (events.js:315:20) at ChildProcess.EventEmitter.emit (domain.js:467:12) at Process.ChildProcess._handle.onexit (internal/child_process.js:277:12) ``` Rust Stacktrace: ``` 0: <unknown> 1: <unknown> 2: <unknown> 3: <unknown> 4: <unknown> 5: <unknown> 6: <unknown> 7: <unknown> 8: <unknown> 9: <unknown> 10: <unknown> 11: <unknown> 12: <unknown> 13: <unknown> 14: <unknown> 15: <unknown> 16: <unknown> 17: <unknown> 18: <unknown> 19: <unknown> 20: <unknown> 21: <unknown> 22: <unknown> 23: <unknown> 24: <unknown> 25: <unknown> 26: <unknown> 27: BaseThreadInitThunk 28: RtlUserThreadStart ```
1.0
Error: [introspection-engine\connectors\sql-introspection-connector\src\introspection_helpers.rs:188:64] called `Option::unwrap()` on a `None` value - <!-- If required, please update the title to be clear and descriptive --> Command: `prisma introspect` Version: `2.21.2` Binary Version: `e421996c87d5f3c8f7eeadd502d4ad402c89464d` Report: https://prisma-errors.netlify.app/report/13243 OS: `x64 win32 10.0.19042` JS Stacktrace: ``` Error: [introspection-engine\connectors\sql-introspection-connector\src\introspection_helpers.rs:188:64] called `Option::unwrap()` on a `None` value at ChildProcess.<anonymous> (D:\PJ\fym2\node_modules\prisma\build\index.js:39953:28) at ChildProcess.emit (events.js:315:20) at ChildProcess.EventEmitter.emit (domain.js:467:12) at Process.ChildProcess._handle.onexit (internal/child_process.js:277:12) ``` Rust Stacktrace: ``` 0: <unknown> 1: <unknown> 2: <unknown> 3: <unknown> 4: <unknown> 5: <unknown> 6: <unknown> 7: <unknown> 8: <unknown> 9: <unknown> 10: <unknown> 11: <unknown> 12: <unknown> 13: <unknown> 14: <unknown> 15: <unknown> 16: <unknown> 17: <unknown> 18: <unknown> 19: <unknown> 20: <unknown> 21: <unknown> 22: <unknown> 23: <unknown> 24: <unknown> 25: <unknown> 26: <unknown> 27: BaseThreadInitThunk 28: RtlUserThreadStart ```
process
error called option unwrap on a none value command prisma introspect version binary version report os js stacktrace error called option unwrap on a none value at childprocess d pj node modules prisma build index js at childprocess emit events js at childprocess eventemitter emit domain js at process childprocess handle onexit internal child process js rust stacktrace basethreadinitthunk rtluserthreadstart
1
10,867
13,639,662,581
IssuesEvent
2020-09-25 11:25:21
googleapis/google-cloud-dotnet
https://api.github.com/repos/googleapis/google-cloud-dotnet
opened
[Storage]: Review and modify V4 URL signing tests...
api: storage priority: p2 type: process
So as to make sure that headers that are now required to be signed, are being so. (all x-goog-* and x-amz-* are now to be signed) Double check that production code is not affected.
1.0
[Storage]: Review and modify V4 URL signing tests... - So as to make sure that headers that are now required to be signed, are being so. (all x-goog-* and x-amz-* are now to be signed) Double check that production code is not affected.
process
review and modify url signing tests so as to make sure that headers that are now required to be signed are being so all x goog and x amz are now to be signed double check that production code is not affected
1
7,008
10,151,140,981
IssuesEvent
2019-08-05 19:31:51
bow-simulation/virtualbow
https://api.github.com/repos/bow-simulation/virtualbow
closed
Add and maintain a changelog
area: software process type: feature
As a more detailed addition to the usual release announcements. https://keepachangelog.com/en/1.0.0/ has a good style suggestion.
1.0
Add and maintain a changelog - As a more detailed addition to the usual release announcements. https://keepachangelog.com/en/1.0.0/ has a good style suggestion.
process
add and maintain a changelog as a more detailed addition to the usual release announcements has a good style suggestion
1
19,028
25,036,907,203
IssuesEvent
2022-11-04 16:47:22
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
Toolchain Resolution with bzlmod
more data needed type: support / not a bug (process) team-ExternalDeps area-Bzlmod
### Description of the bug: Toolchain resolution when using Bazel Modules seems backwards. My understanding is that toolchains resolve in order of whoever registered them first. I've been experimenting with converting our WORKSPACE based bazel rules to use modules, which includes using module extensions to register a custom c++ toolchain. With Bazel Modules though, the intuitive thing is to have them resolve in reverse, the most recently defined toolchain being the one that is selected. ``` INFO: ToolchainResolution: Target platform @txe~0.0.1//toolchain/platforms:linux_x64: Selected execution platform @txe~0.0.1//toolchain/platforms:linux_x64, INFO: ToolchainResolution: Type @bazel_tools//tools/cpp:toolchain_type: target platform @txe~0.0.1//toolchain/platforms:linux_x64: Rejected toolchain @bazel_tools~cc_configure_extension~local_config_cc//:cc-compiler-armeabi-v7a; mismatching values: armv7, android INFO: ToolchainResolution: Type @bazel_tools//tools/cpp:toolchain_type: target platform @txe~0.0.1//toolchain/platforms:linux_x64: execution @txe~0.0.1//toolchain/platforms:linux_x64: Selected toolchain @bazel_tools~cc_configure_extension~local_config_cc//:cc-compiler-k8 INFO: ToolchainResolution: Type @bazel_tools//tools/cpp:toolchain_type: target platform @txe~0.0.1//toolchain/platforms:linux_x64: execution platform @txe~0.0.1//toolchain/platforms:linux_x64: Skipping toolchain @bazel_tools~cc_configure_extension~local_config_cc//:cc-compiler-k8; execution platform already has selected toolchain INFO: ToolchainResolution: Type @bazel_tools//tools/cpp:toolchain_type: target platform @txe~0.0.1//toolchain/platforms:linux_x64: execution @local_config_platform//:host: Selected toolchain @bazel_tools~cc_configure_extension~local_config_cc//:cc-compiler-k8 INFO: ToolchainResolution: Type @bazel_tools//tools/cpp:toolchain_type: target platform @txe~0.0.1//toolchain/platforms:linux_x64: execution platform @txe~0.0.1//toolchain/platforms:linux_x64: Skipping toolchain @txe~0.0.1~toolchains_extension~internal_clang//:internal_cc; execution platform already has selected toolchain ``` ### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. _No response_ ### Which operating system are you running Bazel on? 5.13.0-52-generic ### What is the output of `bazel info release`? development version ### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel. Bazel 6 release building bazel HEAD. (Needed proto rule fixes) ### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ? ```text https://github.com/bazelbuild/bazel.git 97803fda18c43a76f2dfd134e72866decf7977bb 97803fda18c43a76f2dfd134e72866decf7977bb ``` ### Have you found anything relevant by searching the web? _No response_ ### Any other information, logs, or outputs that you want to share? _No response_
1.0
Toolchain Resolution with bzlmod - ### Description of the bug: Toolchain resolution when using Bazel Modules seems backwards. My understanding is that toolchains resolve in order of whoever registered them first. I've been experimenting with converting our WORKSPACE based bazel rules to use modules, which includes using module extensions to register a custom c++ toolchain. With Bazel Modules though, the intuitive thing is to have them resolve in reverse, the most recently defined toolchain being the one that is selected. ``` INFO: ToolchainResolution: Target platform @txe~0.0.1//toolchain/platforms:linux_x64: Selected execution platform @txe~0.0.1//toolchain/platforms:linux_x64, INFO: ToolchainResolution: Type @bazel_tools//tools/cpp:toolchain_type: target platform @txe~0.0.1//toolchain/platforms:linux_x64: Rejected toolchain @bazel_tools~cc_configure_extension~local_config_cc//:cc-compiler-armeabi-v7a; mismatching values: armv7, android INFO: ToolchainResolution: Type @bazel_tools//tools/cpp:toolchain_type: target platform @txe~0.0.1//toolchain/platforms:linux_x64: execution @txe~0.0.1//toolchain/platforms:linux_x64: Selected toolchain @bazel_tools~cc_configure_extension~local_config_cc//:cc-compiler-k8 INFO: ToolchainResolution: Type @bazel_tools//tools/cpp:toolchain_type: target platform @txe~0.0.1//toolchain/platforms:linux_x64: execution platform @txe~0.0.1//toolchain/platforms:linux_x64: Skipping toolchain @bazel_tools~cc_configure_extension~local_config_cc//:cc-compiler-k8; execution platform already has selected toolchain INFO: ToolchainResolution: Type @bazel_tools//tools/cpp:toolchain_type: target platform @txe~0.0.1//toolchain/platforms:linux_x64: execution @local_config_platform//:host: Selected toolchain @bazel_tools~cc_configure_extension~local_config_cc//:cc-compiler-k8 INFO: ToolchainResolution: Type @bazel_tools//tools/cpp:toolchain_type: target platform @txe~0.0.1//toolchain/platforms:linux_x64: execution platform @txe~0.0.1//toolchain/platforms:linux_x64: Skipping toolchain @txe~0.0.1~toolchains_extension~internal_clang//:internal_cc; execution platform already has selected toolchain ``` ### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. _No response_ ### Which operating system are you running Bazel on? 5.13.0-52-generic ### What is the output of `bazel info release`? development version ### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel. Bazel 6 release building bazel HEAD. (Needed proto rule fixes) ### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ? ```text https://github.com/bazelbuild/bazel.git 97803fda18c43a76f2dfd134e72866decf7977bb 97803fda18c43a76f2dfd134e72866decf7977bb ``` ### Have you found anything relevant by searching the web? _No response_ ### Any other information, logs, or outputs that you want to share? _No response_
process
toolchain resolution with bzlmod description of the bug toolchain resolution when using bazel modules seems backwards my understanding is that toolchains resolve in order of whoever registered them first i ve been experimenting with converting our workspace based bazel rules to use modules which includes using module extensions to register a custom c toolchain with bazel modules though the intuitive thing is to have them resolve in reverse the most recently defined toolchain being the one that is selected info toolchainresolution target platform txe toolchain platforms linux selected execution platform txe toolchain platforms linux info toolchainresolution type bazel tools tools cpp toolchain type target platform txe toolchain platforms linux rejected toolchain bazel tools cc configure extension local config cc cc compiler armeabi mismatching values android info toolchainresolution type bazel tools tools cpp toolchain type target platform txe toolchain platforms linux execution txe toolchain platforms linux selected toolchain bazel tools cc configure extension local config cc cc compiler info toolchainresolution type bazel tools tools cpp toolchain type target platform txe toolchain platforms linux execution platform txe toolchain platforms linux skipping toolchain bazel tools cc configure extension local config cc cc compiler execution platform already has selected toolchain info toolchainresolution type bazel tools tools cpp toolchain type target platform txe toolchain platforms linux execution local config platform host selected toolchain bazel tools cc configure extension local config cc cc compiler info toolchainresolution type bazel tools tools cpp toolchain type target platform txe toolchain platforms linux execution platform txe toolchain platforms linux skipping toolchain txe toolchains extension internal clang internal cc execution platform already has selected toolchain what s the simplest easiest way to reproduce this bug please provide a minimal example if possible no response which operating system are you running bazel on generic what is the output of bazel info release development version if bazel info release returns development version or non git tell us how you built bazel bazel release building bazel head needed proto rule fixes what s the output of git remote get url origin git rev parse master git rev parse head text have you found anything relevant by searching the web no response any other information logs or outputs that you want to share no response
1
624
3,090,300,613
IssuesEvent
2015-08-26 05:13:12
e-government-ua/i
https://api.github.com/repos/e-government-ua/i
closed
На главном портале, в форме услуги, при выборе времени в Электронной Очереди - выволить сообщение о ганятости, если возвращается ошибка об этом.
active hi priority In process of testing question test version
прикрутить диалог(хоть даже alert обычный): когда выбираеш время электронной очереди, и возникает ошибка - этого не видно, и визуально как будто-бы ничего не произошло... а уже после сабмита вываливает ошибку, причем другую (типа обьект поломанный) Можно просто выводить алерт с текстом "Ці дата та час вже зайняті. оберіть інші, будь ласка" (желательно после этого произваодить перезагрузку списка с комбобоксе, но пока хотя-бы просто чтоб он оставался не выбранным) П.С.: с бека явы возвращается текст ошибка, с текстом, и по ним можно выводить разные сообщения: "FlowSlotTicket with nID_FlowSlot=123 is bBusyStatic by getnID_Task_Activiti()=4556" -"Ці дата та час вже зайняті. оберіть інші, будь ласка" "FlowSlotTicket with nID_FlowSlot=123 is bBusyTemp by getnID_Task_Activiti()=4556" -"Ці дата та час вже зарезервовані. оберіть інші, будь ласка" При другой ошибке: -"Неможливо вибрати час. Спробуйте обрати інші або пізніше, будь ласка"
1.0
На главном портале, в форме услуги, при выборе времени в Электронной Очереди - выволить сообщение о ганятости, если возвращается ошибка об этом. - прикрутить диалог(хоть даже alert обычный): когда выбираеш время электронной очереди, и возникает ошибка - этого не видно, и визуально как будто-бы ничего не произошло... а уже после сабмита вываливает ошибку, причем другую (типа обьект поломанный) Можно просто выводить алерт с текстом "Ці дата та час вже зайняті. оберіть інші, будь ласка" (желательно после этого произваодить перезагрузку списка с комбобоксе, но пока хотя-бы просто чтоб он оставался не выбранным) П.С.: с бека явы возвращается текст ошибка, с текстом, и по ним можно выводить разные сообщения: "FlowSlotTicket with nID_FlowSlot=123 is bBusyStatic by getnID_Task_Activiti()=4556" -"Ці дата та час вже зайняті. оберіть інші, будь ласка" "FlowSlotTicket with nID_FlowSlot=123 is bBusyTemp by getnID_Task_Activiti()=4556" -"Ці дата та час вже зарезервовані. оберіть інші, будь ласка" При другой ошибке: -"Неможливо вибрати час. Спробуйте обрати інші або пізніше, будь ласка"
process
на главном портале в форме услуги при выборе времени в электронной очереди выволить сообщение о ганятости если возвращается ошибка об этом прикрутить диалог хоть даже alert обычный когда выбираеш время электронной очереди и возникает ошибка этого не видно и визуально как будто бы ничего не произошло а уже после сабмита вываливает ошибку причем другую типа обьект поломанный можно просто выводить алерт с текстом ці дата та час вже зайняті оберіть інші будь ласка желательно после этого произваодить перезагрузку списка с комбобоксе но пока хотя бы просто чтоб он оставался не выбранным п с с бека явы возвращается текст ошибка с текстом и по ним можно выводить разные сообщения flowslotticket with nid flowslot is bbusystatic by getnid task activiti ці дата та час вже зайняті оберіть інші будь ласка flowslotticket with nid flowslot is bbusytemp by getnid task activiti ці дата та час вже зарезервовані оберіть інші будь ласка при другой ошибке неможливо вибрати час спробуйте обрати інші або пізніше будь ласка
1
115,269
4,662,255,467
IssuesEvent
2016-10-05 02:33:21
ClinGen/clincoded
https://api.github.com/repos/ClinGen/clincoded
closed
Fix typos in PP3 description in Predictors Tab
external curator priority: high R7 release ready text change variant curation interface
The description for PP3 in the Predictors Tab has several typos, it currently says: "PP3: Multiple lines of computation evidence support a deleteriuos effect on the gene or gene product (conservation, evolutaionary, splicing impact, etc.) (has caveat)" but should say "PP3: Multiple lines of computational evidence support a deleterious effect on the gene or gene product (conservation, evolutionary, splicing impact, etc.) (has caveat)" Please also change the text for BP4 on the same tab, to be consistent with PP3 it should read: "BP4: Multiple lines of computational evidence suggest no impact on the gene or gene product (conservation, evolutionary, splicing impact, etc.) (has caveat)"
1.0
Fix typos in PP3 description in Predictors Tab - The description for PP3 in the Predictors Tab has several typos, it currently says: "PP3: Multiple lines of computation evidence support a deleteriuos effect on the gene or gene product (conservation, evolutaionary, splicing impact, etc.) (has caveat)" but should say "PP3: Multiple lines of computational evidence support a deleterious effect on the gene or gene product (conservation, evolutionary, splicing impact, etc.) (has caveat)" Please also change the text for BP4 on the same tab, to be consistent with PP3 it should read: "BP4: Multiple lines of computational evidence suggest no impact on the gene or gene product (conservation, evolutionary, splicing impact, etc.) (has caveat)"
non_process
fix typos in description in predictors tab the description for in the predictors tab has several typos it currently says multiple lines of computation evidence support a deleteriuos effect on the gene or gene product conservation evolutaionary splicing impact etc has caveat but should say multiple lines of computational evidence support a deleterious effect on the gene or gene product conservation evolutionary splicing impact etc has caveat please also change the text for on the same tab to be consistent with it should read multiple lines of computational evidence suggest no impact on the gene or gene product conservation evolutionary splicing impact etc has caveat
0
2,766
5,703,524,083
IssuesEvent
2017-04-18 00:14:11
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
System.Diagnostics.Tests.ProcessTests.TestProcessOnRemoteMachineWindows failure
area-System.Diagnostics.Process easy test bug up-for-grabs
System.Diagnostics.Tests.ProcessTests.TestProcessOnRemoteMachineWindows will fail if the machine that it is running on has inaccessible performance counter. We should move this test to outer loop, or make it robust against dirty environment. ``` System.Diagnostics.Tests.ProcessTests.TestProcessOnRemoteMachineWindows [FAIL] System.InvalidOperationException : Couldn't connect to remote machine. ---- System.InvalidOperationException : Couldn't get process information from performance counter. -------- System.InvalidOperationException : Cannot load Counter Name data because an invalid index 'RemoteFX Synth3D VSC VM Transport Channel' was read from the registry. Stack Trace: X:\dev\coreclr\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\ProcessManager.Windows.cs(521,0): at System.Diagnostics.NtProcessManager.GetProcessInfos(String machineName, Boolean isRemoteMachine) X:\dev\coreclr\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\ProcessManager.Windows.cs(521,0): at System.Diagnostics.NtProcessManager.GetProcessInfos(String machineName, Boolean isRemoteMachine) X:\dev\coreclr\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\ProcessManager.Windows.cs(266,0): at System.Diagnostics.NtProcessManager.GetProcessIds(String machineName, Boolean isRemoteMachine) X:\dev\coreclr\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\ProcessManager.Windows.cs(69,0): at System.Diagnostics.ProcessManager.GetProcessIds(String machineName) X:\dev\coreclr\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\ProcessManager.Windows.cs(30,0): at System.Diagnostics.ProcessManager.IsProcessRunning(Int32 processId, String machineName) ```
1.0
System.Diagnostics.Tests.ProcessTests.TestProcessOnRemoteMachineWindows failure - System.Diagnostics.Tests.ProcessTests.TestProcessOnRemoteMachineWindows will fail if the machine that it is running on has inaccessible performance counter. We should move this test to outer loop, or make it robust against dirty environment. ``` System.Diagnostics.Tests.ProcessTests.TestProcessOnRemoteMachineWindows [FAIL] System.InvalidOperationException : Couldn't connect to remote machine. ---- System.InvalidOperationException : Couldn't get process information from performance counter. -------- System.InvalidOperationException : Cannot load Counter Name data because an invalid index 'RemoteFX Synth3D VSC VM Transport Channel' was read from the registry. Stack Trace: X:\dev\coreclr\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\ProcessManager.Windows.cs(521,0): at System.Diagnostics.NtProcessManager.GetProcessInfos(String machineName, Boolean isRemoteMachine) X:\dev\coreclr\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\ProcessManager.Windows.cs(521,0): at System.Diagnostics.NtProcessManager.GetProcessInfos(String machineName, Boolean isRemoteMachine) X:\dev\coreclr\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\ProcessManager.Windows.cs(266,0): at System.Diagnostics.NtProcessManager.GetProcessIds(String machineName, Boolean isRemoteMachine) X:\dev\coreclr\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\ProcessManager.Windows.cs(69,0): at System.Diagnostics.ProcessManager.GetProcessIds(String machineName) X:\dev\coreclr\corefx\src\System.Diagnostics.Process\src\System\Diagnostics\ProcessManager.Windows.cs(30,0): at System.Diagnostics.ProcessManager.IsProcessRunning(Int32 processId, String machineName) ```
process
system diagnostics tests processtests testprocessonremotemachinewindows failure system diagnostics tests processtests testprocessonremotemachinewindows will fail if the machine that it is running on has inaccessible performance counter we should move this test to outer loop or make it robust against dirty environment system diagnostics tests processtests testprocessonremotemachinewindows system invalidoperationexception couldn t connect to remote machine system invalidoperationexception couldn t get process information from performance counter system invalidoperationexception cannot load counter name data because an invalid index remotefx vsc vm transport channel was read from the registry stack trace x dev coreclr corefx src system diagnostics process src system diagnostics processmanager windows cs at system diagnostics ntprocessmanager getprocessinfos string machinename boolean isremotemachine x dev coreclr corefx src system diagnostics process src system diagnostics processmanager windows cs at system diagnostics ntprocessmanager getprocessinfos string machinename boolean isremotemachine x dev coreclr corefx src system diagnostics process src system diagnostics processmanager windows cs at system diagnostics ntprocessmanager getprocessids string machinename boolean isremotemachine x dev coreclr corefx src system diagnostics process src system diagnostics processmanager windows cs at system diagnostics processmanager getprocessids string machinename x dev coreclr corefx src system diagnostics process src system diagnostics processmanager windows cs at system diagnostics processmanager isprocessrunning processid string machinename
1
330,011
28,322,987,215
IssuesEvent
2023-04-11 03:54:35
pravega/pravega
https://api.github.com/repos/pravega/pravega
closed
Consumption Based Retention with Controller and Segment Store restarts (failover)
area/controller area/testing priority/P3
**Problem description** Repeat the same steps as in #5554 with Controller restarting and later segment store restarting during the test. **Problem location** System Tests
1.0
Consumption Based Retention with Controller and Segment Store restarts (failover) - **Problem description** Repeat the same steps as in #5554 with Controller restarting and later segment store restarting during the test. **Problem location** System Tests
non_process
consumption based retention with controller and segment store restarts failover problem description repeat the same steps as in with controller restarting and later segment store restarting during the test problem location system tests
0
662,005
22,100,382,760
IssuesEvent
2022-06-01 13:21:54
OpenNebula/one
https://api.github.com/repos/OpenNebula/one
opened
Sunstone VM Migration shows DS on KVM driver
Category: Sunstone Type: Bug Sponsored Status: Accepted Priority: Normal
**Description** Some VMs in OpenNebula doesn't have the HYPERVISOR attribute, this creates misbehavior in the migration dialog. **To Reproduce** Create 2 VMs, the first one with `HYPERVISOR=KVM` and the second without it, and check the migration dialogs. **Expected behavior** Datastore change on live migration is not available on KVM **Details** - Affected Component: Sunstone - Hypervisor: KVM - Version: [e.g. 5.4.12 or development] **Additional context** Remove any reference to the HYPERVISOR attribute and check this according to the VM HISTORY attribute. <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [ ] Code committed - [ ] Testing - QA - [ ] Documentation (Release notes - resolved issues, compatibility, known issues)
1.0
Sunstone VM Migration shows DS on KVM driver - **Description** Some VMs in OpenNebula doesn't have the HYPERVISOR attribute, this creates misbehavior in the migration dialog. **To Reproduce** Create 2 VMs, the first one with `HYPERVISOR=KVM` and the second without it, and check the migration dialogs. **Expected behavior** Datastore change on live migration is not available on KVM **Details** - Affected Component: Sunstone - Hypervisor: KVM - Version: [e.g. 5.4.12 or development] **Additional context** Remove any reference to the HYPERVISOR attribute and check this according to the VM HISTORY attribute. <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [ ] Code committed - [ ] Testing - QA - [ ] Documentation (Release notes - resolved issues, compatibility, known issues)
non_process
sunstone vm migration shows ds on kvm driver description some vms in opennebula doesn t have the hypervisor attribute this creates misbehavior in the migration dialog to reproduce create vms the first one with hypervisor kvm and the second without it and check the migration dialogs expected behavior datastore change on live migration is not available on kvm details affected component sunstone hypervisor kvm version additional context remove any reference to the hypervisor attribute and check this according to the vm history attribute progress status code committed testing qa documentation release notes resolved issues compatibility known issues
0
118,952
15,382,329,507
IssuesEvent
2021-03-03 00:24:46
hackforla/food-oasis
https://api.github.com/repos/hackforla/food-oasis
reopened
Create Mock Up of Map
design: ui/ux
Overview As a food seeker I need a simple landing page so that I can find food. Action Items - [x] Document the points of friction in a comment below (eg having to click the magnifying glass after entering address) - [x] Create Simple mock up of the map page Resources/Instructions [Mockups for A/B Testing](https://www.figma.com/file/D3oq9QOXl0rFkwrEUJbABs/Fola-Design-System?node-id=2591%3A0) [John's Dev site ](https://devla.foodoasis.net/)
1.0
Create Mock Up of Map - Overview As a food seeker I need a simple landing page so that I can find food. Action Items - [x] Document the points of friction in a comment below (eg having to click the magnifying glass after entering address) - [x] Create Simple mock up of the map page Resources/Instructions [Mockups for A/B Testing](https://www.figma.com/file/D3oq9QOXl0rFkwrEUJbABs/Fola-Design-System?node-id=2591%3A0) [John's Dev site ](https://devla.foodoasis.net/)
non_process
create mock up of map overview as a food seeker i need a simple landing page so that i can find food action items document the points of friction in a comment below eg having to click the magnifying glass after entering address create simple mock up of the map page resources instructions
0
64,918
6,926,718,741
IssuesEvent
2017-11-30 20:09:54
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Ipsec router resource consumption (CPU, memory) is increasing indefinitely after upgrade to Rancher 1.6.11
area/networking kind/bug priority/-1 status/resolved status/to-test version/1.6
**Rancher versions:** rancher/server: 1.6.11 rancher/agent: 1.2.7 **Infrastructure Stack versions:** healthcheck: 0.3.3 ipsec: 0.2.0 network-services: 0.2.7 scheduler: 0.6.3 **Docker version: (`docker version`,`docker info` preferred)** ``` Client: Version: 17.03.2-ce API version: 1.27 Go version: go1.7.5 Git commit: f5ec1e2 Built: Tue Jun 27 03:35:14 2017 OS/Arch: linux/amd64 Server: Version: 17.03.2-ce API version: 1.27 (minimum version 1.12) Go version: go1.7.5 Git commit: f5ec1e2 Built: Tue Jun 27 03:35:14 2017 OS/Arch: linux/amd64 Experimental: false ``` **Operating system and kernel: (`cat /etc/os-release`, `uname -r` preferred)** ``` NAME="Ubuntu" VERSION="16.04.3 LTS (Xenial Xerus)" ID=ubuntu ID_LIKE=debian PRETTY_NAME="Ubuntu 16.04.3 LTS" VERSION_ID="16.04" HOME_URL="http://www.ubuntu.com/" SUPPORT_URL="http://help.ubuntu.com/" BUG_REPORT_URL="http://bugs.launchpad.net/ubuntu/" VERSION_CODENAME=xenial UBUNTU_CODENAME=xenia ``` **Type/provider of hosts: (VirtualBox/Bare-metal/AWS/GCE/DO)** AWS **Setup details: (single node rancher vs. HA rancher, internal DB vs. external DB)** HA rancher (usually only one server is running) on AWS with AWS Aurora RDS **Environment Template: (Cattle/Kubernetes/Swarm/Mesos)** Cattle **Steps to Reproduce:** Not 100% sure about this, here's what we did: - Upgraded to Rancher 1.6.11 on Nov. 15 at night. - Upgraded infrastructure stacks on Nov. 16 in the morning (all default parameters, no modifications) **Results:** - Since then CPU and memory usage constantly increased (see graphs below) on all hosts in all environments connected to the Rancher server. Some are servers are backend servers which have a constant load. - It seems to be related to the ipsec-router container. The currently highest memory usage is 740 MB with CPU load peaks every 2-3 seconds causing 100%+ CPU usage for about 5 seconds (see graph below). - The load and memory usage of the ipsec-router builds up over time (initial memory usage is about 19MB) **Metric graphs** ipsec router metrics (logs of said containers below): ![image](https://user-images.githubusercontent.com/2604142/33032028-d2d06ad4-cdd4-11e7-8906-19634f9dbabf.png) frontend cluster metrics: ![image](https://user-images.githubusercontent.com/2604142/33032043-e31bea8a-cdd4-11e7-970c-8633596b4ce6.png) **Logs** ``` 11/20/2017 9:23:59 AM05[IKE] peer not responding, trying again (1649/0) 11/20/2017 9:23:59 AM05[IKE] initiating IKE_SA conn-54.236.242.153[2] to 54.236.242.153 11/20/2017 9:23:59 AM05[ENC] generating IKE_SA_INIT request 0 [ SA KE No N(NATD_S_IP) N(NATD_D_IP) N(FRAG_SUP) N(HASH_ALG) N(REDIR_SUP) ] 11/20/2017 9:23:59 AM05[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:23:59 AM07[IKE] retransmit 5 of request with message ID 0 11/20/2017 9:23:59 AM07[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:23:59 AM15[IKE] retransmit 1 of request with message ID 0 11/20/2017 9:23:59 AM15[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:23:59 AM13[IKE] retransmit 2 of request with message ID 0 11/20/2017 9:23:59 AM13[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:23:59 AM14[IKE] retransmit 4 of request with message ID 0 11/20/2017 9:23:59 AM14[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:23:59 AM10[IKE] retransmit 3 of request with message ID 0 11/20/2017 9:23:59 AM10[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:23:59 AM15[IKE] retransmit 4 of request with message ID 0 11/20/2017 9:23:59 AM15[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:23:59 AM13[IKE] retransmit 5 of request with message ID 0 11/20/2017 9:23:59 AM13[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:23:59 AM13[IKE] giving up after 5 retransmits 11/20/2017 9:23:59 AM13[IKE] peer not responding, trying again (79/0) 11/20/2017 9:23:59 AM13[IKE] initiating IKE_SA conn-34.200.226.240[3] to 34.200.226.240 11/20/2017 9:23:59 AM13[ENC] generating IKE_SA_INIT request 0 [ SA KE No N(NATD_S_IP) N(NATD_D_IP) N(FRAG_SUP) N(HASH_ALG) N(REDIR_SUP) ] 11/20/2017 9:23:59 AM13[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:23:59 AM04[IKE] retransmit 1 of request with message ID 0 11/20/2017 9:23:59 AM04[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:23:59 AM04[IKE] retransmit 2 of request with message ID 0 11/20/2017 9:23:59 AM04[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:23:59 AM08[IKE] retransmit 5 of request with message ID 0 11/20/2017 9:23:59 AM08[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:23:59 AM11[IKE] retransmit 3 of request with message ID 0 11/20/2017 9:23:59 AM11[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:23:59 AM04[IKE] retransmit 4 of request with message ID 0 11/20/2017 9:23:59 AM04[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:23:59 AM09[IKE] giving up after 5 retransmits 11/20/2017 9:23:59 AM09[IKE] peer not responding, trying again (2173/0) 11/20/2017 9:23:59 AM09[IKE] initiating IKE_SA conn-54.160.220.166[1] to 54.160.220.166 11/20/2017 9:23:59 AM09[ENC] generating IKE_SA_INIT request 0 [ SA KE No N(NATD_S_IP) N(NATD_D_IP) N(FRAG_SUP) N(HASH_ALG) N(REDIR_SUP) ] 11/20/2017 9:23:59 AM09[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:23:59 AM08[IKE] retransmit 1 of request with message ID 0 11/20/2017 9:23:59 AM08[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:27:20 AM06[IKE] retransmit 2 of request with message ID 0 11/20/2017 9:27:20 AM06[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:27:20 AM05[IKE] retransmit 3 of request with message ID 0 11/20/2017 9:27:20 AM05[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:27:20 AM15[IKE] giving up after 5 retransmits 11/20/2017 9:27:20 AM15[IKE] peer not responding, trying again (1650/0) 11/20/2017 9:27:20 AM15[IKE] initiating IKE_SA conn-54.236.242.153[2] to 54.236.242.153 11/20/2017 9:27:20 AM15[ENC] generating IKE_SA_INIT request 0 [ SA KE No N(NATD_S_IP) N(NATD_D_IP) N(FRAG_SUP) N(HASH_ALG) N(REDIR_SUP) ] 11/20/2017 9:27:20 AM15[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:27:20 AM10[IKE] retransmit 5 of request with message ID 0 11/20/2017 9:27:20 AM10[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:27:20 AM14[IKE] retransmit 1 of request with message ID 0 11/20/2017 9:27:20 AM14[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:27:20 AM11[IKE] retransmit 2 of request with message ID 0 11/20/2017 9:27:20 AM11[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:27:20 AM09[IKE] retransmit 4 of request with message ID 0 11/20/2017 9:27:20 AM09[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:27:20 AM07[IKE] retransmit 3 of request with message ID 0 11/20/2017 9:27:20 AM07[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:27:20 AM06[IKE] retransmit 4 of request with message ID 0 11/20/2017 9:27:20 AM06[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:27:20 AM15[IKE] retransmit 5 of request with message ID 0 11/20/2017 9:27:20 AM15[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:27:20 AM10[IKE] giving up after 5 retransmits 11/20/2017 9:27:20 AM10[IKE] peer not responding, trying again (80/0) 11/20/2017 9:27:20 AM10[IKE] initiating IKE_SA conn-34.200.226.240[3] to 34.200.226.240 11/20/2017 9:27:20 AM10[ENC] generating IKE_SA_INIT request 0 [ SA KE No N(NATD_S_IP) N(NATD_D_IP) N(FRAG_SUP) N(HASH_ALG) N(REDIR_SUP) ] 11/20/2017 9:27:20 AM10[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:27:20 AM07[IKE] retransmit 1 of request with message ID 0 11/20/2017 9:27:20 AM07[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:27:20 AM04[IKE] retransmit 2 of request with message ID 0 11/20/2017 9:27:20 AM04[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:27:20 AM14[IKE] retransmit 5 of request with message ID 0 11/20/2017 9:27:20 AM14[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:27:20 AM12[IKE] retransmit 3 of request with message ID 0 11/20/2017 9:27:20 AM12[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:27:20 AM09[IKE] retransmit 4 of request with message ID 0 11/20/2017 9:27:20 AM09[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:27:20 AM15[IKE] giving up after 5 retransmits 11/20/2017 9:27:20 AM15[IKE] peer not responding, trying again (2174/0) 11/20/2017 9:27:20 AM15[IKE] initiating IKE_SA conn-54.160.220.166[1] to 54.160.220.166 11/20/2017 9:27:20 AM15[ENC] generating IKE_SA_INIT request 0 [ SA KE No N(NATD_S_IP) N(NATD_D_IP) N(FRAG_SUP) N(HASH_ALG) N(REDIR_SUP) ] 11/20/2017 9:27:20 AM15[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:27:20 AM14[IKE] retransmit 1 of request with message ID 0 11/20/2017 9:27:20 AM14[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:27:20 AM13[IKE] retransmit 2 of request with message ID 0 11/20/2017 9:27:20 AM13[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:27:20 AM09[IKE] retransmit 3 of request with message ID 0 11/20/2017 9:27:20 AM09[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:27:20 AM14[IKE] giving up after 5 retransmits 11/20/2017 9:27:20 AM14[IKE] peer not responding, trying again (1651/0) 11/20/2017 9:27:20 AM14[IKE] initiating IKE_SA conn-54.236.242.153[2] to 54.236.242.153 11/20/2017 9:27:20 AM14[ENC] generating IKE_SA_INIT request 0 [ SA KE No N(NATD_S_IP) N(NATD_D_IP) N(FRAG_SUP) N(HASH_ALG) N(REDIR_SUP) ] 11/20/2017 9:27:20 AM14[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:27:20 AM04[IKE] retransmit 5 of request with message ID 0 11/20/2017 9:27:20 AM04[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:27:20 AM10[IKE] retransmit 1 of request with message ID 0 11/20/2017 9:27:20 AM10[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:27:20 AM10[IKE] retransmit 2 of request with message ID 0 11/20/2017 9:27:20 AM10[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) ```
1.0
Ipsec router resource consumption (CPU, memory) is increasing indefinitely after upgrade to Rancher 1.6.11 - **Rancher versions:** rancher/server: 1.6.11 rancher/agent: 1.2.7 **Infrastructure Stack versions:** healthcheck: 0.3.3 ipsec: 0.2.0 network-services: 0.2.7 scheduler: 0.6.3 **Docker version: (`docker version`,`docker info` preferred)** ``` Client: Version: 17.03.2-ce API version: 1.27 Go version: go1.7.5 Git commit: f5ec1e2 Built: Tue Jun 27 03:35:14 2017 OS/Arch: linux/amd64 Server: Version: 17.03.2-ce API version: 1.27 (minimum version 1.12) Go version: go1.7.5 Git commit: f5ec1e2 Built: Tue Jun 27 03:35:14 2017 OS/Arch: linux/amd64 Experimental: false ``` **Operating system and kernel: (`cat /etc/os-release`, `uname -r` preferred)** ``` NAME="Ubuntu" VERSION="16.04.3 LTS (Xenial Xerus)" ID=ubuntu ID_LIKE=debian PRETTY_NAME="Ubuntu 16.04.3 LTS" VERSION_ID="16.04" HOME_URL="http://www.ubuntu.com/" SUPPORT_URL="http://help.ubuntu.com/" BUG_REPORT_URL="http://bugs.launchpad.net/ubuntu/" VERSION_CODENAME=xenial UBUNTU_CODENAME=xenia ``` **Type/provider of hosts: (VirtualBox/Bare-metal/AWS/GCE/DO)** AWS **Setup details: (single node rancher vs. HA rancher, internal DB vs. external DB)** HA rancher (usually only one server is running) on AWS with AWS Aurora RDS **Environment Template: (Cattle/Kubernetes/Swarm/Mesos)** Cattle **Steps to Reproduce:** Not 100% sure about this, here's what we did: - Upgraded to Rancher 1.6.11 on Nov. 15 at night. - Upgraded infrastructure stacks on Nov. 16 in the morning (all default parameters, no modifications) **Results:** - Since then CPU and memory usage constantly increased (see graphs below) on all hosts in all environments connected to the Rancher server. Some are servers are backend servers which have a constant load. - It seems to be related to the ipsec-router container. The currently highest memory usage is 740 MB with CPU load peaks every 2-3 seconds causing 100%+ CPU usage for about 5 seconds (see graph below). - The load and memory usage of the ipsec-router builds up over time (initial memory usage is about 19MB) **Metric graphs** ipsec router metrics (logs of said containers below): ![image](https://user-images.githubusercontent.com/2604142/33032028-d2d06ad4-cdd4-11e7-8906-19634f9dbabf.png) frontend cluster metrics: ![image](https://user-images.githubusercontent.com/2604142/33032043-e31bea8a-cdd4-11e7-970c-8633596b4ce6.png) **Logs** ``` 11/20/2017 9:23:59 AM05[IKE] peer not responding, trying again (1649/0) 11/20/2017 9:23:59 AM05[IKE] initiating IKE_SA conn-54.236.242.153[2] to 54.236.242.153 11/20/2017 9:23:59 AM05[ENC] generating IKE_SA_INIT request 0 [ SA KE No N(NATD_S_IP) N(NATD_D_IP) N(FRAG_SUP) N(HASH_ALG) N(REDIR_SUP) ] 11/20/2017 9:23:59 AM05[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:23:59 AM07[IKE] retransmit 5 of request with message ID 0 11/20/2017 9:23:59 AM07[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:23:59 AM15[IKE] retransmit 1 of request with message ID 0 11/20/2017 9:23:59 AM15[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:23:59 AM13[IKE] retransmit 2 of request with message ID 0 11/20/2017 9:23:59 AM13[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:23:59 AM14[IKE] retransmit 4 of request with message ID 0 11/20/2017 9:23:59 AM14[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:23:59 AM10[IKE] retransmit 3 of request with message ID 0 11/20/2017 9:23:59 AM10[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:23:59 AM15[IKE] retransmit 4 of request with message ID 0 11/20/2017 9:23:59 AM15[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:23:59 AM13[IKE] retransmit 5 of request with message ID 0 11/20/2017 9:23:59 AM13[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:23:59 AM13[IKE] giving up after 5 retransmits 11/20/2017 9:23:59 AM13[IKE] peer not responding, trying again (79/0) 11/20/2017 9:23:59 AM13[IKE] initiating IKE_SA conn-34.200.226.240[3] to 34.200.226.240 11/20/2017 9:23:59 AM13[ENC] generating IKE_SA_INIT request 0 [ SA KE No N(NATD_S_IP) N(NATD_D_IP) N(FRAG_SUP) N(HASH_ALG) N(REDIR_SUP) ] 11/20/2017 9:23:59 AM13[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:23:59 AM04[IKE] retransmit 1 of request with message ID 0 11/20/2017 9:23:59 AM04[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:23:59 AM04[IKE] retransmit 2 of request with message ID 0 11/20/2017 9:23:59 AM04[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:23:59 AM08[IKE] retransmit 5 of request with message ID 0 11/20/2017 9:23:59 AM08[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:23:59 AM11[IKE] retransmit 3 of request with message ID 0 11/20/2017 9:23:59 AM11[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:23:59 AM04[IKE] retransmit 4 of request with message ID 0 11/20/2017 9:23:59 AM04[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:23:59 AM09[IKE] giving up after 5 retransmits 11/20/2017 9:23:59 AM09[IKE] peer not responding, trying again (2173/0) 11/20/2017 9:23:59 AM09[IKE] initiating IKE_SA conn-54.160.220.166[1] to 54.160.220.166 11/20/2017 9:23:59 AM09[ENC] generating IKE_SA_INIT request 0 [ SA KE No N(NATD_S_IP) N(NATD_D_IP) N(FRAG_SUP) N(HASH_ALG) N(REDIR_SUP) ] 11/20/2017 9:23:59 AM09[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:23:59 AM08[IKE] retransmit 1 of request with message ID 0 11/20/2017 9:23:59 AM08[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:27:20 AM06[IKE] retransmit 2 of request with message ID 0 11/20/2017 9:27:20 AM06[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:27:20 AM05[IKE] retransmit 3 of request with message ID 0 11/20/2017 9:27:20 AM05[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:27:20 AM15[IKE] giving up after 5 retransmits 11/20/2017 9:27:20 AM15[IKE] peer not responding, trying again (1650/0) 11/20/2017 9:27:20 AM15[IKE] initiating IKE_SA conn-54.236.242.153[2] to 54.236.242.153 11/20/2017 9:27:20 AM15[ENC] generating IKE_SA_INIT request 0 [ SA KE No N(NATD_S_IP) N(NATD_D_IP) N(FRAG_SUP) N(HASH_ALG) N(REDIR_SUP) ] 11/20/2017 9:27:20 AM15[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:27:20 AM10[IKE] retransmit 5 of request with message ID 0 11/20/2017 9:27:20 AM10[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:27:20 AM14[IKE] retransmit 1 of request with message ID 0 11/20/2017 9:27:20 AM14[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:27:20 AM11[IKE] retransmit 2 of request with message ID 0 11/20/2017 9:27:20 AM11[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:27:20 AM09[IKE] retransmit 4 of request with message ID 0 11/20/2017 9:27:20 AM09[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:27:20 AM07[IKE] retransmit 3 of request with message ID 0 11/20/2017 9:27:20 AM07[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:27:20 AM06[IKE] retransmit 4 of request with message ID 0 11/20/2017 9:27:20 AM06[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:27:20 AM15[IKE] retransmit 5 of request with message ID 0 11/20/2017 9:27:20 AM15[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:27:20 AM10[IKE] giving up after 5 retransmits 11/20/2017 9:27:20 AM10[IKE] peer not responding, trying again (80/0) 11/20/2017 9:27:20 AM10[IKE] initiating IKE_SA conn-34.200.226.240[3] to 34.200.226.240 11/20/2017 9:27:20 AM10[ENC] generating IKE_SA_INIT request 0 [ SA KE No N(NATD_S_IP) N(NATD_D_IP) N(FRAG_SUP) N(HASH_ALG) N(REDIR_SUP) ] 11/20/2017 9:27:20 AM10[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:27:20 AM07[IKE] retransmit 1 of request with message ID 0 11/20/2017 9:27:20 AM07[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:27:20 AM04[IKE] retransmit 2 of request with message ID 0 11/20/2017 9:27:20 AM04[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:27:20 AM14[IKE] retransmit 5 of request with message ID 0 11/20/2017 9:27:20 AM14[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:27:20 AM12[IKE] retransmit 3 of request with message ID 0 11/20/2017 9:27:20 AM12[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:27:20 AM09[IKE] retransmit 4 of request with message ID 0 11/20/2017 9:27:20 AM09[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:27:20 AM15[IKE] giving up after 5 retransmits 11/20/2017 9:27:20 AM15[IKE] peer not responding, trying again (2174/0) 11/20/2017 9:27:20 AM15[IKE] initiating IKE_SA conn-54.160.220.166[1] to 54.160.220.166 11/20/2017 9:27:20 AM15[ENC] generating IKE_SA_INIT request 0 [ SA KE No N(NATD_S_IP) N(NATD_D_IP) N(FRAG_SUP) N(HASH_ALG) N(REDIR_SUP) ] 11/20/2017 9:27:20 AM15[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:27:20 AM14[IKE] retransmit 1 of request with message ID 0 11/20/2017 9:27:20 AM14[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:27:20 AM13[IKE] retransmit 2 of request with message ID 0 11/20/2017 9:27:20 AM13[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:27:20 AM09[IKE] retransmit 3 of request with message ID 0 11/20/2017 9:27:20 AM09[NET] sending packet: from 10.42.245.238[500] to 54.160.220.166[500] (500 bytes) 11/20/2017 9:27:20 AM14[IKE] giving up after 5 retransmits 11/20/2017 9:27:20 AM14[IKE] peer not responding, trying again (1651/0) 11/20/2017 9:27:20 AM14[IKE] initiating IKE_SA conn-54.236.242.153[2] to 54.236.242.153 11/20/2017 9:27:20 AM14[ENC] generating IKE_SA_INIT request 0 [ SA KE No N(NATD_S_IP) N(NATD_D_IP) N(FRAG_SUP) N(HASH_ALG) N(REDIR_SUP) ] 11/20/2017 9:27:20 AM14[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:27:20 AM04[IKE] retransmit 5 of request with message ID 0 11/20/2017 9:27:20 AM04[NET] sending packet: from 10.42.245.238[500] to 34.200.226.240[500] (500 bytes) 11/20/2017 9:27:20 AM10[IKE] retransmit 1 of request with message ID 0 11/20/2017 9:27:20 AM10[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) 11/20/2017 9:27:20 AM10[IKE] retransmit 2 of request with message ID 0 11/20/2017 9:27:20 AM10[NET] sending packet: from 10.42.245.238[500] to 54.236.242.153[500] (500 bytes) ```
non_process
ipsec router resource consumption cpu memory is increasing indefinitely after upgrade to rancher rancher versions rancher server rancher agent infrastructure stack versions healthcheck ipsec network services scheduler docker version docker version docker info preferred client version ce api version go version git commit built tue jun os arch linux server version ce api version minimum version go version git commit built tue jun os arch linux experimental false operating system and kernel cat etc os release uname r preferred name ubuntu version lts xenial xerus id ubuntu id like debian pretty name ubuntu lts version id home url support url bug report url version codename xenial ubuntu codename xenia type provider of hosts virtualbox bare metal aws gce do aws setup details single node rancher vs ha rancher internal db vs external db ha rancher usually only one server is running on aws with aws aurora rds environment template cattle kubernetes swarm mesos cattle steps to reproduce not sure about this here s what we did upgraded to rancher on nov at night upgraded infrastructure stacks on nov in the morning all default parameters no modifications results since then cpu and memory usage constantly increased see graphs below on all hosts in all environments connected to the rancher server some are servers are backend servers which have a constant load it seems to be related to the ipsec router container the currently highest memory usage is mb with cpu load peaks every seconds causing cpu usage for about seconds see graph below the load and memory usage of the ipsec router builds up over time initial memory usage is about metric graphs ipsec router metrics logs of said containers below frontend cluster metrics logs peer not responding trying again initiating ike sa conn to generating ike sa init request sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes giving up after retransmits peer not responding trying again initiating ike sa conn to generating ike sa init request sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes giving up after retransmits peer not responding trying again initiating ike sa conn to generating ike sa init request sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes giving up after retransmits peer not responding trying again initiating ike sa conn to generating ike sa init request sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes giving up after retransmits peer not responding trying again initiating ike sa conn to generating ike sa init request sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes giving up after retransmits peer not responding trying again initiating ike sa conn to generating ike sa init request sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes giving up after retransmits peer not responding trying again initiating ike sa conn to generating ike sa init request sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes retransmit of request with message id sending packet from to bytes
0
20,631
27,314,544,265
IssuesEvent
2023-02-24 14:45:34
ESMValGroup/ESMValCore
https://api.github.com/repos/ESMValGroup/ESMValCore
closed
Regridding to reference dataset fails sometimes
bug preprocessor
Not an urgent issue, but something I noticed: Regridding to a reference dataset, e.g. ```yaml regrid_to_ref: regrid: target_grid: GFDL-CM3 scheme: linear ``` invokes a simple `iris.load_cube` https://github.com/ESMValGroup/ESMValTool/blob/43598ee48a327fc3ef2a6e53768324d02ad61863/esmvaltool/preprocessor/_regrid.py#L169 on the **raw** nc file, which may fail in rare cases if the file contains multiple variables (e.g `rsdt` of GFDL-CM3), which leads to errors like this: ``` Traceback (most recent call last): File "<stdin>", line 1, in <module> File "~/anaconda3/envs/esmvaltool/lib/python3.6/site-packages/iris/__init__.py", line 376, in load_cube raise iris.exceptions.ConstraintMismatchError(str(e)) iris.exceptions.ConstraintMismatchError: failed to merge into a single cube. cube.long_name differs: 'Length of average period' != 'End time for average period' cube.var_name differs: 'average_DT' != 'average_T2' cube.units differs: Unit('days') != Unit('days since 1860-01-01 00:00:00', calendar='gregorian') ``` I think we need an advanced loading function here.
1.0
Regridding to reference dataset fails sometimes - Not an urgent issue, but something I noticed: Regridding to a reference dataset, e.g. ```yaml regrid_to_ref: regrid: target_grid: GFDL-CM3 scheme: linear ``` invokes a simple `iris.load_cube` https://github.com/ESMValGroup/ESMValTool/blob/43598ee48a327fc3ef2a6e53768324d02ad61863/esmvaltool/preprocessor/_regrid.py#L169 on the **raw** nc file, which may fail in rare cases if the file contains multiple variables (e.g `rsdt` of GFDL-CM3), which leads to errors like this: ``` Traceback (most recent call last): File "<stdin>", line 1, in <module> File "~/anaconda3/envs/esmvaltool/lib/python3.6/site-packages/iris/__init__.py", line 376, in load_cube raise iris.exceptions.ConstraintMismatchError(str(e)) iris.exceptions.ConstraintMismatchError: failed to merge into a single cube. cube.long_name differs: 'Length of average period' != 'End time for average period' cube.var_name differs: 'average_DT' != 'average_T2' cube.units differs: Unit('days') != Unit('days since 1860-01-01 00:00:00', calendar='gregorian') ``` I think we need an advanced loading function here.
process
regridding to reference dataset fails sometimes not an urgent issue but something i noticed regridding to a reference dataset e g yaml regrid to ref regrid target grid gfdl scheme linear invokes a simple iris load cube on the raw nc file which may fail in rare cases if the file contains multiple variables e g rsdt of gfdl which leads to errors like this traceback most recent call last file line in file envs esmvaltool lib site packages iris init py line in load cube raise iris exceptions constraintmismatcherror str e iris exceptions constraintmismatcherror failed to merge into a single cube cube long name differs length of average period end time for average period cube var name differs average dt average cube units differs unit days unit days since calendar gregorian i think we need an advanced loading function here
1
17,555
23,368,995,529
IssuesEvent
2022-08-10 17:58:48
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
opened
How not to break "out of tree" users
RFC Process
Provide some guarantees, guidelines and a process keeping out of tree users operational while the zephyr project code advances with new technologies, code cleanups and other major code and API changes. Out of tree users are not limited to only drivers, we have users with their own subsystems, architectures, toolchains, SoCs, boards, drivers, driver subsystems etc. Any change in zephyr might break such users if changes are not following a deprecation process, announcements and a grace period given (deprecation period in many cases) to those users to adapt to the new interfaces or upstream code. The process should find the sweet spot which allows the project to advance with its agenda and roadmap while allowing users to adapt to change.
1.0
How not to break "out of tree" users - Provide some guarantees, guidelines and a process keeping out of tree users operational while the zephyr project code advances with new technologies, code cleanups and other major code and API changes. Out of tree users are not limited to only drivers, we have users with their own subsystems, architectures, toolchains, SoCs, boards, drivers, driver subsystems etc. Any change in zephyr might break such users if changes are not following a deprecation process, announcements and a grace period given (deprecation period in many cases) to those users to adapt to the new interfaces or upstream code. The process should find the sweet spot which allows the project to advance with its agenda and roadmap while allowing users to adapt to change.
process
how not to break out of tree users provide some guarantees guidelines and a process keeping out of tree users operational while the zephyr project code advances with new technologies code cleanups and other major code and api changes out of tree users are not limited to only drivers we have users with their own subsystems architectures toolchains socs boards drivers driver subsystems etc any change in zephyr might break such users if changes are not following a deprecation process announcements and a grace period given deprecation period in many cases to those users to adapt to the new interfaces or upstream code the process should find the sweet spot which allows the project to advance with its agenda and roadmap while allowing users to adapt to change
1
785,113
27,598,886,802
IssuesEvent
2023-03-09 08:42:23
radical-cybertools/radical.pilot
https://api.github.com/repos/radical-cybertools/radical.pilot
opened
API simplification proposal
topic:api type:enhancement priority:medium
Our example codes and virtually all application codes currently use the same sequence for application initialization: ```py session = rp.Session() tmgr = rp.TaskManager(session) pmgr = rp.PilotManager(session) pilot = pmgr.submit_pilots(...) tmgr.add_pilots(pilot) tasks = tmgr.submit_tasks(...) ``` With a backward-compatible addition of two calls to the `rp.Session()` object, that initialization can be significantly simplified to: ```py session = rp.Session() pilot = session.submit_pilots(...) tasks = session.submit_tasks(...) ``` The session would internally hold a private `PilotManager` and `TaskManager` and simply proxy those calls.
1.0
API simplification proposal - Our example codes and virtually all application codes currently use the same sequence for application initialization: ```py session = rp.Session() tmgr = rp.TaskManager(session) pmgr = rp.PilotManager(session) pilot = pmgr.submit_pilots(...) tmgr.add_pilots(pilot) tasks = tmgr.submit_tasks(...) ``` With a backward-compatible addition of two calls to the `rp.Session()` object, that initialization can be significantly simplified to: ```py session = rp.Session() pilot = session.submit_pilots(...) tasks = session.submit_tasks(...) ``` The session would internally hold a private `PilotManager` and `TaskManager` and simply proxy those calls.
non_process
api simplification proposal our example codes and virtually all application codes currently use the same sequence for application initialization py session rp session tmgr rp taskmanager session pmgr rp pilotmanager session pilot pmgr submit pilots tmgr add pilots pilot tasks tmgr submit tasks with a backward compatible addition of two calls to the rp session object that initialization can be significantly simplified to py session rp session pilot session submit pilots tasks session submit tasks the session would internally hold a private pilotmanager and taskmanager and simply proxy those calls
0
232,714
25,603,492,830
IssuesEvent
2022-12-01 22:34:50
amplify-education/tmp_SAST_eval_skf-labs
https://api.github.com/repos/amplify-education/tmp_SAST_eval_skf-labs
opened
jquery-3.1.1.min.js: 3 vulnerabilities (highest severity is: 6.1)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-3.1.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js</a></p> <p> <p>Found in HEAD commit: <a href="https://github.com/amplify-education/tmp_SAST_eval_skf-labs/commit/91e1032bf85eea611eeb66711dc8b85719f6752e">91e1032bf85eea611eeb66711dc8b85719f6752e</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (jquery version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2020-11023](https://www.mend.io/vulnerability-database/CVE-2020-11023) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-3.1.1.min.js | Direct | jquery - 3.5.0;jquery-rails - 4.4.0 | &#10060; | | [CVE-2020-11022](https://www.mend.io/vulnerability-database/CVE-2020-11022) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-3.1.1.min.js | Direct | jQuery - 3.5.0 | &#10060; | | [CVE-2019-11358](https://www.mend.io/vulnerability-database/CVE-2019-11358) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-3.1.1.min.js | Direct | jquery - 3.4.0 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11023</summary> ### Vulnerable Library - <b>jquery-3.1.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js</a></p> <p> Dependency Hierarchy: - :x: **jquery-3.1.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/amplify-education/tmp_SAST_eval_skf-labs/commit/91e1032bf85eea611eeb66711dc8b85719f6752e">91e1032bf85eea611eeb66711dc8b85719f6752e</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-11023>CVE-2020-11023</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440">https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jquery - 3.5.0;jquery-rails - 4.4.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11022</summary> ### Vulnerable Library - <b>jquery-3.1.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js</a></p> <p> Dependency Hierarchy: - :x: **jquery-3.1.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/amplify-education/tmp_SAST_eval_skf-labs/commit/91e1032bf85eea611eeb66711dc8b85719f6752e">91e1032bf85eea611eeb66711dc8b85719f6752e</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-11022>CVE-2020-11022</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2019-11358</summary> ### Vulnerable Library - <b>jquery-3.1.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js</a></p> <p> Dependency Hierarchy: - :x: **jquery-3.1.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/amplify-education/tmp_SAST_eval_skf-labs/commit/91e1032bf85eea611eeb66711dc8b85719f6752e">91e1032bf85eea611eeb66711dc8b85719f6752e</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype. <p>Publish Date: 2019-04-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-11358>CVE-2019-11358</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358</a></p> <p>Release Date: 2019-04-20</p> <p>Fix Resolution: jquery - 3.4.0</p> </p> <p></p> </details>
True
jquery-3.1.1.min.js: 3 vulnerabilities (highest severity is: 6.1) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-3.1.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js</a></p> <p> <p>Found in HEAD commit: <a href="https://github.com/amplify-education/tmp_SAST_eval_skf-labs/commit/91e1032bf85eea611eeb66711dc8b85719f6752e">91e1032bf85eea611eeb66711dc8b85719f6752e</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (jquery version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2020-11023](https://www.mend.io/vulnerability-database/CVE-2020-11023) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-3.1.1.min.js | Direct | jquery - 3.5.0;jquery-rails - 4.4.0 | &#10060; | | [CVE-2020-11022](https://www.mend.io/vulnerability-database/CVE-2020-11022) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-3.1.1.min.js | Direct | jQuery - 3.5.0 | &#10060; | | [CVE-2019-11358](https://www.mend.io/vulnerability-database/CVE-2019-11358) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-3.1.1.min.js | Direct | jquery - 3.4.0 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11023</summary> ### Vulnerable Library - <b>jquery-3.1.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js</a></p> <p> Dependency Hierarchy: - :x: **jquery-3.1.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/amplify-education/tmp_SAST_eval_skf-labs/commit/91e1032bf85eea611eeb66711dc8b85719f6752e">91e1032bf85eea611eeb66711dc8b85719f6752e</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-11023>CVE-2020-11023</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440">https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jquery - 3.5.0;jquery-rails - 4.4.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11022</summary> ### Vulnerable Library - <b>jquery-3.1.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js</a></p> <p> Dependency Hierarchy: - :x: **jquery-3.1.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/amplify-education/tmp_SAST_eval_skf-labs/commit/91e1032bf85eea611eeb66711dc8b85719f6752e">91e1032bf85eea611eeb66711dc8b85719f6752e</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-11022>CVE-2020-11022</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2019-11358</summary> ### Vulnerable Library - <b>jquery-3.1.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js</a></p> <p> Dependency Hierarchy: - :x: **jquery-3.1.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/amplify-education/tmp_SAST_eval_skf-labs/commit/91e1032bf85eea611eeb66711dc8b85719f6752e">91e1032bf85eea611eeb66711dc8b85719f6752e</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype. <p>Publish Date: 2019-04-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-11358>CVE-2019-11358</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358</a></p> <p>Release Date: 2019-04-20</p> <p>Fix Resolution: jquery - 3.4.0</p> </p> <p></p> </details>
non_process
jquery min js vulnerabilities highest severity is vulnerable library jquery min js javascript library for dom operations library home page a href found in head commit a href vulnerabilities cve severity cvss dependency type fixed in jquery version remediation available medium jquery min js direct jquery jquery rails medium jquery min js direct jquery medium jquery min js direct jquery details cve vulnerable library jquery min js javascript library for dom operations library home page a href dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details in jquery versions greater than or equal to and before passing html containing elements from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery jquery rails cve vulnerable library jquery min js javascript library for dom operations library home page a href dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery cve vulnerable library jquery min js javascript library for dom operations library home page a href dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details jquery before as used in drupal backdrop cms and other products mishandles jquery extend true because of object prototype pollution if an unsanitized source object contained an enumerable proto property it could extend the native object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery
0
5,725
8,567,919,385
IssuesEvent
2018-11-10 16:35:06
Great-Hill-Corporation/quickBlocks
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
closed
Should be able to return code at a specific block
help-wanted status-inprocess tools-isContract type-enhancement
This command should work: getCode 0x9bd20e68b22f63c651a30f2cdbf6d539e465f83c 2608034 by adding block number to getCode command But--notice--the code created itself and killed itself on the same transaction (and sent money to the etherTip jar.
1.0
Should be able to return code at a specific block - This command should work: getCode 0x9bd20e68b22f63c651a30f2cdbf6d539e465f83c 2608034 by adding block number to getCode command But--notice--the code created itself and killed itself on the same transaction (and sent money to the etherTip jar.
process
should be able to return code at a specific block this command should work getcode by adding block number to getcode command but notice the code created itself and killed itself on the same transaction and sent money to the ethertip jar
1
11,030
13,836,562,719
IssuesEvent
2020-10-14 01:24:01
opendistro-for-elasticsearch/opendistro-build
https://api.github.com/repos/opendistro-for-elasticsearch/opendistro-build
opened
ODFE 1.11.0 based on ES 7.9.1
in process new release
ODFE 1.11.1 based on ES 7.9.1 **Note: Yes, we are still on 7.9.1** Release Engineering / Build Repo Key Changes: * Add startupProbe in Helm charts, thanks @madeleine666 ([#415](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/415)) * Add Kibana pods annotations in Helm charts, thanks @rexbut ([#404](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/404)) * Add Kibana extra volumes in Helm charts, thanks @rexbut ([#419](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/419)) * Add Kibana extra containers field in Helm charts, thanks @rexbut ([#420](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/420)) ------
1.0
ODFE 1.11.0 based on ES 7.9.1 - ODFE 1.11.1 based on ES 7.9.1 **Note: Yes, we are still on 7.9.1** Release Engineering / Build Repo Key Changes: * Add startupProbe in Helm charts, thanks @madeleine666 ([#415](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/415)) * Add Kibana pods annotations in Helm charts, thanks @rexbut ([#404](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/404)) * Add Kibana extra volumes in Helm charts, thanks @rexbut ([#419](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/419)) * Add Kibana extra containers field in Helm charts, thanks @rexbut ([#420](https://github.com/opendistro-for-elasticsearch/opendistro-build/pull/420)) ------
process
odfe based on es odfe based on es note yes we are still on release engineering build repo key changes add startupprobe in helm charts thanks add kibana pods annotations in helm charts thanks rexbut add kibana extra volumes in helm charts thanks rexbut add kibana extra containers field in helm charts thanks rexbut
1
4,916
7,794,177,324
IssuesEvent
2018-06-08 00:56:37
StrikeNP/trac_test
https://api.github.com/repos/StrikeNP/trac_test
closed
GABLS2 rtm, rtp2, thlm, and thlp2 are set to zero when plotgen is run manually (but not for the nightly tests) (Trac #24)
Migrated from Trac enhancement post_processing senkbeil@uwm.edu
Some time ago, in order to test CLUBB's scalars, Brandon changed plotgen so that it outputs scalars in place of rtm and thlm. The nightly plots work great. However, if CLUBB is run manually without outputting scalars, and then plotgen is executed manually, then rtm, thlm, rtp2, and thlp2 are set to zero. For manual runs, typically we don't want to check scalars; we just want to plot standard versions of rtm, thlm, rtp2, and thlp2. I probably forgot to mention this earlier. Is it feasible to insert some nightly flags or re-arrange some code so that the nightly plots test the scalars, but the manual plots simply plot rtm, thlm, rtp2, and thlp2? I believe that this is what is done for other specialized nightly tests, e.g. the restart test and some of the altered grid tests. Perhaps those pieces of code would provide ideas on how to implement separate behavior for nightly and manual runs. However, we have a deadline on the TWP-ICE case, so don't bother with this until TWP-ICE is submitted, unless it is trivial to fix. Attachments: http://carson.math.uwm.edu/trac/clubb/attachment/ticket/24/Plotgen.pdf Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/24 ```json { "status": "closed", "changetime": "2009-09-02T20:37:37", "description": "Some time ago, in order to test CLUBB's scalars, Brandon changed plotgen so that it outputs scalars in place of rtm and thlm. The nightly plots work great.\n\nHowever, if CLUBB is run manually without outputting scalars, and then plotgen is executed manually, then rtm, thlm, rtp2, and thlp2 are set to zero. For manual runs, typically we don't want to check scalars; we just want to plot standard versions of rtm, thlm, rtp2, and thlp2. I probably forgot to mention this earlier.\n\nIs it feasible to insert some nightly flags or re-arrange some code so that the nightly plots test the scalars, but the manual plots simply plot rtm, thlm, rtp2, and thlp2? I believe that this is what is done for other specialized nightly tests, e.g. the restart test and some of the altered grid tests. Perhaps those pieces of code would provide ideas on how to implement separate behavior for nightly and manual runs.\n\nHowever, we have a deadline on the TWP-ICE case, so don't bother with this until TWP-ICE is submitted, unless it is trivial to fix.", "reporter": "vlarson@uwm.edu", "cc": "", "resolution": "Verified by V. Larson", "_ts": "1251923857000000", "component": "post_processing", "summary": "GABLS2 rtm, rtp2, thlm, and thlp2 are set to zero when plotgen is run manually (but not for the nightly tests)", "priority": "minor", "keywords": "scalars, gabls2, nightly plots, rtm, thlm, rtp2, thlp2", "time": "2009-05-13T14:26:46", "milestone": "Plotgen 3.0", "owner": "senkbeil@uwm.edu", "type": "enhancement" } ```
1.0
GABLS2 rtm, rtp2, thlm, and thlp2 are set to zero when plotgen is run manually (but not for the nightly tests) (Trac #24) - Some time ago, in order to test CLUBB's scalars, Brandon changed plotgen so that it outputs scalars in place of rtm and thlm. The nightly plots work great. However, if CLUBB is run manually without outputting scalars, and then plotgen is executed manually, then rtm, thlm, rtp2, and thlp2 are set to zero. For manual runs, typically we don't want to check scalars; we just want to plot standard versions of rtm, thlm, rtp2, and thlp2. I probably forgot to mention this earlier. Is it feasible to insert some nightly flags or re-arrange some code so that the nightly plots test the scalars, but the manual plots simply plot rtm, thlm, rtp2, and thlp2? I believe that this is what is done for other specialized nightly tests, e.g. the restart test and some of the altered grid tests. Perhaps those pieces of code would provide ideas on how to implement separate behavior for nightly and manual runs. However, we have a deadline on the TWP-ICE case, so don't bother with this until TWP-ICE is submitted, unless it is trivial to fix. Attachments: http://carson.math.uwm.edu/trac/clubb/attachment/ticket/24/Plotgen.pdf Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/24 ```json { "status": "closed", "changetime": "2009-09-02T20:37:37", "description": "Some time ago, in order to test CLUBB's scalars, Brandon changed plotgen so that it outputs scalars in place of rtm and thlm. The nightly plots work great.\n\nHowever, if CLUBB is run manually without outputting scalars, and then plotgen is executed manually, then rtm, thlm, rtp2, and thlp2 are set to zero. For manual runs, typically we don't want to check scalars; we just want to plot standard versions of rtm, thlm, rtp2, and thlp2. I probably forgot to mention this earlier.\n\nIs it feasible to insert some nightly flags or re-arrange some code so that the nightly plots test the scalars, but the manual plots simply plot rtm, thlm, rtp2, and thlp2? I believe that this is what is done for other specialized nightly tests, e.g. the restart test and some of the altered grid tests. Perhaps those pieces of code would provide ideas on how to implement separate behavior for nightly and manual runs.\n\nHowever, we have a deadline on the TWP-ICE case, so don't bother with this until TWP-ICE is submitted, unless it is trivial to fix.", "reporter": "vlarson@uwm.edu", "cc": "", "resolution": "Verified by V. Larson", "_ts": "1251923857000000", "component": "post_processing", "summary": "GABLS2 rtm, rtp2, thlm, and thlp2 are set to zero when plotgen is run manually (but not for the nightly tests)", "priority": "minor", "keywords": "scalars, gabls2, nightly plots, rtm, thlm, rtp2, thlp2", "time": "2009-05-13T14:26:46", "milestone": "Plotgen 3.0", "owner": "senkbeil@uwm.edu", "type": "enhancement" } ```
process
rtm thlm and are set to zero when plotgen is run manually but not for the nightly tests trac some time ago in order to test clubb s scalars brandon changed plotgen so that it outputs scalars in place of rtm and thlm the nightly plots work great however if clubb is run manually without outputting scalars and then plotgen is executed manually then rtm thlm and are set to zero for manual runs typically we don t want to check scalars we just want to plot standard versions of rtm thlm and i probably forgot to mention this earlier is it feasible to insert some nightly flags or re arrange some code so that the nightly plots test the scalars but the manual plots simply plot rtm thlm and i believe that this is what is done for other specialized nightly tests e g the restart test and some of the altered grid tests perhaps those pieces of code would provide ideas on how to implement separate behavior for nightly and manual runs however we have a deadline on the twp ice case so don t bother with this until twp ice is submitted unless it is trivial to fix attachments migrated from json status closed changetime description some time ago in order to test clubb s scalars brandon changed plotgen so that it outputs scalars in place of rtm and thlm the nightly plots work great n nhowever if clubb is run manually without outputting scalars and then plotgen is executed manually then rtm thlm and are set to zero for manual runs typically we don t want to check scalars we just want to plot standard versions of rtm thlm and i probably forgot to mention this earlier n nis it feasible to insert some nightly flags or re arrange some code so that the nightly plots test the scalars but the manual plots simply plot rtm thlm and i believe that this is what is done for other specialized nightly tests e g the restart test and some of the altered grid tests perhaps those pieces of code would provide ideas on how to implement separate behavior for nightly and manual runs n nhowever we have a deadline on the twp ice case so don t bother with this until twp ice is submitted unless it is trivial to fix reporter vlarson uwm edu cc resolution verified by v larson ts component post processing summary rtm thlm and are set to zero when plotgen is run manually but not for the nightly tests priority minor keywords scalars nightly plots rtm thlm time milestone plotgen owner senkbeil uwm edu type enhancement
1
450,889
13,021,147,715
IssuesEvent
2020-07-27 05:32:48
ahmedkaludi/accelerated-mobile-pages
https://api.github.com/repos/ahmedkaludi/accelerated-mobile-pages
closed
Autoptimize is causing validation for logged in users, confusing alot of people
NEXT UPDATE [Priority: HIGH] bug
The attribute 'percentage' may not appear in tag 'div'. Debug. 1 71498 The attribute 'target' may not appear in tag 'div'. Debug. 1 74443 The attribute 'target' may not appear in tag 'div'. Debug.
1.0
Autoptimize is causing validation for logged in users, confusing alot of people - The attribute 'percentage' may not appear in tag 'div'. Debug. 1 71498 The attribute 'target' may not appear in tag 'div'. Debug. 1 74443 The attribute 'target' may not appear in tag 'div'. Debug.
non_process
autoptimize is causing validation for logged in users confusing alot of people the attribute percentage may not appear in tag div debug the attribute target may not appear in tag div debug the attribute target may not appear in tag div debug
0
322,371
23,904,567,386
IssuesEvent
2022-09-08 22:31:05
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
closed
Typo in the REST API doc for BlocksResponse
documentation
### Description The API doc for BlocksResponse says: <img width="347" alt="Screenshot 2022-09-02 at 23 46 39" src="https://user-images.githubusercontent.com/16097111/188240038-839abc5a-340e-432c-9a89-2fda9cf4d0ec.png"> but the name of the first field of the data structure is actually `blocks`. ### Steps to reproduce http://previewnet.mirrornode.hedera.com/api/v1/docs/#/blocks/listBlocks ### Additional context _No response_ ### Hedera network previewnet ### Version 0.63.0 ### Operating system _No response_
1.0
Typo in the REST API doc for BlocksResponse - ### Description The API doc for BlocksResponse says: <img width="347" alt="Screenshot 2022-09-02 at 23 46 39" src="https://user-images.githubusercontent.com/16097111/188240038-839abc5a-340e-432c-9a89-2fda9cf4d0ec.png"> but the name of the first field of the data structure is actually `blocks`. ### Steps to reproduce http://previewnet.mirrornode.hedera.com/api/v1/docs/#/blocks/listBlocks ### Additional context _No response_ ### Hedera network previewnet ### Version 0.63.0 ### Operating system _No response_
non_process
typo in the rest api doc for blocksresponse description the api doc for blocksresponse says img width alt screenshot at src but the name of the first field of the data structure is actually blocks steps to reproduce additional context no response hedera network previewnet version operating system no response
0
50,757
21,389,124,133
IssuesEvent
2022-04-21 04:22:45
IBM-Cloud/terraform-provider-ibm
https://api.github.com/repos/IBM-Cloud/terraform-provider-ibm
closed
Limitation on the number of characters in the resource ibm_resource_tag
service/Cloud Databases service/Resource Management service/Global Tagging
### Terraform CLI and Terraform IBM Provider Version ```bash Terraform v1.0.3 on darwin_amd64 + provider registry.terraform.io/ibm-cloud/ibm v1.40.0 ``` ### Error The ibm_resource_tag resource has a limitation in the resource_id parameter of 128 characters, the problem is that there are some ibm cloud resources that their crn is longer than 128 characters. ```hcl resource "ibm_resource_tag" "tag" { resource_id = ibm_satellite_location.location.crn tags = var.tag_names } ``` To reproduce this error you can run the following template: ```hcl resource "ibm_database" "elasticsearch" { resource_group_id = data.ibm_resource_group.resource_group.id name = "test" service = "databases-for-elasticsearch" plan = "standard" location = "eu-de" members_disk_allocation_mb = 61440 members_memory_allocation_mb = 43008 tags = ["one:two"] whitelist { address = "172.168.1.2/32" description = "desc1" } timeouts { create = "120m" update = "120m" delete = "15m" } } data "ibm_resource_tag" "current" { resource_id = ibm_database.elasticsearch.id } resource "ibm_resource_tag" "update" { resource_id = ibm_database.elasticsearch.id tags = ["two:three"] } ``` The output error: ```bash terraform apply ibm_database.elasticsearch: Refreshing state... [id=crn:v1:bluemix:public:databases-for-elasticsearch:eu-de:*/******************:*******************:::] Error: "resource_id" ("crn:v1:bluemix:public:databases-for-elasticsearch:eu-de:*/******************:*******************::") must contain from 1 to 128 characters with data.ibm_resource_tag.current, on main.tf line 59, in data "ibm_resource_tag" "current": 59: resource_id = ibm_database.elasticsearch.id ``` ### Solution The solution would be to increase the resource_id characters to accept more than 128 characters.
3.0
Limitation on the number of characters in the resource ibm_resource_tag - ### Terraform CLI and Terraform IBM Provider Version ```bash Terraform v1.0.3 on darwin_amd64 + provider registry.terraform.io/ibm-cloud/ibm v1.40.0 ``` ### Error The ibm_resource_tag resource has a limitation in the resource_id parameter of 128 characters, the problem is that there are some ibm cloud resources that their crn is longer than 128 characters. ```hcl resource "ibm_resource_tag" "tag" { resource_id = ibm_satellite_location.location.crn tags = var.tag_names } ``` To reproduce this error you can run the following template: ```hcl resource "ibm_database" "elasticsearch" { resource_group_id = data.ibm_resource_group.resource_group.id name = "test" service = "databases-for-elasticsearch" plan = "standard" location = "eu-de" members_disk_allocation_mb = 61440 members_memory_allocation_mb = 43008 tags = ["one:two"] whitelist { address = "172.168.1.2/32" description = "desc1" } timeouts { create = "120m" update = "120m" delete = "15m" } } data "ibm_resource_tag" "current" { resource_id = ibm_database.elasticsearch.id } resource "ibm_resource_tag" "update" { resource_id = ibm_database.elasticsearch.id tags = ["two:three"] } ``` The output error: ```bash terraform apply ibm_database.elasticsearch: Refreshing state... [id=crn:v1:bluemix:public:databases-for-elasticsearch:eu-de:*/******************:*******************:::] Error: "resource_id" ("crn:v1:bluemix:public:databases-for-elasticsearch:eu-de:*/******************:*******************::") must contain from 1 to 128 characters with data.ibm_resource_tag.current, on main.tf line 59, in data "ibm_resource_tag" "current": 59: resource_id = ibm_database.elasticsearch.id ``` ### Solution The solution would be to increase the resource_id characters to accept more than 128 characters.
non_process
limitation on the number of characters in the resource ibm resource tag terraform cli and terraform ibm provider version bash terraform on darwin provider registry terraform io ibm cloud ibm error the ibm resource tag resource has a limitation in the resource id parameter of characters the problem is that there are some ibm cloud resources that their crn is longer than characters hcl resource ibm resource tag tag resource id ibm satellite location location crn tags var tag names to reproduce this error you can run the following template hcl resource ibm database elasticsearch resource group id data ibm resource group resource group id name test service databases for elasticsearch plan standard location eu de members disk allocation mb members memory allocation mb tags whitelist address description timeouts create update delete data ibm resource tag current resource id ibm database elasticsearch id resource ibm resource tag update resource id ibm database elasticsearch id tags the output error bash terraform apply ibm database elasticsearch refreshing state error resource id crn bluemix public databases for elasticsearch eu de must contain from to characters with data ibm resource tag current on main tf line in data ibm resource tag current resource id ibm database elasticsearch id solution the solution would be to increase the resource id characters to accept more than characters
0