Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
16,843
22,094,951,341
IssuesEvent
2022-06-01 09:11:08
camunda/zeebe
https://api.github.com/repos/camunda/zeebe
closed
Extend ProcessInstanceCreation record with start instructions
team/process-automation
The `ProcessInstanceCreation` records are sent by the gateway to the broker. The broker processes it to start the process instance. We'll need to extend it with start instructions to allow starting a process instance from an arbitrary point in the process. In the future, we may add support for local variables and instruction types (e.g. `startBeforeElement`, `startAfterElement`, etc), so the design should take this into account. Make sure to document how we can extend records. ## Out of scope - `localVariables` (root scope variables are already available) - `type` (only `startBeforeElement` will be implemented in the first iteration) - the created element instance keys in the `CREATED` event (not necessary in first iteration)
1.0
Extend ProcessInstanceCreation record with start instructions - The `ProcessInstanceCreation` records are sent by the gateway to the broker. The broker processes it to start the process instance. We'll need to extend it with start instructions to allow starting a process instance from an arbitrary point in the process. In the future, we may add support for local variables and instruction types (e.g. `startBeforeElement`, `startAfterElement`, etc), so the design should take this into account. Make sure to document how we can extend records. ## Out of scope - `localVariables` (root scope variables are already available) - `type` (only `startBeforeElement` will be implemented in the first iteration) - the created element instance keys in the `CREATED` event (not necessary in first iteration)
process
extend processinstancecreation record with start instructions the processinstancecreation records are sent by the gateway to the broker the broker processes it to start the process instance we ll need to extend it with start instructions to allow starting a process instance from an arbitrary point in the process in the future we may add support for local variables and instruction types e g startbeforeelement startafterelement etc so the design should take this into account make sure to document how we can extend records out of scope localvariables root scope variables are already available type only startbeforeelement will be implemented in the first iteration the created element instance keys in the created event not necessary in first iteration
1
30,352
2,723,576,528
IssuesEvent
2015-04-14 13:28:55
CruxFramework/crux-widgets
https://api.github.com/repos/CruxFramework/crux-widgets
closed
Broken Layout on DynaTabs when open a lot of tabs
bug CruxWidgetsLibrary imported Milestone-2.2.1 Priority-Medium
_From [tr_busta...@yahoo.com.br](https://code.google.com/u/115454294030253308352/) on May 04, 2010 13:39:28_ What steps will reproduce the problem? Create a DynaTabs widget with a big number of tabs. If the dynaTbas with is not enough to hold all of them, the layout is broken _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=94_
1.0
Broken Layout on DynaTabs when open a lot of tabs - _From [tr_busta...@yahoo.com.br](https://code.google.com/u/115454294030253308352/) on May 04, 2010 13:39:28_ What steps will reproduce the problem? Create a DynaTabs widget with a big number of tabs. If the dynaTbas with is not enough to hold all of them, the layout is broken _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=94_
non_process
broken layout on dynatabs when open a lot of tabs from on may what steps will reproduce the problem create a dynatabs widget with a big number of tabs if the dynatbas with is not enough to hold all of them the layout is broken original issue
0
9,502
12,489,704,064
IssuesEvent
2020-05-31 20:07:06
Jeffail/benthos
https://api.github.com/repos/Jeffail/benthos
closed
Looking for a way to decode HTML entities in XML file.
enhancement processors
We are using the XML processor https://www.benthos.dev/docs/components/processors/xml Though the conversion fails and returns the same XML back when the XML content contains HTML entities such as `&`, `<`, `–`, `ˆ` etc... So, maybe the XML processor itself could also handle UnescapeString?
1.0
Looking for a way to decode HTML entities in XML file. - We are using the XML processor https://www.benthos.dev/docs/components/processors/xml Though the conversion fails and returns the same XML back when the XML content contains HTML entities such as `&`, `<`, `–`, `ˆ` etc... So, maybe the XML processor itself could also handle UnescapeString?
process
looking for a way to decode html entities in xml file we are using the xml processor though the conversion fails and returns the same xml back when the xml content contains html entities such as amp lt ndash circ etc so maybe the xml processor itself could also handle unescapestring
1
11,535
14,408,577,966
IssuesEvent
2020-12-04 00:07:33
KevCor99/5a
https://api.github.com/repos/KevCor99/5a
opened
complete_size_estimating_template
process-dashboard
-completar el formato de estimacion de LOC con los valores reales obtenidos
1.0
complete_size_estimating_template - -completar el formato de estimacion de LOC con los valores reales obtenidos
process
complete size estimating template completar el formato de estimacion de loc con los valores reales obtenidos
1
19,455
25,737,122,759
IssuesEvent
2022-12-08 01:57:33
brucemiller/LaTeXML
https://api.github.com/repos/brucemiller/LaTeXML
closed
Refactor away from relative mpadded widths
enhancement postprocessing schema math parsing
In the upcoming MathML Core recommendation, mpadded [width](https://www.w3.org/TR/2021/WD-mathml-core-20210816/#attribute-mpadded-width) is restricted to "a valid [length-percentage](https://www.w3.org/TR/2021/WD-mathml-core-20210816/#dfn-length-percentage)". This excludes the relative values such as `width="+5pt"`, `width="-2pt"`. LaTeXML relies on these relative values in MathML.pm and Linebreaker.pm, and we should try to find a different formulation that achieves the same spacing effect. Since the code seems to be dealing with left/right padding, maybe there is a way to translate that into a CSS-near "style" attribute with `padding-left` and `padding-right` values? For some standard cases we could also consider adding specialized spacing classes, which will be a little more configurable on the publisher end.
1.0
Refactor away from relative mpadded widths - In the upcoming MathML Core recommendation, mpadded [width](https://www.w3.org/TR/2021/WD-mathml-core-20210816/#attribute-mpadded-width) is restricted to "a valid [length-percentage](https://www.w3.org/TR/2021/WD-mathml-core-20210816/#dfn-length-percentage)". This excludes the relative values such as `width="+5pt"`, `width="-2pt"`. LaTeXML relies on these relative values in MathML.pm and Linebreaker.pm, and we should try to find a different formulation that achieves the same spacing effect. Since the code seems to be dealing with left/right padding, maybe there is a way to translate that into a CSS-near "style" attribute with `padding-left` and `padding-right` values? For some standard cases we could also consider adding specialized spacing classes, which will be a little more configurable on the publisher end.
process
refactor away from relative mpadded widths in the upcoming mathml core recommendation mpadded is restricted to a valid this excludes the relative values such as width width latexml relies on these relative values in mathml pm and linebreaker pm and we should try to find a different formulation that achieves the same spacing effect since the code seems to be dealing with left right padding maybe there is a way to translate that into a css near style attribute with padding left and padding right values for some standard cases we could also consider adding specialized spacing classes which will be a little more configurable on the publisher end
1
397,258
27,157,169,765
IssuesEvent
2023-02-17 08:54:18
saveourtool/sarif-utils
https://api.github.com/repos/saveourtool/sarif-utils
opened
Make README more detailed and pretty
documentation help wanted
Need to add more detailed description, icons, several section like `supported platforms`, `dependency`, and so on
1.0
Make README more detailed and pretty - Need to add more detailed description, icons, several section like `supported platforms`, `dependency`, and so on
non_process
make readme more detailed and pretty need to add more detailed description icons several section like supported platforms dependency and so on
0
66,030
8,873,366,465
IssuesEvent
2019-01-11 17:54:39
Anonymousdog/displaycameras
https://api.github.com/repos/Anonymousdog/displaycameras
closed
Layout Examples oddity
documentation
I haven't looked at every example layout, but both: layout.conf.1920x1080.3x3 and layout.conf.1920x1200.6on4x4 are commented as being: "# 2x2 screen with 3 off-screen windows." Could be just a labeling error, but just wanted to let you know.
1.0
Layout Examples oddity - I haven't looked at every example layout, but both: layout.conf.1920x1080.3x3 and layout.conf.1920x1200.6on4x4 are commented as being: "# 2x2 screen with 3 off-screen windows." Could be just a labeling error, but just wanted to let you know.
non_process
layout examples oddity i haven t looked at every example layout but both layout conf and layout conf are commented as being screen with off screen windows could be just a labeling error but just wanted to let you know
0
10,247
3,091,556,696
IssuesEvent
2015-08-26 13:48:25
TEAMMATES/repo
https://api.github.com/repos/TEAMMATES/repo
closed
Improve handling of wildcards in verifyHtml
a-Testing p.Medium t.Enhancement
One enhancement that can be made to improve the handling of our "{\*}" values, is to copy the actual value of the wildcard on the actual html into the expected html. Currently, the lines with {\*} always appear different in the diff, making it hard to isolate the real differences on the tests with many wildcards, such as the results page. This might cause the comparison to take a longer time though, so maybe its not worth it. Edit: the other direction of doing this could be easier: converting the hashes into {*}
1.0
Improve handling of wildcards in verifyHtml - One enhancement that can be made to improve the handling of our "{\*}" values, is to copy the actual value of the wildcard on the actual html into the expected html. Currently, the lines with {\*} always appear different in the diff, making it hard to isolate the real differences on the tests with many wildcards, such as the results page. This might cause the comparison to take a longer time though, so maybe its not worth it. Edit: the other direction of doing this could be easier: converting the hashes into {*}
non_process
improve handling of wildcards in verifyhtml one enhancement that can be made to improve the handling of our values is to copy the actual value of the wildcard on the actual html into the expected html currently the lines with always appear different in the diff making it hard to isolate the real differences on the tests with many wildcards such as the results page this might cause the comparison to take a longer time though so maybe its not worth it edit the other direction of doing this could be easier converting the hashes into
0
121,535
12,128,533,217
IssuesEvent
2020-04-22 20:40:55
openoakland/woeip
https://api.github.com/repos/openoakland/woeip
closed
Update pull request and testing instructions
component: documentation status: ready to work on
**Describe the bug** - The pull request template instructions to execute tests are outdated. - They rely on `make local.shell` to enter the docker environment - Should read: 1. `docker-compose exec api /bin/bash` 2. `make test` - Should include instructions for jest tests - [Testing Section](https://github.com/openoakland/woeip/blob/master/.github/contributing.md#testing) of contributing doc should also be updated to include jest tests. - Also, in my environment, I need to separate entering the container with running the make command. Otherwise, my terminal responds `/bin/bash: make validate: No such file or directory` **To Reproduce** Steps to reproduce the behavior: 1. Go to woeip project directory in terminal 2. Enter `docker-compose exec api /bin/bash "make validate"` 3. See `/bin/bash: make validate: No such file or directory` **Expected behavior** The pull request and testing documentation should reflect current instructions to run both the api and web testing suites. **Desktop (please complete the following information):** - Ubuntu 18.04 **Affected Components (For Developers)** Pull request template and Testing Section of Contributing document **Technical Resources (For Developers)** In addition to the api testing commands above, the web testing command is `npm test`. It will also need to command to enter the node container. `docker exec -it woaq-web /bin/sh`
1.0
Update pull request and testing instructions - **Describe the bug** - The pull request template instructions to execute tests are outdated. - They rely on `make local.shell` to enter the docker environment - Should read: 1. `docker-compose exec api /bin/bash` 2. `make test` - Should include instructions for jest tests - [Testing Section](https://github.com/openoakland/woeip/blob/master/.github/contributing.md#testing) of contributing doc should also be updated to include jest tests. - Also, in my environment, I need to separate entering the container with running the make command. Otherwise, my terminal responds `/bin/bash: make validate: No such file or directory` **To Reproduce** Steps to reproduce the behavior: 1. Go to woeip project directory in terminal 2. Enter `docker-compose exec api /bin/bash "make validate"` 3. See `/bin/bash: make validate: No such file or directory` **Expected behavior** The pull request and testing documentation should reflect current instructions to run both the api and web testing suites. **Desktop (please complete the following information):** - Ubuntu 18.04 **Affected Components (For Developers)** Pull request template and Testing Section of Contributing document **Technical Resources (For Developers)** In addition to the api testing commands above, the web testing command is `npm test`. It will also need to command to enter the node container. `docker exec -it woaq-web /bin/sh`
non_process
update pull request and testing instructions describe the bug the pull request template instructions to execute tests are outdated they rely on make local shell to enter the docker environment should read docker compose exec api bin bash make test should include instructions for jest tests of contributing doc should also be updated to include jest tests also in my environment i need to separate entering the container with running the make command otherwise my terminal responds bin bash make validate no such file or directory to reproduce steps to reproduce the behavior go to woeip project directory in terminal enter docker compose exec api bin bash make validate see bin bash make validate no such file or directory expected behavior the pull request and testing documentation should reflect current instructions to run both the api and web testing suites desktop please complete the following information ubuntu affected components for developers pull request template and testing section of contributing document technical resources for developers in addition to the api testing commands above the web testing command is npm test it will also need to command to enter the node container docker exec it woaq web bin sh
0
100,718
8,752,751,416
IssuesEvent
2018-12-14 05:00:21
humera987/FXLabs-Test-Automation
https://api.github.com/repos/humera987/FXLabs-Test-Automation
closed
Testing 14 : ApiV1ProjectsIdAutoSuggestionsActivateSuitenameTcnumberGetPathParamTcnumberNullValue
Testing 14
Project : Testing 14 Job : UAT Env : UAT Region : US_WEST Result : fail Status Code : 500 Headers : {} Endpoint : http://13.56.210.25/api/v1/api/v1/projects/{id}/auto-suggestions/activate/{suiteName}/null Request : Response : Not enough variable values available to expand 'id' Logs : Assertion [@StatusCode != 401] resolved-to [500 != 401] result [Passed]Assertion [@StatusCode != 500] resolved-to [500 != 500] result [Failed]Assertion [@StatusCode != 404] resolved-to [500 != 404] result [Passed]Assertion [@StatusCode != 200] resolved-to [500 != 200] result [Passed] --- FX Bot ---
1.0
Testing 14 : ApiV1ProjectsIdAutoSuggestionsActivateSuitenameTcnumberGetPathParamTcnumberNullValue - Project : Testing 14 Job : UAT Env : UAT Region : US_WEST Result : fail Status Code : 500 Headers : {} Endpoint : http://13.56.210.25/api/v1/api/v1/projects/{id}/auto-suggestions/activate/{suiteName}/null Request : Response : Not enough variable values available to expand 'id' Logs : Assertion [@StatusCode != 401] resolved-to [500 != 401] result [Passed]Assertion [@StatusCode != 500] resolved-to [500 != 500] result [Failed]Assertion [@StatusCode != 404] resolved-to [500 != 404] result [Passed]Assertion [@StatusCode != 200] resolved-to [500 != 200] result [Passed] --- FX Bot ---
non_process
testing project testing job uat env uat region us west result fail status code headers endpoint request response not enough variable values available to expand id logs assertion resolved to result assertion resolved to result assertion resolved to result assertion resolved to result fx bot
0
7,620
10,727,678,988
IssuesEvent
2019-10-28 12:18:19
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
NTR immune receptor activity and innate immune receptor activity
New term request multi-species process
+[Term] +id: GO:0140375 +name: immune receptor activity +is_a: GO:0038023 ! signaling receptor activity +created_by: pg +creation_date: 2019-10-18T14:11:15Z +default-namespace: molecular_function +default-namespace: Receiving a signal and transmitting it in a cell to initiate an immune response. {xref="PMID:31415752", xref="Wikipedia:Immune_receptor"} +[Term] +id: GO:0140376 +name: innate immune receptor activity +def: "Receiving a signal and transmitting it in a cell to initiate an innate immune response." [PMID:28921463, PMID:31415752] +is_a: GO:0140375 ! immune receptor activity To replace BP's response/detection to molecule of x origin (see #17980) --- Children would be: *'complement receptor activity' *'cytokine receptor activity' *'immunoglobulin receptor activity' *'MHC class I receptor activity' *'MHC class Ib receptor activity' *'MHC class II receptor activity' and for 'innate immune receptor activity' * 'pattern recognition receptor activity' @addiehl Is this OK ? I wonder why this term was not created. Thanks, Pascale
1.0
NTR immune receptor activity and innate immune receptor activity - +[Term] +id: GO:0140375 +name: immune receptor activity +is_a: GO:0038023 ! signaling receptor activity +created_by: pg +creation_date: 2019-10-18T14:11:15Z +default-namespace: molecular_function +default-namespace: Receiving a signal and transmitting it in a cell to initiate an immune response. {xref="PMID:31415752", xref="Wikipedia:Immune_receptor"} +[Term] +id: GO:0140376 +name: innate immune receptor activity +def: "Receiving a signal and transmitting it in a cell to initiate an innate immune response." [PMID:28921463, PMID:31415752] +is_a: GO:0140375 ! immune receptor activity To replace BP's response/detection to molecule of x origin (see #17980) --- Children would be: *'complement receptor activity' *'cytokine receptor activity' *'immunoglobulin receptor activity' *'MHC class I receptor activity' *'MHC class Ib receptor activity' *'MHC class II receptor activity' and for 'innate immune receptor activity' * 'pattern recognition receptor activity' @addiehl Is this OK ? I wonder why this term was not created. Thanks, Pascale
process
ntr immune receptor activity and innate immune receptor activity id go name immune receptor activity is a go signaling receptor activity created by pg creation date default namespace molecular function default namespace receiving a signal and transmitting it in a cell to initiate an immune response xref pmid xref wikipedia immune receptor id go name innate immune receptor activity def receiving a signal and transmitting it in a cell to initiate an innate immune response is a go immune receptor activity to replace bp s response detection to molecule of x origin see children would be complement receptor activity cytokine receptor activity immunoglobulin receptor activity mhc class i receptor activity mhc class ib receptor activity mhc class ii receptor activity and for innate immune receptor activity pattern recognition receptor activity addiehl is this ok i wonder why this term was not created thanks pascale
1
132,609
10,760,532,210
IssuesEvent
2019-10-31 18:47:55
PulpQE/pulp-smash
https://api.github.com/repos/PulpQE/pulp-smash
closed
Test that sync is successful when there is ascii characters in the Url
Issue Type: Test Case pulp 2 - closed - wontfix
https://pulp.plan.io/issues/2809 1. Create a repo with ascii characters in the url 2. Sync the repo
1.0
Test that sync is successful when there is ascii characters in the Url - https://pulp.plan.io/issues/2809 1. Create a repo with ascii characters in the url 2. Sync the repo
non_process
test that sync is successful when there is ascii characters in the url create a repo with ascii characters in the url sync the repo
0
20,299
15,227,292,464
IssuesEvent
2021-02-18 10:00:40
argoproj/argo-cd
https://api.github.com/repos/argoproj/argo-cd
closed
Progressing Health status for Bitnami SealedSecret
component:core enhancement good first issue type:usability
# Summary In Bitnami SealedSecrets v0.12, SealedSecrets now has a status field which we can key off to decide if a SealedSecret is fully deployed https://github.com/bitnami-labs/sealed-secrets/blob/master/RELEASE-NOTES.md#status-field > Now the sealed secrets controller updates the Status field of the SealedSecrets resources. This makes it easier for automation like ArgoCD to detect whether (and when) the controller has reacted to changes in the SealedSecret resources and produced a Secret. It also shows an error message in case it fails (many users are not familiar with k8s events and they may find it easier to see the error message in the status). We should implement a Health check to place a SealedSecret in a Progressing state, if the controller has not yet "reacted to changes in the SealedSecret resources and produced a Secret".
True
Progressing Health status for Bitnami SealedSecret - # Summary In Bitnami SealedSecrets v0.12, SealedSecrets now has a status field which we can key off to decide if a SealedSecret is fully deployed https://github.com/bitnami-labs/sealed-secrets/blob/master/RELEASE-NOTES.md#status-field > Now the sealed secrets controller updates the Status field of the SealedSecrets resources. This makes it easier for automation like ArgoCD to detect whether (and when) the controller has reacted to changes in the SealedSecret resources and produced a Secret. It also shows an error message in case it fails (many users are not familiar with k8s events and they may find it easier to see the error message in the status). We should implement a Health check to place a SealedSecret in a Progressing state, if the controller has not yet "reacted to changes in the SealedSecret resources and produced a Secret".
non_process
progressing health status for bitnami sealedsecret summary in bitnami sealedsecrets sealedsecrets now has a status field which we can key off to decide if a sealedsecret is fully deployed now the sealed secrets controller updates the status field of the sealedsecrets resources this makes it easier for automation like argocd to detect whether and when the controller has reacted to changes in the sealedsecret resources and produced a secret it also shows an error message in case it fails many users are not familiar with events and they may find it easier to see the error message in the status we should implement a health check to place a sealedsecret in a progressing state if the controller has not yet reacted to changes in the sealedsecret resources and produced a secret
0
711,612
24,469,643,534
IssuesEvent
2022-10-07 18:26:37
zulip/zulip
https://api.github.com/repos/zulip/zulip
closed
Draft count briefly appears when sending a new message.
bug help wanted area: left-sidebar area: compose in progress priority: high release goal
Hi :) When sending a new message (i.e. pressing `Enter` after having composed a new message), the draft count briefly appears (or increments when it already had been present before). The issue can be reproduced on `dev`. Discussion: https://chat.zulip.org/#narrow/stream/9-issues/topic/draft.20count.20appears.20briefly.20when.20sending.20a.20message Thanks a lot for looking into this! :heart:
1.0
Draft count briefly appears when sending a new message. - Hi :) When sending a new message (i.e. pressing `Enter` after having composed a new message), the draft count briefly appears (or increments when it already had been present before). The issue can be reproduced on `dev`. Discussion: https://chat.zulip.org/#narrow/stream/9-issues/topic/draft.20count.20appears.20briefly.20when.20sending.20a.20message Thanks a lot for looking into this! :heart:
non_process
draft count briefly appears when sending a new message hi when sending a new message i e pressing enter after having composed a new message the draft count briefly appears or increments when it already had been present before the issue can be reproduced on dev discussion thanks a lot for looking into this heart
0
17,104
22,624,484,673
IssuesEvent
2022-06-30 09:26:07
camunda/zeebe
https://api.github.com/repos/camunda/zeebe
closed
Incorrect creation_mode used for process_instance_creations_total metric
kind/bug area/observability team/process-automation
**Describe the bug** <!-- A clear and concise description of what the bug is. --> The `process_instance_creations_total` metric keeps track of the created process instances. It is registered with a `creation_mode` label to distinguish `CREATION_AT_DEFAULT_START_EVENT ` from `CREATION_AT_GIVEN_ELEMENT`. However, when process instances are created at the default start event the metric with `CREATION_AT_GIVEN_ELEMENT` is incremented, and vice-versa for creations at a given element. **To Reproduce** <!-- Steps to reproduce the behavior If possible add a minimal reproducer code sample - when using the Java client: https://github.com/zeebe-io/zeebe-test-template-java --> - run the latest Zeebe version with Prometheus. - create a process instance without start instructions - check the results for `sum(rate(zeebe_process_instance_creations_total[1m])) by (creation_mode)` **Expected behavior** <!-- A clear and concise description of what you expected to happen. --> The correct metrics should be incremented. **Environment:** - OS: ... - Zeebe Version: medic-cw-25 and onwards - Configuration: ...
1.0
Incorrect creation_mode used for process_instance_creations_total metric - **Describe the bug** <!-- A clear and concise description of what the bug is. --> The `process_instance_creations_total` metric keeps track of the created process instances. It is registered with a `creation_mode` label to distinguish `CREATION_AT_DEFAULT_START_EVENT ` from `CREATION_AT_GIVEN_ELEMENT`. However, when process instances are created at the default start event the metric with `CREATION_AT_GIVEN_ELEMENT` is incremented, and vice-versa for creations at a given element. **To Reproduce** <!-- Steps to reproduce the behavior If possible add a minimal reproducer code sample - when using the Java client: https://github.com/zeebe-io/zeebe-test-template-java --> - run the latest Zeebe version with Prometheus. - create a process instance without start instructions - check the results for `sum(rate(zeebe_process_instance_creations_total[1m])) by (creation_mode)` **Expected behavior** <!-- A clear and concise description of what you expected to happen. --> The correct metrics should be incremented. **Environment:** - OS: ... - Zeebe Version: medic-cw-25 and onwards - Configuration: ...
process
incorrect creation mode used for process instance creations total metric describe the bug the process instance creations total metric keeps track of the created process instances it is registered with a creation mode label to distinguish creation at default start event from creation at given element however when process instances are created at the default start event the metric with creation at given element is incremented and vice versa for creations at a given element to reproduce steps to reproduce the behavior if possible add a minimal reproducer code sample when using the java client run the latest zeebe version with prometheus create a process instance without start instructions check the results for sum rate zeebe process instance creations total by creation mode expected behavior the correct metrics should be incremented environment os zeebe version medic cw and onwards configuration
1
342,515
30,625,405,028
IssuesEvent
2023-07-24 11:07:18
keycloak/keycloak
https://api.github.com/repos/keycloak/keycloak
closed
Add Maven Build Cache Extension as opt-in feature
area/ci kind/task team/continuous-testing
### Description When implementing #16312, the plugin Maven Build Cache Extension couldn't be disabled by default and then re-enabled on the CLI for individual users. With version 1.0.1 being available, this is now possible: https://github.com/apache/maven-build-cache-extension/pull/37 Also Quarkus supports no Maven 3.9 https://github.com/quarkusio/quarkus/issues/31011 So this PR will add support for the cache extension.
1.0
Add Maven Build Cache Extension as opt-in feature - ### Description When implementing #16312, the plugin Maven Build Cache Extension couldn't be disabled by default and then re-enabled on the CLI for individual users. With version 1.0.1 being available, this is now possible: https://github.com/apache/maven-build-cache-extension/pull/37 Also Quarkus supports no Maven 3.9 https://github.com/quarkusio/quarkus/issues/31011 So this PR will add support for the cache extension.
non_process
add maven build cache extension as opt in feature description when implementing the plugin maven build cache extension couldn t be disabled by default and then re enabled on the cli for individual users with version being available this is now possible also quarkus supports no maven so this pr will add support for the cache extension
0
427,850
12,399,794,424
IssuesEvent
2020-05-21 06:17:58
magento/magento2
https://api.github.com/repos/magento/magento2
closed
Missaligned fields on the Customer Login form
Component: Customer Fixed in 2.4.x Issue: Clear Description Issue: Confirmed Issue: Format is valid Issue: Ready for Work Priority: P4 Progress: PR in progress Reproduced on 2.4.x Severity: S4 Triage: Internal Triage Done
<!--- Please review our guidelines before adding a new issue: https://github.com/magento/magento2/wiki/Issue-reporting-guidelines Fields marked with (*) are required. Please don't remove the template. --> ### Preconditions (*) <!--- Provide the exact Magento version (example: 2.3.2) and any important information on the environment where bug is reproducible. --> 1. Magento CE 2.3.5-p1 vanilla without sample data is installed. 2. Magento Blank Theme is used. 3. Screen/Browser width >= 768px. ### Steps to reproduce (*) <!--- Important: Provide a set of clear steps to reproduce this bug. We can not provide support without clear instructions on how to reproduce. --> 1. Login to the admin panel. 2. Navigate to the Content -> Design -> Configuration. 3. Open the Global Store and set the applied theme to 'Magento Blank'. 4. Clean the cache(s). 5. Open the storefront and navigate to the 'Sign In' page (/customer/account/login/) 6. See the input fields on the Customer Login form are not stroking with the design of all other forms on the default Blank Theme. ### Expected result (*) <!--- Tell us what do you expect to happen. --> The input fields on the form should be inline with the labels (Like all other forms in the Blank Theme). | Expected Result | | ------------- | | ![Schermafbeelding 2020-04-30 om 11 41 36](https://user-images.githubusercontent.com/6707410/80696688-0d5d2a80-8ad8-11ea-8ab0-5b0ce6c6c99f.png) | ### Actual result (*) <!--- Tell us what happened instead. Include error messages and issues. --> The input fields are shown below the labels and not aligned. | Actual Result | | ------------- | | ![Schermafbeelding 2020-04-30 om 11 48 50](https://user-images.githubusercontent.com/6707410/80697002-996f5200-8ad8-11ea-9191-f0fd32be1e70.png) | ### Additional information On the Magento 2.4-develop instance the issue looks fixed. But this is not stroking with the styling of all the other form fields design of the Magento Blank theme. | Result on 2.4-develop | | ------------- | | ![Schermafbeelding 2020-04-30 om 12 00 23](https://user-images.githubusercontent.com/6707410/80698104-4dbda800-8ada-11ea-802b-bf50adf68787.png) | On the Magento 2.3.5-p1 the issue is probably caused by the following css line: ![Schermafbeelding 2020-04-30 om 12 02 45](https://user-images.githubusercontent.com/6707410/80698318-a0975f80-8ada-11ea-809b-07700daa85bb.png)
1.0
Missaligned fields on the Customer Login form - <!--- Please review our guidelines before adding a new issue: https://github.com/magento/magento2/wiki/Issue-reporting-guidelines Fields marked with (*) are required. Please don't remove the template. --> ### Preconditions (*) <!--- Provide the exact Magento version (example: 2.3.2) and any important information on the environment where bug is reproducible. --> 1. Magento CE 2.3.5-p1 vanilla without sample data is installed. 2. Magento Blank Theme is used. 3. Screen/Browser width >= 768px. ### Steps to reproduce (*) <!--- Important: Provide a set of clear steps to reproduce this bug. We can not provide support without clear instructions on how to reproduce. --> 1. Login to the admin panel. 2. Navigate to the Content -> Design -> Configuration. 3. Open the Global Store and set the applied theme to 'Magento Blank'. 4. Clean the cache(s). 5. Open the storefront and navigate to the 'Sign In' page (/customer/account/login/) 6. See the input fields on the Customer Login form are not stroking with the design of all other forms on the default Blank Theme. ### Expected result (*) <!--- Tell us what do you expect to happen. --> The input fields on the form should be inline with the labels (Like all other forms in the Blank Theme). | Expected Result | | ------------- | | ![Schermafbeelding 2020-04-30 om 11 41 36](https://user-images.githubusercontent.com/6707410/80696688-0d5d2a80-8ad8-11ea-8ab0-5b0ce6c6c99f.png) | ### Actual result (*) <!--- Tell us what happened instead. Include error messages and issues. --> The input fields are shown below the labels and not aligned. | Actual Result | | ------------- | | ![Schermafbeelding 2020-04-30 om 11 48 50](https://user-images.githubusercontent.com/6707410/80697002-996f5200-8ad8-11ea-9191-f0fd32be1e70.png) | ### Additional information On the Magento 2.4-develop instance the issue looks fixed. But this is not stroking with the styling of all the other form fields design of the Magento Blank theme. | Result on 2.4-develop | | ------------- | | ![Schermafbeelding 2020-04-30 om 12 00 23](https://user-images.githubusercontent.com/6707410/80698104-4dbda800-8ada-11ea-802b-bf50adf68787.png) | On the Magento 2.3.5-p1 the issue is probably caused by the following css line: ![Schermafbeelding 2020-04-30 om 12 02 45](https://user-images.githubusercontent.com/6707410/80698318-a0975f80-8ada-11ea-809b-07700daa85bb.png)
non_process
missaligned fields on the customer login form please review our guidelines before adding a new issue fields marked with are required please don t remove the template preconditions provide the exact magento version example and any important information on the environment where bug is reproducible magento ce vanilla without sample data is installed magento blank theme is used screen browser width steps to reproduce important provide a set of clear steps to reproduce this bug we can not provide support without clear instructions on how to reproduce login to the admin panel navigate to the content design configuration open the global store and set the applied theme to magento blank clean the cache s open the storefront and navigate to the sign in page customer account login see the input fields on the customer login form are not stroking with the design of all other forms on the default blank theme expected result the input fields on the form should be inline with the labels like all other forms in the blank theme expected result actual result the input fields are shown below the labels and not aligned actual result additional information on the magento develop instance the issue looks fixed but this is not stroking with the styling of all the other form fields design of the magento blank theme result on develop on the magento the issue is probably caused by the following css line
0
788,101
27,743,019,255
IssuesEvent
2023-03-15 15:23:43
AY2223S2-CS2113-T12-3/tp
https://api.github.com/repos/AY2223S2-CS2113-T12-3/tp
closed
[Task] Add User Interface to procces and run user commands
priority.High severity.High
**Describe the Task** Add User Interface to process user commands inputted from the user. User Interface executes the various commands based on command key word. **Expected behavior** 1. list pu: lists out the list of universities available in Korea. 2. list current: lists out current modules added to user list. 3. add [module code]: adds module to user list, using module code. 4. list [pu name]: lists out modules that are provided by the universities.
1.0
[Task] Add User Interface to procces and run user commands - **Describe the Task** Add User Interface to process user commands inputted from the user. User Interface executes the various commands based on command key word. **Expected behavior** 1. list pu: lists out the list of universities available in Korea. 2. list current: lists out current modules added to user list. 3. add [module code]: adds module to user list, using module code. 4. list [pu name]: lists out modules that are provided by the universities.
non_process
add user interface to procces and run user commands describe the task add user interface to process user commands inputted from the user user interface executes the various commands based on command key word expected behavior list pu lists out the list of universities available in korea list current lists out current modules added to user list add adds module to user list using module code list lists out modules that are provided by the universities
0
690,032
23,643,727,587
IssuesEvent
2022-08-25 19:42:28
lowRISC/opentitan
https://api.github.com/repos/lowRISC/opentitan
closed
[mask_rom/e2e_tests] Test signature verification with SW and OTBN
Component:Software Priority:P2 Type:Task SW:ROM Milestone:V2
Consider testing both SW verification and OTBN verification of the ROM_EXT signature in the Mask ROM e2e tests.
1.0
[mask_rom/e2e_tests] Test signature verification with SW and OTBN - Consider testing both SW verification and OTBN verification of the ROM_EXT signature in the Mask ROM e2e tests.
non_process
test signature verification with sw and otbn consider testing both sw verification and otbn verification of the rom ext signature in the mask rom tests
0
193,638
14,658,787,043
IssuesEvent
2020-12-28 18:47:34
github-vet/rangeloop-pointer-findings
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
closed
songbowei839/smallHttpServer: src/github.com/ugorji/go/codec/values_codecgen_generated_test.go; 11 LoC
fresh small test
Found a possible issue in [songbowei839/smallHttpServer](https://www.github.com/songbowei839/smallHttpServer) at [src/github.com/ugorji/go/codec/values_codecgen_generated_test.go](https://github.com/songbowei839/smallHttpServer/blob/675d00df5de4f365a6736e4f87b94067f3a40d67/src/github.com/ugorji/go/codec/values_codecgen_generated_test.go#L12207-L12217) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > reference to yyv1 is reassigned at line 12215 [Click here to see the code in its original context.](https://github.com/songbowei839/smallHttpServer/blob/675d00df5de4f365a6736e4f87b94067f3a40d67/src/github.com/ugorji/go/codec/values_codecgen_generated_test.go#L12207-L12217) <details> <summary>Click here to show the 11 line(s) of Go which triggered the analyzer.</summary> ```go for yyk1, yyv1 := range v { z.EncWriteMapElemKey() if z.EncBasicHandle().StringToRaw { r.EncodeStringBytesRaw(z.BytesView(string(yyk1))) } else { r.EncodeStringEnc(codecSelferCcUTF819780, string(yyk1)) } z.EncWriteMapElemValue() yy3 := &yyv1 yy3.CodecEncodeSelf(e) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 675d00df5de4f365a6736e4f87b94067f3a40d67
1.0
songbowei839/smallHttpServer: src/github.com/ugorji/go/codec/values_codecgen_generated_test.go; 11 LoC - Found a possible issue in [songbowei839/smallHttpServer](https://www.github.com/songbowei839/smallHttpServer) at [src/github.com/ugorji/go/codec/values_codecgen_generated_test.go](https://github.com/songbowei839/smallHttpServer/blob/675d00df5de4f365a6736e4f87b94067f3a40d67/src/github.com/ugorji/go/codec/values_codecgen_generated_test.go#L12207-L12217) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > reference to yyv1 is reassigned at line 12215 [Click here to see the code in its original context.](https://github.com/songbowei839/smallHttpServer/blob/675d00df5de4f365a6736e4f87b94067f3a40d67/src/github.com/ugorji/go/codec/values_codecgen_generated_test.go#L12207-L12217) <details> <summary>Click here to show the 11 line(s) of Go which triggered the analyzer.</summary> ```go for yyk1, yyv1 := range v { z.EncWriteMapElemKey() if z.EncBasicHandle().StringToRaw { r.EncodeStringBytesRaw(z.BytesView(string(yyk1))) } else { r.EncodeStringEnc(codecSelferCcUTF819780, string(yyk1)) } z.EncWriteMapElemValue() yy3 := &yyv1 yy3.CodecEncodeSelf(e) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 675d00df5de4f365a6736e4f87b94067f3a40d67
non_process
smallhttpserver src github com ugorji go codec values codecgen generated test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message reference to is reassigned at line click here to show the line s of go which triggered the analyzer go for range v z encwritemapelemkey if z encbasichandle stringtoraw r encodestringbytesraw z bytesview string else r encodestringenc string z encwritemapelemvalue codecencodeself e leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
0
800,650
28,373,644,357
IssuesEvent
2023-04-12 18:59:04
DDMAL/CantusDB
https://api.github.com/repos/DDMAL/CantusDB
closed
syllabifying tool should draw on standard text if there is no full text
priority: medium
in OldCantus, editing the syllabification of a chant with no MS Full Text will give you a window where the syllabification is populated with the current syllabification based on the "Standardized" reading: ![image](https://user-images.githubusercontent.com/67451875/230173767-3ce052eb-536a-4a42-ab47-8d6322642571.png) In NewCantus these cases will give you two blank fields, which means you can't see what text it was you were trying to edit! ![image](https://user-images.githubusercontent.com/67451875/230176237-66df7b15-794f-4044-b0e2-8979e2677344.png)
1.0
syllabifying tool should draw on standard text if there is no full text - in OldCantus, editing the syllabification of a chant with no MS Full Text will give you a window where the syllabification is populated with the current syllabification based on the "Standardized" reading: ![image](https://user-images.githubusercontent.com/67451875/230173767-3ce052eb-536a-4a42-ab47-8d6322642571.png) In NewCantus these cases will give you two blank fields, which means you can't see what text it was you were trying to edit! ![image](https://user-images.githubusercontent.com/67451875/230176237-66df7b15-794f-4044-b0e2-8979e2677344.png)
non_process
syllabifying tool should draw on standard text if there is no full text in oldcantus editing the syllabification of a chant with no ms full text will give you a window where the syllabification is populated with the current syllabification based on the standardized reading in newcantus these cases will give you two blank fields which means you can t see what text it was you were trying to edit
0
19,980
26,459,492,807
IssuesEvent
2023-01-16 16:20:26
rladstaetter/LogoRRR
https://api.github.com/repos/rladstaetter/LogoRRR
closed
Update dependencies, compilers, cleanup
release process
LogoRRR uses a rich ecosystem which changes very quickly. In each release there are updates to various 3rd party libraries or build infrastructure which should be reflected in this issue. Code cleanup, general yak shaving ... here is the place to be.
1.0
Update dependencies, compilers, cleanup - LogoRRR uses a rich ecosystem which changes very quickly. In each release there are updates to various 3rd party libraries or build infrastructure which should be reflected in this issue. Code cleanup, general yak shaving ... here is the place to be.
process
update dependencies compilers cleanup logorrr uses a rich ecosystem which changes very quickly in each release there are updates to various party libraries or build infrastructure which should be reflected in this issue code cleanup general yak shaving here is the place to be
1
38,363
8,788,038,622
IssuesEvent
2018-12-20 20:44:17
NREL/EnergyPlus
https://api.github.com/repos/NREL/EnergyPlus
opened
Zone over-conditioned when using AirLoopHVAC + ZoneHVAC:IdealLoadsAirSystem
Defect PriorityHigh SeverityHigh
Issue overview -------------- I'm doing testing on a simple residential building (one conditioned thermal zone) served by either zonal or airloop HVAC equipment. My tests involve adding a ZoneHVAC:IdealLoadsAirSystem to the model with the expectation that the energy use of the original HVAC system is cut in half (because it's UniformLoad scheme). All simulations were run using a build created from [the UniformLoad bugfix PR](https://github.com/NREL/EnergyPlus/pull/7025). Here are my findings:   | Zonal or AirLoop | HVAC Energy w/o IdealLoads | HVAC Energy w/ IdealLoads | Calculated Fraction | Result -- | -- | -- | -- | -- | -- ZoneHVAC:Baseboard:Convective:Water | Zonal | 40.32 | 20.16 | 0.50 | Good ZoneHVAC:PackagedTerminalAirConditioner | Zonal | 41.60 | 21.13 | 0.51 | Good AirLoopHVAC:UnitarySystem (heating coil only) | Zonal | 46.35 | 23.17 | 0.50 | Good AirLoopHVAC:UnitarySystem (cooling coil only) | AirLoop | 18.57 | 12.59 | 0.68 | **Bad** AirLoopHVAC:UnitarySystem (heating coil only) | AirLoop | 35.91 | 25.24 | 0.70 | **Bad** I get near perfect results for zonal equipment (including UnitarySystems used as zonal equipment). But the AirLoopHVAC test results are way off -- around 0.7 instead of 0.5. When I look at the model with the IdealLoads system, it appears to be over-conditioning the zone. ![image](https://user-images.githubusercontent.com/5861765/50309774-52416d80-045d-11e9-89e4-dda50b7a49f2.png) The two IDFs corresponding to the last row of the table are attached: [AirLoopHVAC_withIdealLoads.idf.txt](https://github.com/NREL/EnergyPlus/files/2700404/AirLoopHVAC_withIdealLoads.idf.txt) [AirLoopHVAC_withoutIdealLoads.idf.txt](https://github.com/NREL/EnergyPlus/files/2700408/AirLoopHVAC_withoutIdealLoads.idf.txt) ### Details Some additional details for this issue (if relevant): - Platform (Operating system, version) - Version of EnergyPlus (if using an intermediate build, include SHA) - Unmethours link or helpdesk ticket number ### Checklist Add to this list or remove from it as applicable. This is a simple templated set of guidelines. - [ ] Defect file added (list location of defect file here) - [ ] Ticket added to Pivotal for defect (development team task) - [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
1.0
Zone over-conditioned when using AirLoopHVAC + ZoneHVAC:IdealLoadsAirSystem - Issue overview -------------- I'm doing testing on a simple residential building (one conditioned thermal zone) served by either zonal or airloop HVAC equipment. My tests involve adding a ZoneHVAC:IdealLoadsAirSystem to the model with the expectation that the energy use of the original HVAC system is cut in half (because it's UniformLoad scheme). All simulations were run using a build created from [the UniformLoad bugfix PR](https://github.com/NREL/EnergyPlus/pull/7025). Here are my findings:   | Zonal or AirLoop | HVAC Energy w/o IdealLoads | HVAC Energy w/ IdealLoads | Calculated Fraction | Result -- | -- | -- | -- | -- | -- ZoneHVAC:Baseboard:Convective:Water | Zonal | 40.32 | 20.16 | 0.50 | Good ZoneHVAC:PackagedTerminalAirConditioner | Zonal | 41.60 | 21.13 | 0.51 | Good AirLoopHVAC:UnitarySystem (heating coil only) | Zonal | 46.35 | 23.17 | 0.50 | Good AirLoopHVAC:UnitarySystem (cooling coil only) | AirLoop | 18.57 | 12.59 | 0.68 | **Bad** AirLoopHVAC:UnitarySystem (heating coil only) | AirLoop | 35.91 | 25.24 | 0.70 | **Bad** I get near perfect results for zonal equipment (including UnitarySystems used as zonal equipment). But the AirLoopHVAC test results are way off -- around 0.7 instead of 0.5. When I look at the model with the IdealLoads system, it appears to be over-conditioning the zone. ![image](https://user-images.githubusercontent.com/5861765/50309774-52416d80-045d-11e9-89e4-dda50b7a49f2.png) The two IDFs corresponding to the last row of the table are attached: [AirLoopHVAC_withIdealLoads.idf.txt](https://github.com/NREL/EnergyPlus/files/2700404/AirLoopHVAC_withIdealLoads.idf.txt) [AirLoopHVAC_withoutIdealLoads.idf.txt](https://github.com/NREL/EnergyPlus/files/2700408/AirLoopHVAC_withoutIdealLoads.idf.txt) ### Details Some additional details for this issue (if relevant): - Platform (Operating system, version) - Version of EnergyPlus (if using an intermediate build, include SHA) - Unmethours link or helpdesk ticket number ### Checklist Add to this list or remove from it as applicable. This is a simple templated set of guidelines. - [ ] Defect file added (list location of defect file here) - [ ] Ticket added to Pivotal for defect (development team task) - [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
non_process
zone over conditioned when using airloophvac zonehvac idealloadsairsystem issue overview i m doing testing on a simple residential building one conditioned thermal zone served by either zonal or airloop hvac equipment my tests involve adding a zonehvac idealloadsairsystem to the model with the expectation that the energy use of the original hvac system is cut in half because it s uniformload scheme all simulations were run using a build created from here are my findings   zonal or airloop hvac energy w o idealloads hvac energy w idealloads calculated fraction result zonehvac baseboard convective water zonal good zonehvac packagedterminalairconditioner zonal good airloophvac unitarysystem heating coil only zonal good airloophvac unitarysystem cooling coil only airloop bad airloophvac unitarysystem heating coil only airloop bad i get near perfect results for zonal equipment including unitarysystems used as zonal equipment but the airloophvac test results are way off around instead of when i look at the model with the idealloads system it appears to be over conditioning the zone the two idfs corresponding to the last row of the table are attached details some additional details for this issue if relevant platform operating system version version of energyplus if using an intermediate build include sha unmethours link or helpdesk ticket number checklist add to this list or remove from it as applicable this is a simple templated set of guidelines defect file added list location of defect file here ticket added to pivotal for defect development team task pull request created the pull request will have additional tasks related to reviewing changes that fix this defect
0
249,388
26,923,370,724
IssuesEvent
2023-02-07 12:07:17
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Remove the unsafe-eval source expression from the Kibana CSP by default
Team:Security Feature:Hardening
Today the `unsafe-eval` source expression is "on" by default in the Kibana Content Security Policy (CSP). Users can choose to set `csp.disableUnsafeEval` to `true` to remove `unsafe-eval` from the CSP. Once we land support for inline partials in `@kbn/handlebars`, we can change the default behaviour so `unsafe-eval` is not present in our CSP unless users explicitly opt in.
True
Remove the unsafe-eval source expression from the Kibana CSP by default - Today the `unsafe-eval` source expression is "on" by default in the Kibana Content Security Policy (CSP). Users can choose to set `csp.disableUnsafeEval` to `true` to remove `unsafe-eval` from the CSP. Once we land support for inline partials in `@kbn/handlebars`, we can change the default behaviour so `unsafe-eval` is not present in our CSP unless users explicitly opt in.
non_process
remove the unsafe eval source expression from the kibana csp by default today the unsafe eval source expression is on by default in the kibana content security policy csp users can choose to set csp disableunsafeeval to true to remove unsafe eval from the csp once we land support for inline partials in kbn handlebars we can change the default behaviour so unsafe eval is not present in our csp unless users explicitly opt in
0
56,992
11,698,763,688
IssuesEvent
2020-03-06 14:28:46
eclipse/codewind
https://api.github.com/repos/eclipse/codewind
closed
Pick up appsody 0.5.9 release for 0.10.0
area/appsody area/eclipse-ide area/intellij-ide area/vscode-ide kind/enhancement
<!-- Please fill out the following form to suggest an enhancement. If some fields do not apply to your situation, feel free to skip them.--> **Codewind version:** **OS:** **Che version:** **IDE extension version:** **IDE version:** **Kubernetes cluster:** **Description of the enhancement:** <!-- How would you like to see Codewind improved?--> Pick up appsody `0.5.9` which is scheduled to be released on Mar 4. **Proposed solution:** <!-- Do you have ideas about how your idea could be implemented?-->
1.0
Pick up appsody 0.5.9 release for 0.10.0 - <!-- Please fill out the following form to suggest an enhancement. If some fields do not apply to your situation, feel free to skip them.--> **Codewind version:** **OS:** **Che version:** **IDE extension version:** **IDE version:** **Kubernetes cluster:** **Description of the enhancement:** <!-- How would you like to see Codewind improved?--> Pick up appsody `0.5.9` which is scheduled to be released on Mar 4. **Proposed solution:** <!-- Do you have ideas about how your idea could be implemented?-->
non_process
pick up appsody release for codewind version os che version ide extension version ide version kubernetes cluster description of the enhancement pick up appsody which is scheduled to be released on mar proposed solution
0
762,707
26,727,078,272
IssuesEvent
2023-01-29 20:58:29
asastats/channel
https://api.github.com/repos/asastats/channel
closed
Add Humble's ALGO/Clown farm
feature high priority addressed
Feature link: https://app.humble.sh/farm?id=1024865003 Application ID: 1024865003 Also Clown/OPUL, Clown/YLDY, Planets/Clown
1.0
Add Humble's ALGO/Clown farm - Feature link: https://app.humble.sh/farm?id=1024865003 Application ID: 1024865003 Also Clown/OPUL, Clown/YLDY, Planets/Clown
non_process
add humble s algo clown farm feature link application id also clown opul clown yldy planets clown
0
22,314
3,633,122,986
IssuesEvent
2016-02-11 13:19:09
bridgedotnet/Bridge
https://api.github.com/repos/bridgedotnet/Bridge
closed
try/catch within a finally block in an async function causes error.
defect
The following test case causes an internal compiler error: public async static void Main() { try { await Task.Delay(1000); } finally { Console.WriteLine("finally"); try { throw new Exception("test"); } catch {} } } http://live.bridge.net/#6f0c3d65729583cb9196
1.0
try/catch within a finally block in an async function causes error. - The following test case causes an internal compiler error: public async static void Main() { try { await Task.Delay(1000); } finally { Console.WriteLine("finally"); try { throw new Exception("test"); } catch {} } } http://live.bridge.net/#6f0c3d65729583cb9196
non_process
try catch within a finally block in an async function causes error the following test case causes an internal compiler error public async static void main try await task delay finally console writeline finally try throw new exception test catch
0
1,842
4,647,114,220
IssuesEvent
2016-10-01 09:09:48
AllenFang/react-bootstrap-table
https://api.github.com/repos/AllenFang/react-bootstrap-table
closed
Is there any way to change edit inline to popover edit ?
help wanted inprocess
Is there any way to change edit inline to popover edit ? @AllenFang
1.0
Is there any way to change edit inline to popover edit ? - Is there any way to change edit inline to popover edit ? @AllenFang
process
is there any way to change edit inline to popover edit is there any way to change edit inline to popover edit allenfang
1
9,689
12,691,391,692
IssuesEvent
2020-06-21 16:47:38
kubeflow/community
https://api.github.com/repos/kubeflow/community
closed
[Process/Tools] needed to follow up with area owners about issues
area/engprod kind/feature kind/process lifecycle/stale priority/p1
We need a process or tools to ensure that area owners are responding to issues in a timely fashion. Right now the triage process assigns area labels to issues. But we have no follow up to ensure that the area owners are following up on those issues. @jtfogarty Any thoughts?
1.0
[Process/Tools] needed to follow up with area owners about issues - We need a process or tools to ensure that area owners are responding to issues in a timely fashion. Right now the triage process assigns area labels to issues. But we have no follow up to ensure that the area owners are following up on those issues. @jtfogarty Any thoughts?
process
needed to follow up with area owners about issues we need a process or tools to ensure that area owners are responding to issues in a timely fashion right now the triage process assigns area labels to issues but we have no follow up to ensure that the area owners are following up on those issues jtfogarty any thoughts
1
4,825
7,720,444,002
IssuesEvent
2018-05-23 23:13:01
UnbFeelings/unb-feelings-docs
https://api.github.com/repos/UnbFeelings/unb-feelings-docs
closed
[Não Conformidade] - Relatório de Desempenho
Processo medição
A auditoria do relatório de desempenho foi executada para verificar se a atividade de medição e análise foi realizada. O resultado da auditoria pode ser acessado através da seguinte página: [Auditoria Relatório de Desempenho - Ciclo 2](https://github.com/UnbFeelings/unb-feelings-GQA/wiki/Auditoria-do-Relat%C3%B3rio-de-Desempenho-Ciclo-2). ## Descrição O documento **Relatório de Desempenho** não foi elaborado. ### Recomendações Tendo em vista que a coleta de métricas foi iniciada, basta terminar a coleta para possibilitar a elaboração do documento. ## Detalhes **Autor:** Guilherme Sant'Ana **Tipo:** Medição e Análise **Prazo:** 24/05/18
1.0
[Não Conformidade] - Relatório de Desempenho - A auditoria do relatório de desempenho foi executada para verificar se a atividade de medição e análise foi realizada. O resultado da auditoria pode ser acessado através da seguinte página: [Auditoria Relatório de Desempenho - Ciclo 2](https://github.com/UnbFeelings/unb-feelings-GQA/wiki/Auditoria-do-Relat%C3%B3rio-de-Desempenho-Ciclo-2). ## Descrição O documento **Relatório de Desempenho** não foi elaborado. ### Recomendações Tendo em vista que a coleta de métricas foi iniciada, basta terminar a coleta para possibilitar a elaboração do documento. ## Detalhes **Autor:** Guilherme Sant'Ana **Tipo:** Medição e Análise **Prazo:** 24/05/18
process
relatório de desempenho a auditoria do relatório de desempenho foi executada para verificar se a atividade de medição e análise foi realizada o resultado da auditoria pode ser acessado através da seguinte página descrição o documento relatório de desempenho não foi elaborado recomendações tendo em vista que a coleta de métricas foi iniciada basta terminar a coleta para possibilitar a elaboração do documento detalhes autor guilherme sant ana tipo medição e análise prazo
1
34,898
14,541,484,652
IssuesEvent
2020-12-15 14:37:34
hashicorp/terraform-provider-aws
https://api.github.com/repos/hashicorp/terraform-provider-aws
closed
r/aws_instance,r/aws_launch_template: Support for AWS Nitro Enclaves
enhancement service/ec2
<!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description AWS Nitro Enclaves are now generally available. ### New or Affected Resource(s) <!--- Please list the new or affected resources and data sources. ---> #### Resources * [`aws_instance`](https://registry.terraform.io/providers/hashicorp/aws/latest/docs/resources/instance) * [`aws_launch_template`](https://registry.terraform.io/providers/hashicorp/aws/latest/docs/resources/launch_template) #### Data sources * [`aws_instance`](https://registry.terraform.io/providers/hashicorp/aws/latest/docs/data-sources/instance) * [`aws_launch_template`](https://registry.terraform.io/providers/hashicorp/aws/latest/docs/data-sources/launch_template) ### Potential Terraform Configuration <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl resource "aws_instance" "example" { enclave_options { enabled = true } } ``` ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example: * https://aws.amazon.com/about-aws/whats-new/2018/04/introducing-amazon-ec2-fleet/ ---> [Announcement](https://aws.amazon.com/about-aws/whats-new/2020/10/announcing-general-availability-of-aws-nitro-enclaves/). [Blog post](https://aws.amazon.com/blogs/aws/aws-nitro-enclaves-isolated-ec2-environments-to-process-confidential-data/). [User Guide](https://docs.aws.amazon.com/enclaves/latest/user/nitro-enclave.html). Requires [AWS SDK v1.35.17](https://github.com/aws/aws-sdk-go/releases/tag/v1.35.17): * https://github.com/terraform-providers/terraform-provider-aws/pull/15905
1.0
r/aws_instance,r/aws_launch_template: Support for AWS Nitro Enclaves - <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description AWS Nitro Enclaves are now generally available. ### New or Affected Resource(s) <!--- Please list the new or affected resources and data sources. ---> #### Resources * [`aws_instance`](https://registry.terraform.io/providers/hashicorp/aws/latest/docs/resources/instance) * [`aws_launch_template`](https://registry.terraform.io/providers/hashicorp/aws/latest/docs/resources/launch_template) #### Data sources * [`aws_instance`](https://registry.terraform.io/providers/hashicorp/aws/latest/docs/data-sources/instance) * [`aws_launch_template`](https://registry.terraform.io/providers/hashicorp/aws/latest/docs/data-sources/launch_template) ### Potential Terraform Configuration <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl resource "aws_instance" "example" { enclave_options { enabled = true } } ``` ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example: * https://aws.amazon.com/about-aws/whats-new/2018/04/introducing-amazon-ec2-fleet/ ---> [Announcement](https://aws.amazon.com/about-aws/whats-new/2020/10/announcing-general-availability-of-aws-nitro-enclaves/). [Blog post](https://aws.amazon.com/blogs/aws/aws-nitro-enclaves-isolated-ec2-environments-to-process-confidential-data/). [User Guide](https://docs.aws.amazon.com/enclaves/latest/user/nitro-enclave.html). Requires [AWS SDK v1.35.17](https://github.com/aws/aws-sdk-go/releases/tag/v1.35.17): * https://github.com/terraform-providers/terraform-provider-aws/pull/15905
non_process
r aws instance r aws launch template support for aws nitro enclaves community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description aws nitro enclaves are now generally available new or affected resource s resources data sources potential terraform configuration hcl resource aws instance example enclave options enabled true references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation for example requires
0
50,001
13,187,305,097
IssuesEvent
2020-08-13 02:59:35
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
[clsim] I3CLSimLightSourceToStepConverterFlasher hangs (Trac #2426)
Incomplete Migration Migrated from Trac combo simulation defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2426">https://code.icecube.wisc.edu/ticket/2426</a>, reported by jvansanten and owned by jvansanten</em></summary> <p> ```json { "status": "closed", "changetime": "2020-06-24T12:32:17", "description": "Steps to reproduce:\n\n{{{\nclsim/resources/scripts/flasher/generateTestFlashes.py\nclsim/resources/scripts/flasher/applyCLSim.py -i test_flashes.i3\n}}}\n\nand observe the second process not exiting, and also not using any CPU time.\n\nThis happens `I3CLSimLightSourceToStepConverterFlasher` can only run in the main thread, as it attempts to call `I3CLSimRandomValueIceCubeFlasherTimeProfile.SampleFromDistribution()` and waits forever trying to acquire the GIL that the main thread never releases. This clashes with its use within I3CLSimLightSourceToStepConverterAsync.\n\nThere are two solutions: either drop the GIL in `I3Tray::Execute()` (deja vu, anyone?) or port I3CLSimRandomValueIceCubeFlasherTimeProfile to C++. The latter is probably easier in the short term.", "reporter": "jvansanten", "cc": "fiedl", "resolution": "fixed", "_ts": "1593001937450890", "component": "combo simulation", "summary": "[clsim] I3CLSimLightSourceToStepConverterFlasher hangs", "priority": "major", "keywords": "", "time": "2020-04-23T19:42:44", "milestone": "Autumnal Equinox 2020", "owner": "jvansanten", "type": "defect" } ``` </p> </details>
1.0
[clsim] I3CLSimLightSourceToStepConverterFlasher hangs (Trac #2426) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2426">https://code.icecube.wisc.edu/ticket/2426</a>, reported by jvansanten and owned by jvansanten</em></summary> <p> ```json { "status": "closed", "changetime": "2020-06-24T12:32:17", "description": "Steps to reproduce:\n\n{{{\nclsim/resources/scripts/flasher/generateTestFlashes.py\nclsim/resources/scripts/flasher/applyCLSim.py -i test_flashes.i3\n}}}\n\nand observe the second process not exiting, and also not using any CPU time.\n\nThis happens `I3CLSimLightSourceToStepConverterFlasher` can only run in the main thread, as it attempts to call `I3CLSimRandomValueIceCubeFlasherTimeProfile.SampleFromDistribution()` and waits forever trying to acquire the GIL that the main thread never releases. This clashes with its use within I3CLSimLightSourceToStepConverterAsync.\n\nThere are two solutions: either drop the GIL in `I3Tray::Execute()` (deja vu, anyone?) or port I3CLSimRandomValueIceCubeFlasherTimeProfile to C++. The latter is probably easier in the short term.", "reporter": "jvansanten", "cc": "fiedl", "resolution": "fixed", "_ts": "1593001937450890", "component": "combo simulation", "summary": "[clsim] I3CLSimLightSourceToStepConverterFlasher hangs", "priority": "major", "keywords": "", "time": "2020-04-23T19:42:44", "milestone": "Autumnal Equinox 2020", "owner": "jvansanten", "type": "defect" } ``` </p> </details>
non_process
hangs trac migrated from json status closed changetime description steps to reproduce n n nclsim resources scripts flasher generatetestflashes py nclsim resources scripts flasher applyclsim py i test flashes n n nand observe the second process not exiting and also not using any cpu time n nthis happens can only run in the main thread as it attempts to call samplefromdistribution and waits forever trying to acquire the gil that the main thread never releases this clashes with its use within n nthere are two solutions either drop the gil in execute deja vu anyone or port to c the latter is probably easier in the short term reporter jvansanten cc fiedl resolution fixed ts component combo simulation summary hangs priority major keywords time milestone autumnal equinox owner jvansanten type defect
0
4,671
5,225,434,402
IssuesEvent
2017-01-27 18:16:57
couchbaselabs/mobile-testkit
https://api.github.com/repos/couchbaselabs/mobile-testkit
closed
Allow 'generate_clusters_from_pool.py' to take a target pool file.
icebox infrastructure
This will allow the user to store multiple configs for easier switching After this is resolved: ``` python libraries/utilities/generate_clusters_from_pool.py --target='resource/pool-local.json' ``` or ``` python libraries/utilities/generate_clusters_from_pool.py --target='resource/pool-aws.json' ``` instead of relying on a 'resources/pool.json'
1.0
Allow 'generate_clusters_from_pool.py' to take a target pool file. - This will allow the user to store multiple configs for easier switching After this is resolved: ``` python libraries/utilities/generate_clusters_from_pool.py --target='resource/pool-local.json' ``` or ``` python libraries/utilities/generate_clusters_from_pool.py --target='resource/pool-aws.json' ``` instead of relying on a 'resources/pool.json'
non_process
allow generate clusters from pool py to take a target pool file this will allow the user to store multiple configs for easier switching after this is resolved python libraries utilities generate clusters from pool py target resource pool local json or python libraries utilities generate clusters from pool py target resource pool aws json instead of relying on a resources pool json
0
501,099
14,521,156,279
IssuesEvent
2020-12-14 06:53:07
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.cpuid.com - desktop site instead of mobile site
browser-fenix engine-gecko ml-needsdiagnosis-false ml-probability-high priority-normal
<!-- @browser: Firefox Mobile 84.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:84.0) Gecko/84.0 Firefox/84.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/63573 --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.cpuid.com/ **Browser / Version**: Firefox Mobile 84.0 **Operating System**: Android 8.1.0 **Tested Another Browser**: No **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2020/12/cdf33b7e-3446-4fc2-827e-9aec42826825.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201206192040</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: true</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/12/2398d716-1b1e-49c0-b736-13111fd9e097) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.cpuid.com - desktop site instead of mobile site - <!-- @browser: Firefox Mobile 84.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:84.0) Gecko/84.0 Firefox/84.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/63573 --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.cpuid.com/ **Browser / Version**: Firefox Mobile 84.0 **Operating System**: Android 8.1.0 **Tested Another Browser**: No **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2020/12/cdf33b7e-3446-4fc2-827e-9aec42826825.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201206192040</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: true</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/12/2398d716-1b1e-49c0-b736-13111fd9e097) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
desktop site instead of mobile site url browser version firefox mobile operating system android tested another browser no problem type desktop site instead of mobile site description desktop site instead of mobile site steps to reproduce view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen true mixed active content blocked true mixed passive content blocked false tracking content blocked false from with ❤️
0
137,922
18,769,545,026
IssuesEvent
2021-11-06 15:28:22
samqws-marketing/box_box-ui-elements
https://api.github.com/repos/samqws-marketing/box_box-ui-elements
opened
CVE-2021-37701 (High) detected in tar-4.4.13.tgz
security vulnerability
## CVE-2021-37701 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-4.4.13.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.13.tgz">https://registry.npmjs.org/tar/-/tar-4.4.13.tgz</a></p> <p>Path to dependency file: box_box-ui-elements/package.json</p> <p>Path to vulnerable library: box_box-ui-elements/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - storybook-5.3.9.tgz (Root Library) - cli-5.3.9.tgz - pkg-add-deps-0.1.0.tgz - pacote-9.5.12.tgz - :x: **tar-4.4.13.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/box_box-ui-elements/commit/4fc776e2b95c8b497f6994cb2165365562ae1f82">4fc776e2b95c8b497f6994cb2165365562ae1f82</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 4.4.16, 5.0.8, and 6.1.7 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory, where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems. The cache checking logic used both `\` and `/` characters as path separators, however `\` is a valid filename character on posix systems. By first creating a directory, and then replacing that directory with a symlink, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. Additionally, a similar confusion could arise on case-insensitive filesystems. If a tar archive contained a directory at `FOO`, followed by a symbolic link named `foo`, then on case-insensitive file systems, the creation of the symbolic link would remove the directory from the filesystem, but _not_ from the internal directory cache, as it would not be treated as a cache hit. A subsequent file entry within the `FOO` directory would then be placed in the target of the symbolic link, thinking that the directory had already been created. These issues were addressed in releases 4.4.16, 5.0.8 and 6.1.7. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-9r2w-394v-53qc. <p>Publish Date: 2021-08-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37701>CVE-2021-37701</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc">https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc</a></p> <p>Release Date: 2021-08-31</p> <p>Fix Resolution: tar - 4.4.16, 5.0.8, 6.1.7</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"tar","packageVersion":"4.4.13","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"storybook:5.3.9;@storybook/cli:5.3.9;pkg-add-deps:0.1.0;pacote:9.5.12;tar:4.4.13","isMinimumFixVersionAvailable":true,"minimumFixVersion":"tar - 4.4.16, 5.0.8, 6.1.7"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-37701","vulnerabilityDetails":"The npm package \"tar\" (aka node-tar) before versions 4.4.16, 5.0.8, and 6.1.7 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory, where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems. The cache checking logic used both `\\` and `/` characters as path separators, however `\\` is a valid filename character on posix systems. By first creating a directory, and then replacing that directory with a symlink, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. Additionally, a similar confusion could arise on case-insensitive filesystems. If a tar archive contained a directory at `FOO`, followed by a symbolic link named `foo`, then on case-insensitive file systems, the creation of the symbolic link would remove the directory from the filesystem, but _not_ from the internal directory cache, as it would not be treated as a cache hit. A subsequent file entry within the `FOO` directory would then be placed in the target of the symbolic link, thinking that the directory had already been created. These issues were addressed in releases 4.4.16, 5.0.8 and 6.1.7. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-9r2w-394v-53qc.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37701","cvss3Severity":"high","cvss3Score":"8.6","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Changed","C":"High","UI":"Required","AV":"Local","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-37701 (High) detected in tar-4.4.13.tgz - ## CVE-2021-37701 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-4.4.13.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.13.tgz">https://registry.npmjs.org/tar/-/tar-4.4.13.tgz</a></p> <p>Path to dependency file: box_box-ui-elements/package.json</p> <p>Path to vulnerable library: box_box-ui-elements/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - storybook-5.3.9.tgz (Root Library) - cli-5.3.9.tgz - pkg-add-deps-0.1.0.tgz - pacote-9.5.12.tgz - :x: **tar-4.4.13.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/box_box-ui-elements/commit/4fc776e2b95c8b497f6994cb2165365562ae1f82">4fc776e2b95c8b497f6994cb2165365562ae1f82</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 4.4.16, 5.0.8, and 6.1.7 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory, where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems. The cache checking logic used both `\` and `/` characters as path separators, however `\` is a valid filename character on posix systems. By first creating a directory, and then replacing that directory with a symlink, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. Additionally, a similar confusion could arise on case-insensitive filesystems. If a tar archive contained a directory at `FOO`, followed by a symbolic link named `foo`, then on case-insensitive file systems, the creation of the symbolic link would remove the directory from the filesystem, but _not_ from the internal directory cache, as it would not be treated as a cache hit. A subsequent file entry within the `FOO` directory would then be placed in the target of the symbolic link, thinking that the directory had already been created. These issues were addressed in releases 4.4.16, 5.0.8 and 6.1.7. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-9r2w-394v-53qc. <p>Publish Date: 2021-08-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37701>CVE-2021-37701</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc">https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc</a></p> <p>Release Date: 2021-08-31</p> <p>Fix Resolution: tar - 4.4.16, 5.0.8, 6.1.7</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"tar","packageVersion":"4.4.13","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"storybook:5.3.9;@storybook/cli:5.3.9;pkg-add-deps:0.1.0;pacote:9.5.12;tar:4.4.13","isMinimumFixVersionAvailable":true,"minimumFixVersion":"tar - 4.4.16, 5.0.8, 6.1.7"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-37701","vulnerabilityDetails":"The npm package \"tar\" (aka node-tar) before versions 4.4.16, 5.0.8, and 6.1.7 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory, where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems. The cache checking logic used both `\\` and `/` characters as path separators, however `\\` is a valid filename character on posix systems. By first creating a directory, and then replacing that directory with a symlink, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. Additionally, a similar confusion could arise on case-insensitive filesystems. If a tar archive contained a directory at `FOO`, followed by a symbolic link named `foo`, then on case-insensitive file systems, the creation of the symbolic link would remove the directory from the filesystem, but _not_ from the internal directory cache, as it would not be treated as a cache hit. A subsequent file entry within the `FOO` directory would then be placed in the target of the symbolic link, thinking that the directory had already been created. These issues were addressed in releases 4.4.16, 5.0.8 and 6.1.7. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-9r2w-394v-53qc.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37701","cvss3Severity":"high","cvss3Score":"8.6","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Changed","C":"High","UI":"Required","AV":"Local","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in tar tgz cve high severity vulnerability vulnerable library tar tgz tar for node library home page a href path to dependency file box box ui elements package json path to vulnerable library box box ui elements node modules tar package json dependency hierarchy storybook tgz root library cli tgz pkg add deps tgz pacote tgz x tar tgz vulnerable library found in head commit a href found in base branch master vulnerability details the npm package tar aka node tar before versions and has an arbitrary file creation overwrite and arbitrary code execution vulnerability node tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted this is in part achieved by ensuring that extracted directories are not symlinks additionally in order to prevent unnecessary stat calls to determine whether a given path is a directory paths are cached when directories are created this logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems the cache checking logic used both and characters as path separators however is a valid filename character on posix systems by first creating a directory and then replacing that directory with a symlink it was thus possible to bypass node tar symlink checks on directories essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location thus allowing arbitrary file creation and overwrite additionally a similar confusion could arise on case insensitive filesystems if a tar archive contained a directory at foo followed by a symbolic link named foo then on case insensitive file systems the creation of the symbolic link would remove the directory from the filesystem but not from the internal directory cache as it would not be treated as a cache hit a subsequent file entry within the foo directory would then be placed in the target of the symbolic link thinking that the directory had already been created these issues were addressed in releases and the branch of node tar has been deprecated and did not receive patches for these issues if you are still using a release we recommend you update to a more recent version of node tar if this is not possible a workaround is available in the referenced ghsa publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree storybook storybook cli pkg add deps pacote tar isminimumfixversionavailable true minimumfixversion tar basebranches vulnerabilityidentifier cve vulnerabilitydetails the npm package tar aka node tar before versions and has an arbitrary file creation overwrite and arbitrary code execution vulnerability node tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted this is in part achieved by ensuring that extracted directories are not symlinks additionally in order to prevent unnecessary stat calls to determine whether a given path is a directory paths are cached when directories are created this logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems the cache checking logic used both and characters as path separators however is a valid filename character on posix systems by first creating a directory and then replacing that directory with a symlink it was thus possible to bypass node tar symlink checks on directories essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location thus allowing arbitrary file creation and overwrite additionally a similar confusion could arise on case insensitive filesystems if a tar archive contained a directory at foo followed by a symbolic link named foo then on case insensitive file systems the creation of the symbolic link would remove the directory from the filesystem but not from the internal directory cache as it would not be treated as a cache hit a subsequent file entry within the foo directory would then be placed in the target of the symbolic link thinking that the directory had already been created these issues were addressed in releases and the branch of node tar has been deprecated and did not receive patches for these issues if you are still using a release we recommend you update to a more recent version of node tar if this is not possible a workaround is available in the referenced ghsa vulnerabilityurl
0
19,178
25,284,745,727
IssuesEvent
2022-11-16 18:19:51
OpenDataScotland/the_od_bods
https://api.github.com/repos/OpenDataScotland/the_od_bods
opened
Add Scottish Parliament data as a source
data processing back end new source
**Link to the data source** Data : https://data.parliament.scot/#/datasets API: https://data.parliament.scot/#/api-list A link to the front page and/or examples of webpages where you can find datasets. Please do not link directly to a file url e.g. https://data.parliament.scot/#/home **Data source type (if known)** Unknown **Organization(s) the data belongs to** Scottish Parliament **What licences are applied to the data being published? (if known)** It was using a version of OGL3 in 2020. "being released under "Scottish Parliament Licence is aligned with OGLv3.0" But the link from https://data.parliament.scot/#/about to the reuse conditions is now broken. **Additional information** We don't have a Scottish Parliament yet set up as an ODS Organisation
1.0
Add Scottish Parliament data as a source - **Link to the data source** Data : https://data.parliament.scot/#/datasets API: https://data.parliament.scot/#/api-list A link to the front page and/or examples of webpages where you can find datasets. Please do not link directly to a file url e.g. https://data.parliament.scot/#/home **Data source type (if known)** Unknown **Organization(s) the data belongs to** Scottish Parliament **What licences are applied to the data being published? (if known)** It was using a version of OGL3 in 2020. "being released under "Scottish Parliament Licence is aligned with OGLv3.0" But the link from https://data.parliament.scot/#/about to the reuse conditions is now broken. **Additional information** We don't have a Scottish Parliament yet set up as an ODS Organisation
process
add scottish parliament data as a source link to the data source data api a link to the front page and or examples of webpages where you can find datasets please do not link directly to a file url e g data source type if known unknown organization s the data belongs to scottish parliament what licences are applied to the data being published if known it was using a version of in being released under scottish parliament licence is aligned with but the link from to the reuse conditions is now broken additional information we don t have a scottish parliament yet set up as an ods organisation
1
4,985
7,816,997,481
IssuesEvent
2018-06-13 07:35:34
lbassin/okty
https://api.github.com/repos/lbassin/okty
closed
Containers with same ID
bug in process
**Describe the bug** There is nothing preventing user to save an existing container id **To Reproduce** Steps to reproduce the behavior: 1. Add new container with id 'bar' 2. Add new container with id 'foo' 3. Edit the first container 'bar' and set his id to 'foo' 4. Two containers exist with the same id but only one is exported **Expected behavior** Prevent the save action if id is already used
1.0
Containers with same ID - **Describe the bug** There is nothing preventing user to save an existing container id **To Reproduce** Steps to reproduce the behavior: 1. Add new container with id 'bar' 2. Add new container with id 'foo' 3. Edit the first container 'bar' and set his id to 'foo' 4. Two containers exist with the same id but only one is exported **Expected behavior** Prevent the save action if id is already used
process
containers with same id describe the bug there is nothing preventing user to save an existing container id to reproduce steps to reproduce the behavior add new container with id bar add new container with id foo edit the first container bar and set his id to foo two containers exist with the same id but only one is exported expected behavior prevent the save action if id is already used
1
10,983
13,783,312,939
IssuesEvent
2020-10-08 19:00:57
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
opened
Migrate from CircleCI to GitHub Actions
enhancement process test
**Problem** CircleCI does not support docker in docker. As a result, we have to manually spin up external resources like databases and configure the app to use them for tests instead of our built-in TestContainers. This makes the tests results brittle and different from local execution. **Solution** Migrate from CircleCI to GitHub Actions. Tasks can be moved one at a time: * build_maven: * Description: Runs maven package * on: push branch, push tag * Depends: none * Change: Move to GitHub Actions * build_rest: * Description: Runs npm test * on: push branch, push tag * Depends: none * Change: [Integrate REST API build into Maven · Issue #761 · hashgraph/hedera-mirror-node · GitHub](https://github.com/hashgraph/hedera-mirror-node/issues/761) * release_artifacts: * Description: Stores release artifacts in CircleCI * on: push tag * Depends: build_maven, build_rest, helm * Change: Move to GitHub Action, use actions/upload-artifact@v2. Later use release-drafter. Can remove if we no longer need jars. * publish_images: * Description: Publishes images to GCR using mvn deploy * on: push master, push tag * Depends: build_maven, build_rest, helm * Change: Move to GitHub Action, use [gcloud Container Registry Docker Login · Actions · GitHub Marketplace · GitHub](https://github.com/marketplace/actions/gcloud-container-registry-docker-login) * perf_maven: * Description: Runs performance tests via maven * on: push branch, push tag * Depends: nothing * Changes: Keep in CircleCI for now. Move to nightly GH job in future * cleanup_images: * Description: Cleans old images on GCR * on: push master * Depends: publish_images * Changes: Keep in CircleCI for now. Move to nightly GH job in future * write_spring_override_config: Not needed if we use test containers **Alternatives** **Additional Context**
1.0
Migrate from CircleCI to GitHub Actions - **Problem** CircleCI does not support docker in docker. As a result, we have to manually spin up external resources like databases and configure the app to use them for tests instead of our built-in TestContainers. This makes the tests results brittle and different from local execution. **Solution** Migrate from CircleCI to GitHub Actions. Tasks can be moved one at a time: * build_maven: * Description: Runs maven package * on: push branch, push tag * Depends: none * Change: Move to GitHub Actions * build_rest: * Description: Runs npm test * on: push branch, push tag * Depends: none * Change: [Integrate REST API build into Maven · Issue #761 · hashgraph/hedera-mirror-node · GitHub](https://github.com/hashgraph/hedera-mirror-node/issues/761) * release_artifacts: * Description: Stores release artifacts in CircleCI * on: push tag * Depends: build_maven, build_rest, helm * Change: Move to GitHub Action, use actions/upload-artifact@v2. Later use release-drafter. Can remove if we no longer need jars. * publish_images: * Description: Publishes images to GCR using mvn deploy * on: push master, push tag * Depends: build_maven, build_rest, helm * Change: Move to GitHub Action, use [gcloud Container Registry Docker Login · Actions · GitHub Marketplace · GitHub](https://github.com/marketplace/actions/gcloud-container-registry-docker-login) * perf_maven: * Description: Runs performance tests via maven * on: push branch, push tag * Depends: nothing * Changes: Keep in CircleCI for now. Move to nightly GH job in future * cleanup_images: * Description: Cleans old images on GCR * on: push master * Depends: publish_images * Changes: Keep in CircleCI for now. Move to nightly GH job in future * write_spring_override_config: Not needed if we use test containers **Alternatives** **Additional Context**
process
migrate from circleci to github actions problem circleci does not support docker in docker as a result we have to manually spin up external resources like databases and configure the app to use them for tests instead of our built in testcontainers this makes the tests results brittle and different from local execution solution migrate from circleci to github actions tasks can be moved one at a time build maven description runs maven package on push branch push tag depends none change move to github actions build rest description runs npm test on push branch push tag depends none change release artifacts description stores release artifacts in circleci on push tag depends build maven build rest helm change move to github action use actions upload artifact later use release drafter can remove if we no longer need jars publish images description publishes images to gcr using mvn deploy on push master push tag depends build maven build rest helm change move to github action use perf maven description runs performance tests via maven on push branch push tag depends nothing changes keep in circleci for now move to nightly gh job in future cleanup images description cleans old images on gcr on push master depends publish images changes keep in circleci for now move to nightly gh job in future write spring override config not needed if we use test containers alternatives additional context
1
19,805
26,188,048,933
IssuesEvent
2023-01-03 04:59:03
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
AssignSystemIdentity option in the Set-AzAutomationAccount does not exists
automation/svc triaged cxp process-automation/subsvc Pri1 docs-bug
[Enter feedback here] AssignSystemIdentity option in the Set-AzAutomationAccount does not work and this open does not exists in the Microsoft doc as well: https://learn.microsoft.com/en-us/powershell/module/az.automation/set-azautomationaccount?view=azps-9.2.0 There is a documentation bug in this page --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: d9ce2612-b600-3fca-3315-a7836ef91c96 * Version Independent ID: 78766eed-c3c6-ce60-7620-17b99f3d9d5e * Content: [Using a system-assigned managed identity for an Azure Automation account](https://learn.microsoft.com/en-us/azure/automation/enable-managed-identity-for-automation) * Content Source: [articles/automation/enable-managed-identity-for-automation.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/enable-managed-identity-for-automation.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @SnehaSudhirG * Microsoft Alias: **sudhirsneha**
1.0
AssignSystemIdentity option in the Set-AzAutomationAccount does not exists - [Enter feedback here] AssignSystemIdentity option in the Set-AzAutomationAccount does not work and this open does not exists in the Microsoft doc as well: https://learn.microsoft.com/en-us/powershell/module/az.automation/set-azautomationaccount?view=azps-9.2.0 There is a documentation bug in this page --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: d9ce2612-b600-3fca-3315-a7836ef91c96 * Version Independent ID: 78766eed-c3c6-ce60-7620-17b99f3d9d5e * Content: [Using a system-assigned managed identity for an Azure Automation account](https://learn.microsoft.com/en-us/azure/automation/enable-managed-identity-for-automation) * Content Source: [articles/automation/enable-managed-identity-for-automation.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/enable-managed-identity-for-automation.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @SnehaSudhirG * Microsoft Alias: **sudhirsneha**
process
assignsystemidentity option in the set azautomationaccount does not exists assignsystemidentity option in the set azautomationaccount does not work and this open does not exists in the microsoft doc as well there is a documentation bug in this page document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login snehasudhirg microsoft alias sudhirsneha
1
117,461
4,716,556,441
IssuesEvent
2016-10-16 04:16:35
CS2103AUG2016-W11-C3/main
https://api.github.com/repos/CS2103AUG2016-W11-C3/main
opened
User Story: Add a floating task
priority.high type.story
Description: User add a task without a deadline Solution/Feature: add Command
1.0
User Story: Add a floating task - Description: User add a task without a deadline Solution/Feature: add Command
non_process
user story add a floating task description user add a task without a deadline solution feature add command
0
800
3,277,610,107
IssuesEvent
2015-10-27 02:03:59
meteor/meteor
https://api.github.com/repos/meteor/meteor
closed
Release 1.2.1 - Release candidate out!
in progress Project:Release Process
This is going to be a relatively small release with no huge headline features, just some nice improvements! - Faster collection inserts on the server - New `ecmascript-runtime` package to replace `ecmascript-collections`, which provides more ES2015 polyfills! - Support for named pipes as a port in production mode, needed for IIS on Windows - See more here: https://github.com/meteor/meteor/blob/release-1.2.1/History.md#v121- Release branch is [`release-1.2.1`](https://github.com/meteor/meteor/tree/release-1.2.1). Tasks are being tracked in the [Release 1.2.1 milestone](https://github.com/meteor/meteor/milestones/Release%201.2.1). ### Release candidate Try the RC with `meteor --release 1.2.1-rc.4`! We haven't done QA yet, but want to get this into people's hands ASAP. ### Todos - [x] Update History.md - [x] Establish what QA needs to be done - [x] Add issues for documentation - [x] Finish all tasks in 1.2.1 milestone - [x] Add https://github.com/meteor/meteor/commit/c9297037900ecd43f8c267023a8736230506c6e7 to History.md - @glasser ### QA to be done - [x] Run package tests in all browsers to check new ES2015 polyfills - in particular, IE8/9. - [x] Do Auth QA with Google to check that the loginHint option didn't break anything
1.0
Release 1.2.1 - Release candidate out! - This is going to be a relatively small release with no huge headline features, just some nice improvements! - Faster collection inserts on the server - New `ecmascript-runtime` package to replace `ecmascript-collections`, which provides more ES2015 polyfills! - Support for named pipes as a port in production mode, needed for IIS on Windows - See more here: https://github.com/meteor/meteor/blob/release-1.2.1/History.md#v121- Release branch is [`release-1.2.1`](https://github.com/meteor/meteor/tree/release-1.2.1). Tasks are being tracked in the [Release 1.2.1 milestone](https://github.com/meteor/meteor/milestones/Release%201.2.1). ### Release candidate Try the RC with `meteor --release 1.2.1-rc.4`! We haven't done QA yet, but want to get this into people's hands ASAP. ### Todos - [x] Update History.md - [x] Establish what QA needs to be done - [x] Add issues for documentation - [x] Finish all tasks in 1.2.1 milestone - [x] Add https://github.com/meteor/meteor/commit/c9297037900ecd43f8c267023a8736230506c6e7 to History.md - @glasser ### QA to be done - [x] Run package tests in all browsers to check new ES2015 polyfills - in particular, IE8/9. - [x] Do Auth QA with Google to check that the loginHint option didn't break anything
process
release release candidate out this is going to be a relatively small release with no huge headline features just some nice improvements faster collection inserts on the server new ecmascript runtime package to replace ecmascript collections which provides more polyfills support for named pipes as a port in production mode needed for iis on windows see more here release branch is tasks are being tracked in the release candidate try the rc with meteor release rc we haven t done qa yet but want to get this into people s hands asap todos update history md establish what qa needs to be done add issues for documentation finish all tasks in milestone add to history md glasser qa to be done run package tests in all browsers to check new polyfills in particular do auth qa with google to check that the loginhint option didn t break anything
1
165,148
12,830,055,980
IssuesEvent
2020-07-07 00:56:47
LiVinson/taboo-game
https://api.github.com/repos/LiVinson/taboo-game
closed
Waiting Room UI & Unit Tests
Styling Tests
Waiting Room components: - Gamecode card (Taboo Card) - Players Card: Separated by Unassigned, Team 1 and Team 2 - 2 buttons: regular, primary, and large regular
1.0
Waiting Room UI & Unit Tests - Waiting Room components: - Gamecode card (Taboo Card) - Players Card: Separated by Unassigned, Team 1 and Team 2 - 2 buttons: regular, primary, and large regular
non_process
waiting room ui unit tests waiting room components gamecode card taboo card players card separated by unassigned team and team buttons regular primary and large regular
0
106,420
16,673,361,323
IssuesEvent
2021-06-07 13:36:09
VivekBuzruk/Hygieia
https://api.github.com/repos/VivekBuzruk/Hygieia
closed
CVE-2013-7285 (High) detected in xstream-1.4.9.jar - autoclosed
security vulnerability
## CVE-2013-7285 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.9.jar</b></p></summary> <p>XStream is a serialization library from Java objects to XML and back.</p> <p>Library home page: <a href="http://x-stream.github.io">http://x-stream.github.io</a></p> <p>Path to dependency file: Hygieia/UI-tests/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.9/xstream-1.4.9.jar</p> <p> Dependency Hierarchy: - serenity-core-1.2.2.jar (Root Library) - :x: **xstream-1.4.9.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/VivekBuzruk/Hygieia/commits/3c4f119e4343cf7fa276bb4756361b926902248e">3c4f119e4343cf7fa276bb4756361b926902248e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Xstream API versions up to 1.4.6 and version 1.4.10, if the security framework has not been initialized, may allow a remote attacker to run arbitrary shell commands by manipulating the processed input stream when unmarshaling XML or any supported format. e.g. JSON. <p>Publish Date: 2019-05-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2013-7285>CVE-2013-7285</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2013-7285">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2013-7285</a></p> <p>Release Date: 2019-05-15</p> <p>Fix Resolution: 1.4.7,1.4.11</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2013-7285 (High) detected in xstream-1.4.9.jar - autoclosed - ## CVE-2013-7285 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.9.jar</b></p></summary> <p>XStream is a serialization library from Java objects to XML and back.</p> <p>Library home page: <a href="http://x-stream.github.io">http://x-stream.github.io</a></p> <p>Path to dependency file: Hygieia/UI-tests/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.9/xstream-1.4.9.jar</p> <p> Dependency Hierarchy: - serenity-core-1.2.2.jar (Root Library) - :x: **xstream-1.4.9.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/VivekBuzruk/Hygieia/commits/3c4f119e4343cf7fa276bb4756361b926902248e">3c4f119e4343cf7fa276bb4756361b926902248e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Xstream API versions up to 1.4.6 and version 1.4.10, if the security framework has not been initialized, may allow a remote attacker to run arbitrary shell commands by manipulating the processed input stream when unmarshaling XML or any supported format. e.g. JSON. <p>Publish Date: 2019-05-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2013-7285>CVE-2013-7285</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2013-7285">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2013-7285</a></p> <p>Release Date: 2019-05-15</p> <p>Fix Resolution: 1.4.7,1.4.11</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in xstream jar autoclosed cve high severity vulnerability vulnerable library xstream jar xstream is a serialization library from java objects to xml and back library home page a href path to dependency file hygieia ui tests pom xml path to vulnerable library home wss scanner repository com thoughtworks xstream xstream xstream jar dependency hierarchy serenity core jar root library x xstream jar vulnerable library found in head commit a href found in base branch master vulnerability details xstream api versions up to and version if the security framework has not been initialized may allow a remote attacker to run arbitrary shell commands by manipulating the processed input stream when unmarshaling xml or any supported format e g json publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
14,493
17,604,277,308
IssuesEvent
2021-08-17 15:12:35
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
[FEATURE][processing] Add "convert to curves" algorithm
Automatic new feature Processing Alg 3.14
Original commit: https://github.com/qgis/QGIS/commit/9939142ba9000f0526ae7196eec9e4f8e52ee77a by olivierdalang Converts a linear geometry type to the corresponding curved geometry type, attemping to identify segments in the original geometries which can be replaced by arcs. The distance tolerance parameter specifies the maximum distance allowed between the original location of vertices and where they would fall on the converted curved geometries. This algorithm only consider a segments as suitable for replacing with an arc if the points are all regularly spaced on the candidate arc. The angle tolerance parameter specifies the maximum angular deviation (in degrees) allowed when testing for regular point spacing. Already curved geometries will be retained without change.
1.0
[FEATURE][processing] Add "convert to curves" algorithm - Original commit: https://github.com/qgis/QGIS/commit/9939142ba9000f0526ae7196eec9e4f8e52ee77a by olivierdalang Converts a linear geometry type to the corresponding curved geometry type, attemping to identify segments in the original geometries which can be replaced by arcs. The distance tolerance parameter specifies the maximum distance allowed between the original location of vertices and where they would fall on the converted curved geometries. This algorithm only consider a segments as suitable for replacing with an arc if the points are all regularly spaced on the candidate arc. The angle tolerance parameter specifies the maximum angular deviation (in degrees) allowed when testing for regular point spacing. Already curved geometries will be retained without change.
process
add convert to curves algorithm original commit by olivierdalang converts a linear geometry type to the corresponding curved geometry type attemping to identify segments in the original geometries which can be replaced by arcs the distance tolerance parameter specifies the maximum distance allowed between the original location of vertices and where they would fall on the converted curved geometries this algorithm only consider a segments as suitable for replacing with an arc if the points are all regularly spaced on the candidate arc the angle tolerance parameter specifies the maximum angular deviation in degrees allowed when testing for regular point spacing already curved geometries will be retained without change
1
407,258
27,608,265,585
IssuesEvent
2023-03-09 14:24:01
hashicorp/terraform-provider-null
https://api.github.com/repos/hashicorp/terraform-provider-null
closed
Documentation examples and splat syntax for null_resource
documentation
Hello, just writing to mention the examples for null_resource in documentation should be updated to reflect newer splat syntax of `[*].` instead of `.*.`. https://www.terraform.io/docs/providers/null/resource.html https://www.terraform.io/docs/configuration/expressions.html#splat-expressions
1.0
Documentation examples and splat syntax for null_resource - Hello, just writing to mention the examples for null_resource in documentation should be updated to reflect newer splat syntax of `[*].` instead of `.*.`. https://www.terraform.io/docs/providers/null/resource.html https://www.terraform.io/docs/configuration/expressions.html#splat-expressions
non_process
documentation examples and splat syntax for null resource hello just writing to mention the examples for null resource in documentation should be updated to reflect newer splat syntax of instead of
0
195,237
6,905,711,923
IssuesEvent
2017-11-27 08:31:27
NHMD/svampeatlas
https://api.github.com/repos/NHMD/svampeatlas
closed
Overordnede statistikker/metrics
Priority: High
Ønsker fra TL: visning af brugere med flest records, flest arter etc i tabel. Kræver jobs der kører om natten til beregning. Kunne også være "Hvor mange arter findes der i DK?" osv. Man kunne også bruge gbifs kortservices til at vise data pr år: http://api.gbif.org/v1/map/density/tile?x={x}&y={y}&z={z}&type=DATASET&key=84d26682-f762-11e1-a439-00145eb45e9a&layer=OBS_NO_YEAR&layer=SP_NO_YEAR&layer=OTH_NO_YEAR&layer=OBS_1900_1910&layer=SP_1900_1910&layer=OTH_1900_1910&layer=OBS_1910_1920&layer=SP_1910_1920&layer=OTH_1910_1920&layer=OBS_1920_1930&layer=SP_1920_1930&layer=OTH_1920_1930&layer=OBS_1930_1940&layer=SP_1930_1940&layer=OTH_1930_1940&layer=OBS_1940_1950&layer=SP_1940_1950&layer=OTH_1940_1950&layer=OBS_1950_1960&layer=SP_1950_1960&layer=OTH_1950_1960&layer=OBS_1960_1970&layer=SP_1960_1970&layer=OTH_1960_1970&layer=OBS_1970_1980&layer=SP_1970_1980&layer=OTH_1970_1980&layer=OBS_1980_1990&layer=SP_1980_1990&layer=OTH_1980_1990&layer=OBS_1990_2000&layer=SP_1990_2000&layer=OTH_1990_2000&layer=OBS_2010_2020&layer=SP_2010_2020&layer=OTH_2010_2020&layer=LIVING&layer=FOSSIL&palette=yellows_reds Desuden kan der laves en nye godkendte obs på ugentligt plan ved at bruge GBIFs processeringsstatistikker: http://api.gbif.org/v1/dataset/84d26682-f762-11e1-a439-00145eb45e9a/process?limit=52
1.0
Overordnede statistikker/metrics - Ønsker fra TL: visning af brugere med flest records, flest arter etc i tabel. Kræver jobs der kører om natten til beregning. Kunne også være "Hvor mange arter findes der i DK?" osv. Man kunne også bruge gbifs kortservices til at vise data pr år: http://api.gbif.org/v1/map/density/tile?x={x}&y={y}&z={z}&type=DATASET&key=84d26682-f762-11e1-a439-00145eb45e9a&layer=OBS_NO_YEAR&layer=SP_NO_YEAR&layer=OTH_NO_YEAR&layer=OBS_1900_1910&layer=SP_1900_1910&layer=OTH_1900_1910&layer=OBS_1910_1920&layer=SP_1910_1920&layer=OTH_1910_1920&layer=OBS_1920_1930&layer=SP_1920_1930&layer=OTH_1920_1930&layer=OBS_1930_1940&layer=SP_1930_1940&layer=OTH_1930_1940&layer=OBS_1940_1950&layer=SP_1940_1950&layer=OTH_1940_1950&layer=OBS_1950_1960&layer=SP_1950_1960&layer=OTH_1950_1960&layer=OBS_1960_1970&layer=SP_1960_1970&layer=OTH_1960_1970&layer=OBS_1970_1980&layer=SP_1970_1980&layer=OTH_1970_1980&layer=OBS_1980_1990&layer=SP_1980_1990&layer=OTH_1980_1990&layer=OBS_1990_2000&layer=SP_1990_2000&layer=OTH_1990_2000&layer=OBS_2010_2020&layer=SP_2010_2020&layer=OTH_2010_2020&layer=LIVING&layer=FOSSIL&palette=yellows_reds Desuden kan der laves en nye godkendte obs på ugentligt plan ved at bruge GBIFs processeringsstatistikker: http://api.gbif.org/v1/dataset/84d26682-f762-11e1-a439-00145eb45e9a/process?limit=52
non_process
overordnede statistikker metrics ønsker fra tl visning af brugere med flest records flest arter etc i tabel kræver jobs der kører om natten til beregning kunne også være hvor mange arter findes der i dk osv man kunne også bruge gbifs kortservices til at vise data pr år desuden kan der laves en nye godkendte obs på ugentligt plan ved at bruge gbifs processeringsstatistikker
0
13,727
3,184,883,268
IssuesEvent
2015-09-27 19:42:31
Glucosio/android
https://api.github.com/repos/Glucosio/android
reopened
Terms of Use.
blocker bug copy design
In Hello we should remove the period behind Terms of Use and maybe consider a Glucosio pink line underneath it. Lets have @elioqoshi give some design feedback one last time and look at the app for any slight styling advice he might have.
1.0
Terms of Use. - In Hello we should remove the period behind Terms of Use and maybe consider a Glucosio pink line underneath it. Lets have @elioqoshi give some design feedback one last time and look at the app for any slight styling advice he might have.
non_process
terms of use in hello we should remove the period behind terms of use and maybe consider a glucosio pink line underneath it lets have elioqoshi give some design feedback one last time and look at the app for any slight styling advice he might have
0
6,937
10,102,519,865
IssuesEvent
2019-07-29 11:25:12
linnovate/root
https://api.github.com/repos/linnovate/root
opened
task from meeting still appears in list after meeting change
Process bug Tasks
create a task assign it a meeting change the meeting to a different one go to tasks from meeting filter and click on both the meetings the task shows up in both of them ![image](https://user-images.githubusercontent.com/38312178/62044885-a2ab5580-b20c-11e9-9f65-01377aa0592d.png)
1.0
task from meeting still appears in list after meeting change - create a task assign it a meeting change the meeting to a different one go to tasks from meeting filter and click on both the meetings the task shows up in both of them ![image](https://user-images.githubusercontent.com/38312178/62044885-a2ab5580-b20c-11e9-9f65-01377aa0592d.png)
process
task from meeting still appears in list after meeting change create a task assign it a meeting change the meeting to a different one go to tasks from meeting filter and click on both the meetings the task shows up in both of them
1
9,163
12,219,338,954
IssuesEvent
2020-05-01 21:28:07
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
closed
Auto Clean up old master images in GCR
P2 enhancement process
**Problem** As part of CI we will push images to GCR on master merges. This will create a build up of images that we'll want to remove. **Solution** Update circle ci config to remove images older than 7 days **Additional Context** Current logic exists to do this but we saw discrepancies in the time filter at the precision of a few hours. . In some tests when filtering on images older than a few hours ago the console would return an expected subset, upon running the delete command all images with that tag were removed. Seems to not occur when accuracy of timing is in days and not hours. For now logic is left as an echo to see intentions but not actually remove images.
1.0
Auto Clean up old master images in GCR - **Problem** As part of CI we will push images to GCR on master merges. This will create a build up of images that we'll want to remove. **Solution** Update circle ci config to remove images older than 7 days **Additional Context** Current logic exists to do this but we saw discrepancies in the time filter at the precision of a few hours. . In some tests when filtering on images older than a few hours ago the console would return an expected subset, upon running the delete command all images with that tag were removed. Seems to not occur when accuracy of timing is in days and not hours. For now logic is left as an echo to see intentions but not actually remove images.
process
auto clean up old master images in gcr problem as part of ci we will push images to gcr on master merges this will create a build up of images that we ll want to remove solution update circle ci config to remove images older than days additional context current logic exists to do this but we saw discrepancies in the time filter at the precision of a few hours in some tests when filtering on images older than a few hours ago the console would return an expected subset upon running the delete command all images with that tag were removed seems to not occur when accuracy of timing is in days and not hours for now logic is left as an echo to see intentions but not actually remove images
1
17,238
6,394,860,283
IssuesEvent
2017-08-04 11:34:32
yahoo/fili
https://api.github.com/repos/yahoo/fili
closed
Maven build warings: duplicate dependency declaration
BUILD
> [WARNING] > [WARNING] Some problems were encountered while building the effective model for com.yahoo.fili:fili-security:jar:0.9-SNAPSHOT > [WARNING] 'dependencies.dependency.(groupId:artifactId:type:classifier)' must be unique: com.yahoo.fili:fili-core:jar -> duplicate declaration of version (?) @ com.yahoo.fili:fili-se > curity:[unknown-version], C:\Work\IDEA-17\other\fili\fili-security\pom.xml, line 45, column 21 > [WARNING]
1.0
Maven build warings: duplicate dependency declaration - > [WARNING] > [WARNING] Some problems were encountered while building the effective model for com.yahoo.fili:fili-security:jar:0.9-SNAPSHOT > [WARNING] 'dependencies.dependency.(groupId:artifactId:type:classifier)' must be unique: com.yahoo.fili:fili-core:jar -> duplicate declaration of version (?) @ com.yahoo.fili:fili-se > curity:[unknown-version], C:\Work\IDEA-17\other\fili\fili-security\pom.xml, line 45, column 21 > [WARNING]
non_process
maven build warings duplicate dependency declaration some problems were encountered while building the effective model for com yahoo fili fili security jar snapshot dependencies dependency groupid artifactid type classifier must be unique com yahoo fili fili core jar duplicate declaration of version com yahoo fili fili se curity c work idea other fili fili security pom xml line column
0
139,899
5,394,146,695
IssuesEvent
2017-02-27 01:25:10
elementary/icons
https://api.github.com/repos/elementary/icons
closed
actions/document-open-recent isn't consistent across sizes
Priority: Low Status: Confirmed
32px and up is a symlink to folder-recent ## <bountysource-plugin> Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/29254200-actions-document-open-recent-isn-t-consistent-across-sizes?utm_campaign=plugin&utm_content=tracker%2F27377189&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F27377189&utm_medium=issues&utm_source=github). </bountysource-plugin>
1.0
actions/document-open-recent isn't consistent across sizes - 32px and up is a symlink to folder-recent ## <bountysource-plugin> Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/29254200-actions-document-open-recent-isn-t-consistent-across-sizes?utm_campaign=plugin&utm_content=tracker%2F27377189&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F27377189&utm_medium=issues&utm_source=github). </bountysource-plugin>
non_process
actions document open recent isn t consistent across sizes and up is a symlink to folder recent want to back this issue we accept bounties via
0
13,851
16,613,436,028
IssuesEvent
2021-06-02 14:08:12
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
Add command to release process to ensure release binary is downloadable
process: release stage: needs review type: chore
Add a command for the release process that tests the binary url for the newly released version of Cypress that ensures it is available and downloadable on each operating system.
1.0
Add command to release process to ensure release binary is downloadable - Add a command for the release process that tests the binary url for the newly released version of Cypress that ensures it is available and downloadable on each operating system.
process
add command to release process to ensure release binary is downloadable add a command for the release process that tests the binary url for the newly released version of cypress that ensures it is available and downloadable on each operating system
1
14
2,496,238,503
IssuesEvent
2015-01-06 18:03:06
vivo-isf/vivo-isf-ontology
https://api.github.com/repos/vivo-isf/vivo-isf-ontology
closed
menopause
biological_process imported
_From [vasil...@ohsu.edu](https://code.google.com/u/108803237899917466626/) on November 12, 2012 17:39:28_ GO:0042697 parent: reproductive process \<a href="http://purl.obolibrary.org/obo/GO_0022414" rel="nofollow">http://purl.obolibrary.org/obo/GO_0022414</a>&#13; &#13; \<a href="http://ohsu.eagle-i.net/i/0000012b-00d1-5c5b-79a3-373680000000" rel="nofollow">http://ohsu.eagle-i.net/i/0000012b-00d1-5c5b-79a3-373680000000</a> _Original issue: http://code.google.com/p/eagle-i/issues/detail?id=165_
1.0
menopause - _From [vasil...@ohsu.edu](https://code.google.com/u/108803237899917466626/) on November 12, 2012 17:39:28_ GO:0042697 parent: reproductive process \<a href="http://purl.obolibrary.org/obo/GO_0022414" rel="nofollow">http://purl.obolibrary.org/obo/GO_0022414</a>&#13; &#13; \<a href="http://ohsu.eagle-i.net/i/0000012b-00d1-5c5b-79a3-373680000000" rel="nofollow">http://ohsu.eagle-i.net/i/0000012b-00d1-5c5b-79a3-373680000000</a> _Original issue: http://code.google.com/p/eagle-i/issues/detail?id=165_
process
menopause from on november go parent reproductive process original issue
1
7,896
11,083,387,005
IssuesEvent
2019-12-13 14:22:21
Open-EO/openeo-processes
https://api.github.com/repos/Open-EO/openeo-processes
closed
Load assets (e.g. GeoJSON) from user workspace
new process
In the old version is was allowed to pass files to zonal_statistics (now aggregate_polygon) and other processes. This is mostly not allowed any longer, but you can pass a vector data cube or a GeoJSON. So we probably need a process to load files (e.g. GeoJSON files) from the user workspace.
1.0
Load assets (e.g. GeoJSON) from user workspace - In the old version is was allowed to pass files to zonal_statistics (now aggregate_polygon) and other processes. This is mostly not allowed any longer, but you can pass a vector data cube or a GeoJSON. So we probably need a process to load files (e.g. GeoJSON files) from the user workspace.
process
load assets e g geojson from user workspace in the old version is was allowed to pass files to zonal statistics now aggregate polygon and other processes this is mostly not allowed any longer but you can pass a vector data cube or a geojson so we probably need a process to load files e g geojson files from the user workspace
1
14,491
17,604,190,959
IssuesEvent
2021-08-17 15:07:22
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
Request in QGIS ([needs-docs] Show 'template layer' constraints info in the Refactor fields algorithm's UI)
Processing Alg 3.14 GUI
### Request for documentation From pull request QGIS/qgis#31387 Author: @gacarrillor QGIS version: 3.14 **[needs-docs] Show 'template layer' constraints info in the Refactor fields algorithm's UI** ### PR Description: ## Description Dialog: Refactor fields algorithm's UI. Objective: Highlight field rows when the 'template layer' field has constraints. **TODO** + [x] Add an orange color (same as used in attribute form warnings) to the whole row if a template layer field has constraints. + [x] Adjust text in help panel to let users know why some rows are coloured. + [x] Remove NOT NULL column. + [x] New column "Template properties." + Which content to show? + "Constraints active"? + What the tooltip should show? + Constraints, widget configuration, ...? + [x] Replace "Load fields from layer" by "Load fields from template layer." **Screenshots** ![image](https://user-images.githubusercontent.com/652785/75100644-6079a500-559e-11ea-91b6-f99fa05451de.png) Fix #34458 ## Checklist <!-- Reviewing is a process done by project maintainers, mostly on a volunteer basis. We try to keep the overhead as small as possible and appreciate if you help us to do so by completing the following items. Feel free to ask in a comment if you have troubles with any of them. --> - [x] Commit messages are descriptive and explain the rationale for changes - [x] I have read the [QGIS Coding Standards](https://docs.qgis.org/testing/en/docs/developers_guide/codingstandards.html) and this PR complies with them - [x] I have run [the `scripts/prepare-commit.sh` script](https://github.com/qgis/QGIS/blob/master/.github/CONTRIBUTING.md#contributing-to-qgis) before each commit - [x] I have evaluated whether it is appropriate for this PR to be backported, backport requests are left as label or comment. ### Commits tagged with [need-docs] or [FEATURE]
1.0
Request in QGIS ([needs-docs] Show 'template layer' constraints info in the Refactor fields algorithm's UI) - ### Request for documentation From pull request QGIS/qgis#31387 Author: @gacarrillor QGIS version: 3.14 **[needs-docs] Show 'template layer' constraints info in the Refactor fields algorithm's UI** ### PR Description: ## Description Dialog: Refactor fields algorithm's UI. Objective: Highlight field rows when the 'template layer' field has constraints. **TODO** + [x] Add an orange color (same as used in attribute form warnings) to the whole row if a template layer field has constraints. + [x] Adjust text in help panel to let users know why some rows are coloured. + [x] Remove NOT NULL column. + [x] New column "Template properties." + Which content to show? + "Constraints active"? + What the tooltip should show? + Constraints, widget configuration, ...? + [x] Replace "Load fields from layer" by "Load fields from template layer." **Screenshots** ![image](https://user-images.githubusercontent.com/652785/75100644-6079a500-559e-11ea-91b6-f99fa05451de.png) Fix #34458 ## Checklist <!-- Reviewing is a process done by project maintainers, mostly on a volunteer basis. We try to keep the overhead as small as possible and appreciate if you help us to do so by completing the following items. Feel free to ask in a comment if you have troubles with any of them. --> - [x] Commit messages are descriptive and explain the rationale for changes - [x] I have read the [QGIS Coding Standards](https://docs.qgis.org/testing/en/docs/developers_guide/codingstandards.html) and this PR complies with them - [x] I have run [the `scripts/prepare-commit.sh` script](https://github.com/qgis/QGIS/blob/master/.github/CONTRIBUTING.md#contributing-to-qgis) before each commit - [x] I have evaluated whether it is appropriate for this PR to be backported, backport requests are left as label or comment. ### Commits tagged with [need-docs] or [FEATURE]
process
request in qgis show template layer constraints info in the refactor fields algorithm s ui request for documentation from pull request qgis qgis author gacarrillor qgis version show template layer constraints info in the refactor fields algorithm s ui pr description description dialog refactor fields algorithm s ui objective highlight field rows when the template layer field has constraints todo add an orange color same as used in attribute form warnings to the whole row if a template layer field has constraints adjust text in help panel to let users know why some rows are coloured remove not null column new column template properties which content to show constraints active what the tooltip should show constraints widget configuration replace load fields from layer by load fields from template layer screenshots fix checklist reviewing is a process done by project maintainers mostly on a volunteer basis we try to keep the overhead as small as possible and appreciate if you help us to do so by completing the following items feel free to ask in a comment if you have troubles with any of them commit messages are descriptive and explain the rationale for changes i have read the and this pr complies with them i have run before each commit i have evaluated whether it is appropriate for this pr to be backported backport requests are left as label or comment commits tagged with or
1
5,247
3,917,001,688
IssuesEvent
2016-04-21 05:57:07
kolliSuman/issues
https://api.github.com/repos/kolliSuman/issues
closed
QA_Affine Transformation_Usability_smk
Category: Usability Developed By: VLEAD Release Number: Production Severity: S2 Status: Open
Defect Description : In the landing page/home page of "Affine Transformation" experiment, the name of the experiment is not present in the screen instead the experiment name should be displayed on the screen at the top of the header options Actual Result : In the landing page/home page of "Affine Transformation" experiment, the name of the experiment is not present in the screen Environment : OS: Windows 7, Ubuntu-16.04,Centos-6 Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0 Bandwidth : 100Mbps Hardware Configuration:8GBRAM , Processor:i5 Test Step Link: https://github.com/Virtual-Labs/image-processing-iiith/blob/master/test-cases/integration_test-cases/Affine%20Transformation/Affine%20Transformation_01_Usability_smk.org
True
QA_Affine Transformation_Usability_smk - Defect Description : In the landing page/home page of "Affine Transformation" experiment, the name of the experiment is not present in the screen instead the experiment name should be displayed on the screen at the top of the header options Actual Result : In the landing page/home page of "Affine Transformation" experiment, the name of the experiment is not present in the screen Environment : OS: Windows 7, Ubuntu-16.04,Centos-6 Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0 Bandwidth : 100Mbps Hardware Configuration:8GBRAM , Processor:i5 Test Step Link: https://github.com/Virtual-Labs/image-processing-iiith/blob/master/test-cases/integration_test-cases/Affine%20Transformation/Affine%20Transformation_01_Usability_smk.org
non_process
qa affine transformation usability smk defect description in the landing page home page of affine transformation experiment the name of the experiment is not present in the screen instead the experiment name should be displayed on the screen at the top of the header options actual result in the landing page home page of affine transformation experiment the name of the experiment is not present in the screen environment os windows ubuntu centos browsers firefox chrome chromium bandwidth hardware configuration processor test step link
0
3,703
6,730,313,826
IssuesEvent
2017-10-18 00:05:13
coala/corobo
https://api.github.com/repos/coala/corobo
closed
Fix labelling anamoly
process/pending review process/wip
There are some PRs in review queue which are having both labels - process/wip as well as process/pending review. IMO corobo should use only latest label and remove the older one. Opened by @prnvdixit at [gitter/coala/coala](https://gitter.im/coala/coala?at=59e668fe614889d475c8f18f)
2.0
Fix labelling anamoly - There are some PRs in review queue which are having both labels - process/wip as well as process/pending review. IMO corobo should use only latest label and remove the older one. Opened by @prnvdixit at [gitter/coala/coala](https://gitter.im/coala/coala?at=59e668fe614889d475c8f18f)
process
fix labelling anamoly there are some prs in review queue which are having both labels process wip as well as process pending review imo corobo should use only latest label and remove the older one opened by prnvdixit at
1
705,437
24,234,769,234
IssuesEvent
2022-09-26 21:48:31
certbot/certbot
https://api.github.com/repos/certbot/certbot
closed
Certbot-dns-cloudflare 1.30.0 doesn't work with cloudflare 2.10.1
easy-fix has pr area: dns priority: significant
If you're having trouble using Certbot and aren't sure you've found a bug or request for a new feature, please first try asking for help at https://community.letsencrypt.org/. There is a much larger community there of people familiar with the project who will be able to more quickly answer your questions. ## My operating system is (include version): Gentoo Linux ## I installed Certbot with (snap, OS package manager, pip, certbot-auto, etc): package manager ## I ran this command and it produced this output: certbot renew --dry-run -v Saving debug log to /var/log/letsencrypt/letsencrypt.log - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - Processing /etc/letsencrypt/renewal/phoenix591.com.conf - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - Certificate not due for renewal, but simulating renewal for dry run Plugins selected: Authenticator dns-cloudflare, Installer None Simulating renewal of an existing certificate for *.phoenix591.com and phoenix591.com Performing the following challenges: dns-01 challenge for phoenix591.com dns-01 challenge for phoenix591.com Cleaning up challenges Failed to renew certificate phoenix591.com with error: Unable to determine zone_id for phoenix591.com using zone names: ['phoenix591.com', 'com']. The error from Cloudflare was: 0 confused info - both key and token defined. - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - All simulated renewals failed. The following certificates could not be renewed: /etc/letsencrypt/live/phoenix591.com/fullchain.pem (failure) - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - 1 renew failure(s), 0 parse failure(s) Ask for help or search for solutions at https://community.letsencrypt.org. See the logfile /var/log/letsencrypt/letsencrypt.log or re-run Certbot with -v for more details. Downgrading cloudflare to 2.9.10 fixes the error. ## Certbot's behavior differed from what I expected because: it threw out this error when the credential file is simply dns_cloudflare_api_token = (token) ## Here is a Certbot log showing the issue (if available): [letsencrypt.log](https://github.com/certbot/certbot/files/9591153/letsencrypt.log) ###### Logs are stored in `/var/log/letsencrypt` by default. Feel free to redact domains, e-mail and IP addresses as you see fit. ## Here is the relevant nginx server block or Apache virtualhost for the domain I am configuring: N/A
1.0
Certbot-dns-cloudflare 1.30.0 doesn't work with cloudflare 2.10.1 - If you're having trouble using Certbot and aren't sure you've found a bug or request for a new feature, please first try asking for help at https://community.letsencrypt.org/. There is a much larger community there of people familiar with the project who will be able to more quickly answer your questions. ## My operating system is (include version): Gentoo Linux ## I installed Certbot with (snap, OS package manager, pip, certbot-auto, etc): package manager ## I ran this command and it produced this output: certbot renew --dry-run -v Saving debug log to /var/log/letsencrypt/letsencrypt.log - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - Processing /etc/letsencrypt/renewal/phoenix591.com.conf - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - Certificate not due for renewal, but simulating renewal for dry run Plugins selected: Authenticator dns-cloudflare, Installer None Simulating renewal of an existing certificate for *.phoenix591.com and phoenix591.com Performing the following challenges: dns-01 challenge for phoenix591.com dns-01 challenge for phoenix591.com Cleaning up challenges Failed to renew certificate phoenix591.com with error: Unable to determine zone_id for phoenix591.com using zone names: ['phoenix591.com', 'com']. The error from Cloudflare was: 0 confused info - both key and token defined. - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - All simulated renewals failed. The following certificates could not be renewed: /etc/letsencrypt/live/phoenix591.com/fullchain.pem (failure) - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - 1 renew failure(s), 0 parse failure(s) Ask for help or search for solutions at https://community.letsencrypt.org. See the logfile /var/log/letsencrypt/letsencrypt.log or re-run Certbot with -v for more details. Downgrading cloudflare to 2.9.10 fixes the error. ## Certbot's behavior differed from what I expected because: it threw out this error when the credential file is simply dns_cloudflare_api_token = (token) ## Here is a Certbot log showing the issue (if available): [letsencrypt.log](https://github.com/certbot/certbot/files/9591153/letsencrypt.log) ###### Logs are stored in `/var/log/letsencrypt` by default. Feel free to redact domains, e-mail and IP addresses as you see fit. ## Here is the relevant nginx server block or Apache virtualhost for the domain I am configuring: N/A
non_process
certbot dns cloudflare doesn t work with cloudflare if you re having trouble using certbot and aren t sure you ve found a bug or request for a new feature please first try asking for help at there is a much larger community there of people familiar with the project who will be able to more quickly answer your questions my operating system is include version gentoo linux i installed certbot with snap os package manager pip certbot auto etc package manager i ran this command and it produced this output certbot renew dry run v saving debug log to var log letsencrypt letsencrypt log processing etc letsencrypt renewal com conf certificate not due for renewal but simulating renewal for dry run plugins selected authenticator dns cloudflare installer none simulating renewal of an existing certificate for com and com performing the following challenges dns challenge for com dns challenge for com cleaning up challenges failed to renew certificate com with error unable to determine zone id for com using zone names the error from cloudflare was confused info both key and token defined all simulated renewals failed the following certificates could not be renewed etc letsencrypt live com fullchain pem failure renew failure s parse failure s ask for help or search for solutions at see the logfile var log letsencrypt letsencrypt log or re run certbot with v for more details downgrading cloudflare to fixes the error certbot s behavior differed from what i expected because it threw out this error when the credential file is simply dns cloudflare api token token here is a certbot log showing the issue if available logs are stored in var log letsencrypt by default feel free to redact domains e mail and ip addresses as you see fit here is the relevant nginx server block or apache virtualhost for the domain i am configuring n a
0
49,105
20,603,981,573
IssuesEvent
2022-03-06 17:53:40
PreMiD/Presences
https://api.github.com/repos/PreMiD/Presences
opened
Website request
Service Request
### Website name FSOnline ### Website URL http://www.filmeserialeonline.org/ ### Website logo FSOnline ### Prerequisites - [ ] It is a paid service - [ ] It displays NSFW content - [X] It is region restricted ### Description It's an website where a lot of romanians watching movies, tv series, that are out in cinemas or had been.
1.0
Website request - ### Website name FSOnline ### Website URL http://www.filmeserialeonline.org/ ### Website logo FSOnline ### Prerequisites - [ ] It is a paid service - [ ] It displays NSFW content - [X] It is region restricted ### Description It's an website where a lot of romanians watching movies, tv series, that are out in cinemas or had been.
non_process
website request website name fsonline website url website logo fsonline prerequisites it is a paid service it displays nsfw content it is region restricted description it s an website where a lot of romanians watching movies tv series that are out in cinemas or had been
0
3,522
6,564,736,468
IssuesEvent
2017-09-08 03:50:08
triplea-game/triplea
https://api.github.com/repos/triplea-game/triplea
closed
Bugs, Features, Usability problems
category: dev & admin process close pending confirmation discussion
I'd like to be sure we're on the same page for tracking and why. First, context and definitions, I've been thinking of the three terms as: - bug: something is not working correctly - usability problem: it works, but the feature is arguably broken, it's hard to use. If we were to test the feature today we would not accept the functionality, it is incomplete or confusing. - feature request: this is 'wish list' thing, something brand new, not necessarily any improvement but something new to add to the game. We decided to move feature requests to forum for a couple of reasons: 1. community engagement on the topic 2. our bug queue is a task queue, we do not want to track things in that queue that are going to sit around and linger. Wish list items are valuable to track, but we do not want those items clouding up the things that are broken or not working well. It's also fundamentally slightly different tasks for a developer to do a bug fix or a UX improvement, compared to building something new. Bugs and usability problems we could track in forums, I raised the topic and it seemed not have been a popular proposal. Forums I do believe are weak when dealing with a high volume of issues, they are good for long conversational threads, not detailed task tracking (at least I have not seen how that would work well). The above definitions do seem clear to me. If a usability problem is really hard to fix and is very low value, and is just sitting around for a long time, and likely to be casually fixed when we do something else, then it's probably a good thing to "ice box". I'm not sure if we gain anything at all by moving something so it can then just sit in the forums. In general I think we should stop the practice of asking for something to be moved and just ice-box stuff here in-place. If bugs are opened in forums, then let's just tag it and ask that bugs be created next time in github (or maybe not even bother since long term we clearly want bug reporting to be more built in). -------------------------------- So mainly I want to clear up the definitions and expectations. We can certainly choose to change how we label things and how we track, it is open for debate! Though, we should balance our efficiency and what we need vs the pain of reporting problems. My philosophy is basically things that "suck" about TripleA, whether broken or just seemingly broken, should be tracked and are important to fix because we do not want TripleA to suck. On the other hand, idea and wish list are very useful things, for a number of reasons, though as established we do not want to track those items with everything else. What we track in github is ideally actionable, a developer can pick up an issue, read it and after 5 minutes know what to do and can start work. Usability problems are tricky, likely a proposed fix would have to be opened in forums or github issues first, or maybe just floated in a PR with risk that it may be rejected.
1.0
Bugs, Features, Usability problems - I'd like to be sure we're on the same page for tracking and why. First, context and definitions, I've been thinking of the three terms as: - bug: something is not working correctly - usability problem: it works, but the feature is arguably broken, it's hard to use. If we were to test the feature today we would not accept the functionality, it is incomplete or confusing. - feature request: this is 'wish list' thing, something brand new, not necessarily any improvement but something new to add to the game. We decided to move feature requests to forum for a couple of reasons: 1. community engagement on the topic 2. our bug queue is a task queue, we do not want to track things in that queue that are going to sit around and linger. Wish list items are valuable to track, but we do not want those items clouding up the things that are broken or not working well. It's also fundamentally slightly different tasks for a developer to do a bug fix or a UX improvement, compared to building something new. Bugs and usability problems we could track in forums, I raised the topic and it seemed not have been a popular proposal. Forums I do believe are weak when dealing with a high volume of issues, they are good for long conversational threads, not detailed task tracking (at least I have not seen how that would work well). The above definitions do seem clear to me. If a usability problem is really hard to fix and is very low value, and is just sitting around for a long time, and likely to be casually fixed when we do something else, then it's probably a good thing to "ice box". I'm not sure if we gain anything at all by moving something so it can then just sit in the forums. In general I think we should stop the practice of asking for something to be moved and just ice-box stuff here in-place. If bugs are opened in forums, then let's just tag it and ask that bugs be created next time in github (or maybe not even bother since long term we clearly want bug reporting to be more built in). -------------------------------- So mainly I want to clear up the definitions and expectations. We can certainly choose to change how we label things and how we track, it is open for debate! Though, we should balance our efficiency and what we need vs the pain of reporting problems. My philosophy is basically things that "suck" about TripleA, whether broken or just seemingly broken, should be tracked and are important to fix because we do not want TripleA to suck. On the other hand, idea and wish list are very useful things, for a number of reasons, though as established we do not want to track those items with everything else. What we track in github is ideally actionable, a developer can pick up an issue, read it and after 5 minutes know what to do and can start work. Usability problems are tricky, likely a proposed fix would have to be opened in forums or github issues first, or maybe just floated in a PR with risk that it may be rejected.
process
bugs features usability problems i d like to be sure we re on the same page for tracking and why first context and definitions i ve been thinking of the three terms as bug something is not working correctly usability problem it works but the feature is arguably broken it s hard to use if we were to test the feature today we would not accept the functionality it is incomplete or confusing feature request this is wish list thing something brand new not necessarily any improvement but something new to add to the game we decided to move feature requests to forum for a couple of reasons community engagement on the topic our bug queue is a task queue we do not want to track things in that queue that are going to sit around and linger wish list items are valuable to track but we do not want those items clouding up the things that are broken or not working well it s also fundamentally slightly different tasks for a developer to do a bug fix or a ux improvement compared to building something new bugs and usability problems we could track in forums i raised the topic and it seemed not have been a popular proposal forums i do believe are weak when dealing with a high volume of issues they are good for long conversational threads not detailed task tracking at least i have not seen how that would work well the above definitions do seem clear to me if a usability problem is really hard to fix and is very low value and is just sitting around for a long time and likely to be casually fixed when we do something else then it s probably a good thing to ice box i m not sure if we gain anything at all by moving something so it can then just sit in the forums in general i think we should stop the practice of asking for something to be moved and just ice box stuff here in place if bugs are opened in forums then let s just tag it and ask that bugs be created next time in github or maybe not even bother since long term we clearly want bug reporting to be more built in so mainly i want to clear up the definitions and expectations we can certainly choose to change how we label things and how we track it is open for debate though we should balance our efficiency and what we need vs the pain of reporting problems my philosophy is basically things that suck about triplea whether broken or just seemingly broken should be tracked and are important to fix because we do not want triplea to suck on the other hand idea and wish list are very useful things for a number of reasons though as established we do not want to track those items with everything else what we track in github is ideally actionable a developer can pick up an issue read it and after minutes know what to do and can start work usability problems are tricky likely a proposed fix would have to be opened in forums or github issues first or maybe just floated in a pr with risk that it may be rejected
1
13,850
16,613,268,030
IssuesEvent
2021-06-02 13:58:41
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
how to define variable via code and show in UI for user to provide value
Pri1 devops-cicd-process/tech devops/prod doc-enhancement support-request
[I found it is difficult to find: how to 'define' a variable, which will appear in the UI of the azure pipeline, so that user can have some input... maybe it should be runtime variable.... I tried to define the variable in the UI, and then save the YAML file, but it still did not back fill to the yml to help me how i can programatically define a variable, for user to provide value at run time.] --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: dd7e0bd3-1f7d-d7b6-cc72-5ef63c31b46a * Version Independent ID: dae87abd-b73d-9120-bcdb-6097d4b40f2a * Content: [Define variables - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/variables?tabs=yaml%2Cbatch&view=azure-devops) * Content Source: [docs/pipelines/process/variables.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/variables.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
how to define variable via code and show in UI for user to provide value - [I found it is difficult to find: how to 'define' a variable, which will appear in the UI of the azure pipeline, so that user can have some input... maybe it should be runtime variable.... I tried to define the variable in the UI, and then save the YAML file, but it still did not back fill to the yml to help me how i can programatically define a variable, for user to provide value at run time.] --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: dd7e0bd3-1f7d-d7b6-cc72-5ef63c31b46a * Version Independent ID: dae87abd-b73d-9120-bcdb-6097d4b40f2a * Content: [Define variables - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/variables?tabs=yaml%2Cbatch&view=azure-devops) * Content Source: [docs/pipelines/process/variables.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/variables.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
how to define variable via code and show in ui for user to provide value document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id bcdb content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
16,639
21,707,262,995
IssuesEvent
2022-05-10 10:46:04
sjmog/smartflix
https://api.github.com/repos/sjmog/smartflix
opened
Testing external API calls
04-background-processing Rails/Testing external APIs
Previously, we’ve combined background processing and connecting to an external API. But, we haven't tested any of this yet. Apart from being a threat to the stability of our application, a lack of proper testing means we can only validate the application behaviour by manually connecting to the external API, and making requests. Testing APIs can be tricky. We don’t want to make real HTTP requests every single time we run tests because it’d be inefficient and expensive. What we’re going to do is to stub the API response first by using **VCR** records called **cassettes**. > We've assumed you're using RSpec. If not, you can adapt this ticket to your test runner! ## To complete this challenge, you will need to - [ ] Install and configure [Webmock](https://github.com/bblimke/webmock). - [ ] Configure Webmock to make use of [VCR](https://github.com/vcr/vcr). - [ ] Update your acceptance test for a show's `show` page to avoid making the API request when the test is run. (Try to avoid changing the controller action!) ## Tips - For more information about Webmock, VCR, and using them together, consider reading [this tutorial article from Honeybadger](https://www.honeybadger.io/blog/ruby-external-api-test/), and/or this [guide from Code with Jason](https://www.codewithjason.com/vcr-webmock-hello-world-tutorial/).
1.0
Testing external API calls - Previously, we’ve combined background processing and connecting to an external API. But, we haven't tested any of this yet. Apart from being a threat to the stability of our application, a lack of proper testing means we can only validate the application behaviour by manually connecting to the external API, and making requests. Testing APIs can be tricky. We don’t want to make real HTTP requests every single time we run tests because it’d be inefficient and expensive. What we’re going to do is to stub the API response first by using **VCR** records called **cassettes**. > We've assumed you're using RSpec. If not, you can adapt this ticket to your test runner! ## To complete this challenge, you will need to - [ ] Install and configure [Webmock](https://github.com/bblimke/webmock). - [ ] Configure Webmock to make use of [VCR](https://github.com/vcr/vcr). - [ ] Update your acceptance test for a show's `show` page to avoid making the API request when the test is run. (Try to avoid changing the controller action!) ## Tips - For more information about Webmock, VCR, and using them together, consider reading [this tutorial article from Honeybadger](https://www.honeybadger.io/blog/ruby-external-api-test/), and/or this [guide from Code with Jason](https://www.codewithjason.com/vcr-webmock-hello-world-tutorial/).
process
testing external api calls previously we’ve combined background processing and connecting to an external api but we haven t tested any of this yet apart from being a threat to the stability of our application a lack of proper testing means we can only validate the application behaviour by manually connecting to the external api and making requests testing apis can be tricky we don’t want to make real http requests every single time we run tests because it’d be inefficient and expensive what we’re going to do is to stub the api response first by using vcr records called cassettes we ve assumed you re using rspec if not you can adapt this ticket to your test runner to complete this challenge you will need to install and configure configure webmock to make use of update your acceptance test for a show s show page to avoid making the api request when the test is run try to avoid changing the controller action tips for more information about webmock vcr and using them together consider reading and or this
1
54,834
7,926,737,693
IssuesEvent
2018-07-06 04:09:54
ansible/galaxy
https://api.github.com/repos/ansible/galaxy
closed
Role import somewhat broken since a few days ago
area/documentation priority/high type/bug
<!--- Verify first that your issue/request is not already reported on GitHub. --> ## Bug Report ##### SUMMARY My role NimbleStorage.Ansinimble won't import my releases and make it the latest anymore (https://galaxy.ansible.com/NimbleStorage/Ansinimble). ##### STEPS TO REPRODUCE ``` mmattsson@ln70-mmattsson:~/code/Ansinimble$ ansible-galaxy install NimbleStorage.Ansinimble - downloading role 'Ansinimble', owned by NimbleStorage - downloading role from https://github.com/NimbleStorage/ansinimble/archive/ansinimble-0.11.tar.gz - extracting NimbleStorage.Ansinimble to /auto/home.nas02/mmattsson/.ansible/roles/NimbleStorage.Ansinimble - NimbleStorage.Ansinimble (ansinimble-0.11) was installed successfully mmattsson@ln70-mmattsson:~/code/Ansinimble$ cat /auto/home.nas02/mmattsson/.ansible/roles/NimbleStorage.Ansinimble/VERSION 0.11.0 mmattsson@ln70-mmattsson:~/code/Ansinimble$ ansible-galaxy remove NimbleStorage.Ansinimble - successfully removed NimbleStorage.Ansinimble mmattsson@ln70-mmattsson:~/code/Ansinimble$ ansible-galaxy install NimbleStorage.Ansinimble,master - downloading role 'Ansinimble', owned by NimbleStorage - downloading role from https://github.com/NimbleStorage/ansinimble/archive/master.tar.gz - extracting NimbleStorage.Ansinimble to /auto/home.nas02/mmattsson/.ansible/roles/NimbleStorage.Ansinimble - NimbleStorage.Ansinimble (master) was installed successfully mmattsson@ln70-mmattsson:~/code/Ansinimble$ cat /auto/home.nas02/mmattsson/.ansible/roles/NimbleStorage.Ansinimble/VERSION 0.12 ``` ##### EXPECTED RESULTS Ansible Galaxy `install` would install the latest release from my repo. `ansinimble-0.12`. This wasn't an issue 10 days ago. ##### ACTUAL RESULTS Installed `ansinible-0.11`.
1.0
Role import somewhat broken since a few days ago - <!--- Verify first that your issue/request is not already reported on GitHub. --> ## Bug Report ##### SUMMARY My role NimbleStorage.Ansinimble won't import my releases and make it the latest anymore (https://galaxy.ansible.com/NimbleStorage/Ansinimble). ##### STEPS TO REPRODUCE ``` mmattsson@ln70-mmattsson:~/code/Ansinimble$ ansible-galaxy install NimbleStorage.Ansinimble - downloading role 'Ansinimble', owned by NimbleStorage - downloading role from https://github.com/NimbleStorage/ansinimble/archive/ansinimble-0.11.tar.gz - extracting NimbleStorage.Ansinimble to /auto/home.nas02/mmattsson/.ansible/roles/NimbleStorage.Ansinimble - NimbleStorage.Ansinimble (ansinimble-0.11) was installed successfully mmattsson@ln70-mmattsson:~/code/Ansinimble$ cat /auto/home.nas02/mmattsson/.ansible/roles/NimbleStorage.Ansinimble/VERSION 0.11.0 mmattsson@ln70-mmattsson:~/code/Ansinimble$ ansible-galaxy remove NimbleStorage.Ansinimble - successfully removed NimbleStorage.Ansinimble mmattsson@ln70-mmattsson:~/code/Ansinimble$ ansible-galaxy install NimbleStorage.Ansinimble,master - downloading role 'Ansinimble', owned by NimbleStorage - downloading role from https://github.com/NimbleStorage/ansinimble/archive/master.tar.gz - extracting NimbleStorage.Ansinimble to /auto/home.nas02/mmattsson/.ansible/roles/NimbleStorage.Ansinimble - NimbleStorage.Ansinimble (master) was installed successfully mmattsson@ln70-mmattsson:~/code/Ansinimble$ cat /auto/home.nas02/mmattsson/.ansible/roles/NimbleStorage.Ansinimble/VERSION 0.12 ``` ##### EXPECTED RESULTS Ansible Galaxy `install` would install the latest release from my repo. `ansinimble-0.12`. This wasn't an issue 10 days ago. ##### ACTUAL RESULTS Installed `ansinible-0.11`.
non_process
role import somewhat broken since a few days ago verify first that your issue request is not already reported on github bug report summary my role nimblestorage ansinimble won t import my releases and make it the latest anymore steps to reproduce mmattsson mmattsson code ansinimble ansible galaxy install nimblestorage ansinimble downloading role ansinimble owned by nimblestorage downloading role from extracting nimblestorage ansinimble to auto home mmattsson ansible roles nimblestorage ansinimble nimblestorage ansinimble ansinimble was installed successfully mmattsson mmattsson code ansinimble cat auto home mmattsson ansible roles nimblestorage ansinimble version mmattsson mmattsson code ansinimble ansible galaxy remove nimblestorage ansinimble successfully removed nimblestorage ansinimble mmattsson mmattsson code ansinimble ansible galaxy install nimblestorage ansinimble master downloading role ansinimble owned by nimblestorage downloading role from extracting nimblestorage ansinimble to auto home mmattsson ansible roles nimblestorage ansinimble nimblestorage ansinimble master was installed successfully mmattsson mmattsson code ansinimble cat auto home mmattsson ansible roles nimblestorage ansinimble version expected results ansible galaxy install would install the latest release from my repo ansinimble this wasn t an issue days ago actual results installed ansinible
0
86,242
16,869,757,427
IssuesEvent
2021-06-22 01:42:39
parcel-bundler/parcel
https://api.github.com/repos/parcel-bundler/parcel
closed
Uncaught ReferenceError: globalThis is not defined
:bug: Bug ✨ Parcel 2 🐡 JS Codegen
# 🐛 bug report <!--- Provide a general summary of the issue here --> Compiling a package with Node (or Electron-renderer) context and using it triggers this error in some cases. ``` ReferenceError: globalThis is not defined at /packages/linter-ui-default/dist/index.js:1:7989 at Se (/packages/linter-ui-default/dist/index.js:1:8142) at oi (/packages/linter-ui-default/dist/index.js:1:44368) at _e (/packages/linter-ui-default/dist/index.js:1:15274) at _e (/packages/linter-ui-default/dist/index.js:1:7948) at st (/packages/linter-ui-default/dist/index.js:1:15266) at Qn.e.forEach.e (/packages/linter-ui-default/dist/index.js:1:83011) at Array.forEach (<anonymous>) at /packages/linter-ui-default/dist/index.js:1:82976) at mn.updateTooltip (/packages/linter-ui-default/dist/index.js:1:89363) at textEditor.onDidChangeCursorPosition (/packages/linter-ui-default/dist/index.js:1:88651) at /packages/linter-ui-default/dist/index.js:1:41580) at /packages/linter-ui-default/dist/index.js:1:41857) at /packages/linter-ui-default/dist/index.js:1:41746) ``` ## 🎛 Configuration (.babelrc, package.json, cli command) <!--- If describing a bug, tell us what your babel configuration looks like --> ```js "engines": { "atom": ">=1.19.0 <2.0.0", "electron": ">=5.x" }, "targets": { "main": { "context": "node", "includeNodeModules": { "atom": false, "electron": false, "atom-package-deps": false, "disposable-event": false }, "outputFormat": "commonjs", "isLibrary": true } }, ``` ``` "build": "cross-env NODE_ENV=production parcel build --target main ./lib/index.ts", ``` ## 🤔 Expected Behavior globalThis should not be needed? <!--- Tell us what should happen --> ## 😯 Current Behavior Parcel generates `globalThis` <!--- Tell us what happens instead of the expected behavior --> <!--- If you are seeing an error, please include the full error message and stack trace --> ## 💁 Possible Solution Use unbundled JS <!--- Not obligatory, but suggest a fix/reason for the bug --> ## 🔦 Context https://github.com/steelbrain/linter-ui-default/issues/617 <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most useful in the real world --> ## 💻 Code Sample https://github.com/steelbrain/linter-ui-default <!-- Please provide a code repository, gist, code snippet or sample files to reproduce the issue --> ## 🌍 Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> | Software | Version(s) | | ---------------- | ---------- | | Parcel | "parcel": "2.0.0-nightly.562", | Node | 12 | npm/Yarn | 6 | Operating System | win 10 <!-- Love parcel? Please consider supporting our collective: 👉 https://opencollective.com/parcel/donate -->
1.0
Uncaught ReferenceError: globalThis is not defined - # 🐛 bug report <!--- Provide a general summary of the issue here --> Compiling a package with Node (or Electron-renderer) context and using it triggers this error in some cases. ``` ReferenceError: globalThis is not defined at /packages/linter-ui-default/dist/index.js:1:7989 at Se (/packages/linter-ui-default/dist/index.js:1:8142) at oi (/packages/linter-ui-default/dist/index.js:1:44368) at _e (/packages/linter-ui-default/dist/index.js:1:15274) at _e (/packages/linter-ui-default/dist/index.js:1:7948) at st (/packages/linter-ui-default/dist/index.js:1:15266) at Qn.e.forEach.e (/packages/linter-ui-default/dist/index.js:1:83011) at Array.forEach (<anonymous>) at /packages/linter-ui-default/dist/index.js:1:82976) at mn.updateTooltip (/packages/linter-ui-default/dist/index.js:1:89363) at textEditor.onDidChangeCursorPosition (/packages/linter-ui-default/dist/index.js:1:88651) at /packages/linter-ui-default/dist/index.js:1:41580) at /packages/linter-ui-default/dist/index.js:1:41857) at /packages/linter-ui-default/dist/index.js:1:41746) ``` ## 🎛 Configuration (.babelrc, package.json, cli command) <!--- If describing a bug, tell us what your babel configuration looks like --> ```js "engines": { "atom": ">=1.19.0 <2.0.0", "electron": ">=5.x" }, "targets": { "main": { "context": "node", "includeNodeModules": { "atom": false, "electron": false, "atom-package-deps": false, "disposable-event": false }, "outputFormat": "commonjs", "isLibrary": true } }, ``` ``` "build": "cross-env NODE_ENV=production parcel build --target main ./lib/index.ts", ``` ## 🤔 Expected Behavior globalThis should not be needed? <!--- Tell us what should happen --> ## 😯 Current Behavior Parcel generates `globalThis` <!--- Tell us what happens instead of the expected behavior --> <!--- If you are seeing an error, please include the full error message and stack trace --> ## 💁 Possible Solution Use unbundled JS <!--- Not obligatory, but suggest a fix/reason for the bug --> ## 🔦 Context https://github.com/steelbrain/linter-ui-default/issues/617 <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most useful in the real world --> ## 💻 Code Sample https://github.com/steelbrain/linter-ui-default <!-- Please provide a code repository, gist, code snippet or sample files to reproduce the issue --> ## 🌍 Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> | Software | Version(s) | | ---------------- | ---------- | | Parcel | "parcel": "2.0.0-nightly.562", | Node | 12 | npm/Yarn | 6 | Operating System | win 10 <!-- Love parcel? Please consider supporting our collective: 👉 https://opencollective.com/parcel/donate -->
non_process
uncaught referenceerror globalthis is not defined 🐛 bug report compiling a package with node or electron renderer context and using it triggers this error in some cases referenceerror globalthis is not defined at packages linter ui default dist index js at se packages linter ui default dist index js at oi packages linter ui default dist index js at e packages linter ui default dist index js at e packages linter ui default dist index js at st packages linter ui default dist index js at qn e foreach e packages linter ui default dist index js at array foreach at packages linter ui default dist index js at mn updatetooltip packages linter ui default dist index js at texteditor ondidchangecursorposition packages linter ui default dist index js at packages linter ui default dist index js at packages linter ui default dist index js at packages linter ui default dist index js 🎛 configuration babelrc package json cli command js engines atom electron x targets main context node includenodemodules atom false electron false atom package deps false disposable event false outputformat commonjs islibrary true build cross env node env production parcel build target main lib index ts 🤔 expected behavior globalthis should not be needed 😯 current behavior parcel generates globalthis 💁 possible solution use unbundled js 🔦 context 💻 code sample 🌍 your environment software version s parcel parcel nightly node npm yarn operating system win love parcel please consider supporting our collective 👉
0
15,602
19,724,012,274
IssuesEvent
2022-01-13 18:02:17
googleapis/python-compute
https://api.github.com/repos/googleapis/python-compute
closed
Release as GA
type: process api: compute
[GA release template](https://github.com/googleapis/google-cloud-common/issues/287) ## Required - [ ] 28 days elapsed since last beta release with new API surface - [ ] Server API is GA - [ ] Package API is stable, and we can commit to backward compatibility - [ ] All dependencies are GA
1.0
Release as GA - [GA release template](https://github.com/googleapis/google-cloud-common/issues/287) ## Required - [ ] 28 days elapsed since last beta release with new API surface - [ ] Server API is GA - [ ] Package API is stable, and we can commit to backward compatibility - [ ] All dependencies are GA
process
release as ga required days elapsed since last beta release with new api surface server api is ga package api is stable and we can commit to backward compatibility all dependencies are ga
1
12,183
14,742,088,173
IssuesEvent
2021-01-07 11:40:45
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Speed Ez - Child Account balance to Parent account | Parent:1326
anc-process anp-1 ant-bug ant-child/secondary ant-support
In GitLab by @kdjstudios on Mar 13, 2019, 13:36 **Submitted by:** <joann@speedez.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/7514811 **Server:** External **Client/Site:** Speed Ez **Account:** Multiple **Issue:** On Parent Account 1002 I have balances in there from their former accounts 5390 and 5394. We cannot post a payment on the Child Accounts. If you look at the Parent Account, those amounts which are $740.62 for account 5390 and $289.41 for account 5394 are NOT included on the Parent Account Balances. The Parent Account balance is CORRECT but those other amounts are just sitting there not attached to any balance. If we put in the payment on the Parent Account, it takes it off of the Parent Account Balance which is not correct. How should we handle this? Also, I need Account 5407 attached to this Parent Account, but, not till you figure out what we need to do about the balances because there is a balance on 5407.
1.0
Speed Ez - Child Account balance to Parent account | Parent:1326 - In GitLab by @kdjstudios on Mar 13, 2019, 13:36 **Submitted by:** <joann@speedez.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/7514811 **Server:** External **Client/Site:** Speed Ez **Account:** Multiple **Issue:** On Parent Account 1002 I have balances in there from their former accounts 5390 and 5394. We cannot post a payment on the Child Accounts. If you look at the Parent Account, those amounts which are $740.62 for account 5390 and $289.41 for account 5394 are NOT included on the Parent Account Balances. The Parent Account balance is CORRECT but those other amounts are just sitting there not attached to any balance. If we put in the payment on the Parent Account, it takes it off of the Parent Account Balance which is not correct. How should we handle this? Also, I need Account 5407 attached to this Parent Account, but, not till you figure out what we need to do about the balances because there is a balance on 5407.
process
speed ez child account balance to parent account parent in gitlab by kdjstudios on mar submitted by helpdesk server external client site speed ez account multiple issue on parent account i have balances in there from their former accounts and we cannot post a payment on the child accounts if you look at the parent account those amounts which are for account and for account are not included on the parent account balances the parent account balance is correct but those other amounts are just sitting there not attached to any balance if we put in the payment on the parent account it takes it off of the parent account balance which is not correct how should we handle this also i need account attached to this parent account but not till you figure out what we need to do about the balances because there is a balance on
1
49,067
6,007,679,953
IssuesEvent
2017-06-06 04:38:11
lexibrent/certificate-transparency
https://api.github.com/repos/lexibrent/certificate-transparency
closed
Change PSQL library to Psycopg for performance reasons
enhancement needs-deployment-testing
Also, use socket-based connections to Postgres rather than TCP/IP.
1.0
Change PSQL library to Psycopg for performance reasons - Also, use socket-based connections to Postgres rather than TCP/IP.
non_process
change psql library to psycopg for performance reasons also use socket based connections to postgres rather than tcp ip
0
42,001
10,738,203,995
IssuesEvent
2019-10-29 14:25:28
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
UDT.getQualifiedName() doesn't return schema qualification for generated tables
C: Functionality E: All Editions P: High R: Fixed T: Defect
Generated `Table` and `UDT` instances override `getSchema()`, but that method is not being called by `getQualifiedName()`. This seems easier to fix than https://github.com/jOOQ/jOOQ/issues/7172, even if the two issues are caused by the same problem, which is why this will be fixed separately. Fixing this will fix the regression mentioned here: https://github.com/jOOQ/jOOQ/issues/7496#issuecomment-546048221
1.0
UDT.getQualifiedName() doesn't return schema qualification for generated tables - Generated `Table` and `UDT` instances override `getSchema()`, but that method is not being called by `getQualifiedName()`. This seems easier to fix than https://github.com/jOOQ/jOOQ/issues/7172, even if the two issues are caused by the same problem, which is why this will be fixed separately. Fixing this will fix the regression mentioned here: https://github.com/jOOQ/jOOQ/issues/7496#issuecomment-546048221
non_process
udt getqualifiedname doesn t return schema qualification for generated tables generated table and udt instances override getschema but that method is not being called by getqualifiedname this seems easier to fix than even if the two issues are caused by the same problem which is why this will be fixed separately fixing this will fix the regression mentioned here
0
15,319
19,426,539,025
IssuesEvent
2021-12-21 06:37:32
nion-software/nionswift
https://api.github.com/repos/nion-software/nionswift
closed
Mappable processing items (Fourier windows) do not compute properly
type - bug level - easy f - processing
An error is triggered if there is no graphic on the source image.
1.0
Mappable processing items (Fourier windows) do not compute properly - An error is triggered if there is no graphic on the source image.
process
mappable processing items fourier windows do not compute properly an error is triggered if there is no graphic on the source image
1
89,345
15,828,981,395
IssuesEvent
2021-04-06 10:34:06
bci-oss/keycloak
https://api.github.com/repos/bci-oss/keycloak
opened
CVE-2016-2141 (High) detected in jgroups-3.2.7.Final.jar
security vulnerability
## CVE-2016-2141 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jgroups-3.2.7.Final.jar</b></p></summary> <p>Reliable cluster communication toolkit</p> <p>Library home page: <a href="http://www.jgroups.org">http://www.jgroups.org</a></p> <p>Path to dependency file: keycloak/adapters/saml/as7-eap6/subsystem/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/jgroups/jgroups/3.2.7.Final/jgroups-3.2.7.Final.jar,/home/wss-scanner/.m2/repository/org/jgroups/jgroups/3.2.7.Final/jgroups-3.2.7.Final.jar</p> <p> Dependency Hierarchy: - jboss-as-web-7.2.0.Final.jar (Root Library) - jboss-as-clustering-registry-7.2.0.Final.jar - jboss-as-clustering-infinispan-7.2.0.Final.jar - jboss-as-clustering-jgroups-7.2.0.Final.jar - :x: **jgroups-3.2.7.Final.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/bci-oss/keycloak/commit/4ece64b821b25688d9fbfacd091a88584ffc90e0">4ece64b821b25688d9fbfacd091a88584ffc90e0</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> JGroups before 4.0 does not require the proper headers for the ENCRYPT and AUTH protocols from nodes joining the cluster, which allows remote attackers to bypass security restrictions and send and receive messages within the cluster via unspecified vectors. <p>Publish Date: 2016-06-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-2141>CVE-2016-2141</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-2141">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-2141</a></p> <p>Release Date: 2016-06-30</p> <p>Fix Resolution: 4.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.jgroups","packageName":"jgroups","packageVersion":"3.2.7.Final","packageFilePaths":["/adapters/saml/as7-eap6/subsystem/pom.xml","/adapters/oidc/as7-eap6/as7-subsystem/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.jboss.as:jboss-as-web:7.2.0.Final;org.jboss.as:jboss-as-clustering-registry:7.2.0.Final;org.jboss.as:jboss-as-clustering-infinispan:7.2.0.Final;org.jboss.as:jboss-as-clustering-jgroups:7.2.0.Final;org.jgroups:jgroups:3.2.7.Final","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2016-2141","vulnerabilityDetails":"JGroups before 4.0 does not require the proper headers for the ENCRYPT and AUTH protocols from nodes joining the cluster, which allows remote attackers to bypass security restrictions and send and receive messages within the cluster via unspecified vectors.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-2141","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2016-2141 (High) detected in jgroups-3.2.7.Final.jar - ## CVE-2016-2141 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jgroups-3.2.7.Final.jar</b></p></summary> <p>Reliable cluster communication toolkit</p> <p>Library home page: <a href="http://www.jgroups.org">http://www.jgroups.org</a></p> <p>Path to dependency file: keycloak/adapters/saml/as7-eap6/subsystem/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/jgroups/jgroups/3.2.7.Final/jgroups-3.2.7.Final.jar,/home/wss-scanner/.m2/repository/org/jgroups/jgroups/3.2.7.Final/jgroups-3.2.7.Final.jar</p> <p> Dependency Hierarchy: - jboss-as-web-7.2.0.Final.jar (Root Library) - jboss-as-clustering-registry-7.2.0.Final.jar - jboss-as-clustering-infinispan-7.2.0.Final.jar - jboss-as-clustering-jgroups-7.2.0.Final.jar - :x: **jgroups-3.2.7.Final.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/bci-oss/keycloak/commit/4ece64b821b25688d9fbfacd091a88584ffc90e0">4ece64b821b25688d9fbfacd091a88584ffc90e0</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> JGroups before 4.0 does not require the proper headers for the ENCRYPT and AUTH protocols from nodes joining the cluster, which allows remote attackers to bypass security restrictions and send and receive messages within the cluster via unspecified vectors. <p>Publish Date: 2016-06-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-2141>CVE-2016-2141</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-2141">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-2141</a></p> <p>Release Date: 2016-06-30</p> <p>Fix Resolution: 4.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.jgroups","packageName":"jgroups","packageVersion":"3.2.7.Final","packageFilePaths":["/adapters/saml/as7-eap6/subsystem/pom.xml","/adapters/oidc/as7-eap6/as7-subsystem/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.jboss.as:jboss-as-web:7.2.0.Final;org.jboss.as:jboss-as-clustering-registry:7.2.0.Final;org.jboss.as:jboss-as-clustering-infinispan:7.2.0.Final;org.jboss.as:jboss-as-clustering-jgroups:7.2.0.Final;org.jgroups:jgroups:3.2.7.Final","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2016-2141","vulnerabilityDetails":"JGroups before 4.0 does not require the proper headers for the ENCRYPT and AUTH protocols from nodes joining the cluster, which allows remote attackers to bypass security restrictions and send and receive messages within the cluster via unspecified vectors.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-2141","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in jgroups final jar cve high severity vulnerability vulnerable library jgroups final jar reliable cluster communication toolkit library home page a href path to dependency file keycloak adapters saml subsystem pom xml path to vulnerable library home wss scanner repository org jgroups jgroups final jgroups final jar home wss scanner repository org jgroups jgroups final jgroups final jar dependency hierarchy jboss as web final jar root library jboss as clustering registry final jar jboss as clustering infinispan final jar jboss as clustering jgroups final jar x jgroups final jar vulnerable library found in head commit a href found in base branch master vulnerability details jgroups before does not require the proper headers for the encrypt and auth protocols from nodes joining the cluster which allows remote attackers to bypass security restrictions and send and receive messages within the cluster via unspecified vectors publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org jboss as jboss as web final org jboss as jboss as clustering registry final org jboss as jboss as clustering infinispan final org jboss as jboss as clustering jgroups final org jgroups jgroups final isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails jgroups before does not require the proper headers for the encrypt and auth protocols from nodes joining the cluster which allows remote attackers to bypass security restrictions and send and receive messages within the cluster via unspecified vectors vulnerabilityurl
0
256,236
22,042,353,386
IssuesEvent
2022-05-29 14:51:07
SAA-SDT/eac-cpf-schema
https://api.github.com/repos/SAA-SDT/eac-cpf-schema
closed
<address>
Element Tested by Schema Team
## Address - keep name and scope - add optional attributes `@audience` `@conventationDeclarationReference` `@localTypeDeclarationReference` `@maintenanceEventReference` `@scriptOfElement` `@sourceReference` ## Creator of issue 1. Silke Jagodzinski 2. TS-EAS: EAC-CPF subgroup 3. silkejagodzinski@gmail.com ## Related issues / documents ## EAD3 Reconciliation **Additional EAD 3 attributes**: altrender - Optional audience - Optional (values limited to: external, internal) script - optional ## Context **Summary**: A postal or other address. **May contain**: addressLine **May occur within**: place **Attributes**: localType, xml:id, xml:lang - all optional **Availability**: Optional, Repeatable ## Solution documentation: **May contain**: `<addressLine>` **May occur within**: `<place>` **Attributes**: `@audience` - optional (values limited to: external, internal) `@conventationDeclarationReference` - optional `@id` - optional `@languageOfElement` - optional `@localType` - optional `@localTypeDeclarationReference` - optional `@maintenanceEventReference` - optional `@scriptOfElement` - optional `@sourceReference` - optional **Availability**: Optional, Repeatable - `<address>` is ideally bundled with `<placeName>` - use `<address>` to encode a physical or analog address vs. `<contact>` to encode digital addresses and contact information ## Example encoding ``` <place> <placeName>name of the place</placeName> <address audience="external" conventionDeclarationReference="conventiondeclaration1" id="address1" languageOfElement="en" localType="localAdress" localTypeDeclarationReference="localTypeDeclaration1" maintenanceEventReference="maintenancevent1" scriptOfElement="lat" sourceReference="source1" target="element> <addressLine>distinct address information</addressLine> <addressLine>distinct address information</addressLine> </address> </place> ```
1.0
<address> - ## Address - keep name and scope - add optional attributes `@audience` `@conventationDeclarationReference` `@localTypeDeclarationReference` `@maintenanceEventReference` `@scriptOfElement` `@sourceReference` ## Creator of issue 1. Silke Jagodzinski 2. TS-EAS: EAC-CPF subgroup 3. silkejagodzinski@gmail.com ## Related issues / documents ## EAD3 Reconciliation **Additional EAD 3 attributes**: altrender - Optional audience - Optional (values limited to: external, internal) script - optional ## Context **Summary**: A postal or other address. **May contain**: addressLine **May occur within**: place **Attributes**: localType, xml:id, xml:lang - all optional **Availability**: Optional, Repeatable ## Solution documentation: **May contain**: `<addressLine>` **May occur within**: `<place>` **Attributes**: `@audience` - optional (values limited to: external, internal) `@conventationDeclarationReference` - optional `@id` - optional `@languageOfElement` - optional `@localType` - optional `@localTypeDeclarationReference` - optional `@maintenanceEventReference` - optional `@scriptOfElement` - optional `@sourceReference` - optional **Availability**: Optional, Repeatable - `<address>` is ideally bundled with `<placeName>` - use `<address>` to encode a physical or analog address vs. `<contact>` to encode digital addresses and contact information ## Example encoding ``` <place> <placeName>name of the place</placeName> <address audience="external" conventionDeclarationReference="conventiondeclaration1" id="address1" languageOfElement="en" localType="localAdress" localTypeDeclarationReference="localTypeDeclaration1" maintenanceEventReference="maintenancevent1" scriptOfElement="lat" sourceReference="source1" target="element> <addressLine>distinct address information</addressLine> <addressLine>distinct address information</addressLine> </address> </place> ```
non_process
address keep name and scope add optional attributes audience conventationdeclarationreference localtypedeclarationreference maintenanceeventreference scriptofelement sourcereference creator of issue silke jagodzinski ts eas eac cpf subgroup silkejagodzinski gmail com related issues documents reconciliation additional ead attributes altrender optional audience optional values limited to external internal script optional context summary a postal or other address may contain addressline may occur within place attributes localtype xml id xml lang all optional availability optional repeatable solution documentation may contain may occur within attributes audience optional values limited to external internal conventationdeclarationreference optional id optional languageofelement optional localtype optional localtypedeclarationreference optional maintenanceeventreference optional scriptofelement optional sourcereference optional availability optional repeatable is ideally bundled with use to encode a physical or analog address vs to encode digital addresses and contact information example encoding name of the place distinct address information distinct address information
0
107,475
9,214,351,398
IssuesEvent
2019-03-10 19:11:31
SunwellTracker/issues
https://api.github.com/repos/SunwellTracker/issues
closed
All Hail Roanauk! Missing NPC
Works locally OR Requires testing
Decription: Missing NPC How it works: NPC Missing How it should work: NPC should be there https://www.wowhead.com/quest=12140/all-hail-roanauk Source (you should point out proofs of your report, please give us some source): http://prntscr.com/lygqa9
1.0
All Hail Roanauk! Missing NPC - Decription: Missing NPC How it works: NPC Missing How it should work: NPC should be there https://www.wowhead.com/quest=12140/all-hail-roanauk Source (you should point out proofs of your report, please give us some source): http://prntscr.com/lygqa9
non_process
all hail roanauk missing npc decription missing npc how it works npc missing how it should work npc should be there source you should point out proofs of your report please give us some source
0
2,005
4,819,351,284
IssuesEvent
2016-11-04 18:59:03
Azure/azure-event-hubs-java
https://api.github.com/repos/Azure/azure-event-hubs-java
closed
EPH stopped receiving from all partitions after hitting IllegalArgumentException couple of times.
EventProcessorHost
_From @serkantkaraca on July 26, 2016 22:22_ - Single process - Single host - 4 partitions EH Host ran quite fine for 3 hours in our stress tests and then hit multiple IllegalArgumentExceptions then completely stopped delivering messages. <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: Jul 25, 2016 8:32:04 PM com.microsoft.azure.servicebus.MessagingFactory$RunReactor run <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: WARNING: UnHandled exception while processing events in reactor: <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: java.lang.IllegalArgumentException <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.engine.impl.EventImpl.dispatch(EventImpl.java:112) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.reactor.impl.ReactorImpl.dispatch(ReactorImpl.java:307) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.reactor.impl.ReactorImpl.process(ReactorImpl.java:276) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: com.microsoft.azure.servicebus.MessagingFactory$RunReactor.run(MessagingFactory.java:355) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: java.lang.Thread.run(Unknown Source)Cause: null <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: java.nio.Buffer.position(Unknown Source) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.engine.impl.ssl.SimpleSslTransportWrapper.pop(SimpleSslTransportWrapper.java:411) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.engine.impl.ssl.SslImpl$UnsecureClientAwareTransportWrapper.pop(SslImpl.java:185) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.engine.impl.TransportImpl.pop(TransportImpl.java:1476) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.reactor.impl.SelectorImpl.select(SelectorImpl.java:145) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.reactor.impl.IOHandler.handleQuiesced(IOHandler.java:57) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.reactor.impl.IOHandler.onUnhandled(IOHandler.java:381) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.engine.BaseHandler.onReactorQuiesced(BaseHandler.java:87) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.engine.BaseHandler.handle(BaseHandler.java:206) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.engine.impl.EventImpl.dispatch(EventImpl.java:108) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.reactor.impl.ReactorImpl.dispatch(ReactorImpl.java:307) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.reactor.impl.ReactorImpl.process(ReactorImpl.java:276) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: com.microsoft.azure.servicebus.MessagingFactory$RunReactor.run(MessagingFactory.java:355) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: java.lang.Thread.run(Unknown Source) _Copied from original issue: Azure/azure-event-hubs#200_
1.0
EPH stopped receiving from all partitions after hitting IllegalArgumentException couple of times. - _From @serkantkaraca on July 26, 2016 22:22_ - Single process - Single host - 4 partitions EH Host ran quite fine for 3 hours in our stress tests and then hit multiple IllegalArgumentExceptions then completely stopped delivering messages. <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: Jul 25, 2016 8:32:04 PM com.microsoft.azure.servicebus.MessagingFactory$RunReactor run <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: WARNING: UnHandled exception while processing events in reactor: <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: java.lang.IllegalArgumentException <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.engine.impl.EventImpl.dispatch(EventImpl.java:112) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.reactor.impl.ReactorImpl.dispatch(ReactorImpl.java:307) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.reactor.impl.ReactorImpl.process(ReactorImpl.java:276) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: com.microsoft.azure.servicebus.MessagingFactory$RunReactor.run(MessagingFactory.java:355) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: java.lang.Thread.run(Unknown Source)Cause: null <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: java.nio.Buffer.position(Unknown Source) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.engine.impl.ssl.SimpleSslTransportWrapper.pop(SimpleSslTransportWrapper.java:411) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.engine.impl.ssl.SslImpl$UnsecureClientAwareTransportWrapper.pop(SslImpl.java:185) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.engine.impl.TransportImpl.pop(TransportImpl.java:1476) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.reactor.impl.SelectorImpl.select(SelectorImpl.java:145) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.reactor.impl.IOHandler.handleQuiesced(IOHandler.java:57) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.reactor.impl.IOHandler.onUnhandled(IOHandler.java:381) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.engine.BaseHandler.onReactorQuiesced(BaseHandler.java:87) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.engine.BaseHandler.handle(BaseHandler.java:206) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.engine.impl.EventImpl.dispatch(EventImpl.java:108) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.reactor.impl.ReactorImpl.dispatch(ReactorImpl.java:307) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: org.apache.qpid.proton.reactor.impl.ReactorImpl.process(ReactorImpl.java:276) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: com.microsoft.azure.servicebus.MessagingFactory$RunReactor.run(MessagingFactory.java:355) <8084.20> 07/25/2016 20:32:04 - Error: JAVA-Receiver_0<stderr>: java.lang.Thread.run(Unknown Source) _Copied from original issue: Azure/azure-event-hubs#200_
process
eph stopped receiving from all partitions after hitting illegalargumentexception couple of times from serkantkaraca on july single process single host partitions eh host ran quite fine for hours in our stress tests and then hit multiple illegalargumentexceptions then completely stopped delivering messages error java receiver jul pm com microsoft azure servicebus messagingfactory runreactor run error java receiver warning unhandled exception while processing events in reactor error java receiver java lang illegalargumentexception error java receiver org apache qpid proton engine impl eventimpl dispatch eventimpl java error java receiver org apache qpid proton reactor impl reactorimpl dispatch reactorimpl java error java receiver org apache qpid proton reactor impl reactorimpl process reactorimpl java error java receiver com microsoft azure servicebus messagingfactory runreactor run messagingfactory java error java receiver java lang thread run unknown source cause null error java receiver java nio buffer position unknown source error java receiver org apache qpid proton engine impl ssl simplessltransportwrapper pop simplessltransportwrapper java error java receiver org apache qpid proton engine impl ssl sslimpl unsecureclientawaretransportwrapper pop sslimpl java error java receiver org apache qpid proton engine impl transportimpl pop transportimpl java error java receiver org apache qpid proton reactor impl selectorimpl select selectorimpl java error java receiver org apache qpid proton reactor impl iohandler handlequiesced iohandler java error java receiver org apache qpid proton reactor impl iohandler onunhandled iohandler java error java receiver org apache qpid proton engine basehandler onreactorquiesced basehandler java error java receiver org apache qpid proton engine basehandler handle basehandler java error java receiver org apache qpid proton engine impl eventimpl dispatch eventimpl java error java receiver org apache qpid proton reactor impl reactorimpl dispatch reactorimpl java error java receiver org apache qpid proton reactor impl reactorimpl process reactorimpl java error java receiver com microsoft azure servicebus messagingfactory runreactor run messagingfactory java error java receiver java lang thread run unknown source copied from original issue azure azure event hubs
1
5,727
8,568,987,796
IssuesEvent
2018-11-11 04:41:36
gfrebello/qs-trip-planning-procedure
https://api.github.com/repos/gfrebello/qs-trip-planning-procedure
closed
Update Home Page
Priority:Medium Process:Implement Requirement
An update to the home page would be welcome. The first version is not very appealing. Moreover, some standard flights could be offered directly to the user to make the experience (and even testing for development) easier further on. https://github.com/facebook/react-native/issues/2481 https://stackoverflow.com/questions/15757036/creating-a-zoom-effect-on-an-image-on-hover-using-css
1.0
Update Home Page - An update to the home page would be welcome. The first version is not very appealing. Moreover, some standard flights could be offered directly to the user to make the experience (and even testing for development) easier further on. https://github.com/facebook/react-native/issues/2481 https://stackoverflow.com/questions/15757036/creating-a-zoom-effect-on-an-image-on-hover-using-css
process
update home page an update to the home page would be welcome the first version is not very appealing moreover some standard flights could be offered directly to the user to make the experience and even testing for development easier further on
1
13,128
15,527,049,191
IssuesEvent
2021-03-13 04:00:20
hasura/ask-me-anything
https://api.github.com/repos/hasura/ask-me-anything
closed
Can a read-only database be connected to with Hasura?
next-up-for-ama processing-for-shortvid question
Yes. However there needs to be a metadata database, which can be connected using the environment variable HASURA_GRAPHQL_METADATA_DATABASE_URL. Any other databases can then be connected using what configuration in regards to read, write, or read-only.
1.0
Can a read-only database be connected to with Hasura? - Yes. However there needs to be a metadata database, which can be connected using the environment variable HASURA_GRAPHQL_METADATA_DATABASE_URL. Any other databases can then be connected using what configuration in regards to read, write, or read-only.
process
can a read only database be connected to with hasura yes however there needs to be a metadata database which can be connected using the environment variable hasura graphql metadata database url any other databases can then be connected using what configuration in regards to read write or read only
1
17,682
23,517,052,379
IssuesEvent
2022-08-18 22:53:38
LLNL/axom
https://api.github.com/repos/LLNL/axom
opened
New symlink breaks `git difftool` (for me)
bug question Software process usability
The new symlink added in #901 breaks `git difftool` for me. The symlink lives in axom's `scripts` directory and points to a script in `blt` #### Reproducer: ``` >git difftool -d develop fatal: could not open '<tmp>/left/scripts/make_local_branch_from_fork_pr.sh' for writing: Not a directory ``` Should we remove the symlink, and rely on knowledge of its existence in `blt`? Alternatively, we could add the symlink file to .gitignore ``` # in .gitignore ... scripts/make_local_branch_from_fork_pr.sh ```
1.0
New symlink breaks `git difftool` (for me) - The new symlink added in #901 breaks `git difftool` for me. The symlink lives in axom's `scripts` directory and points to a script in `blt` #### Reproducer: ``` >git difftool -d develop fatal: could not open '<tmp>/left/scripts/make_local_branch_from_fork_pr.sh' for writing: Not a directory ``` Should we remove the symlink, and rely on knowledge of its existence in `blt`? Alternatively, we could add the symlink file to .gitignore ``` # in .gitignore ... scripts/make_local_branch_from_fork_pr.sh ```
process
new symlink breaks git difftool for me the new symlink added in breaks git difftool for me the symlink lives in axom s scripts directory and points to a script in blt reproducer git difftool d develop fatal could not open left scripts make local branch from fork pr sh for writing not a directory should we remove the symlink and rely on knowledge of its existence in blt alternatively we could add the symlink file to gitignore in gitignore scripts make local branch from fork pr sh
1
2,760
5,695,581,362
IssuesEvent
2017-04-16 00:22:44
nodejs/node
https://api.github.com/repos/nodejs/node
closed
child_process.exec() fails with spaces in absolute or relative path to binary file
child_process doc good first contribution
- **Version**: v5.10.1 - **Platform**: Darwin Joran.local 15.4.0 Darwin Kernel Version 15.4.0: Fri Feb 26 22:08:05 PST 2016; root:xnu-3248.40.184~3/RELEASE_X86_64 x86_64 - **Subsystem**: child_process Using exec() to execute an absolute path to a binary, with spaces in the absolute path, e.g. `/Users/Joran/test script.sh` fails: ``` { [Error: Command failed: /Users/Joran/test script.sh /bin/sh: /Users/Joran/test: No such file or directory ] killed: false, code: 127, signal: null, cmd: '/Users/Joran/test script.sh' } '' '/bin/sh: /Users/Joran/test: No such file or directory\n' ```
1.0
child_process.exec() fails with spaces in absolute or relative path to binary file - - **Version**: v5.10.1 - **Platform**: Darwin Joran.local 15.4.0 Darwin Kernel Version 15.4.0: Fri Feb 26 22:08:05 PST 2016; root:xnu-3248.40.184~3/RELEASE_X86_64 x86_64 - **Subsystem**: child_process Using exec() to execute an absolute path to a binary, with spaces in the absolute path, e.g. `/Users/Joran/test script.sh` fails: ``` { [Error: Command failed: /Users/Joran/test script.sh /bin/sh: /Users/Joran/test: No such file or directory ] killed: false, code: 127, signal: null, cmd: '/Users/Joran/test script.sh' } '' '/bin/sh: /Users/Joran/test: No such file or directory\n' ```
process
child process exec fails with spaces in absolute or relative path to binary file version platform darwin joran local darwin kernel version fri feb pst root xnu release subsystem child process using exec to execute an absolute path to a binary with spaces in the absolute path e g users joran test script sh fails error command failed users joran test script sh bin sh users joran test no such file or directory killed false code signal null cmd users joran test script sh bin sh users joran test no such file or directory n
1
312,456
26,866,095,090
IssuesEvent
2023-02-04 00:03:35
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
closed
Fix averages_and_variances.test_numpy_nanmean
NumPy Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4010131450/jobs/6886279521" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4010131450/jobs/6886279521" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4010131450/jobs/6886279521" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4085755865/jobs/7044184797" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix averages_and_variances.test_numpy_nanmean - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4010131450/jobs/6886279521" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4010131450/jobs/6886279521" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4010131450/jobs/6886279521" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4085755865/jobs/7044184797" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
non_process
fix averages and variances test numpy nanmean tensorflow img src torch img src numpy img src jax img src
0
2,174
5,027,565,604
IssuesEvent
2016-12-15 15:52:10
Alfresco/alfresco-ng2-components
https://api.github.com/repos/Alfresco/alfresco-ng2-components
opened
Cannot check checkbox on form within user task
browser: all bug comp: activiti-processList
When clicking on the date picker within a form attached to a user task within a process it does not display 1. Import below app 2. Start process 3. Complete start form attached to start event 4. Go to active tasks and click on user task 5. Click checkbox **Expected results** Checkbox is checked **Actual results** Checkbox is not checked [no typeaheads.zip](https://github.com/Alfresco/alfresco-ng2-components/files/655039/no.typeaheads.zip)
1.0
Cannot check checkbox on form within user task - When clicking on the date picker within a form attached to a user task within a process it does not display 1. Import below app 2. Start process 3. Complete start form attached to start event 4. Go to active tasks and click on user task 5. Click checkbox **Expected results** Checkbox is checked **Actual results** Checkbox is not checked [no typeaheads.zip](https://github.com/Alfresco/alfresco-ng2-components/files/655039/no.typeaheads.zip)
process
cannot check checkbox on form within user task when clicking on the date picker within a form attached to a user task within a process it does not display import below app start process complete start form attached to start event go to active tasks and click on user task click checkbox expected results checkbox is checked actual results checkbox is not checked
1
154,713
12,226,160,761
IssuesEvent
2020-05-03 09:33:16
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: jepsen/sets/parts-start-kill-2 failed
C-test-failure O-roachtest O-robot branch-release-19.2 release-blocker
[(roachtest).jepsen/sets/parts-start-kill-2 failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1916055&tab=buildLog) on [release-19.2@8085a7edbb294a78b6a7550f05522ec3088754ad](https://github.com/cockroachdb/cockroach/commits/8085a7edbb294a78b6a7550f05522ec3088754ad): ``` | main.runJepsen.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/jepsen.go:159 | main.runJepsen.func3 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/jepsen.go:180 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1357 Wraps: (2) 2 safe details enclosed Wraps: (3) output in run_092617.624_n6_bash Wraps: (4) /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1916055-1588487054-44-n6cpu4:6 -- bash -e -c "\ | cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \ | ~/lein run test \ | --tarball file://${PWD}/cockroach.tgz \ | --username ${USER} \ | --ssh-private-key ~/.ssh/id_rsa \ | --os ubuntu \ | --time-limit 300 \ | --concurrency 30 \ | --recovery-time 25 \ | --test-count 1 \ | -n 10.128.0.227 -n 10.128.1.8 -n 10.128.0.236 -n 10.128.0.120 -n 10.128.0.143 \ | --test sets --nemesis parts --nemesis2 start-kill-2 \ | > invoke.log 2>&1 \ | " returned | stderr: | Error: SSH_PROBLEM: exit status 255 | (1) SSH_PROBLEM | Wraps: (2) Node 6. Command with error: | | ``` | | bash -e -c "\ | | cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \ | | ~/lein run test \ | | --tarball file://${PWD}/cockroach.tgz \ | | --username ${USER} \ | | --ssh-private-key ~/.ssh/id_rsa \ | | --os ubuntu \ | | --time-limit 300 \ | | --concurrency 30 \ | | --recovery-time 25 \ | | --test-count 1 \ | | -n 10.128.0.227 -n 10.128.1.8 -n 10.128.0.236 -n 10.128.0.120 -n 10.128.0.143 \ | | --test sets --nemesis parts --nemesis2 start-kill-2 \ | | > invoke.log 2>&1 \ | | " | | ``` | Wraps: (3) exit status 255 | Error types: (1) errors.SSH (2) *hintdetail.withDetail (3) *exec.ExitError | | stdout: Wraps: (5) exit status 10 Error types: (1) *withstack.withStack (2) *safedetails.withSafeDetails (3) *errutil.withMessage (4) *main.withCommandDetails (5) *exec.ExitError ``` <details><summary>More</summary><p> Artifacts: [/jepsen/sets/parts-start-kill-2](https://teamcity.cockroachdb.com/viewLog.html?buildId=1916055&tab=artifacts#/jepsen/sets/parts-start-kill-2) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Ajepsen%2Fsets%2Fparts-start-kill-2.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
2.0
roachtest: jepsen/sets/parts-start-kill-2 failed - [(roachtest).jepsen/sets/parts-start-kill-2 failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1916055&tab=buildLog) on [release-19.2@8085a7edbb294a78b6a7550f05522ec3088754ad](https://github.com/cockroachdb/cockroach/commits/8085a7edbb294a78b6a7550f05522ec3088754ad): ``` | main.runJepsen.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/jepsen.go:159 | main.runJepsen.func3 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/jepsen.go:180 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1357 Wraps: (2) 2 safe details enclosed Wraps: (3) output in run_092617.624_n6_bash Wraps: (4) /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1916055-1588487054-44-n6cpu4:6 -- bash -e -c "\ | cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \ | ~/lein run test \ | --tarball file://${PWD}/cockroach.tgz \ | --username ${USER} \ | --ssh-private-key ~/.ssh/id_rsa \ | --os ubuntu \ | --time-limit 300 \ | --concurrency 30 \ | --recovery-time 25 \ | --test-count 1 \ | -n 10.128.0.227 -n 10.128.1.8 -n 10.128.0.236 -n 10.128.0.120 -n 10.128.0.143 \ | --test sets --nemesis parts --nemesis2 start-kill-2 \ | > invoke.log 2>&1 \ | " returned | stderr: | Error: SSH_PROBLEM: exit status 255 | (1) SSH_PROBLEM | Wraps: (2) Node 6. Command with error: | | ``` | | bash -e -c "\ | | cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \ | | ~/lein run test \ | | --tarball file://${PWD}/cockroach.tgz \ | | --username ${USER} \ | | --ssh-private-key ~/.ssh/id_rsa \ | | --os ubuntu \ | | --time-limit 300 \ | | --concurrency 30 \ | | --recovery-time 25 \ | | --test-count 1 \ | | -n 10.128.0.227 -n 10.128.1.8 -n 10.128.0.236 -n 10.128.0.120 -n 10.128.0.143 \ | | --test sets --nemesis parts --nemesis2 start-kill-2 \ | | > invoke.log 2>&1 \ | | " | | ``` | Wraps: (3) exit status 255 | Error types: (1) errors.SSH (2) *hintdetail.withDetail (3) *exec.ExitError | | stdout: Wraps: (5) exit status 10 Error types: (1) *withstack.withStack (2) *safedetails.withSafeDetails (3) *errutil.withMessage (4) *main.withCommandDetails (5) *exec.ExitError ``` <details><summary>More</summary><p> Artifacts: [/jepsen/sets/parts-start-kill-2](https://teamcity.cockroachdb.com/viewLog.html?buildId=1916055&tab=artifacts#/jepsen/sets/parts-start-kill-2) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Ajepsen%2Fsets%2Fparts-start-kill-2.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
non_process
roachtest jepsen sets parts start kill failed on main runjepsen home agent work go src github com cockroachdb cockroach pkg cmd roachtest jepsen go main runjepsen home agent work go src github com cockroachdb cockroach pkg cmd roachtest jepsen go runtime goexit usr local go src runtime asm s wraps safe details enclosed wraps output in run bash wraps home agent work go src github com cockroachdb cockroach bin roachprod run teamcity bash e c cd mnt jepsen cockroachdb set eo pipefail lein run test tarball file pwd cockroach tgz username user ssh private key ssh id rsa os ubuntu time limit concurrency recovery time test count n n n n n test sets nemesis parts start kill invoke log returned stderr error ssh problem exit status ssh problem wraps node command with error bash e c cd mnt jepsen cockroachdb set eo pipefail lein run test tarball file pwd cockroach tgz username user ssh private key ssh id rsa os ubuntu time limit concurrency recovery time test count n n n n n test sets nemesis parts start kill invoke log wraps exit status error types errors ssh hintdetail withdetail exec exiterror stdout wraps exit status error types withstack withstack safedetails withsafedetails errutil withmessage main withcommanddetails exec exiterror more artifacts powered by
0
19,520
25,831,087,661
IssuesEvent
2022-12-12 16:08:46
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
closed
Move dependency check to separate workflow
bug process
### Description We moved the check for vulnerable dependencies inline into the individual project's build. This causes all projects to fail when there's a single vulnerable dependency. We added the GRADLE_ARGS secret to allow temporarily removing that task but we did not realize that dependabot has a separate secret and outside contributors can't access secrets at all. ### Steps to reproduce Submit a PR from a fork ### Additional context _No response_ ### Hedera network other ### Version main ### Operating system None
1.0
Move dependency check to separate workflow - ### Description We moved the check for vulnerable dependencies inline into the individual project's build. This causes all projects to fail when there's a single vulnerable dependency. We added the GRADLE_ARGS secret to allow temporarily removing that task but we did not realize that dependabot has a separate secret and outside contributors can't access secrets at all. ### Steps to reproduce Submit a PR from a fork ### Additional context _No response_ ### Hedera network other ### Version main ### Operating system None
process
move dependency check to separate workflow description we moved the check for vulnerable dependencies inline into the individual project s build this causes all projects to fail when there s a single vulnerable dependency we added the gradle args secret to allow temporarily removing that task but we did not realize that dependabot has a separate secret and outside contributors can t access secrets at all steps to reproduce submit a pr from a fork additional context no response hedera network other version main operating system none
1
20,856
27,635,651,552
IssuesEvent
2023-03-10 14:16:03
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
spanmetricsprocessor: Set resource attributes where possible
enhancement processor/spanmetrics connector/spanmetrics
### Component(s) processor/spanmetrics ### Is your feature request related to a problem? Please describe. I was using the `spanmetricsprocessor` with `prometheusexporter` and was surprised to find the `service_name` as a label instead of `job`. Digging deeper, this is because the processor only sets metric attributes but not resource attributes. ### Describe the solution you'd like I think we should set resource attributes _if we are aggregating by a resource attribute_. For example, if we are aggregating by `service.name` and `service.instance.id`, the dimensions in the metrics emitted should be part of resource attributes and not metric attributes. ### Describe alternatives you've considered Use the transform processor to move the metric attributes to resource attributes. ### Additional context _No response_
1.0
spanmetricsprocessor: Set resource attributes where possible - ### Component(s) processor/spanmetrics ### Is your feature request related to a problem? Please describe. I was using the `spanmetricsprocessor` with `prometheusexporter` and was surprised to find the `service_name` as a label instead of `job`. Digging deeper, this is because the processor only sets metric attributes but not resource attributes. ### Describe the solution you'd like I think we should set resource attributes _if we are aggregating by a resource attribute_. For example, if we are aggregating by `service.name` and `service.instance.id`, the dimensions in the metrics emitted should be part of resource attributes and not metric attributes. ### Describe alternatives you've considered Use the transform processor to move the metric attributes to resource attributes. ### Additional context _No response_
process
spanmetricsprocessor set resource attributes where possible component s processor spanmetrics is your feature request related to a problem please describe i was using the spanmetricsprocessor with prometheusexporter and was surprised to find the service name as a label instead of job digging deeper this is because the processor only sets metric attributes but not resource attributes describe the solution you d like i think we should set resource attributes if we are aggregating by a resource attribute for example if we are aggregating by service name and service instance id the dimensions in the metrics emitted should be part of resource attributes and not metric attributes describe alternatives you ve considered use the transform processor to move the metric attributes to resource attributes additional context no response
1
8,641
11,787,894,365
IssuesEvent
2020-03-17 14:44:43
kubeflow/testing
https://api.github.com/repos/kubeflow/testing
opened
label sync needs to auto-sync the config to the cluster
area/engprod kind/process priority/p1
Our label sync process currently relies on manually updating the config map when things change. https://github.com/kubeflow/testing/tree/master/label_sync Lets try to use gitops and anthos CM for this. https://cloud.google.com/anthos-config-management/docs/quickstart
1.0
label sync needs to auto-sync the config to the cluster - Our label sync process currently relies on manually updating the config map when things change. https://github.com/kubeflow/testing/tree/master/label_sync Lets try to use gitops and anthos CM for this. https://cloud.google.com/anthos-config-management/docs/quickstart
process
label sync needs to auto sync the config to the cluster our label sync process currently relies on manually updating the config map when things change lets try to use gitops and anthos cm for this
1
169,640
6,413,145,670
IssuesEvent
2017-08-08 06:39:51
GoogleCloudPlatform/google-cloud-python
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-python
reopened
Should we broaden the 'admin' flag pattern?
api: bigquery api: bigtable api: pubsub api: storage priority: p2+ type: question
The Bigtable client takes an `admin` flag, defaulting to `False`: if it isn't set to `True`, the client avoids adding scopes for the instance / cluster admin APIs, and blocks creation of the corresponding stubs. It has the effect of a "safety" knob, preventing a user (likely at the REPL prompt?) from accidentally creating / deleting "expensive" resources. The client also takes a `readonly` flag, defaulting to `False`, and uses it to decide which of two scopes to use. Logically, if such a pattern is a good idea, it would be equally good to have such protection for the "expensive"-ish resources in other APIs: storage buckets, bigquery datasets / tables(?), pubsub topics / subscriptions (feel free to add others as they occur to you). As an alternative, perhaps we need to make it more obvious how to configure clients with reduced scopes, maybe even requiring a non-default flag to set the more permissive ones.
1.0
Should we broaden the 'admin' flag pattern? - The Bigtable client takes an `admin` flag, defaulting to `False`: if it isn't set to `True`, the client avoids adding scopes for the instance / cluster admin APIs, and blocks creation of the corresponding stubs. It has the effect of a "safety" knob, preventing a user (likely at the REPL prompt?) from accidentally creating / deleting "expensive" resources. The client also takes a `readonly` flag, defaulting to `False`, and uses it to decide which of two scopes to use. Logically, if such a pattern is a good idea, it would be equally good to have such protection for the "expensive"-ish resources in other APIs: storage buckets, bigquery datasets / tables(?), pubsub topics / subscriptions (feel free to add others as they occur to you). As an alternative, perhaps we need to make it more obvious how to configure clients with reduced scopes, maybe even requiring a non-default flag to set the more permissive ones.
non_process
should we broaden the admin flag pattern the bigtable client takes an admin flag defaulting to false if it isn t set to true the client avoids adding scopes for the instance cluster admin apis and blocks creation of the corresponding stubs it has the effect of a safety knob preventing a user likely at the repl prompt from accidentally creating deleting expensive resources the client also takes a readonly flag defaulting to false and uses it to decide which of two scopes to use logically if such a pattern is a good idea it would be equally good to have such protection for the expensive ish resources in other apis storage buckets bigquery datasets tables pubsub topics subscriptions feel free to add others as they occur to you as an alternative perhaps we need to make it more obvious how to configure clients with reduced scopes maybe even requiring a non default flag to set the more permissive ones
0
5,438
8,302,614,071
IssuesEvent
2018-09-21 14:59:56
GoogleCloudPlatform/golang-samples
https://api.github.com/repos/GoogleCloudPlatform/golang-samples
closed
Package context available in the standard library as of Go 1.7
priority: p3 type: process
Shall we make a sweep update for package context in all the subdirectories? Replacing `golang.org/x/net/context` with `context`. "As of Go 1.7 this package is available in the standard library under the name context. " (Source: [godoc](https://godoc.org/context)) Go 1.7 was released in Aug 2016.
1.0
Package context available in the standard library as of Go 1.7 - Shall we make a sweep update for package context in all the subdirectories? Replacing `golang.org/x/net/context` with `context`. "As of Go 1.7 this package is available in the standard library under the name context. " (Source: [godoc](https://godoc.org/context)) Go 1.7 was released in Aug 2016.
process
package context available in the standard library as of go shall we make a sweep update for package context in all the subdirectories replacing golang org x net context with context as of go this package is available in the standard library under the name context source go was released in aug
1
10,270
13,125,300,601
IssuesEvent
2020-08-06 06:21:40
spring-projects/spring-hateoas
https://api.github.com/repos/spring-projects/spring-hateoas
closed
How to custom Wrapped Responses which wrap RepresentationModel?
process: waiting for feedback
RESTful Service Best Practices has a Wrapped Responses section , which introduce so-called JSEND response. and responses has with the following properties: > • code – contains the HTTP response status code as an integer. • status – contains the text: “success”, “fail”, or “error”. Where “fail” is for HTTP status response values from 500-599, “error” is for statuses 400-499, and “success” is for everything else (e.g. 1XX, 2XX and 3XX responses). • message – only used for “fail” and “error” statuses to contain the error message. For internationalization (i18n) purposes, this could contain a message number or code, either alone or contained within delimiters. • data – that contains the response body. In the case of “error” or “fail” statuses, this contains the cause, or exception name. data should be the RepresentationModel. for example: the original response `{ "id": 1, "name": "Steven Spielberg", "_embedded": { "movieList": [ { "id": "1", "title": "ET" }, { "id": "2", "title": "Jaws" } ] } }` if we enabled JSEND response , then the response should be: ` { "code": 200, "message": "success", "data": { "id": 1, "name": "Steven Spielberg", "_embedded": { "movieList": [ { "id": "1", "title": "ET" }, { "id": "2", "title": "Jaws" } ] } } } ` wrapper class can be provided by developer
1.0
How to custom Wrapped Responses which wrap RepresentationModel? - RESTful Service Best Practices has a Wrapped Responses section , which introduce so-called JSEND response. and responses has with the following properties: > • code – contains the HTTP response status code as an integer. • status – contains the text: “success”, “fail”, or “error”. Where “fail” is for HTTP status response values from 500-599, “error” is for statuses 400-499, and “success” is for everything else (e.g. 1XX, 2XX and 3XX responses). • message – only used for “fail” and “error” statuses to contain the error message. For internationalization (i18n) purposes, this could contain a message number or code, either alone or contained within delimiters. • data – that contains the response body. In the case of “error” or “fail” statuses, this contains the cause, or exception name. data should be the RepresentationModel. for example: the original response `{ "id": 1, "name": "Steven Spielberg", "_embedded": { "movieList": [ { "id": "1", "title": "ET" }, { "id": "2", "title": "Jaws" } ] } }` if we enabled JSEND response , then the response should be: ` { "code": 200, "message": "success", "data": { "id": 1, "name": "Steven Spielberg", "_embedded": { "movieList": [ { "id": "1", "title": "ET" }, { "id": "2", "title": "Jaws" } ] } } } ` wrapper class can be provided by developer
process
how to custom wrapped responses which wrap representationmodel restful service best practices has a wrapped responses section which introduce so called jsend response and responses has with the following properties • code – contains the http response status code as an integer • status – contains the text “success” “fail” or “error” where “fail” is for http status response values from “error” is for statuses and “success” is for everything else e g and responses • message – only used for “fail” and “error” statuses to contain the error message for internationalization purposes this could contain a message number or code either alone or contained within delimiters • data – that contains the response body in the case of “error” or “fail” statuses this contains the cause or exception name data should be the representationmodel for example the original response id name steven spielberg embedded movielist id title et id title jaws if we enabled jsend response then the response should be code message success data id name steven spielberg embedded movielist id title et id title jaws wrapper class can be provided by developer
1
76,211
9,398,651,239
IssuesEvent
2019-04-08 12:48:33
linnovate/root
https://api.github.com/repos/linnovate/root
reopened
option to delete deleted items using multiple select
2.0.6.1 Fixed Search design
go to show deleted items enter multiple select mode the delete option appears
1.0
option to delete deleted items using multiple select - go to show deleted items enter multiple select mode the delete option appears
non_process
option to delete deleted items using multiple select go to show deleted items enter multiple select mode the delete option appears
0
84,287
10,368,908,849
IssuesEvent
2019-09-07 20:59:58
tobiasanker/libKitsunePersistence
https://api.github.com/repos/tobiasanker/libKitsunePersistence
closed
Add logger
Documentation Feature
The boost-library already provides a good logger. But I would like it to have all persist-functions collected into this lib. So a wrapper have to be implemented, which use internally the logger of the boost-lib.
1.0
Add logger - The boost-library already provides a good logger. But I would like it to have all persist-functions collected into this lib. So a wrapper have to be implemented, which use internally the logger of the boost-lib.
non_process
add logger the boost library already provides a good logger but i would like it to have all persist functions collected into this lib so a wrapper have to be implemented which use internally the logger of the boost lib
0
18,182
24,234,200,332
IssuesEvent
2022-09-26 21:12:36
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
Integrated Terminal Profile (PowerShell 7) launches current working directory with extra quotation marks
bug confirmation-pending terminal-process
Type: <b>Bug</b> It seems that launching a custom integrated terminal profile may wrap the `cwd` with an extra set of quotation marks. I did not notice the behavior with the baseline terminals. ### Steps to Reproduce 1. Set `terminal.integrated.shellIntegration.enabled` to `true` 2. Set integrate terminal profile settings as such (using [PowerShell 7](https://docs.microsoft.com/en-us/powershell/scripting/install/installing-powershell-on-windows?WT.mc_id=THOMASMAURER-blog-thmaure&view=powershell-7), unsure if other nonstandard terminals are impacted) ```json { "terminal.integrated.profiles.windows": { "PowerShell 7": { "path": ["F:\\Program Files\\PowerShell\\7\\pwsh.exe"], "icon": "terminal-powershell" }, "PowerShell": null, "Command Prompt": null, "Git Bash": null }, "terminal.integrated.defaultProfile.windows": "PowerShell 7" } ``` 3. Open integrated terminal 4. Note that the current directory shows an extra `"` ![image](https://user-images.githubusercontent.com/79218/188028809-bbdeb9d1-48d3-4975-b9e6-606a7659e8b3.png) 5. Try to split the terminal and receive an error ![image](https://user-images.githubusercontent.com/79218/188028835-86b67412-4376-42a2-8b53-c27370d0bf82.png) Turning on Trace logging, the following shows up in the `Pty Host` logs ```json { "shellLaunchConfig": { "executable": "F:\\Program Files\\PowerShell\\7\\pwsh.exe", "icon": { "id": "terminal-powershell" }, "color": null, "cwd": "\"G:\\Projects\\temp\"", "parentTerminalId": 1, "useShellEnvironment": true } } ``` Note the `cwd` parameter which is wrapped in `\"` I had also tried while running `code --disable-extensions` and had the same behavior --- VS Code version: Code 1.71.0 (784b0177c56c607789f9638da7b6bf3230d47a8c, 2022-09-01T07:36:10.600Z) OS version: Windows_NT x64 10.0.22000 Modes: Sandboxed: No <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|Intel(R) Core(TM) i9-9900K CPU @ 3.60GHz (16 x 3600)| |GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: disabled_off<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_renderer: enabled_on<br>video_decode: enabled<br>video_encode: enabled<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: disabled_off| |Load (avg)|undefined| |Memory (System)|31.91GB (17.33GB free)| |Process Argv|--crash-reporter-id 69185b9e-c8fd-4b5f-8db6-6a831fddcff2| |Screen Reader|no| |VM|0%| </details><details><summary>Extensions (19)</summary> Extension|Author (truncated)|Version ---|---|--- vscode-eslint|dba|2.2.6 xml|Dot|2.5.1 gitlens|eam|12.2.1 vsc-material-theme|Equ|33.2.0 vsc-material-theme-icons|equ|1.2.2 prettier-vscode|esb|9.8.0 auto-rename-tag|for|0.1.10 vscode-pull-request-github|Git|0.50.0 rainbow-csv|mec|2.4.0 remote-containers|ms-|0.251.0 remote-wsl|ms-|0.66.3 vscode-remote-extensionpack|ms-|0.21.0 powershell|ms-|2022.8.5 vsliveshare|ms-|1.0.5705 material-icon-theme|PKi|4.20.0 vscode-thunder-client|ran|1.18.7 code-spell-checker|str|2.7.2 vscode-icons|vsc|11.16.0 markdown-all-in-one|yzh|3.4.3 </details><details> <summary>A/B Experiments</summary> ``` vsliv368cf:30146710 vsreu685:30147344 python383:30185418 vspor879:30202332 vspor708:30202333 vspor363:30204092 vslsvsres303:30308271 pythonvspyl392:30443607 vserr242cf:30382550 pythontb:30283811 vsjup518:30340749 pythonptprofiler:30281270 vsdfh931:30280409 vshan820:30294714 vstes263:30335439 pythondataviewer:30285071 vscod805:30301674 binariesv615:30325510 bridge0708:30335490 bridge0723:30353136 cmake_vspar411cf:30557515 vsaa593:30376534 pythonvs932:30410667 cppdebug:30492333 pylanb8912:30545647 vsclangdc:30486549 c4g48928:30535728 hb751961:30553087 dsvsc012:30540252 azure-dev_surveyone:30548225 i497e931:30553904 ``` </details> <!-- generated by issue reporter -->
1.0
Integrated Terminal Profile (PowerShell 7) launches current working directory with extra quotation marks - Type: <b>Bug</b> It seems that launching a custom integrated terminal profile may wrap the `cwd` with an extra set of quotation marks. I did not notice the behavior with the baseline terminals. ### Steps to Reproduce 1. Set `terminal.integrated.shellIntegration.enabled` to `true` 2. Set integrate terminal profile settings as such (using [PowerShell 7](https://docs.microsoft.com/en-us/powershell/scripting/install/installing-powershell-on-windows?WT.mc_id=THOMASMAURER-blog-thmaure&view=powershell-7), unsure if other nonstandard terminals are impacted) ```json { "terminal.integrated.profiles.windows": { "PowerShell 7": { "path": ["F:\\Program Files\\PowerShell\\7\\pwsh.exe"], "icon": "terminal-powershell" }, "PowerShell": null, "Command Prompt": null, "Git Bash": null }, "terminal.integrated.defaultProfile.windows": "PowerShell 7" } ``` 3. Open integrated terminal 4. Note that the current directory shows an extra `"` ![image](https://user-images.githubusercontent.com/79218/188028809-bbdeb9d1-48d3-4975-b9e6-606a7659e8b3.png) 5. Try to split the terminal and receive an error ![image](https://user-images.githubusercontent.com/79218/188028835-86b67412-4376-42a2-8b53-c27370d0bf82.png) Turning on Trace logging, the following shows up in the `Pty Host` logs ```json { "shellLaunchConfig": { "executable": "F:\\Program Files\\PowerShell\\7\\pwsh.exe", "icon": { "id": "terminal-powershell" }, "color": null, "cwd": "\"G:\\Projects\\temp\"", "parentTerminalId": 1, "useShellEnvironment": true } } ``` Note the `cwd` parameter which is wrapped in `\"` I had also tried while running `code --disable-extensions` and had the same behavior --- VS Code version: Code 1.71.0 (784b0177c56c607789f9638da7b6bf3230d47a8c, 2022-09-01T07:36:10.600Z) OS version: Windows_NT x64 10.0.22000 Modes: Sandboxed: No <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|Intel(R) Core(TM) i9-9900K CPU @ 3.60GHz (16 x 3600)| |GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: disabled_off<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_renderer: enabled_on<br>video_decode: enabled<br>video_encode: enabled<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: disabled_off| |Load (avg)|undefined| |Memory (System)|31.91GB (17.33GB free)| |Process Argv|--crash-reporter-id 69185b9e-c8fd-4b5f-8db6-6a831fddcff2| |Screen Reader|no| |VM|0%| </details><details><summary>Extensions (19)</summary> Extension|Author (truncated)|Version ---|---|--- vscode-eslint|dba|2.2.6 xml|Dot|2.5.1 gitlens|eam|12.2.1 vsc-material-theme|Equ|33.2.0 vsc-material-theme-icons|equ|1.2.2 prettier-vscode|esb|9.8.0 auto-rename-tag|for|0.1.10 vscode-pull-request-github|Git|0.50.0 rainbow-csv|mec|2.4.0 remote-containers|ms-|0.251.0 remote-wsl|ms-|0.66.3 vscode-remote-extensionpack|ms-|0.21.0 powershell|ms-|2022.8.5 vsliveshare|ms-|1.0.5705 material-icon-theme|PKi|4.20.0 vscode-thunder-client|ran|1.18.7 code-spell-checker|str|2.7.2 vscode-icons|vsc|11.16.0 markdown-all-in-one|yzh|3.4.3 </details><details> <summary>A/B Experiments</summary> ``` vsliv368cf:30146710 vsreu685:30147344 python383:30185418 vspor879:30202332 vspor708:30202333 vspor363:30204092 vslsvsres303:30308271 pythonvspyl392:30443607 vserr242cf:30382550 pythontb:30283811 vsjup518:30340749 pythonptprofiler:30281270 vsdfh931:30280409 vshan820:30294714 vstes263:30335439 pythondataviewer:30285071 vscod805:30301674 binariesv615:30325510 bridge0708:30335490 bridge0723:30353136 cmake_vspar411cf:30557515 vsaa593:30376534 pythonvs932:30410667 cppdebug:30492333 pylanb8912:30545647 vsclangdc:30486549 c4g48928:30535728 hb751961:30553087 dsvsc012:30540252 azure-dev_surveyone:30548225 i497e931:30553904 ``` </details> <!-- generated by issue reporter -->
process
integrated terminal profile powershell launches current working directory with extra quotation marks type bug it seems that launching a custom integrated terminal profile may wrap the cwd with an extra set of quotation marks i did not notice the behavior with the baseline terminals steps to reproduce set terminal integrated shellintegration enabled to true set integrate terminal profile settings as such using unsure if other nonstandard terminals are impacted json terminal integrated profiles windows powershell path icon terminal powershell powershell null command prompt null git bash null terminal integrated defaultprofile windows powershell open integrated terminal note that the current directory shows an extra try to split the terminal and receive an error turning on trace logging the following shows up in the pty host logs json shelllaunchconfig executable f program files powershell pwsh exe icon id terminal powershell color null cwd g projects temp parentterminalid useshellenvironment true note the cwd parameter which is wrapped in i had also tried while running code disable extensions and had the same behavior vs code version code os version windows nt modes sandboxed no system info item value cpus intel r core tm cpu x gpu status canvas enabled canvas oop rasterization disabled off direct rendering display compositor disabled off ok gpu compositing enabled multiple raster threads enabled on opengl enabled on rasterization enabled raw draw disabled off ok skia renderer enabled on video decode enabled video encode enabled vulkan disabled off webgl enabled enabled webgpu disabled off load avg undefined memory system free process argv crash reporter id screen reader no vm extensions extension author truncated version vscode eslint dba xml dot gitlens eam vsc material theme equ vsc material theme icons equ prettier vscode esb auto rename tag for vscode pull request github git rainbow csv mec remote containers ms remote wsl ms vscode remote extensionpack ms powershell ms vsliveshare ms material icon theme pki vscode thunder client ran code spell checker str vscode icons vsc markdown all in one yzh a b experiments pythontb pythonptprofiler pythondataviewer cmake cppdebug vsclangdc azure dev surveyone
1
12,288
14,816,243,862
IssuesEvent
2021-01-14 08:44:10
modi-w/AutoVersionsDB
https://api.github.com/repos/modi-w/AutoVersionsDB
opened
Add indication for scripts that executed virtually
area-Core area-UI good first issue process-ready-for-implementation type-enhancement up-for-grab
**The Problem** When a developer ran the "virtual" process, it means that some of the scripts file not actually executed. It is not clear on the UI and not on the Console app, which file executed virtually and which actually executed. **Solution** Add indication for scripts that executed virtually. **Action Items:** 1. Add the indication from the DB on the related runtimefile classes 2. Add the indication on the script files grids on the UI - maybe by another column. 3. Add the indication on the console related methods.
1.0
Add indication for scripts that executed virtually - **The Problem** When a developer ran the "virtual" process, it means that some of the scripts file not actually executed. It is not clear on the UI and not on the Console app, which file executed virtually and which actually executed. **Solution** Add indication for scripts that executed virtually. **Action Items:** 1. Add the indication from the DB on the related runtimefile classes 2. Add the indication on the script files grids on the UI - maybe by another column. 3. Add the indication on the console related methods.
process
add indication for scripts that executed virtually the problem when a developer ran the virtual process it means that some of the scripts file not actually executed it is not clear on the ui and not on the console app which file executed virtually and which actually executed solution add indication for scripts that executed virtually action items add the indication from the db on the related runtimefile classes add the indication on the script files grids on the ui maybe by another column add the indication on the console related methods
1
16,776
21,958,771,211
IssuesEvent
2022-05-24 14:13:23
ethereum/EIPs
https://api.github.com/repos/ethereum/EIPs
closed
Require at least one Github handle in author field
type: EIP1 (Process) stale
For a while now, @MicahZoltu has enforced that at least one author lists their github handle in the author field. This allows them to interact with the auto merge bot. EIP-1 should be updated to reflect this.
1.0
Require at least one Github handle in author field - For a while now, @MicahZoltu has enforced that at least one author lists their github handle in the author field. This allows them to interact with the auto merge bot. EIP-1 should be updated to reflect this.
process
require at least one github handle in author field for a while now micahzoltu has enforced that at least one author lists their github handle in the author field this allows them to interact with the auto merge bot eip should be updated to reflect this
1
14,943
18,410,408,672
IssuesEvent
2021-10-13 04:33:34
edmobe/android-video-magnification
https://api.github.com/repos/edmobe/android-video-magnification
closed
OB-1005 El error en el procesamiento de signos vitales es mayor que el 5%
video-processing obstacle
El algoritmo de detección de picos por FFT no es suficiente para conseguir el error esperado.
1.0
OB-1005 El error en el procesamiento de signos vitales es mayor que el 5% - El algoritmo de detección de picos por FFT no es suficiente para conseguir el error esperado.
process
ob el error en el procesamiento de signos vitales es mayor que el el algoritmo de detección de picos por fft no es suficiente para conseguir el error esperado
1
15,580
19,704,106,851
IssuesEvent
2022-01-12 19:49:29
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Automatically adding the agent to a VM
doc-enhancement devops/prod Pri2 devops-cicd-process/tech needs-sme
Hello, The instructions explain how to use the devops user interface to add a VM resource (the Add a resource section). However, we use ARM templates and devops pipelines to define our VMs, so we would really prefer to add the VM resource as part of an automated process, without needing manual steps afterwards. The docs for 'classic' deployment groups included info about how to do that, here: https://docs.microsoft.com/en-us/azure/devops/pipelines/release/deployment-groups/howto-provision-deployment-group-agents?view=azure-devops#use-the-arm-template-deployment-task Can you tell me how to achive the same thing when working with devops Environments? Thanks, Matt --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 91d0d31f-81ee-c024-db7e-daddbf525f71 * Version Independent ID: 330f1649-386c-d0aa-5f96-b8343a1480d3 * Content: [Environment - Virtual machine resource - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments-virtual-machines?view=azure-devops) * Content Source: [docs/pipelines/process/environments-virtual-machines.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/environments-virtual-machines.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
Automatically adding the agent to a VM - Hello, The instructions explain how to use the devops user interface to add a VM resource (the Add a resource section). However, we use ARM templates and devops pipelines to define our VMs, so we would really prefer to add the VM resource as part of an automated process, without needing manual steps afterwards. The docs for 'classic' deployment groups included info about how to do that, here: https://docs.microsoft.com/en-us/azure/devops/pipelines/release/deployment-groups/howto-provision-deployment-group-agents?view=azure-devops#use-the-arm-template-deployment-task Can you tell me how to achive the same thing when working with devops Environments? Thanks, Matt --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 91d0d31f-81ee-c024-db7e-daddbf525f71 * Version Independent ID: 330f1649-386c-d0aa-5f96-b8343a1480d3 * Content: [Environment - Virtual machine resource - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments-virtual-machines?view=azure-devops) * Content Source: [docs/pipelines/process/environments-virtual-machines.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/environments-virtual-machines.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
automatically adding the agent to a vm hello the instructions explain how to use the devops user interface to add a vm resource the add a resource section however we use arm templates and devops pipelines to define our vms so we would really prefer to add the vm resource as part of an automated process without needing manual steps afterwards the docs for classic deployment groups included info about how to do that here can you tell me how to achive the same thing when working with devops environments thanks matt document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
34,643
6,360,034,536
IssuesEvent
2017-07-31 09:02:37
playframework/playframework
https://api.github.com/repos/playframework/playframework
closed
PlayMinimalJava not officially documented
documentation
Play Java 2.6 I've noticed from the some of the Play 2.6 source that there is a new Java Plugin (PlayMinimalJava) which allows applications to remove dependencies against play-forms (and in particular spring), however this plugin isn't formally documented and hence official support for this feature is unclear. Assuming that 'PlayMinimalJava' is officially supported, can we please get official documentation for it.
1.0
PlayMinimalJava not officially documented - Play Java 2.6 I've noticed from the some of the Play 2.6 source that there is a new Java Plugin (PlayMinimalJava) which allows applications to remove dependencies against play-forms (and in particular spring), however this plugin isn't formally documented and hence official support for this feature is unclear. Assuming that 'PlayMinimalJava' is officially supported, can we please get official documentation for it.
non_process
playminimaljava not officially documented play java i ve noticed from the some of the play source that there is a new java plugin playminimaljava which allows applications to remove dependencies against play forms and in particular spring however this plugin isn t formally documented and hence official support for this feature is unclear assuming that playminimaljava is officially supported can we please get official documentation for it
0
79,202
10,113,916,044
IssuesEvent
2019-07-30 17:54:14
material-components/material-components-ios
https://api.github.com/repos/material-components/material-components-ios
closed
Document our support contract for APIs
[Documentation] skill:Technical writing type:Documentation
<!-- Auto-generated content below, do not modify --> --- #### Internal data - Associated internal bug: [b/117179043](http://b/117179043)
2.0
Document our support contract for APIs - <!-- Auto-generated content below, do not modify --> --- #### Internal data - Associated internal bug: [b/117179043](http://b/117179043)
non_process
document our support contract for apis internal data associated internal bug
0
381,547
11,276,564,517
IssuesEvent
2020-01-14 23:38:23
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
Postgres money type breaks visualization (no avg(money))
Administration/Metadata & Sync Database/Postgres Priority:P2 Type:Bug
**Describe the bug** I have a table with a money type and in the x-ray view, the visual component says there is an error. **Logs** 12-11 14:24:35 WARN middleware.process-userland-query :: Query failure {:status :failed, :class org.postgresql.util.PSQLException, :error "ERROR: function avg(money) does not exist\n Hint: No function matches the given name and argument types. You might need to add explicit type casts.\n Position: 259", :stacktrace **To Reproduce** Create a table with some money type in it. Add a row or two (I guess). Use metabase. **Expected behavior** Should not give an error **Severity** Critical, I had to change the type to numeric. **Additional context** From the log, doing a 'avg(money::numeric)' would probably solve it.
1.0
Postgres money type breaks visualization (no avg(money)) - **Describe the bug** I have a table with a money type and in the x-ray view, the visual component says there is an error. **Logs** 12-11 14:24:35 WARN middleware.process-userland-query :: Query failure {:status :failed, :class org.postgresql.util.PSQLException, :error "ERROR: function avg(money) does not exist\n Hint: No function matches the given name and argument types. You might need to add explicit type casts.\n Position: 259", :stacktrace **To Reproduce** Create a table with some money type in it. Add a row or two (I guess). Use metabase. **Expected behavior** Should not give an error **Severity** Critical, I had to change the type to numeric. **Additional context** From the log, doing a 'avg(money::numeric)' would probably solve it.
non_process
postgres money type breaks visualization no avg money describe the bug i have a table with a money type and in the x ray view the visual component says there is an error logs  middleware process userland query query failure status failed class org postgresql util psqlexception error error function avg money does not exist n hint no function matches the given name and argument types you might need to add explicit type casts n position stacktrace to reproduce create a table with some money type in it add a row or two i guess use metabase expected behavior should not give an error severity critical i had to change the type to numeric additional context from the log doing a avg money numeric would probably solve it
0
409,396
27,737,371,386
IssuesEvent
2023-03-15 12:11:07
Arquisoft/lomap_es5a
https://api.github.com/repos/Arquisoft/lomap_es5a
closed
Docs: Actualizar diagrama de secuencia 3 (Crear un punto) apart. nº. 6 (Runtime View)
documentation invalid
## Cambios Cambiar los diagramas 6.1.3, 6.2.3 y 6.3.3, actualizando la arquitectura. En este caso, los datos del punto de interés no se almacenan en base de datos sino en el POD. [Enlace a la documentación](https://arquisoft.github.io/lomap_es5a/#section-runtime-view)
1.0
Docs: Actualizar diagrama de secuencia 3 (Crear un punto) apart. nº. 6 (Runtime View) - ## Cambios Cambiar los diagramas 6.1.3, 6.2.3 y 6.3.3, actualizando la arquitectura. En este caso, los datos del punto de interés no se almacenan en base de datos sino en el POD. [Enlace a la documentación](https://arquisoft.github.io/lomap_es5a/#section-runtime-view)
non_process
docs actualizar diagrama de secuencia crear un punto apart nº runtime view cambios cambiar los diagramas y actualizando la arquitectura en este caso los datos del punto de interés no se almacenan en base de datos sino en el pod
0
58,062
6,572,669,250
IssuesEvent
2017-09-11 04:15:27
brave/browser-laptop
https://api.github.com/repos/brave/browser-laptop
closed
Manual Test Run for Win64 0.18.x Hotfix 2 (Chromium Update)
OS/Windows release-notes/exclude tests
## Installer 1. [x] Check that installer is close to the size of last release. 2. [x] Check signature: If OS Run `spctl --assess --verbose /Applications/Brave.app/` and make sure it returns `accepted`. If Windows right click on the installer exe and go to Properties, go to the Digital Signatures tab and double click on the signature. Make sure it says "The digital signature is OK" in the popup window. 3. [x] Check the libchromiumcontent version in about:brave and make sure it is EXACTLY as expected. ## Printing 1. [x] Test that you can print a PDF ## Widevine/Netflix test 1. [x] Test that you can log into Netflix and start a show. ## Performance test _Each start should take less than 7 seconds_ 1. [x] Enable only sync (new sync group). 2. [ ] Enable only sync with a large sync group (many entries). 3. [x] Enable only payments. 4. [x] Only import a large set of bookmarks. 5. [x] Combine sync, payments, and a large set of bookmarks. ## Ledger 1. [ ] Create a wallet with a value other than $5 selected in the monthly budget dropdown. Click on the 'Add Funds' button and check that Coinbase transactions are blocked. 2. [x] Remove all `ledger-*.json` files from `~/Library/Application\ Support/Brave/`. Go to the Payments tab in about:preferences, enable payments, click on `create wallet`. Check that the `add funds` button appears after a wallet is created. 3. [ ] Click on `add funds` and verify that adding funds through Coinbase increases the account balance. 4. [ ] Repeat the step above but add funds by scanning the QR code in a mobile bitcoin app instead of through Coinbase. 5. [x] Visit nytimes.com for a few seconds and make sure it shows up in the Payments table. 6. [x] Check that disabling payments and enabling them again does not lose state. ## Sync 1. [x] Verify you are able to sync two devices using the secret code 2. [x] Visit a site on device 1 and change shield setting, ensure that the saved site preference is synced to device 2 3. [x] Enable Browsing history sync on device 1, ensure the history is shown on device 2 4. [x] Import/Add bookmarks on device 1, ensure it is synced on device 2 5. [x] Ensure imported bookmark folder structure is maintained on device 2 6. [x] Ensure bookmark favicons are shown after sync ## Data 1. [x] Make sure that data from the last version appears in the new version OK. 2. [x] Test that the previous version's cookies are preserved in the next version. ## Bookmarks 1. [x] Test that creating a bookmark on the bookmarks toolbar works 2. [x] Test that creating a bookmark folder on the bookmarks toolbar works 3. [x] Test that moving a bookmark into a folder by drag and drop on the bookmarks folder works 4. [x] Test that clicking a bookmark in the toolbar loads the bookmark. 5. [x] Test that clicking a bookmark in a bookmark toolbar folder loads the bookmark. ## Context menus 1. [x] Make sure context menu items in the URL bar work 2. [x] Make sure context menu items on content work with no selected text. 3. [x] Make sure context menu items on content work with selected text. 4. [x] Make sure context menu items on content work inside an editable control on `about:styles` (input, textarea, or contenteditable). ## Find on page 1. [x] Ensure search box is shown with shortcut 2. [x] Test successful find 3. [x] Test forward and backward find navigation 4. [x] Test failed find shows 0 results 5. [x] Test match case find ## Geolocation 1. [x] Check that https://developer.mozilla.org/en-US/docs/Web/API/Geolocation/Using_geolocation works ## Site hacks 1. [x] Test https://www.twitch.tv/adobe sub-page loads a video and you can play it ## Downloads 1. [x] Test downloading a file works and that all actions on the download item works. ## Fullscreen 1. [x] Test that entering full screen window works View -> Toggle Full Screen. And exit back (Not Esc). 2. [x] Test that entering HTML5 full screen works. And Esc to go back. (youtube.com) ## Tabs, Pinning and Tear off tabs 1. [x] Test that tabs are pinnable 2. [x] Test that tabs are unpinnable 3. [x] Test that tabs are draggable to same tabset 4. [x] Test that tabs are draggable to alternate tabset 5. [x] Test that tabs can be detached to create a new window 6. [x] Test that you are able to reattach a tab to an existing window 7. [x] Test that you can quickly switch tabs ## Zoom 1. [x] Test zoom in / out shortcut works 2. [x] Test hamburger menu zooms. 3. [x] Test zoom saved when you close the browser and restore on a single site. 4. [x] Test zoom saved when you navigate within a single origin site. 5. [x] Test that navigating to a different origin resets the zoom ## Content tests 1. [x] Go to https://brianbondy.com/ and click on the twitter icon on the top right. Test that context menus work in the new twitter tab. 2. [x] Load twitter and click on a tweet so the popup div shows. Click to dismiss and repeat with another div. Make sure it shows. 3. [x] Go to http://www.bennish.net/web-notifications.html and test that clicking on 'Show' pops up a notification asking for permission. Make sure that clicking 'Deny' leads to no notifications being shown. 4. [x] Go to https://trac.torproject.org/projects/tor/login and make sure that the password can be saved. Make sure the saved password shows up in `about:passwords`. Then reload https://trac.torproject.org/projects/tor/login and make sure the password is autofilled. 5. [ ] Open a github issue and type some misspellings, make sure they are underlined. 6. [ ] Make sure that right clicking on a word with suggestions gives a suggestion and that clicking on the suggestion replaces the text. 7. [x] Make sure that Command + Click (Control + Click on Windows, Control + Click on Ubuntu) on a link opens a new tab but does NOT switch to it. Click on it and make sure it is already loaded. 8. [x] Open an email on http://mail.google.com/ or inbox.google.com and click on a link. Make sure it works. 9. [x] Test that PDF is loaded at http://www.orimi.com/pdf-test.pdf 10. [x] Test that https://mixed-script.badssl.com/ shows up as grey not red (no mixed content scripts are run). ## Flash tests 1. [x] Turn on Flash in about:preferences#security. Test that clicking on 'Install Flash' banner on myspace.com shows a notification to allow Flash and that the banner disappears when 'Allow' is clicked. 2. [x] Test that flash placeholder appears on http://www.homestarrunner.com ## Autofill tests 1. [x] Test that autofill works on http://www.roboform.com/filling-test-all-fields ## Session storage Do not forget to make a backup of your entire `~/Library/Application\ Support/Brave` folder. 1. [x] Temporarily move away your `~/Library/Application\ Support/Brave/session-store-1` and test that clean session storage works. (`%appdata%\Brave in Windows`, `./config/brave` in Ubuntu) 2. [x] Test that windows and tabs restore when closed, including active tab. 3. [x] Move away your entire `~/Library/Application\ Support/Brave` folder (`%appdata%\Brave in Windows`, `./config/brave` in Ubuntu) ## Cookie and Cache 1. [x] Make a backup of your profile, turn on all clearing in preferences and shut down. Make sure when you bring the browser back up everything is gone that is specified. 2. [x] Go to http://samy.pl/evercookie/ and set an evercookie. Check that going to prefs, clearing site data and cache, and going back to the Evercookie site does not remember the old evercookie value. ## Update tests 1. [ ] Test that updating using `BRAVE_UPDATE_VERSION=0.8.3` env variable works correctly.
1.0
Manual Test Run for Win64 0.18.x Hotfix 2 (Chromium Update) - ## Installer 1. [x] Check that installer is close to the size of last release. 2. [x] Check signature: If OS Run `spctl --assess --verbose /Applications/Brave.app/` and make sure it returns `accepted`. If Windows right click on the installer exe and go to Properties, go to the Digital Signatures tab and double click on the signature. Make sure it says "The digital signature is OK" in the popup window. 3. [x] Check the libchromiumcontent version in about:brave and make sure it is EXACTLY as expected. ## Printing 1. [x] Test that you can print a PDF ## Widevine/Netflix test 1. [x] Test that you can log into Netflix and start a show. ## Performance test _Each start should take less than 7 seconds_ 1. [x] Enable only sync (new sync group). 2. [ ] Enable only sync with a large sync group (many entries). 3. [x] Enable only payments. 4. [x] Only import a large set of bookmarks. 5. [x] Combine sync, payments, and a large set of bookmarks. ## Ledger 1. [ ] Create a wallet with a value other than $5 selected in the monthly budget dropdown. Click on the 'Add Funds' button and check that Coinbase transactions are blocked. 2. [x] Remove all `ledger-*.json` files from `~/Library/Application\ Support/Brave/`. Go to the Payments tab in about:preferences, enable payments, click on `create wallet`. Check that the `add funds` button appears after a wallet is created. 3. [ ] Click on `add funds` and verify that adding funds through Coinbase increases the account balance. 4. [ ] Repeat the step above but add funds by scanning the QR code in a mobile bitcoin app instead of through Coinbase. 5. [x] Visit nytimes.com for a few seconds and make sure it shows up in the Payments table. 6. [x] Check that disabling payments and enabling them again does not lose state. ## Sync 1. [x] Verify you are able to sync two devices using the secret code 2. [x] Visit a site on device 1 and change shield setting, ensure that the saved site preference is synced to device 2 3. [x] Enable Browsing history sync on device 1, ensure the history is shown on device 2 4. [x] Import/Add bookmarks on device 1, ensure it is synced on device 2 5. [x] Ensure imported bookmark folder structure is maintained on device 2 6. [x] Ensure bookmark favicons are shown after sync ## Data 1. [x] Make sure that data from the last version appears in the new version OK. 2. [x] Test that the previous version's cookies are preserved in the next version. ## Bookmarks 1. [x] Test that creating a bookmark on the bookmarks toolbar works 2. [x] Test that creating a bookmark folder on the bookmarks toolbar works 3. [x] Test that moving a bookmark into a folder by drag and drop on the bookmarks folder works 4. [x] Test that clicking a bookmark in the toolbar loads the bookmark. 5. [x] Test that clicking a bookmark in a bookmark toolbar folder loads the bookmark. ## Context menus 1. [x] Make sure context menu items in the URL bar work 2. [x] Make sure context menu items on content work with no selected text. 3. [x] Make sure context menu items on content work with selected text. 4. [x] Make sure context menu items on content work inside an editable control on `about:styles` (input, textarea, or contenteditable). ## Find on page 1. [x] Ensure search box is shown with shortcut 2. [x] Test successful find 3. [x] Test forward and backward find navigation 4. [x] Test failed find shows 0 results 5. [x] Test match case find ## Geolocation 1. [x] Check that https://developer.mozilla.org/en-US/docs/Web/API/Geolocation/Using_geolocation works ## Site hacks 1. [x] Test https://www.twitch.tv/adobe sub-page loads a video and you can play it ## Downloads 1. [x] Test downloading a file works and that all actions on the download item works. ## Fullscreen 1. [x] Test that entering full screen window works View -> Toggle Full Screen. And exit back (Not Esc). 2. [x] Test that entering HTML5 full screen works. And Esc to go back. (youtube.com) ## Tabs, Pinning and Tear off tabs 1. [x] Test that tabs are pinnable 2. [x] Test that tabs are unpinnable 3. [x] Test that tabs are draggable to same tabset 4. [x] Test that tabs are draggable to alternate tabset 5. [x] Test that tabs can be detached to create a new window 6. [x] Test that you are able to reattach a tab to an existing window 7. [x] Test that you can quickly switch tabs ## Zoom 1. [x] Test zoom in / out shortcut works 2. [x] Test hamburger menu zooms. 3. [x] Test zoom saved when you close the browser and restore on a single site. 4. [x] Test zoom saved when you navigate within a single origin site. 5. [x] Test that navigating to a different origin resets the zoom ## Content tests 1. [x] Go to https://brianbondy.com/ and click on the twitter icon on the top right. Test that context menus work in the new twitter tab. 2. [x] Load twitter and click on a tweet so the popup div shows. Click to dismiss and repeat with another div. Make sure it shows. 3. [x] Go to http://www.bennish.net/web-notifications.html and test that clicking on 'Show' pops up a notification asking for permission. Make sure that clicking 'Deny' leads to no notifications being shown. 4. [x] Go to https://trac.torproject.org/projects/tor/login and make sure that the password can be saved. Make sure the saved password shows up in `about:passwords`. Then reload https://trac.torproject.org/projects/tor/login and make sure the password is autofilled. 5. [ ] Open a github issue and type some misspellings, make sure they are underlined. 6. [ ] Make sure that right clicking on a word with suggestions gives a suggestion and that clicking on the suggestion replaces the text. 7. [x] Make sure that Command + Click (Control + Click on Windows, Control + Click on Ubuntu) on a link opens a new tab but does NOT switch to it. Click on it and make sure it is already loaded. 8. [x] Open an email on http://mail.google.com/ or inbox.google.com and click on a link. Make sure it works. 9. [x] Test that PDF is loaded at http://www.orimi.com/pdf-test.pdf 10. [x] Test that https://mixed-script.badssl.com/ shows up as grey not red (no mixed content scripts are run). ## Flash tests 1. [x] Turn on Flash in about:preferences#security. Test that clicking on 'Install Flash' banner on myspace.com shows a notification to allow Flash and that the banner disappears when 'Allow' is clicked. 2. [x] Test that flash placeholder appears on http://www.homestarrunner.com ## Autofill tests 1. [x] Test that autofill works on http://www.roboform.com/filling-test-all-fields ## Session storage Do not forget to make a backup of your entire `~/Library/Application\ Support/Brave` folder. 1. [x] Temporarily move away your `~/Library/Application\ Support/Brave/session-store-1` and test that clean session storage works. (`%appdata%\Brave in Windows`, `./config/brave` in Ubuntu) 2. [x] Test that windows and tabs restore when closed, including active tab. 3. [x] Move away your entire `~/Library/Application\ Support/Brave` folder (`%appdata%\Brave in Windows`, `./config/brave` in Ubuntu) ## Cookie and Cache 1. [x] Make a backup of your profile, turn on all clearing in preferences and shut down. Make sure when you bring the browser back up everything is gone that is specified. 2. [x] Go to http://samy.pl/evercookie/ and set an evercookie. Check that going to prefs, clearing site data and cache, and going back to the Evercookie site does not remember the old evercookie value. ## Update tests 1. [ ] Test that updating using `BRAVE_UPDATE_VERSION=0.8.3` env variable works correctly.
non_process
manual test run for x hotfix chromium update installer check that installer is close to the size of last release check signature if os run spctl assess verbose applications brave app and make sure it returns accepted if windows right click on the installer exe and go to properties go to the digital signatures tab and double click on the signature make sure it says the digital signature is ok in the popup window check the libchromiumcontent version in about brave and make sure it is exactly as expected printing test that you can print a pdf widevine netflix test test that you can log into netflix and start a show performance test each start should take less than seconds enable only sync new sync group enable only sync with a large sync group many entries enable only payments only import a large set of bookmarks combine sync payments and a large set of bookmarks ledger create a wallet with a value other than selected in the monthly budget dropdown click on the add funds button and check that coinbase transactions are blocked remove all ledger json files from library application support brave go to the payments tab in about preferences enable payments click on create wallet check that the add funds button appears after a wallet is created click on add funds and verify that adding funds through coinbase increases the account balance repeat the step above but add funds by scanning the qr code in a mobile bitcoin app instead of through coinbase visit nytimes com for a few seconds and make sure it shows up in the payments table check that disabling payments and enabling them again does not lose state sync verify you are able to sync two devices using the secret code visit a site on device and change shield setting ensure that the saved site preference is synced to device enable browsing history sync on device ensure the history is shown on device import add bookmarks on device ensure it is synced on device ensure imported bookmark folder structure is maintained on device ensure bookmark favicons are shown after sync data make sure that data from the last version appears in the new version ok test that the previous version s cookies are preserved in the next version bookmarks test that creating a bookmark on the bookmarks toolbar works test that creating a bookmark folder on the bookmarks toolbar works test that moving a bookmark into a folder by drag and drop on the bookmarks folder works test that clicking a bookmark in the toolbar loads the bookmark test that clicking a bookmark in a bookmark toolbar folder loads the bookmark context menus make sure context menu items in the url bar work make sure context menu items on content work with no selected text make sure context menu items on content work with selected text make sure context menu items on content work inside an editable control on about styles input textarea or contenteditable find on page ensure search box is shown with shortcut test successful find test forward and backward find navigation test failed find shows results test match case find geolocation check that works site hacks test sub page loads a video and you can play it downloads test downloading a file works and that all actions on the download item works fullscreen test that entering full screen window works view toggle full screen and exit back not esc test that entering full screen works and esc to go back youtube com tabs pinning and tear off tabs test that tabs are pinnable test that tabs are unpinnable test that tabs are draggable to same tabset test that tabs are draggable to alternate tabset test that tabs can be detached to create a new window test that you are able to reattach a tab to an existing window test that you can quickly switch tabs zoom test zoom in out shortcut works test hamburger menu zooms test zoom saved when you close the browser and restore on a single site test zoom saved when you navigate within a single origin site test that navigating to a different origin resets the zoom content tests go to and click on the twitter icon on the top right test that context menus work in the new twitter tab load twitter and click on a tweet so the popup div shows click to dismiss and repeat with another div make sure it shows go to and test that clicking on show pops up a notification asking for permission make sure that clicking deny leads to no notifications being shown go to and make sure that the password can be saved make sure the saved password shows up in about passwords then reload and make sure the password is autofilled open a github issue and type some misspellings make sure they are underlined make sure that right clicking on a word with suggestions gives a suggestion and that clicking on the suggestion replaces the text make sure that command click control click on windows control click on ubuntu on a link opens a new tab but does not switch to it click on it and make sure it is already loaded open an email on or inbox google com and click on a link make sure it works test that pdf is loaded at test that shows up as grey not red no mixed content scripts are run flash tests turn on flash in about preferences security test that clicking on install flash banner on myspace com shows a notification to allow flash and that the banner disappears when allow is clicked test that flash placeholder appears on autofill tests test that autofill works on session storage do not forget to make a backup of your entire library application support brave folder temporarily move away your library application support brave session store and test that clean session storage works appdata brave in windows config brave in ubuntu test that windows and tabs restore when closed including active tab move away your entire library application support brave folder appdata brave in windows config brave in ubuntu cookie and cache make a backup of your profile turn on all clearing in preferences and shut down make sure when you bring the browser back up everything is gone that is specified go to and set an evercookie check that going to prefs clearing site data and cache and going back to the evercookie site does not remember the old evercookie value update tests test that updating using brave update version env variable works correctly
0
270,467
8,460,828,050
IssuesEvent
2018-10-22 19:58:48
robotframework/robotframework
https://api.github.com/repos/robotframework/robotframework
closed
Deprecate using for loops with other separators than exact `IN`, `IN RANGE`, `IN ZIP` and `IN ENUMERATE`
deprecation enhancement priority: medium
When creating a for loop in Robot Framework data, the separator between the loop variables and the loop items is, depending on the for loop flavor, either `IN`, `IN RANGE`, `IN ZIP` or `IN ENUMERATE`. Currently this separator is both case and space insensitive, so also values like `in`, `INZIP` or `i n r a n g e` would all work. This makes it harder to create tooling that can reliably parse Robot's test data, and also complicates Robot's own test data parser. There are no real benefits. Due to the problems and lack of benefits, a decision has been made to require the for loop markers to use exactly format `IN`, `IN RANGE`, `IN ZIP` and `IN ENUMERATE`. An added benefit is that the all uppercase style matches the new possibility to create a for loop using all uppercase `FOR` without the extra colon (#2990). This idea was originally discussed as part of #2783 that proposed also other enhancements to the for loop syntax.
1.0
Deprecate using for loops with other separators than exact `IN`, `IN RANGE`, `IN ZIP` and `IN ENUMERATE` - When creating a for loop in Robot Framework data, the separator between the loop variables and the loop items is, depending on the for loop flavor, either `IN`, `IN RANGE`, `IN ZIP` or `IN ENUMERATE`. Currently this separator is both case and space insensitive, so also values like `in`, `INZIP` or `i n r a n g e` would all work. This makes it harder to create tooling that can reliably parse Robot's test data, and also complicates Robot's own test data parser. There are no real benefits. Due to the problems and lack of benefits, a decision has been made to require the for loop markers to use exactly format `IN`, `IN RANGE`, `IN ZIP` and `IN ENUMERATE`. An added benefit is that the all uppercase style matches the new possibility to create a for loop using all uppercase `FOR` without the extra colon (#2990). This idea was originally discussed as part of #2783 that proposed also other enhancements to the for loop syntax.
non_process
deprecate using for loops with other separators than exact in in range in zip and in enumerate when creating a for loop in robot framework data the separator between the loop variables and the loop items is depending on the for loop flavor either in in range in zip or in enumerate currently this separator is both case and space insensitive so also values like in inzip or i n r a n g e would all work this makes it harder to create tooling that can reliably parse robot s test data and also complicates robot s own test data parser there are no real benefits due to the problems and lack of benefits a decision has been made to require the for loop markers to use exactly format in in range in zip and in enumerate an added benefit is that the all uppercase style matches the new possibility to create a for loop using all uppercase for without the extra colon this idea was originally discussed as part of that proposed also other enhancements to the for loop syntax
0
18,338
24,460,272,208
IssuesEvent
2022-10-07 10:28:24
Open-Data-Product-Initiative/open-data-product-spec
https://api.github.com/repos/Open-Data-Product-Initiative/open-data-product-spec
closed
dataAccess as a separate component and renaming
enhancement processed
dataAccess and technical implementation should be kept separate for clarity. Thus, dataOps (formerly dataPipeline) has been implemented as a separate component.
1.0
dataAccess as a separate component and renaming - dataAccess and technical implementation should be kept separate for clarity. Thus, dataOps (formerly dataPipeline) has been implemented as a separate component.
process
dataaccess as a separate component and renaming dataaccess and technical implementation should be kept separate for clarity thus dataops formerly datapipeline has been implemented as a separate component
1