Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 1 744 | labels stringlengths 4 574 | body stringlengths 9 211k | index stringclasses 10 values | text_combine stringlengths 96 211k | label stringclasses 2 values | text stringlengths 96 188k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
1,725 | 4,381,729,581 | IssuesEvent | 2016-08-06 12:21:14 | mlr-org/mlr | https://api.github.com/repos/mlr-org/mlr | closed | impute() does not handle non missing values in columns | bug enhancement preprocessing | impute should be able to do nothing if no missing values are present:
```r
library(mlr)
iris.m = iris
imp = impute(data = iris.m,
target = "Species",
cols = list(Sepal.Width = imputeLearner(makeLearner("regr.rpart"))))
#Error in impute(data = iris.m, target = "Species", cols = list(Sepal.Width = imputeLearner(makeLearner("regr.rpart")))) :
# Assertion on 'newdata' failed: Must have at least 1 rows, but has 0 rows.
| 1.0 | impute() does not handle non missing values in columns - impute should be able to do nothing if no missing values are present:
```r
library(mlr)
iris.m = iris
imp = impute(data = iris.m,
target = "Species",
cols = list(Sepal.Width = imputeLearner(makeLearner("regr.rpart"))))
#Error in impute(data = iris.m, target = "Species", cols = list(Sepal.Width = imputeLearner(makeLearner("regr.rpart")))) :
# Assertion on 'newdata' failed: Must have at least 1 rows, but has 0 rows.
| process | impute does not handle non missing values in columns impute should be able to do nothing if no missing values are present r library mlr iris m iris imp impute data iris m target species cols list sepal width imputelearner makelearner regr rpart error in impute data iris m target species cols list sepal width imputelearner makelearner regr rpart assertion on newdata failed must have at least rows but has rows | 1 |
373,978 | 11,053,615,754 | IssuesEvent | 2019-12-10 11:46:17 | wso2/product-is | https://api.github.com/repos/wso2/product-is | opened | Cannot set Access URL for a service provider through the management console | Priority/High Severity/Critical Type/Bug | Steps to reproduce.
**Case I**
1. Create a service provider named as `test`
2. Set `Access URL` as `https://www.google.com/` in the next page.
3. Click on update.
(Update seems to complete without any errors and directed to SP listing page)
4. Click on edit of the created `test` service provider.
5. `Access URL` shown as empty.
**Case II**
1. Create a service provider named as `test`
2. Set `Access URL` as `https://www.google.com/` in the next page.
3. Click on `Discoverable Application` checkbox to make it enable.
4. Click on update.
Update get failure with the following error message,
`Error while updating Service Provider : Invalid application configuration for application: 'test' of tenantDomain: carbon.super. A valid accessURL needs to be defined if an application is marked as discoverable.` | 1.0 | Cannot set Access URL for a service provider through the management console - Steps to reproduce.
**Case I**
1. Create a service provider named as `test`
2. Set `Access URL` as `https://www.google.com/` in the next page.
3. Click on update.
(Update seems to complete without any errors and directed to SP listing page)
4. Click on edit of the created `test` service provider.
5. `Access URL` shown as empty.
**Case II**
1. Create a service provider named as `test`
2. Set `Access URL` as `https://www.google.com/` in the next page.
3. Click on `Discoverable Application` checkbox to make it enable.
4. Click on update.
Update get failure with the following error message,
`Error while updating Service Provider : Invalid application configuration for application: 'test' of tenantDomain: carbon.super. A valid accessURL needs to be defined if an application is marked as discoverable.` | non_process | cannot set access url for a service provider through the management console steps to reproduce case i create a service provider named as test set access url as in the next page click on update update seems to complete without any errors and directed to sp listing page click on edit of the created test service provider access url shown as empty case ii create a service provider named as test set access url as in the next page click on discoverable application checkbox to make it enable click on update update get failure with the following error message error while updating service provider invalid application configuration for application test of tenantdomain carbon super a valid accessurl needs to be defined if an application is marked as discoverable | 0 |
42,927 | 17,371,996,106 | IssuesEvent | 2021-07-30 15:08:34 | Azure/azure-cli | https://api.github.com/repos/Azure/azure-cli | closed | az webapp deployment source config-zip: Deployment endpoint responded with status code 202. But fail to run? | Service Attention Web Apps | I'm running `az webapp deployment source config-zip --resource-group XX --name XX --src XX-package.zip` on a Linux Web Service instance.
It's a node project. My zip contain a package.json with a npm start command that I expect Azure to run.
```
...
Getting scm site credentials for zip deployment
Starting zip deployment. This operation can take a while to complete ...
Deployment endpoint responded with status code 202
{
"active": true,
"author": "N/A",
"author_email": "N/A",
"complete": true,
"deployer": "Push-Deployer",
"end_time": "2021-07-30T09:41:17.9438556Z",
"id": "a8063d7cc33943e9b852cd5a12a1184d",
"is_readonly": true,
"is_temp": false,
"last_success_end_time": "2021-07-30T09:41:17.9438556Z",
"log_url": "https:/xx.scm.azurewebsites.net/api/deployments/latest/log",
"message": "Created via a push deployment",
"progress": "",
"received_time": "2021-07-30T09:41:13.0359297Z",
"site_name": "xx",
"start_time": "2021-07-30T09:41:14.2349318Z",
"status": 4,
"status_text": "",
"url": "https://xx.scm.azurewebsites.net/api/deployments/latest"
}
```
But then? The app do not seem to run.
| 1.0 | az webapp deployment source config-zip: Deployment endpoint responded with status code 202. But fail to run? - I'm running `az webapp deployment source config-zip --resource-group XX --name XX --src XX-package.zip` on a Linux Web Service instance.
It's a node project. My zip contain a package.json with a npm start command that I expect Azure to run.
```
...
Getting scm site credentials for zip deployment
Starting zip deployment. This operation can take a while to complete ...
Deployment endpoint responded with status code 202
{
"active": true,
"author": "N/A",
"author_email": "N/A",
"complete": true,
"deployer": "Push-Deployer",
"end_time": "2021-07-30T09:41:17.9438556Z",
"id": "a8063d7cc33943e9b852cd5a12a1184d",
"is_readonly": true,
"is_temp": false,
"last_success_end_time": "2021-07-30T09:41:17.9438556Z",
"log_url": "https:/xx.scm.azurewebsites.net/api/deployments/latest/log",
"message": "Created via a push deployment",
"progress": "",
"received_time": "2021-07-30T09:41:13.0359297Z",
"site_name": "xx",
"start_time": "2021-07-30T09:41:14.2349318Z",
"status": 4,
"status_text": "",
"url": "https://xx.scm.azurewebsites.net/api/deployments/latest"
}
```
But then? The app do not seem to run.
| non_process | az webapp deployment source config zip deployment endpoint responded with status code but fail to run i m running az webapp deployment source config zip resource group xx name xx src xx package zip on a linux web service instance it s a node project my zip contain a package json with a npm start command that i expect azure to run getting scm site credentials for zip deployment starting zip deployment this operation can take a while to complete deployment endpoint responded with status code active true author n a author email n a complete true deployer push deployer end time id is readonly true is temp false last success end time log url https xx scm azurewebsites net api deployments latest log message created via a push deployment progress received time site name xx start time status status text url but then the app do not seem to run | 0 |
2,253 | 5,088,654,207 | IssuesEvent | 2017-01-01 00:06:33 | sw4j-org/tool-jpa-processor | https://api.github.com/repos/sw4j-org/tool-jpa-processor | opened | Handle @SecondaryTable Annotation | annotation processor task | Handle the `@SecondaryTable` annotation for an entity.
See [JSR 338: Java Persistence API, Version 2.1](http://download.oracle.com/otn-pub/jcp/persistence-2_1-fr-eval-spec/JavaPersistence.pdf)
- 11.1.46 SecondaryTable Annotation
| 1.0 | Handle @SecondaryTable Annotation - Handle the `@SecondaryTable` annotation for an entity.
See [JSR 338: Java Persistence API, Version 2.1](http://download.oracle.com/otn-pub/jcp/persistence-2_1-fr-eval-spec/JavaPersistence.pdf)
- 11.1.46 SecondaryTable Annotation
| process | handle secondarytable annotation handle the secondarytable annotation for an entity see secondarytable annotation | 1 |
27,056 | 6,813,308,172 | IssuesEvent | 2017-11-06 08:45:09 | BTDF/DeploymentFramework | https://api.github.com/repos/BTDF/DeploymentFramework | closed | Feature: More extensibilty on WIX targets | CodePlexMigrationInitiated enhancement Impact: Low MSI Creation and WiX Release 5.5 | Are you considering to update the framework to provide extensibility to WIX targets? Here are few things I am looking at right now...
I am against to change the framework files for custom menu items as we have one build server for several applications. Changing framework files screws up other applications. Also, any future updates to framework will loose the customization. If you can provide more extensibility to WIX (i know right now this is in .js file) ie, moving wix xml-composition to targets with extensibility, we will have lot of flexibility in customizing the MSI.
Use of a property (something like InstallerFileName) for MSI name. [TFA -- available in BTDF v5.0]
Setting File/Product Versions for assemblies (Alternatively I am changing the each project file before compile the solution) [TFA -- this must be done before compilation, so it cannot be handled by the Deployment Framework]
Ability to set custom properties (like Version, Source, BuildServer) on MSI file. I don't know if that is possible with WIX. Right now I am using 'PackageComments' with concatenating all properties I need.
#### This work item was migrated from CodePlex
CodePlex work item ID: '8159'
Assigned to: 'tfabraham'
Vote count: '3'
| 1.0 | Feature: More extensibilty on WIX targets - Are you considering to update the framework to provide extensibility to WIX targets? Here are few things I am looking at right now...
I am against to change the framework files for custom menu items as we have one build server for several applications. Changing framework files screws up other applications. Also, any future updates to framework will loose the customization. If you can provide more extensibility to WIX (i know right now this is in .js file) ie, moving wix xml-composition to targets with extensibility, we will have lot of flexibility in customizing the MSI.
Use of a property (something like InstallerFileName) for MSI name. [TFA -- available in BTDF v5.0]
Setting File/Product Versions for assemblies (Alternatively I am changing the each project file before compile the solution) [TFA -- this must be done before compilation, so it cannot be handled by the Deployment Framework]
Ability to set custom properties (like Version, Source, BuildServer) on MSI file. I don't know if that is possible with WIX. Right now I am using 'PackageComments' with concatenating all properties I need.
#### This work item was migrated from CodePlex
CodePlex work item ID: '8159'
Assigned to: 'tfabraham'
Vote count: '3'
| non_process | feature more extensibilty on wix targets are you considering to update the framework to provide extensibility to wix targets here are few things i am looking at right now i am against to change the framework files for custom menu items as we have one build server for several applications changing framework files screws up other applications also any future updates to framework will loose the customization if you can provide more extensibility to wix i know right now this is in js file ie moving wix xml composition to targets with extensibility we will have lot of flexibility in customizing the msi use of a property something like installerfilename for msi name setting file product versions for assemblies alternatively i am changing the each project file before compile the solution ability to set custom properties like version source buildserver on msi file i don t know if that is possible with wix right now i am using packagecomments with concatenating all properties i need this work item was migrated from codeplex codeplex work item id assigned to tfabraham vote count | 0 |
4,296 | 7,192,448,899 | IssuesEvent | 2018-02-03 03:41:35 | Great-Hill-Corporation/quickBlocks | https://api.github.com/repos/Great-Hill-Corporation/quickBlocks | closed | ethslurp needs an Etherscan API key to work. | apps-ethslurp status-inprocess type-enhancement | We need to test the process of how a user is warned that they need an Etherscan API key and how to go about doing that. Referring to #248. | 1.0 | ethslurp needs an Etherscan API key to work. - We need to test the process of how a user is warned that they need an Etherscan API key and how to go about doing that. Referring to #248. | process | ethslurp needs an etherscan api key to work we need to test the process of how a user is warned that they need an etherscan api key and how to go about doing that referring to | 1 |
17,553 | 23,367,332,901 | IssuesEvent | 2022-08-10 16:28:04 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | Process: maintainer involvement in triaging issues | Process | This is a re-hash of the initial discussion in #46759.
Following some offline discussion with @nashif, we agreed that it would be worthwhile to start sending requests for action from maintainers to periodically triage open bugs and issues in parallel to the ongoing broader discussions about release criteria that turned into https://github.com/zephyrproject-rtos/zephyr/issues/38566.
To resolve in particular:
- Triaging of issues, especially low priority issues: what to use? "known issue" label? "milestone" metadata?
- Any other triaging for initial effort? | 1.0 | Process: maintainer involvement in triaging issues - This is a re-hash of the initial discussion in #46759.
Following some offline discussion with @nashif, we agreed that it would be worthwhile to start sending requests for action from maintainers to periodically triage open bugs and issues in parallel to the ongoing broader discussions about release criteria that turned into https://github.com/zephyrproject-rtos/zephyr/issues/38566.
To resolve in particular:
- Triaging of issues, especially low priority issues: what to use? "known issue" label? "milestone" metadata?
- Any other triaging for initial effort? | process | process maintainer involvement in triaging issues this is a re hash of the initial discussion in following some offline discussion with nashif we agreed that it would be worthwhile to start sending requests for action from maintainers to periodically triage open bugs and issues in parallel to the ongoing broader discussions about release criteria that turned into to resolve in particular triaging of issues especially low priority issues what to use known issue label milestone metadata any other triaging for initial effort | 1 |
145,445 | 5,575,654,671 | IssuesEvent | 2017-03-28 03:00:10 | YaleSTC/vesta | https://api.github.com/repos/YaleSTC/vesta | opened | Improve enrollment errors / flashes | complexity: 3 priority: 4 type: bug | * [ ] Silent failures when users already exist (check for no flashes?)
* [ ] Show successes/failures even after timeout | 1.0 | Improve enrollment errors / flashes - * [ ] Silent failures when users already exist (check for no flashes?)
* [ ] Show successes/failures even after timeout | non_process | improve enrollment errors flashes silent failures when users already exist check for no flashes show successes failures even after timeout | 0 |
9,671 | 12,676,767,184 | IssuesEvent | 2020-06-19 06:11:45 | feast-dev/feast | https://api.github.com/repos/feast-dev/feast | closed | Link to BigQuery permissions for service account is broken | kind/process kind/question | The helm charts README mentioned that
```
In production, it is advised to give only the required permissions for the the service account, versus editor role which is very permissive.```
however the link to the required permissions is broken.
Would be great to understand which permissions are required in production environment.
| 1.0 | Link to BigQuery permissions for service account is broken - The helm charts README mentioned that
```
In production, it is advised to give only the required permissions for the the service account, versus editor role which is very permissive.```
however the link to the required permissions is broken.
Would be great to understand which permissions are required in production environment.
| process | link to bigquery permissions for service account is broken the helm charts readme mentioned that in production it is advised to give only the required permissions for the the service account versus editor role which is very permissive however the link to the required permissions is broken would be great to understand which permissions are required in production environment | 1 |
197,894 | 14,948,419,258 | IssuesEvent | 2021-01-26 10:03:16 | pints-team/functional-testing | https://api.github.com/repos/pints-team/functional-testing | opened | Add "simulation-based calibration" tests | new test | https://github.com/pints-team/pints/issues/843
https://arxiv.org/pdf/1804.06788.pdf
Sean Talts, Michael Betancourt, Daniel Simpson, Aki Vehtari, Andrew
Gelman (2020) Validating Bayesian Inference Algorithms with Simulation-Based Calibration | 1.0 | Add "simulation-based calibration" tests - https://github.com/pints-team/pints/issues/843
https://arxiv.org/pdf/1804.06788.pdf
Sean Talts, Michael Betancourt, Daniel Simpson, Aki Vehtari, Andrew
Gelman (2020) Validating Bayesian Inference Algorithms with Simulation-Based Calibration | non_process | add simulation based calibration tests sean talts michael betancourt daniel simpson aki vehtari andrew gelman validating bayesian inference algorithms with simulation based calibration | 0 |
22,296 | 30,851,719,419 | IssuesEvent | 2023-08-02 17:15:38 | googleapis/python-bigquery | https://api.github.com/repos/googleapis/python-bigquery | closed | Please add tests for the BigQuery v1 branch | api: bigquery status: will not fix type: process | There is an issue reported that `google-api-core` version 2.11 breaks compatibility with the `google-cloud-bigquery` v1 branch. See https://github.com/googleapis/python-api-core/issues/477 . Creating an issue to enable tests for the v1 branch if possible. | 1.0 | Please add tests for the BigQuery v1 branch - There is an issue reported that `google-api-core` version 2.11 breaks compatibility with the `google-cloud-bigquery` v1 branch. See https://github.com/googleapis/python-api-core/issues/477 . Creating an issue to enable tests for the v1 branch if possible. | process | please add tests for the bigquery branch there is an issue reported that google api core version breaks compatibility with the google cloud bigquery branch see creating an issue to enable tests for the branch if possible | 1 |
19,559 | 25,883,173,720 | IssuesEvent | 2022-12-14 12:46:15 | gradle/gradle | https://api.github.com/repos/gradle/gradle | closed | Module export rules needed with java17 defined via options.compilerArg are not applied to annotation processor | a:bug in:modular-java in:annotation-processing | <!--- Provide a brief summary of the issue in the title above -->
### Expected Behavior
There should be a way to specify module export rules for those to apply to annotation processor
### Current Behavior
only able to set module rules via global gradle JVM gradle.properies
### Context
We have a happy codebase which works fine under java 11. JDK17 imposes additional module export requirments around javac internal API usages by our annotation processor, we are unable to set such rules locally just for specific compilation task. As a workaround we set export rules globally. Both problematic and workaround versions are attached.
How we believe it export rules should be added does not get the job done see attached project
```
compileJava {
options.compilerArgs += ["--add-exports=jdk.compiler/com.sun.tools.javac.api=ALL-UNNAMED",
"--add-exports=jdk.compiler/com.sun.tools.javac.file=ALL-UNNAMED",
"--add-exports=jdk.compiler/com.sun.tools.javac.main=ALL-UNNAMED",
"--add-exports=jdk.compiler/com.sun.tools.javac.model=ALL-UNNAMED",
"--add-exports=jdk.compiler/com.sun.tools.javac.parser=ALL-UNNAMED",
"--add-exports=jdk.compiler/com.sun.tools.javac.processing=ALL-UNNAMED",
"--add-exports=jdk.compiler/com.sun.tools.javac.tree=ALL-UNNAMED",
"--add-exports=jdk.compiler/com.sun.tools.javac.util=ALL-UNNAMED",
"--add-exports=jdk.compiler/com.sun.tools.javac.code=ALL-UNNAMED",
"--add-exports=jdk.compiler/com.sun.tools.javac.comp=ALL-UNNAMED",
"--add-exports=jdk.compiler/sun.nio.ch=ALL-UNNAMED"]
}
```
how we work around the issue
```
org.gradle.jvmargs= --add-exports=jdk.compiler/com.sun.tools.javac.api=ALL-UNNAMED \
--add-exports=jdk.compiler/com.sun.tools.javac.file=ALL-UNNAMED \
--add-exports=jdk.compiler/com.sun.tools.javac.main=ALL-UNNAMED \
--add-exports=jdk.compiler/com.sun.tools.javac.model=ALL-UNNAMED \
--add-exports=jdk.compiler/com.sun.tools.javac.parser=ALL-UNNAMED \
--add-exports=jdk.compiler/com.sun.tools.javac.processing=ALL-UNNAMED \
--add-exports=jdk.compiler/com.sun.tools.javac.tree=ALL-UNNAMED \
--add-exports=jdk.compiler/com.sun.tools.javac.util=ALL-UNNAMED \
--add-exports=jdk.compiler/com.sun.tools.javac.code=ALL-UNNAMED \
--add-exports=jdk.compiler/com.sun.tools.javac.comp=ALL-UNNAMED \
--add-exports=jdk.compiler/sun.nio.ch=ALL-UNNAMED \
--add-exports=java.base/sun.nio.ch=ALL-UNNAMED
```
[devexperts-17-workaround.zip](https://github.com/gradle/gradle/files/9822041/devexperts-17-workaround.zip)
[devexperts-17-issue.zip](https://github.com/gradle/gradle/files/9822042/devexperts-17-issue.zip)
This adds steps-to-reproduce for https://github.com/gradle/gradle/issues/21310
This is a duplicate of https://github.com/gradle/gradle/issues/22065 where we have failed to express ourselves
| 1.0 | Module export rules needed with java17 defined via options.compilerArg are not applied to annotation processor - <!--- Provide a brief summary of the issue in the title above -->
### Expected Behavior
There should be a way to specify module export rules for those to apply to annotation processor
### Current Behavior
only able to set module rules via global gradle JVM gradle.properies
### Context
We have a happy codebase which works fine under java 11. JDK17 imposes additional module export requirments around javac internal API usages by our annotation processor, we are unable to set such rules locally just for specific compilation task. As a workaround we set export rules globally. Both problematic and workaround versions are attached.
How we believe it export rules should be added does not get the job done see attached project
```
compileJava {
options.compilerArgs += ["--add-exports=jdk.compiler/com.sun.tools.javac.api=ALL-UNNAMED",
"--add-exports=jdk.compiler/com.sun.tools.javac.file=ALL-UNNAMED",
"--add-exports=jdk.compiler/com.sun.tools.javac.main=ALL-UNNAMED",
"--add-exports=jdk.compiler/com.sun.tools.javac.model=ALL-UNNAMED",
"--add-exports=jdk.compiler/com.sun.tools.javac.parser=ALL-UNNAMED",
"--add-exports=jdk.compiler/com.sun.tools.javac.processing=ALL-UNNAMED",
"--add-exports=jdk.compiler/com.sun.tools.javac.tree=ALL-UNNAMED",
"--add-exports=jdk.compiler/com.sun.tools.javac.util=ALL-UNNAMED",
"--add-exports=jdk.compiler/com.sun.tools.javac.code=ALL-UNNAMED",
"--add-exports=jdk.compiler/com.sun.tools.javac.comp=ALL-UNNAMED",
"--add-exports=jdk.compiler/sun.nio.ch=ALL-UNNAMED"]
}
```
how we work around the issue
```
org.gradle.jvmargs= --add-exports=jdk.compiler/com.sun.tools.javac.api=ALL-UNNAMED \
--add-exports=jdk.compiler/com.sun.tools.javac.file=ALL-UNNAMED \
--add-exports=jdk.compiler/com.sun.tools.javac.main=ALL-UNNAMED \
--add-exports=jdk.compiler/com.sun.tools.javac.model=ALL-UNNAMED \
--add-exports=jdk.compiler/com.sun.tools.javac.parser=ALL-UNNAMED \
--add-exports=jdk.compiler/com.sun.tools.javac.processing=ALL-UNNAMED \
--add-exports=jdk.compiler/com.sun.tools.javac.tree=ALL-UNNAMED \
--add-exports=jdk.compiler/com.sun.tools.javac.util=ALL-UNNAMED \
--add-exports=jdk.compiler/com.sun.tools.javac.code=ALL-UNNAMED \
--add-exports=jdk.compiler/com.sun.tools.javac.comp=ALL-UNNAMED \
--add-exports=jdk.compiler/sun.nio.ch=ALL-UNNAMED \
--add-exports=java.base/sun.nio.ch=ALL-UNNAMED
```
[devexperts-17-workaround.zip](https://github.com/gradle/gradle/files/9822041/devexperts-17-workaround.zip)
[devexperts-17-issue.zip](https://github.com/gradle/gradle/files/9822042/devexperts-17-issue.zip)
This adds steps-to-reproduce for https://github.com/gradle/gradle/issues/21310
This is a duplicate of https://github.com/gradle/gradle/issues/22065 where we have failed to express ourselves
| process | module export rules needed with defined via options compilerarg are not applied to annotation processor expected behavior there should be a way to specify module export rules for those to apply to annotation processor current behavior only able to set module rules via global gradle jvm gradle properies context we have a happy codebase which works fine under java imposes additional module export requirments around javac internal api usages by our annotation processor we are unable to set such rules locally just for specific compilation task as a workaround we set export rules globally both problematic and workaround versions are attached how we believe it export rules should be added does not get the job done see attached project compilejava options compilerargs add exports jdk compiler com sun tools javac api all unnamed add exports jdk compiler com sun tools javac file all unnamed add exports jdk compiler com sun tools javac main all unnamed add exports jdk compiler com sun tools javac model all unnamed add exports jdk compiler com sun tools javac parser all unnamed add exports jdk compiler com sun tools javac processing all unnamed add exports jdk compiler com sun tools javac tree all unnamed add exports jdk compiler com sun tools javac util all unnamed add exports jdk compiler com sun tools javac code all unnamed add exports jdk compiler com sun tools javac comp all unnamed add exports jdk compiler sun nio ch all unnamed how we work around the issue org gradle jvmargs add exports jdk compiler com sun tools javac api all unnamed add exports jdk compiler com sun tools javac file all unnamed add exports jdk compiler com sun tools javac main all unnamed add exports jdk compiler com sun tools javac model all unnamed add exports jdk compiler com sun tools javac parser all unnamed add exports jdk compiler com sun tools javac processing all unnamed add exports jdk compiler com sun tools javac tree all unnamed add exports jdk compiler com sun tools javac util all unnamed add exports jdk compiler com sun tools javac code all unnamed add exports jdk compiler com sun tools javac comp all unnamed add exports jdk compiler sun nio ch all unnamed add exports java base sun nio ch all unnamed this adds steps to reproduce for this is a duplicate of where we have failed to express ourselves | 1 |
694,935 | 23,836,788,050 | IssuesEvent | 2022-09-06 06:51:12 | space-wizards/space-station-14 | https://api.github.com/repos/space-wizards/space-station-14 | closed | Traitor selection needs vibechecked | Issue: Bug Priority: 1-Urgent Difficulty: 2-Medium | ## Description
Seems to tend to give it to people who have antags all off. | 1.0 | Traitor selection needs vibechecked - ## Description
Seems to tend to give it to people who have antags all off. | non_process | traitor selection needs vibechecked description seems to tend to give it to people who have antags all off | 0 |
8,137 | 2,611,454,071 | IssuesEvent | 2015-02-27 05:01:13 | chrsmith/hedgewars | https://api.github.com/repos/chrsmith/hedgewars | closed | Unlimited Attacks: turn ends if no more enemy hedgehogs and when your hog is using rope, flying saucer, etc. | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Run fight with weapon set containing rope or flying saucer and game mode
with Unlimited Attacks mode.
2. Kill all enemy hedgehog and use rope or flying saucer.
What is the expected output? What do you see instead?
The turn ends after few seconds. The situation might be annoying in some cases,
for example when you have to escape from dynamite, and the only way is to use
flying saucer (and the safe place is far away).
What version of the product are you using? On what operating system?
0.9.14.1 on Windows XP SP2
Please provide any additional information below.
```
Original issue reported on code.google.com by `adibiaz...@gmail.com` on 22 Dec 2010 at 5:36
* Merged into: #129 | 1.0 | Unlimited Attacks: turn ends if no more enemy hedgehogs and when your hog is using rope, flying saucer, etc. - ```
What steps will reproduce the problem?
1. Run fight with weapon set containing rope or flying saucer and game mode
with Unlimited Attacks mode.
2. Kill all enemy hedgehog and use rope or flying saucer.
What is the expected output? What do you see instead?
The turn ends after few seconds. The situation might be annoying in some cases,
for example when you have to escape from dynamite, and the only way is to use
flying saucer (and the safe place is far away).
What version of the product are you using? On what operating system?
0.9.14.1 on Windows XP SP2
Please provide any additional information below.
```
Original issue reported on code.google.com by `adibiaz...@gmail.com` on 22 Dec 2010 at 5:36
* Merged into: #129 | non_process | unlimited attacks turn ends if no more enemy hedgehogs and when your hog is using rope flying saucer etc what steps will reproduce the problem run fight with weapon set containing rope or flying saucer and game mode with unlimited attacks mode kill all enemy hedgehog and use rope or flying saucer what is the expected output what do you see instead the turn ends after few seconds the situation might be annoying in some cases for example when you have to escape from dynamite and the only way is to use flying saucer and the safe place is far away what version of the product are you using on what operating system on windows xp please provide any additional information below original issue reported on code google com by adibiaz gmail com on dec at merged into | 0 |
10,914 | 13,690,855,782 | IssuesEvent | 2020-09-30 14:52:14 | prisma/prisma-engines | https://api.github.com/repos/prisma/prisma-engines | closed | Postgres: serial native type and `@default(autoincrement())` must not be possible at the same time | engines/data model parser process/candidate team/engines | The following schema should be invalid:
```
datasource test {
provider = "postgresql"
url = ...
}
model Blog {
id Int @id
serial Int @default(autoincrement()) @test.Serial
}
```
The native types `serial` is a special type in Postgres that translates to an Integer column with an auto incrementing counter as a default. So it means exactly the same thing as `@default(autoincrement())`. In fact the migration engine translates `Int @default(autoincrement())` to the `serial` type underneath. Hence the parser should validate this and ask the user to remove one of the two attributes on Postgres. | 1.0 | Postgres: serial native type and `@default(autoincrement())` must not be possible at the same time - The following schema should be invalid:
```
datasource test {
provider = "postgresql"
url = ...
}
model Blog {
id Int @id
serial Int @default(autoincrement()) @test.Serial
}
```
The native types `serial` is a special type in Postgres that translates to an Integer column with an auto incrementing counter as a default. So it means exactly the same thing as `@default(autoincrement())`. In fact the migration engine translates `Int @default(autoincrement())` to the `serial` type underneath. Hence the parser should validate this and ask the user to remove one of the two attributes on Postgres. | process | postgres serial native type and default autoincrement must not be possible at the same time the following schema should be invalid datasource test provider postgresql url model blog id int id serial int default autoincrement test serial the native types serial is a special type in postgres that translates to an integer column with an auto incrementing counter as a default so it means exactly the same thing as default autoincrement in fact the migration engine translates int default autoincrement to the serial type underneath hence the parser should validate this and ask the user to remove one of the two attributes on postgres | 1 |
69,459 | 22,358,988,244 | IssuesEvent | 2022-06-15 18:26:06 | hpi-swa-teaching/SVGMorph | https://api.github.com/repos/hpi-swa-teaching/SVGMorph | opened | Show elemets that do not have a color specified | defect User Story | As a viewer, I would be able to see all elements of an SVG that should be visible according to the SVG standard. This includes elements without a color specified, since the default color should be black, which is not implemented at the moment. | 1.0 | Show elemets that do not have a color specified - As a viewer, I would be able to see all elements of an SVG that should be visible according to the SVG standard. This includes elements without a color specified, since the default color should be black, which is not implemented at the moment. | non_process | show elemets that do not have a color specified as a viewer i would be able to see all elements of an svg that should be visible according to the svg standard this includes elements without a color specified since the default color should be black which is not implemented at the moment | 0 |
347,208 | 10,426,551,865 | IssuesEvent | 2019-09-16 17:51:31 | googleapis/google-cloud-python | https://api.github.com/repos/googleapis/google-cloud-python | closed | Vision: Operation returned from 'purge_product: raises TypeError in 'done'. | api: vision external priority: p1 type: bug |
#### Environment details
1. Vision API, Product Search, Purge Product
2. OSX
3. 2.7.14
4. google-cloud-vision 0.39.0
#### Steps to reproduce
1. Run tests or either of the purge product samples from this PR: https://github.com/GoogleCloudPlatform/python-docs-samples/pull/2349
1. The LRO fails at `operation.done()`
#### Code example
see PR linked above.
#### Stack trace
```
Traceback (most recent call last):
File "/Users/yuhanliu/projects/dpe-github/python-docs-samples/vision/cloud-client/product_search/product_in_product_set_management_test.py", line 88, in test_purge_products_in_product_set
PROJECT_ID, LOCATION, PRODUCT_SET_ID, force=True)
File "/Users/yuhanliu/projects/dpe-github/python-docs-samples/vision/cloud-client/product_search/product_in_product_set_management.py", line 153, in purge_products_in_product_set
if operation.done():
File "/Users/yuhanliu/.virtualenvs/vision-ga/lib/python2.7/site-packages/google/api_core/operation.py", line 162, in done
self._refresh_and_update()
File "/Users/yuhanliu/.virtualenvs/vision-ga/lib/python2.7/site-packages/google/api_core/operation.py", line 154, in _refresh_and_update
self._set_result_from_operation()
File "/Users/yuhanliu/.virtualenvs/vision-ga/lib/python2.7/site-packages/google/api_core/operation.py", line 131, in _set_result_from_operation
self._result_type, self._operation.response
File "/Users/yuhanliu/.virtualenvs/vision-ga/lib/python2.7/site-packages/google/api_core/protobuf_helpers.py", line 69, in from_any_pb
any_pb.__class__.__name__, pb_type.__name__
TypeError: Could not convert Any to Empty
```
| 1.0 | Vision: Operation returned from 'purge_product: raises TypeError in 'done'. -
#### Environment details
1. Vision API, Product Search, Purge Product
2. OSX
3. 2.7.14
4. google-cloud-vision 0.39.0
#### Steps to reproduce
1. Run tests or either of the purge product samples from this PR: https://github.com/GoogleCloudPlatform/python-docs-samples/pull/2349
1. The LRO fails at `operation.done()`
#### Code example
see PR linked above.
#### Stack trace
```
Traceback (most recent call last):
File "/Users/yuhanliu/projects/dpe-github/python-docs-samples/vision/cloud-client/product_search/product_in_product_set_management_test.py", line 88, in test_purge_products_in_product_set
PROJECT_ID, LOCATION, PRODUCT_SET_ID, force=True)
File "/Users/yuhanliu/projects/dpe-github/python-docs-samples/vision/cloud-client/product_search/product_in_product_set_management.py", line 153, in purge_products_in_product_set
if operation.done():
File "/Users/yuhanliu/.virtualenvs/vision-ga/lib/python2.7/site-packages/google/api_core/operation.py", line 162, in done
self._refresh_and_update()
File "/Users/yuhanliu/.virtualenvs/vision-ga/lib/python2.7/site-packages/google/api_core/operation.py", line 154, in _refresh_and_update
self._set_result_from_operation()
File "/Users/yuhanliu/.virtualenvs/vision-ga/lib/python2.7/site-packages/google/api_core/operation.py", line 131, in _set_result_from_operation
self._result_type, self._operation.response
File "/Users/yuhanliu/.virtualenvs/vision-ga/lib/python2.7/site-packages/google/api_core/protobuf_helpers.py", line 69, in from_any_pb
any_pb.__class__.__name__, pb_type.__name__
TypeError: Could not convert Any to Empty
```
| non_process | vision operation returned from purge product raises typeerror in done environment details vision api product search purge product osx google cloud vision steps to reproduce run tests or either of the purge product samples from this pr the lro fails at operation done code example see pr linked above stack trace traceback most recent call last file users yuhanliu projects dpe github python docs samples vision cloud client product search product in product set management test py line in test purge products in product set project id location product set id force true file users yuhanliu projects dpe github python docs samples vision cloud client product search product in product set management py line in purge products in product set if operation done file users yuhanliu virtualenvs vision ga lib site packages google api core operation py line in done self refresh and update file users yuhanliu virtualenvs vision ga lib site packages google api core operation py line in refresh and update self set result from operation file users yuhanliu virtualenvs vision ga lib site packages google api core operation py line in set result from operation self result type self operation response file users yuhanliu virtualenvs vision ga lib site packages google api core protobuf helpers py line in from any pb any pb class name pb type name typeerror could not convert any to empty | 0 |
7,557 | 10,678,541,611 | IssuesEvent | 2019-10-21 17:29:06 | qgis/QGIS | https://api.github.com/repos/qgis/QGIS | opened | batch mode: cannot autofill "column" if if is an raster input | Bug High Priority Processing Regression | QGIS 3.4.12 on Ubuntu:
Open a Processing tool (native or else) that uses an input raster layer.
Fill the first input with a raster taken from the project or file system
Add new lines to the batch mode.
Try to fill the new lines by double clicking the input parameter name/column > do not work.
Does not happen with vector inputs. | 1.0 | batch mode: cannot autofill "column" if if is an raster input - QGIS 3.4.12 on Ubuntu:
Open a Processing tool (native or else) that uses an input raster layer.
Fill the first input with a raster taken from the project or file system
Add new lines to the batch mode.
Try to fill the new lines by double clicking the input parameter name/column > do not work.
Does not happen with vector inputs. | process | batch mode cannot autofill column if if is an raster input qgis on ubuntu open a processing tool native or else that uses an input raster layer fill the first input with a raster taken from the project or file system add new lines to the batch mode try to fill the new lines by double clicking the input parameter name column do not work does not happen with vector inputs | 1 |
349,420 | 31,801,062,987 | IssuesEvent | 2023-09-13 11:08:31 | wazuh/wazuh | https://api.github.com/repos/wazuh/wazuh | closed | Release 4.6.0 - Alpha 1 - E2E UX tests - Configuration assessment | type/test level/subtask | ## End-to-End (E2E) Testing Guideline
- **Documentation:** Always consult the development documentation for the current stage tag at [this link](https://documentation-dev.wazuh.com/v4.6.0-alpha1/index.html). Be careful because some of the description steps might refer to a current version in production, always navigate using the current development documention for the stage under test.
- **Test Requirements:** Ensure your test comprehensively includes a full stack and agent/s deployment as per the Deployment requirements, detailing the machine OS, installed version, and revision.
- **Deployment Options:** While deployments can be local (using VMs, Vagrant, or Docker) or on the aws-dev account, opt for local deployments when feasible. For AWS access, coordinate with the CICD team through [this link](https://github.com/wazuh/internal-devel-requests/issues/new?assignees=&labels=level%2Ftask%2C+request%2Foperational%2C+type%2Fchange&projects=&template=operational--request.md&title=%3CTitle%3E).
- **External Accounts:** If tests require third-party accounts (e.g., GitHub, Azure, AWS, GCP), request the necessary access through the CICD team [here](https://github.com/wazuh/internal-devel-requests/issues/new?assignees=&labels=level%2Ftask%2C+request%2Foperational%2C+type%2Fchange&projects=&template=operational--request.md&title=%3CTitle%3E).
- **Alerts:** Every test should generate a minimum of one end-to-end alert, from the agent to the dashboard, irrespective of test type.
- **Multi-node Testing:** For multi-node wazuh-manager tests, ensure agents are connected to both workers and the master node.
- **Package Verification:** Use the pre-release package that matches the current TAG you're testing. Confirm its version and revision.
- **Filebeat Errors:** If you encounter errors with Filebeat during testing, refer to [this Slack discussion](https://wazuh-team.slack.com/archives/C03BDG0K6JC/p1672168163537809) for insights and resolutions.
- **Known Issues:** Familiarize yourself with previously reported issues in the Known Issues section. This helps in identifying already recognized errors during testing.
- **Reporting New Issues:** Any new errors discovered during testing that aren't listed under Known Issues should be reported. Communicate these to QA via the c-release Slack channel.
- **Test Conduct:** It's imperative to be thorough in your testing, offering enough detail for reviewers. Incomplete tests might necessitate a redo.
- **Documentation Feedback:** Encountering documentation gaps, unclear guidelines, or anything that disrupts the testing or UX? Open an issue, especially if it's not listed under Known Issues.
- **Format:** If this is your first time doing this, refer to the format (but not necessarily the content, as it may vary) of previous E2E tests, here you have an example https://github.com/wazuh/wazuh/issues/13994.
- **Status and completion:** Change the issue status within your team project accordingly. Once you finish testing and write the conclusions, move it to Pending review and notify the @wazuh/framework team via Slack using the [c-release channel](https://wazuh-team.slack.com/archives/C02A737S5MJ). Beware that the reviewers might request additional information or task repetitions.
- **For reviewers:** Please move the issue to Pending final review and notify via Slack using the same thread if everything is ok, otherwise, perform an issue update with the requested changes and move it to On hold, increase the review_cycles in the team project by one and notify the issue assignee via Slack using the same thread.
For the conclusions and the issue testing and updates, use the following legend:
**Status legend**
- 🟢 All checks passed
- 🟡 Found a known issue
- 🔴 Found a new error
## Deployment requirements
| Component | Installation | Type | OS |
|----------|--------------|------|----|
| Indexer | [Installation assistant](https://documentation-dev.wazuh.com/v4.6.0-alpha1/installation-guide/wazuh-indexer/installation-assistant.html) | Multi node | Fedora 38 x86_64 |
| Server | [Installation assistant](https://documentation-dev.wazuh.com/v4.6.0-alpha1/installation-guide/wazuh-indexer/installation-assistant.html) | Single node | Fedora 38 x86_64 |
| Dashboard | [Installation assistant](https://documentation-dev.wazuh.com/v4.6.0-alpha1/installation-guide/wazuh-indexer/installation-assistant.html) | - | Fedora 38 x86_64 |
| Agent | [Wazuh WUI one-liner deploy using FQDN](https://documentation-dev.wazuh.com/v4.6.0-alpha1/_images/deploy-new-agent-from-ui1.png) | - | Windows server 2016 x86_64, Debian 10 x86_64, Fedora 38 x86_64, macOS Ventura arm|
## Test description
For the selected Wazuh Agent OS:
- Check that their respective SCA policies are applied properly
- Use and activate a custom policy for any of them
- Disable an used policy and confirm it is not used anymore
- Push SCA config through centralized config and check it applies properly
## Known issues
- https://github.com/wazuh/wazuh-packages/issues/1422
- https://github.com/wazuh/wazuh/issues/12347
## Conclusions
Summarize the errors detected (Known Issues included). Illustrate using the table below, removing current examples:
| **Status** | **Test** | **Failure type** | **Notes** |
|----------------|-------------|---------------------|----------------|
| 🟡 | macOS Ventura agent installation with Wazuh WUI one-liner deploy using FQDN | The `WAZUH_MANAGER` variable did not work on the installation | #18933 |
| :yellow_circle: | Push SCA config through centralized config and check it applies properly with Fedora | Fedora operating system does not have an SCA policy | #18306 |
| :yellow_circle: | Push SCA config through centralized config and check it applies properly with Windows Server 2016 | Failure when attempting to create an SCA policy with Windows Server 2016 | #18981 |
## Feedback
We value your feedback. Please provide insights on your testing experience.
- Was the testing guideline clear? Were there any ambiguities?
Everything was clear and without any ambiguity
- Did you face any challenges not covered by the guideline?
No
- Suggestions for improvement:
## Reviewers validation
The criteria for completing this task is based on the validation of the conclusions and the test results by all reviewers.
All the checkboxes below must be marked in order to close this issue.
- [ ] @davidjiglesias
- [ ] @wazuh/security-warriors | 1.0 | Release 4.6.0 - Alpha 1 - E2E UX tests - Configuration assessment - ## End-to-End (E2E) Testing Guideline
- **Documentation:** Always consult the development documentation for the current stage tag at [this link](https://documentation-dev.wazuh.com/v4.6.0-alpha1/index.html). Be careful because some of the description steps might refer to a current version in production, always navigate using the current development documention for the stage under test.
- **Test Requirements:** Ensure your test comprehensively includes a full stack and agent/s deployment as per the Deployment requirements, detailing the machine OS, installed version, and revision.
- **Deployment Options:** While deployments can be local (using VMs, Vagrant, or Docker) or on the aws-dev account, opt for local deployments when feasible. For AWS access, coordinate with the CICD team through [this link](https://github.com/wazuh/internal-devel-requests/issues/new?assignees=&labels=level%2Ftask%2C+request%2Foperational%2C+type%2Fchange&projects=&template=operational--request.md&title=%3CTitle%3E).
- **External Accounts:** If tests require third-party accounts (e.g., GitHub, Azure, AWS, GCP), request the necessary access through the CICD team [here](https://github.com/wazuh/internal-devel-requests/issues/new?assignees=&labels=level%2Ftask%2C+request%2Foperational%2C+type%2Fchange&projects=&template=operational--request.md&title=%3CTitle%3E).
- **Alerts:** Every test should generate a minimum of one end-to-end alert, from the agent to the dashboard, irrespective of test type.
- **Multi-node Testing:** For multi-node wazuh-manager tests, ensure agents are connected to both workers and the master node.
- **Package Verification:** Use the pre-release package that matches the current TAG you're testing. Confirm its version and revision.
- **Filebeat Errors:** If you encounter errors with Filebeat during testing, refer to [this Slack discussion](https://wazuh-team.slack.com/archives/C03BDG0K6JC/p1672168163537809) for insights and resolutions.
- **Known Issues:** Familiarize yourself with previously reported issues in the Known Issues section. This helps in identifying already recognized errors during testing.
- **Reporting New Issues:** Any new errors discovered during testing that aren't listed under Known Issues should be reported. Communicate these to QA via the c-release Slack channel.
- **Test Conduct:** It's imperative to be thorough in your testing, offering enough detail for reviewers. Incomplete tests might necessitate a redo.
- **Documentation Feedback:** Encountering documentation gaps, unclear guidelines, or anything that disrupts the testing or UX? Open an issue, especially if it's not listed under Known Issues.
- **Format:** If this is your first time doing this, refer to the format (but not necessarily the content, as it may vary) of previous E2E tests, here you have an example https://github.com/wazuh/wazuh/issues/13994.
- **Status and completion:** Change the issue status within your team project accordingly. Once you finish testing and write the conclusions, move it to Pending review and notify the @wazuh/framework team via Slack using the [c-release channel](https://wazuh-team.slack.com/archives/C02A737S5MJ). Beware that the reviewers might request additional information or task repetitions.
- **For reviewers:** Please move the issue to Pending final review and notify via Slack using the same thread if everything is ok, otherwise, perform an issue update with the requested changes and move it to On hold, increase the review_cycles in the team project by one and notify the issue assignee via Slack using the same thread.
For the conclusions and the issue testing and updates, use the following legend:
**Status legend**
- 🟢 All checks passed
- 🟡 Found a known issue
- 🔴 Found a new error
## Deployment requirements
| Component | Installation | Type | OS |
|----------|--------------|------|----|
| Indexer | [Installation assistant](https://documentation-dev.wazuh.com/v4.6.0-alpha1/installation-guide/wazuh-indexer/installation-assistant.html) | Multi node | Fedora 38 x86_64 |
| Server | [Installation assistant](https://documentation-dev.wazuh.com/v4.6.0-alpha1/installation-guide/wazuh-indexer/installation-assistant.html) | Single node | Fedora 38 x86_64 |
| Dashboard | [Installation assistant](https://documentation-dev.wazuh.com/v4.6.0-alpha1/installation-guide/wazuh-indexer/installation-assistant.html) | - | Fedora 38 x86_64 |
| Agent | [Wazuh WUI one-liner deploy using FQDN](https://documentation-dev.wazuh.com/v4.6.0-alpha1/_images/deploy-new-agent-from-ui1.png) | - | Windows server 2016 x86_64, Debian 10 x86_64, Fedora 38 x86_64, macOS Ventura arm|
## Test description
For the selected Wazuh Agent OS:
- Check that their respective SCA policies are applied properly
- Use and activate a custom policy for any of them
- Disable an used policy and confirm it is not used anymore
- Push SCA config through centralized config and check it applies properly
## Known issues
- https://github.com/wazuh/wazuh-packages/issues/1422
- https://github.com/wazuh/wazuh/issues/12347
## Conclusions
Summarize the errors detected (Known Issues included). Illustrate using the table below, removing current examples:
| **Status** | **Test** | **Failure type** | **Notes** |
|----------------|-------------|---------------------|----------------|
| 🟡 | macOS Ventura agent installation with Wazuh WUI one-liner deploy using FQDN | The `WAZUH_MANAGER` variable did not work on the installation | #18933 |
| :yellow_circle: | Push SCA config through centralized config and check it applies properly with Fedora | Fedora operating system does not have an SCA policy | #18306 |
| :yellow_circle: | Push SCA config through centralized config and check it applies properly with Windows Server 2016 | Failure when attempting to create an SCA policy with Windows Server 2016 | #18981 |
## Feedback
We value your feedback. Please provide insights on your testing experience.
- Was the testing guideline clear? Were there any ambiguities?
Everything was clear and without any ambiguity
- Did you face any challenges not covered by the guideline?
No
- Suggestions for improvement:
## Reviewers validation
The criteria for completing this task is based on the validation of the conclusions and the test results by all reviewers.
All the checkboxes below must be marked in order to close this issue.
- [ ] @davidjiglesias
- [ ] @wazuh/security-warriors | non_process | release alpha ux tests configuration assessment end to end testing guideline documentation always consult the development documentation for the current stage tag at be careful because some of the description steps might refer to a current version in production always navigate using the current development documention for the stage under test test requirements ensure your test comprehensively includes a full stack and agent s deployment as per the deployment requirements detailing the machine os installed version and revision deployment options while deployments can be local using vms vagrant or docker or on the aws dev account opt for local deployments when feasible for aws access coordinate with the cicd team through external accounts if tests require third party accounts e g github azure aws gcp request the necessary access through the cicd team alerts every test should generate a minimum of one end to end alert from the agent to the dashboard irrespective of test type multi node testing for multi node wazuh manager tests ensure agents are connected to both workers and the master node package verification use the pre release package that matches the current tag you re testing confirm its version and revision filebeat errors if you encounter errors with filebeat during testing refer to for insights and resolutions known issues familiarize yourself with previously reported issues in the known issues section this helps in identifying already recognized errors during testing reporting new issues any new errors discovered during testing that aren t listed under known issues should be reported communicate these to qa via the c release slack channel test conduct it s imperative to be thorough in your testing offering enough detail for reviewers incomplete tests might necessitate a redo documentation feedback encountering documentation gaps unclear guidelines or anything that disrupts the testing or ux open an issue especially if it s not listed under known issues format if this is your first time doing this refer to the format but not necessarily the content as it may vary of previous tests here you have an example status and completion change the issue status within your team project accordingly once you finish testing and write the conclusions move it to pending review and notify the wazuh framework team via slack using the beware that the reviewers might request additional information or task repetitions for reviewers please move the issue to pending final review and notify via slack using the same thread if everything is ok otherwise perform an issue update with the requested changes and move it to on hold increase the review cycles in the team project by one and notify the issue assignee via slack using the same thread for the conclusions and the issue testing and updates use the following legend status legend 🟢 all checks passed 🟡 found a known issue 🔴 found a new error deployment requirements component installation type os indexer multi node fedora server single node fedora dashboard fedora agent windows server debian fedora macos ventura arm test description for the selected wazuh agent os check that their respective sca policies are applied properly use and activate a custom policy for any of them disable an used policy and confirm it is not used anymore push sca config through centralized config and check it applies properly known issues conclusions summarize the errors detected known issues included illustrate using the table below removing current examples status test failure type notes 🟡 macos ventura agent installation with wazuh wui one liner deploy using fqdn the wazuh manager variable did not work on the installation yellow circle push sca config through centralized config and check it applies properly with fedora fedora operating system does not have an sca policy yellow circle push sca config through centralized config and check it applies properly with windows server failure when attempting to create an sca policy with windows server feedback we value your feedback please provide insights on your testing experience was the testing guideline clear were there any ambiguities everything was clear and without any ambiguity did you face any challenges not covered by the guideline no suggestions for improvement reviewers validation the criteria for completing this task is based on the validation of the conclusions and the test results by all reviewers all the checkboxes below must be marked in order to close this issue davidjiglesias wazuh security warriors | 0 |
3,466 | 6,546,017,393 | IssuesEvent | 2017-09-04 08:32:18 | threefoldfoundation/app_backend | https://api.github.com/repos/threefoldfoundation/app_backend | closed | Make TFF app interface with IYO verification levels | priority_major process_duplicate type_feature | In IYO, there are different verification levels.
Within TFF app, certain functionalities require different verification levels.
TFF app should be able to check verification level of user and ask to initiate improved verification if needed.
E.g., a hoster should be address verified (ID level 2). | 1.0 | Make TFF app interface with IYO verification levels - In IYO, there are different verification levels.
Within TFF app, certain functionalities require different verification levels.
TFF app should be able to check verification level of user and ask to initiate improved verification if needed.
E.g., a hoster should be address verified (ID level 2). | process | make tff app interface with iyo verification levels in iyo there are different verification levels within tff app certain functionalities require different verification levels tff app should be able to check verification level of user and ask to initiate improved verification if needed e g a hoster should be address verified id level | 1 |
79,157 | 9,843,015,484 | IssuesEvent | 2019-06-18 10:35:41 | openSUSE/open-build-service | https://api.github.com/repos/openSUSE/open-build-service | closed | Workers part of /monitor suboptimal | Bootstrap :rocket: Design Frontend | I think the workers part of the monitor page deserves a little more love - the progress bars have rather dull colours and the font is too tiny to read (while it was enlarged everywhere else during bootstrap port).
And it doesn't take all available space and shows a rather strange '#' for non building workers. I was quite confused what it's about before I switched back to old webui to verify.


| 1.0 | Workers part of /monitor suboptimal - I think the workers part of the monitor page deserves a little more love - the progress bars have rather dull colours and the font is too tiny to read (while it was enlarged everywhere else during bootstrap port).
And it doesn't take all available space and shows a rather strange '#' for non building workers. I was quite confused what it's about before I switched back to old webui to verify.


| non_process | workers part of monitor suboptimal i think the workers part of the monitor page deserves a little more love the progress bars have rather dull colours and the font is too tiny to read while it was enlarged everywhere else during bootstrap port and it doesn t take all available space and shows a rather strange for non building workers i was quite confused what it s about before i switched back to old webui to verify | 0 |
17,177 | 22,755,106,059 | IssuesEvent | 2022-07-07 15:54:08 | hashgraph/hedera-json-rpc-relay | https://api.github.com/repos/hashgraph/hedera-json-rpc-relay | closed | Add acceptance test support for getLogs | enhancement limechain P2 process | ### Problem
#103 added support for getLogs.
However , this preceded the acceptance tests so support was not added
### Solution
Add a tests that calls `eth_getLogs` with coverage for a combination of the following params
- fromBlock
- toBlock
- address
- topics
- blockhash
### Alternatives
_No response_ | 1.0 | Add acceptance test support for getLogs - ### Problem
#103 added support for getLogs.
However , this preceded the acceptance tests so support was not added
### Solution
Add a tests that calls `eth_getLogs` with coverage for a combination of the following params
- fromBlock
- toBlock
- address
- topics
- blockhash
### Alternatives
_No response_ | process | add acceptance test support for getlogs problem added support for getlogs however this preceded the acceptance tests so support was not added solution add a tests that calls eth getlogs with coverage for a combination of the following params fromblock toblock address topics blockhash alternatives no response | 1 |
10,846 | 13,626,409,685 | IssuesEvent | 2020-09-24 10:59:15 | eclipse/microprofile-config | https://api.github.com/repos/eclipse/microprofile-config | closed | Adopt Maven Formatter Plugin | process change :busts_in_silhouette: | We should probably use [Maven Formatter Plugin](https://github.com/revelc/formatter-maven-plugin) for consistent code styling and to avoid reformats on code changes. | 1.0 | Adopt Maven Formatter Plugin - We should probably use [Maven Formatter Plugin](https://github.com/revelc/formatter-maven-plugin) for consistent code styling and to avoid reformats on code changes. | process | adopt maven formatter plugin we should probably use for consistent code styling and to avoid reformats on code changes | 1 |
287,763 | 21,672,886,658 | IssuesEvent | 2022-05-08 08:34:22 | BastienFaivre/ExchangeHub | https://api.github.com/repos/BastienFaivre/ExchangeHub | opened | Write text about user feedback | documentation | As stated on the submission page: _Text files (gladly with pictures) on documented user feedback and how it was acted upon_ | 1.0 | Write text about user feedback - As stated on the submission page: _Text files (gladly with pictures) on documented user feedback and how it was acted upon_ | non_process | write text about user feedback as stated on the submission page text files gladly with pictures on documented user feedback and how it was acted upon | 0 |
3,467 | 6,548,910,711 | IssuesEvent | 2017-09-05 02:49:32 | Great-Hill-Corporation/quickBlocks | https://api.github.com/repos/Great-Hill-Corporation/quickBlocks | closed | ethprice initialization | apps-ethPrice status-inprocess type-bug | Remove ~/.quickBlocks/cache/prices folder (i.e. brand new user)
Run test cases for ethprice (they all fail)
| 1.0 | ethprice initialization - Remove ~/.quickBlocks/cache/prices folder (i.e. brand new user)
Run test cases for ethprice (they all fail)
| process | ethprice initialization remove quickblocks cache prices folder i e brand new user run test cases for ethprice they all fail | 1 |
7,159 | 10,308,215,477 | IssuesEvent | 2019-08-29 10:24:07 | natario1/CameraView | https://api.github.com/repos/natario1/CameraView | closed | E/Frame: Frame is dead! time: -1 lastTime: 1565808924463 | about:frame processing is:bug status:has pr | ### Describe the bug
When i use frame processor in ML Kit Face detection.. in long time, always frame is dead.. is there a listener when frame is dead?
- CameraView version: *2.0.0-rc2r*
- Reproducible in official demo app: *yes/no* (i dont understand this)
- Device / Android version: *Redmi note 4, Nougat*
### Expected behavior
- I want a listener when frame is dead
- i want the frame not dead. because, my tracking in ml kit is stoping for about 5 second and then works again.
### Logs
E/Frame: Frame is dead! time: -1 lastTime: 1565811012759
W/CameraCallbacks: dispatchFrame: Error during processor implementation. Can happen when camera is closed while processors are running. java.lang.RuntimeException: You should not access a released frame. If this frame was passed to a FrameProcessor, you can only use its contents synchronously,for the duration of the process() method.
| 1.0 | E/Frame: Frame is dead! time: -1 lastTime: 1565808924463 - ### Describe the bug
When i use frame processor in ML Kit Face detection.. in long time, always frame is dead.. is there a listener when frame is dead?
- CameraView version: *2.0.0-rc2r*
- Reproducible in official demo app: *yes/no* (i dont understand this)
- Device / Android version: *Redmi note 4, Nougat*
### Expected behavior
- I want a listener when frame is dead
- i want the frame not dead. because, my tracking in ml kit is stoping for about 5 second and then works again.
### Logs
E/Frame: Frame is dead! time: -1 lastTime: 1565811012759
W/CameraCallbacks: dispatchFrame: Error during processor implementation. Can happen when camera is closed while processors are running. java.lang.RuntimeException: You should not access a released frame. If this frame was passed to a FrameProcessor, you can only use its contents synchronously,for the duration of the process() method.
| process | e frame frame is dead time lasttime describe the bug when i use frame processor in ml kit face detection in long time always frame is dead is there a listener when frame is dead cameraview version reproducible in official demo app yes no i dont understand this device android version redmi note nougat expected behavior i want a listener when frame is dead i want the frame not dead because my tracking in ml kit is stoping for about second and then works again logs e frame frame is dead time lasttime w cameracallbacks dispatchframe error during processor implementation can happen when camera is closed while processors are running java lang runtimeexception you should not access a released frame if this frame was passed to a frameprocessor you can only use its contents synchronously for the duration of the process method | 1 |
79,499 | 15,586,159,350 | IssuesEvent | 2021-03-18 01:18:26 | jrshutske/unit-conversion-api | https://api.github.com/repos/jrshutske/unit-conversion-api | opened | CVE-2020-35491 (High) detected in jackson-databind-2.9.8.jar | security vulnerability | ## CVE-2020-35491 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /unit-conversion-api/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-actuator-2.1.2.RELEASE.jar (Root Library)
- spring-boot-actuator-autoconfigure-2.1.2.RELEASE.jar
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.datasources.SharedPoolDataSource.
<p>Publish Date: 2020-12-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35491>CVE-2020-35491</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2986">https://github.com/FasterXML/jackson-databind/issues/2986</a></p>
<p>Release Date: 2020-12-17</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-35491 (High) detected in jackson-databind-2.9.8.jar - ## CVE-2020-35491 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /unit-conversion-api/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-actuator-2.1.2.RELEASE.jar (Root Library)
- spring-boot-actuator-autoconfigure-2.1.2.RELEASE.jar
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.datasources.SharedPoolDataSource.
<p>Publish Date: 2020-12-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35491>CVE-2020-35491</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2986">https://github.com/FasterXML/jackson-databind/issues/2986</a></p>
<p>Release Date: 2020-12-17</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file unit conversion api pom xml path to vulnerable library root repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter actuator release jar root library spring boot actuator autoconfigure release jar x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache commons datasources sharedpooldatasource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource | 0 |
135,586 | 30,318,680,793 | IssuesEvent | 2023-07-10 17:27:15 | bevyengine/bevy | https://api.github.com/repos/bevyengine/bevy | closed | Typo in bevy_reflect_derive | D-Good-First-Issue C-Code-Quality A-Reflection | `bevy_reflect_derive` has a little [typo](https://github.com/ClayenKitten/bevy/blob/1b51053f1958d750cdddffb99828bd0aacb61a78/crates/bevy_reflect/bevy_reflect_derive/src/type_path.rs#L76) in `NamedTypePathDef` enum variant. It is currently named `Primtive` and, obviously, should be changed to `Primitive`.
Usage of the enum is tightly coupled with macro, where I don't have much experience, so I decided to leave it for more experienced contributor just in case. | 1.0 | Typo in bevy_reflect_derive - `bevy_reflect_derive` has a little [typo](https://github.com/ClayenKitten/bevy/blob/1b51053f1958d750cdddffb99828bd0aacb61a78/crates/bevy_reflect/bevy_reflect_derive/src/type_path.rs#L76) in `NamedTypePathDef` enum variant. It is currently named `Primtive` and, obviously, should be changed to `Primitive`.
Usage of the enum is tightly coupled with macro, where I don't have much experience, so I decided to leave it for more experienced contributor just in case. | non_process | typo in bevy reflect derive bevy reflect derive has a little in namedtypepathdef enum variant it is currently named primtive and obviously should be changed to primitive usage of the enum is tightly coupled with macro where i don t have much experience so i decided to leave it for more experienced contributor just in case | 0 |
4,625 | 7,468,800,826 | IssuesEvent | 2018-04-02 20:18:43 | w3c/w3process | https://api.github.com/repos/w3c/w3process | closed | Can't update process as if it were a Rec | Process2019Candidate | There is no call for implementations, etc, so offering the Rec track as an alternative path makes no sense. | 1.0 | Can't update process as if it were a Rec - There is no call for implementations, etc, so offering the Rec track as an alternative path makes no sense. | process | can t update process as if it were a rec there is no call for implementations etc so offering the rec track as an alternative path makes no sense | 1 |
15,365 | 19,538,432,749 | IssuesEvent | 2021-12-31 13:30:23 | fadeoutsoftware/WASDI | https://api.github.com/repos/fadeoutsoftware/WASDI | closed | Automatic Autoban -> interface -> include agricolture areas | bug app / processor | There is an error using that flag. Check and solve.
| 1.0 | Automatic Autoban -> interface -> include agricolture areas - There is an error using that flag. Check and solve.
| process | automatic autoban interface include agricolture areas there is an error using that flag check and solve | 1 |
80,616 | 15,586,301,994 | IssuesEvent | 2021-03-18 01:37:57 | benchmarkdebricked/spring-boot | https://api.github.com/repos/benchmarkdebricked/spring-boot | opened | CVE-2020-27782 (High) detected in undertow-core-2.0.20.Final.jar | security vulnerability | ## CVE-2020-27782 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>undertow-core-2.0.20.Final.jar</b></p></summary>
<p>Undertow</p>
<p>Library home page: <a href="http://www.jboss.org/">http://www.jboss.org/</a></p>
<p>Path to dependency file: /spring-boot/spring-boot-project/spring-boot/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar,/root/.m2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar,/root/.m2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar,2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar,/root/.m2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar,/root/.m2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar</p>
<p>
Dependency Hierarchy:
- undertow-servlet-2.0.20.Final.jar (Root Library)
- :x: **undertow-core-2.0.20.Final.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in the Undertow AJP connector. Malicious requests and abrupt connection closes could be triggered by an attacker using query strings with non-RFC compliant characters resulting in a denial of service. The highest threat from this vulnerability is to system availability. This affects Undertow 2.1.5.SP1, 2.0.33.SP2, and 2.2.3.SP1.
<p>Publish Date: 2021-02-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27782>CVE-2020-27782</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-27782">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-27782</a></p>
<p>Release Date: 2021-02-23</p>
<p>Fix Resolution: io.undertow:undertow-core:2.2.4.Final</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-27782 (High) detected in undertow-core-2.0.20.Final.jar - ## CVE-2020-27782 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>undertow-core-2.0.20.Final.jar</b></p></summary>
<p>Undertow</p>
<p>Library home page: <a href="http://www.jboss.org/">http://www.jboss.org/</a></p>
<p>Path to dependency file: /spring-boot/spring-boot-project/spring-boot/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar,/root/.m2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar,/root/.m2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar,2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar,/root/.m2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar,/root/.m2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar</p>
<p>
Dependency Hierarchy:
- undertow-servlet-2.0.20.Final.jar (Root Library)
- :x: **undertow-core-2.0.20.Final.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in the Undertow AJP connector. Malicious requests and abrupt connection closes could be triggered by an attacker using query strings with non-RFC compliant characters resulting in a denial of service. The highest threat from this vulnerability is to system availability. This affects Undertow 2.1.5.SP1, 2.0.33.SP2, and 2.2.3.SP1.
<p>Publish Date: 2021-02-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27782>CVE-2020-27782</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-27782">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-27782</a></p>
<p>Release Date: 2021-02-23</p>
<p>Fix Resolution: io.undertow:undertow-core:2.2.4.Final</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve high detected in undertow core final jar cve high severity vulnerability vulnerable library undertow core final jar undertow library home page a href path to dependency file spring boot spring boot project spring boot pom xml path to vulnerable library root repository io undertow undertow core final undertow core final jar root repository io undertow undertow core final undertow core final jar root repository io undertow undertow core final undertow core final jar repository io undertow undertow core final undertow core final jar root repository io undertow undertow core final undertow core final jar root repository io undertow undertow core final undertow core final jar dependency hierarchy undertow servlet final jar root library x undertow core final jar vulnerable library vulnerability details a flaw was found in the undertow ajp connector malicious requests and abrupt connection closes could be triggered by an attacker using query strings with non rfc compliant characters resulting in a denial of service the highest threat from this vulnerability is to system availability this affects undertow and publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io undertow undertow core final step up your open source security game with whitesource | 0 |
12,424 | 14,924,138,177 | IssuesEvent | 2021-01-23 22:13:53 | eddieantonio/predictive-text-studio | https://api.github.com/repos/eddieantonio/predictive-text-studio | closed | Remove headers from uploaded wordlists | data-backing data-processing good first issue 🔥 High priority | Remove headers such as "word" and "count" headers in the first row of all uploaded wordlists (either Google Sheets or Excel).
This should be done automatically. | 1.0 | Remove headers from uploaded wordlists - Remove headers such as "word" and "count" headers in the first row of all uploaded wordlists (either Google Sheets or Excel).
This should be done automatically. | process | remove headers from uploaded wordlists remove headers such as word and count headers in the first row of all uploaded wordlists either google sheets or excel this should be done automatically | 1 |
74,823 | 14,346,736,069 | IssuesEvent | 2020-11-29 02:31:00 | fps-game-jam-org/game-off-2020 | https://api.github.com/repos/fps-game-jam-org/game-off-2020 | closed | Make the SceneChanger make its GameObject be a singleton | code fix | Right now, if you change into a scene that has a Scene Changer GameObject, you'll keep your old Scene Changer and get the new one. This is not useful. There should be exactly 1 Scene Changer in the scene at any time. | 1.0 | Make the SceneChanger make its GameObject be a singleton - Right now, if you change into a scene that has a Scene Changer GameObject, you'll keep your old Scene Changer and get the new one. This is not useful. There should be exactly 1 Scene Changer in the scene at any time. | non_process | make the scenechanger make its gameobject be a singleton right now if you change into a scene that has a scene changer gameobject you ll keep your old scene changer and get the new one this is not useful there should be exactly scene changer in the scene at any time | 0 |
21,001 | 27,868,901,112 | IssuesEvent | 2023-03-21 12:13:55 | Living-with-machines/MapReader | https://api.github.com/repos/Living-with-machines/MapReader | closed | MapReader and Binder | maps post-processing | @ChristinaLast What do you think if we have more than one Binder link? Does it make sense to have:
- [ ] Example for Maps
- [x] Example for Plant images
For the latter, we have a link now:
https://mybinder.org/v2/gh/Living-with-machines/MapReader/main?labpath=examples%2Fquick_start%2Fquick_start.ipynb
I just tested it, and it works, but I want to also add all the cells/codes of that notebook to CI.
Related issue: https://github.com/Living-with-machines/MapReader/issues/28 | 1.0 | MapReader and Binder - @ChristinaLast What do you think if we have more than one Binder link? Does it make sense to have:
- [ ] Example for Maps
- [x] Example for Plant images
For the latter, we have a link now:
https://mybinder.org/v2/gh/Living-with-machines/MapReader/main?labpath=examples%2Fquick_start%2Fquick_start.ipynb
I just tested it, and it works, but I want to also add all the cells/codes of that notebook to CI.
Related issue: https://github.com/Living-with-machines/MapReader/issues/28 | process | mapreader and binder christinalast what do you think if we have more than one binder link does it make sense to have example for maps example for plant images for the latter we have a link now i just tested it and it works but i want to also add all the cells codes of that notebook to ci related issue | 1 |
17,936 | 23,933,688,546 | IssuesEvent | 2022-09-10 23:19:59 | GregTechCEu/gt-ideas | https://api.github.com/repos/GregTechCEu/gt-ideas | opened | Sea Water Purification | processing chain | ## Details
It should be no surprise that water obtained from the sea is full of impurities and should be purified before being used for industrial purposes. A process known as Multi-stage flash distillation is one of the most widely used methods of desalinating water (It is used to process 26% of today's desalinated water)
## Products
Main Product: Water
Side Product(s): Brine
## Steps
Sea Water + Steam -> Water + Brine (Not sure how much of each liquid should be consumed or produced)

## Yield
Idk
## Sources
https://en.wikipedia.org/wiki/Multi-stage_flash_distillation
| 1.0 | Sea Water Purification - ## Details
It should be no surprise that water obtained from the sea is full of impurities and should be purified before being used for industrial purposes. A process known as Multi-stage flash distillation is one of the most widely used methods of desalinating water (It is used to process 26% of today's desalinated water)
## Products
Main Product: Water
Side Product(s): Brine
## Steps
Sea Water + Steam -> Water + Brine (Not sure how much of each liquid should be consumed or produced)

## Yield
Idk
## Sources
https://en.wikipedia.org/wiki/Multi-stage_flash_distillation
| process | sea water purification details it should be no surprise that water obtained from the sea is full of impurities and should be purified before being used for industrial purposes a process known as multi stage flash distillation is one of the most widely used methods of desalinating water it is used to process of today s desalinated water products main product water side product s brine steps sea water steam water brine not sure how much of each liquid should be consumed or produced yield idk sources | 1 |
157,574 | 6,008,670,950 | IssuesEvent | 2017-06-06 08:31:55 | GeekyAnts/NativeBase | https://api.github.com/repos/GeekyAnts/NativeBase | closed | How about the performance of nativebase? | 1 high priority performance | The sideslip menu on the Android is obviously slow to switch, worrying about performance problems | 1.0 | How about the performance of nativebase? - The sideslip menu on the Android is obviously slow to switch, worrying about performance problems | non_process | how about the performance of nativebase the sideslip menu on the android is obviously slow to switch worrying about performance problems | 0 |
11,449 | 14,270,367,545 | IssuesEvent | 2020-11-21 06:34:24 | Altinn/altinn-studio | https://api.github.com/repos/Altinn/altinn-studio | opened | Implement functional signing support | area/process kind/user-story solution/app-backend solution/app-frontend ux | ## Description
Altinn 3 need to support a signing task. This signing task lets user "sign" parts of or all data submitted as part of an instance.
## Screenshots
> Screenshots or links to Figma (make sure your sketch is public)
## Considerations
### Signing task
### Signing view
### Signature storage
### Signature API
## Acceptance criteria
> Describe criteria here (i.e. What is allowed/not allowed (negative tesing), validations, error messages and warnings etc.)
## Specification tasks
- [ ] Development tasks are defined
- [ ] Test design / decide test need
## Development tasks
> Add tasks here
## Definition of done
Verify that this issue meets [DoD](https://confluence.brreg.no/display/T3KP/Definition+of+Done#DefinitionofDone-DoD%E2%80%93utvikling) (Only for project members) before closing.
- [ ] Documentation is updated (if relevant)
- [ ] Technical documentation (docs.altinn.studio)
- [ ] User documentation (altinn.github.io/docs)
- [ ] QA
- [ ] Manual test is complete (if relevant)
- [ ] Automated test is implemented (if relevant)
- [ ] All tasks in this userstory are closed (i.e. remaining tasks are moved to other user stories or marked obsolete)
| 1.0 | Implement functional signing support - ## Description
Altinn 3 need to support a signing task. This signing task lets user "sign" parts of or all data submitted as part of an instance.
## Screenshots
> Screenshots or links to Figma (make sure your sketch is public)
## Considerations
### Signing task
### Signing view
### Signature storage
### Signature API
## Acceptance criteria
> Describe criteria here (i.e. What is allowed/not allowed (negative tesing), validations, error messages and warnings etc.)
## Specification tasks
- [ ] Development tasks are defined
- [ ] Test design / decide test need
## Development tasks
> Add tasks here
## Definition of done
Verify that this issue meets [DoD](https://confluence.brreg.no/display/T3KP/Definition+of+Done#DefinitionofDone-DoD%E2%80%93utvikling) (Only for project members) before closing.
- [ ] Documentation is updated (if relevant)
- [ ] Technical documentation (docs.altinn.studio)
- [ ] User documentation (altinn.github.io/docs)
- [ ] QA
- [ ] Manual test is complete (if relevant)
- [ ] Automated test is implemented (if relevant)
- [ ] All tasks in this userstory are closed (i.e. remaining tasks are moved to other user stories or marked obsolete)
| process | implement functional signing support description altinn need to support a signing task this signing task lets user sign parts of or all data submitted as part of an instance screenshots screenshots or links to figma make sure your sketch is public considerations signing task signing view signature storage signature api acceptance criteria describe criteria here i e what is allowed not allowed negative tesing validations error messages and warnings etc specification tasks development tasks are defined test design decide test need development tasks add tasks here definition of done verify that this issue meets only for project members before closing documentation is updated if relevant technical documentation docs altinn studio user documentation altinn github io docs qa manual test is complete if relevant automated test is implemented if relevant all tasks in this userstory are closed i e remaining tasks are moved to other user stories or marked obsolete | 1 |
13,349 | 15,810,843,838 | IssuesEvent | 2021-04-05 00:12:59 | darktable-org/darktable | https://api.github.com/repos/darktable-org/darktable | closed | Noise profiles are wrong | bug: wip difficulty: average no-issue-activity scope: image processing understood: clear | **Describe the bug**
Noise profiles aim at having a constant variance of 1 after the generalized anscombe transform of denoiseprofiled, so that the denoising algorithms can work on gaussian noise of variance 1.
Yet, if we check the variance obtained on a flat image at 100% zoom with the compute variance feature, I get variance values of about 14-15 for fujifilm xt3, and about 10 for lumix fz1000.
The noiseprofiling tool assumes that noise is only fine grain, which is not true in practice.
Also, it uses an input color profile of REC709, which (I guess) modifies the values, so the noiseprofiles are not made on the real data we get after demosaic.
**To Reproduce**
Steps to reproduce the behavior:
1. Shoot a defocused picture of a flat uniform image (e.g. https://artphotoheritage.files.wordpress.com/2011/06/gris-neutre.png ) at several ISO values.
2. In darktablerc, set plugins/darkroom/denoiseprofile/show_compute_variance_mode to "true"
3. Launch darktable
4. For each of the considered pictures:
4.1 Move denoiseprofile before colorin
4.2 Put denoiseprofile in "compute variance" mode
4.3 Zoom to 100% in the center of the image (to minimize the contribution of vignetting in the variance)
4.4 The displayed variance values are not close to 1... however, it looks like they are almost the same for all ISO values.
**Expected behavior**
At step 4.3, we should see variances close to 1.
**Screenshots**

**How bad is it to have wrong profiles?**
It is bad for making presets that works well on different cameras if the variance values we get differ from one camera to another.
It does not change much in terms of denoising quality if the user uses the strength parameter to fit its need, as even with a wrong profile the generalized anscombe transform will succeed in stabilizing the variance.
Let me go into further details:
The generalized anscombe we do is (approximately) X -> 2\*sqrt(X\/a)
The "a" number is supposed to be chosen such as a\*mean = variance
V[ 2\*sqrt(X\/a) ] is approximately equal to:
V [ 2\*sqrt(mean\/a) + 1\/(a\*sqrt(mean\/a)) \* (X-mean) ] (Taylor expansion of sqrt around the mean, see https://en.wikipedia.org/wiki/Taylor_expansions_for_the_moments_of_functions_of_random_variables )
which gives V[ 2\*sqrt(X\/a) ] is approximately equal to (a\/(a\*a\*mean))\*V[X] = 1\/(a\*mean)\*V[X] = 1
Thanks to the fact that a*mean = variance, we can get a variance of 1 after the transform.
Now, what happens if we make mistakes and take a value which gives a\*mean=c\*variance?
We can go through the same computations, but we will get V[X] = 1\/c at the end.
So, the good new is that the variance remains constant!
The question is, is the factor between the wrong profile and a true one (i.e. the number "c") always the same? Is our current noiseprofiling tool consistent in the errors it makes?
If so, then it is not necessary to correct anything, we simply get profiles which allows to have a variance of 1\/c instead of 1, which is not problematic.
However, if the number "c" changes from one camera to another, or from one ISO value to another, then we have an issue: the denoiseprofile will behave differently from one camera to another.
From my first experiments, it _seems_ that the "c" does not change much from one ISO value to another, yet I get different values between my 2 cameras.
What I would like to also check is if all bayer cameras have the same "c", and all xtrans cameras have the same "c". In that case, we would still be able to fix the issue without changing the noise profiling tool, and without having to redo all noise profiles.
Otherwise, I guess the only option will be to do a new profiling tool, and to redo all noise profiles. | 1.0 | Noise profiles are wrong - **Describe the bug**
Noise profiles aim at having a constant variance of 1 after the generalized anscombe transform of denoiseprofiled, so that the denoising algorithms can work on gaussian noise of variance 1.
Yet, if we check the variance obtained on a flat image at 100% zoom with the compute variance feature, I get variance values of about 14-15 for fujifilm xt3, and about 10 for lumix fz1000.
The noiseprofiling tool assumes that noise is only fine grain, which is not true in practice.
Also, it uses an input color profile of REC709, which (I guess) modifies the values, so the noiseprofiles are not made on the real data we get after demosaic.
**To Reproduce**
Steps to reproduce the behavior:
1. Shoot a defocused picture of a flat uniform image (e.g. https://artphotoheritage.files.wordpress.com/2011/06/gris-neutre.png ) at several ISO values.
2. In darktablerc, set plugins/darkroom/denoiseprofile/show_compute_variance_mode to "true"
3. Launch darktable
4. For each of the considered pictures:
4.1 Move denoiseprofile before colorin
4.2 Put denoiseprofile in "compute variance" mode
4.3 Zoom to 100% in the center of the image (to minimize the contribution of vignetting in the variance)
4.4 The displayed variance values are not close to 1... however, it looks like they are almost the same for all ISO values.
**Expected behavior**
At step 4.3, we should see variances close to 1.
**Screenshots**

**How bad is it to have wrong profiles?**
It is bad for making presets that works well on different cameras if the variance values we get differ from one camera to another.
It does not change much in terms of denoising quality if the user uses the strength parameter to fit its need, as even with a wrong profile the generalized anscombe transform will succeed in stabilizing the variance.
Let me go into further details:
The generalized anscombe we do is (approximately) X -> 2\*sqrt(X\/a)
The "a" number is supposed to be chosen such as a\*mean = variance
V[ 2\*sqrt(X\/a) ] is approximately equal to:
V [ 2\*sqrt(mean\/a) + 1\/(a\*sqrt(mean\/a)) \* (X-mean) ] (Taylor expansion of sqrt around the mean, see https://en.wikipedia.org/wiki/Taylor_expansions_for_the_moments_of_functions_of_random_variables )
which gives V[ 2\*sqrt(X\/a) ] is approximately equal to (a\/(a\*a\*mean))\*V[X] = 1\/(a\*mean)\*V[X] = 1
Thanks to the fact that a*mean = variance, we can get a variance of 1 after the transform.
Now, what happens if we make mistakes and take a value which gives a\*mean=c\*variance?
We can go through the same computations, but we will get V[X] = 1\/c at the end.
So, the good new is that the variance remains constant!
The question is, is the factor between the wrong profile and a true one (i.e. the number "c") always the same? Is our current noiseprofiling tool consistent in the errors it makes?
If so, then it is not necessary to correct anything, we simply get profiles which allows to have a variance of 1\/c instead of 1, which is not problematic.
However, if the number "c" changes from one camera to another, or from one ISO value to another, then we have an issue: the denoiseprofile will behave differently from one camera to another.
From my first experiments, it _seems_ that the "c" does not change much from one ISO value to another, yet I get different values between my 2 cameras.
What I would like to also check is if all bayer cameras have the same "c", and all xtrans cameras have the same "c". In that case, we would still be able to fix the issue without changing the noise profiling tool, and without having to redo all noise profiles.
Otherwise, I guess the only option will be to do a new profiling tool, and to redo all noise profiles. | process | noise profiles are wrong describe the bug noise profiles aim at having a constant variance of after the generalized anscombe transform of denoiseprofiled so that the denoising algorithms can work on gaussian noise of variance yet if we check the variance obtained on a flat image at zoom with the compute variance feature i get variance values of about for fujifilm and about for lumix the noiseprofiling tool assumes that noise is only fine grain which is not true in practice also it uses an input color profile of which i guess modifies the values so the noiseprofiles are not made on the real data we get after demosaic to reproduce steps to reproduce the behavior shoot a defocused picture of a flat uniform image e g at several iso values in darktablerc set plugins darkroom denoiseprofile show compute variance mode to true launch darktable for each of the considered pictures move denoiseprofile before colorin put denoiseprofile in compute variance mode zoom to in the center of the image to minimize the contribution of vignetting in the variance the displayed variance values are not close to however it looks like they are almost the same for all iso values expected behavior at step we should see variances close to screenshots how bad is it to have wrong profiles it is bad for making presets that works well on different cameras if the variance values we get differ from one camera to another it does not change much in terms of denoising quality if the user uses the strength parameter to fit its need as even with a wrong profile the generalized anscombe transform will succeed in stabilizing the variance let me go into further details the generalized anscombe we do is approximately x sqrt x a the a number is supposed to be chosen such as a mean variance v is approximately equal to v taylor expansion of sqrt around the mean see which gives v is approximately equal to a a a mean v a mean v thanks to the fact that a mean variance we can get a variance of after the transform now what happens if we make mistakes and take a value which gives a mean c variance we can go through the same computations but we will get v c at the end so the good new is that the variance remains constant the question is is the factor between the wrong profile and a true one i e the number c always the same is our current noiseprofiling tool consistent in the errors it makes if so then it is not necessary to correct anything we simply get profiles which allows to have a variance of c instead of which is not problematic however if the number c changes from one camera to another or from one iso value to another then we have an issue the denoiseprofile will behave differently from one camera to another from my first experiments it seems that the c does not change much from one iso value to another yet i get different values between my cameras what i would like to also check is if all bayer cameras have the same c and all xtrans cameras have the same c in that case we would still be able to fix the issue without changing the noise profiling tool and without having to redo all noise profiles otherwise i guess the only option will be to do a new profiling tool and to redo all noise profiles | 1 |
15,048 | 18,762,689,258 | IssuesEvent | 2021-11-05 18:29:25 | ORNL-AMO/AMO-Tools-Suite | https://api.github.com/repos/ORNL-AMO/AMO-Tools-Suite | closed | Available Heat Algorithm | Needs Verification Process Heating | The calculations for calculating sensible heat of the fuel needs to be added to `GasCompositions::ProcessHeatPropertiesResults GasCompositions::getProcessHeatProperties`
Don't need to use the separate method, but the math should be added from lines 154 - 161.
replace `fuelTemp` with `ambientAirTempF`, do what you need to do to make it match the "style" of the rest of getProcessHeatProperties
```
double GasCompositions::calculateSensibleHeat(const double fuelTemp) {
double specificHeatFuel = 0;
for ( auto const & comp : gasses ) {
specificHeatFuel += comp.second->compByWeight * (comp.second->specificHeat(520) / comp.second->molecularWeight);
}
return 1 * specificHeatFuel * (fuelTemp - 32);
}
```
Whatever variable you use for sensible heat needs to be added to Line 131
```
const double availableHeat = (100 * (WhateverSensibleHeat + heatValueFuel + preHeatedAirEff + preHeatedAirMoistureEff) - (H2OHeatContent + CO2HeatContent + N2HeatContent + O2HeatContent)) / (100 * heatValueFuel);
```
A few comments to make the code more understandable:
I know I missed some that should be called out, so if there is anything you can think of to help make it more intuitive, please add.
before ...
line 47: temperature conversions
line 54: calculate average Cp across temperature range of combustion air to flue gas
line 84: something, not 100% sure what is right here, it is kind of used for several things
line 111: calculate heat in combustion air moisture
before whatever you add for sensible heat: calculate sensible heat
before density: calculate final values
| 1.0 | Available Heat Algorithm - The calculations for calculating sensible heat of the fuel needs to be added to `GasCompositions::ProcessHeatPropertiesResults GasCompositions::getProcessHeatProperties`
Don't need to use the separate method, but the math should be added from lines 154 - 161.
replace `fuelTemp` with `ambientAirTempF`, do what you need to do to make it match the "style" of the rest of getProcessHeatProperties
```
double GasCompositions::calculateSensibleHeat(const double fuelTemp) {
double specificHeatFuel = 0;
for ( auto const & comp : gasses ) {
specificHeatFuel += comp.second->compByWeight * (comp.second->specificHeat(520) / comp.second->molecularWeight);
}
return 1 * specificHeatFuel * (fuelTemp - 32);
}
```
Whatever variable you use for sensible heat needs to be added to Line 131
```
const double availableHeat = (100 * (WhateverSensibleHeat + heatValueFuel + preHeatedAirEff + preHeatedAirMoistureEff) - (H2OHeatContent + CO2HeatContent + N2HeatContent + O2HeatContent)) / (100 * heatValueFuel);
```
A few comments to make the code more understandable:
I know I missed some that should be called out, so if there is anything you can think of to help make it more intuitive, please add.
before ...
line 47: temperature conversions
line 54: calculate average Cp across temperature range of combustion air to flue gas
line 84: something, not 100% sure what is right here, it is kind of used for several things
line 111: calculate heat in combustion air moisture
before whatever you add for sensible heat: calculate sensible heat
before density: calculate final values
| process | available heat algorithm the calculations for calculating sensible heat of the fuel needs to be added to gascompositions processheatpropertiesresults gascompositions getprocessheatproperties don t need to use the separate method but the math should be added from lines replace fueltemp with ambientairtempf do what you need to do to make it match the style of the rest of getprocessheatproperties double gascompositions calculatesensibleheat const double fueltemp double specificheatfuel for auto const comp gasses specificheatfuel comp second compbyweight comp second specificheat comp second molecularweight return specificheatfuel fueltemp whatever variable you use for sensible heat needs to be added to line const double availableheat whateversensibleheat heatvaluefuel preheatedaireff preheatedairmoistureeff heatvaluefuel a few comments to make the code more understandable i know i missed some that should be called out so if there is anything you can think of to help make it more intuitive please add before line temperature conversions line calculate average cp across temperature range of combustion air to flue gas line something not sure what is right here it is kind of used for several things line calculate heat in combustion air moisture before whatever you add for sensible heat calculate sensible heat before density calculate final values | 1 |
78,389 | 27,492,835,486 | IssuesEvent | 2023-03-04 20:49:08 | DependencyTrack/dependency-track | https://api.github.com/repos/DependencyTrack/dependency-track | opened | javax.jdo.JDOUserException: Field org.dependencytrack.model.Component.repositoryMeta is not marked as persistent so cannot be queried | defect in triage | ### Current Behavior
When opening the Policy Violations tab, the PolicyViolationResource throws the following error and no violations are shown. I think this bug was introduced somewhere last week...
``
javax.jdo.JDOUserException: Field org.dependencytrack.model.Component.repositoryMeta is not marked as persistent so cannot be queried
at org.datanucleus.api.jdo.JDOAdapter.getJDOExceptionForNucleusException(JDOAdapter.java:698)
at org.datanucleus.api.jdo.JDOQuery.executeInternal(JDOQuery.java:456)
at org.datanucleus.api.jdo.JDOQuery.execute(JDOQuery.java:276)
at alpine.persistence.AbstractAlpineQueryManager.execute(AbstractAlpineQueryManager.java:174)
at org.dependencytrack.persistence.PolicyQueryManager.getPolicyViolations(PolicyQueryManager.java:280)
at org.dependencytrack.persistence.QueryManager.getPolicyViolations(QueryManager.java:604)
at org.dependencytrack.resources.v1.PolicyViolationResource.getViolationsByProject(PolicyViolationResource.java:102)
### Steps to Reproduce
1.Open the Policy Violations tab for a project in the frontend
### Expected Behavior
No Exception is thrown and violations are shown (if any)
### Dependency-Track Version
4.8.0-SNAPSHOT
### Dependency-Track Distribution
Executable WAR
### Database Server
H2
### Database Server Version
_No response_
### Browser
Google Chrome
### Checklist
- [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues)
- [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported | 1.0 | javax.jdo.JDOUserException: Field org.dependencytrack.model.Component.repositoryMeta is not marked as persistent so cannot be queried - ### Current Behavior
When opening the Policy Violations tab, the PolicyViolationResource throws the following error and no violations are shown. I think this bug was introduced somewhere last week...
``
javax.jdo.JDOUserException: Field org.dependencytrack.model.Component.repositoryMeta is not marked as persistent so cannot be queried
at org.datanucleus.api.jdo.JDOAdapter.getJDOExceptionForNucleusException(JDOAdapter.java:698)
at org.datanucleus.api.jdo.JDOQuery.executeInternal(JDOQuery.java:456)
at org.datanucleus.api.jdo.JDOQuery.execute(JDOQuery.java:276)
at alpine.persistence.AbstractAlpineQueryManager.execute(AbstractAlpineQueryManager.java:174)
at org.dependencytrack.persistence.PolicyQueryManager.getPolicyViolations(PolicyQueryManager.java:280)
at org.dependencytrack.persistence.QueryManager.getPolicyViolations(QueryManager.java:604)
at org.dependencytrack.resources.v1.PolicyViolationResource.getViolationsByProject(PolicyViolationResource.java:102)
### Steps to Reproduce
1.Open the Policy Violations tab for a project in the frontend
### Expected Behavior
No Exception is thrown and violations are shown (if any)
### Dependency-Track Version
4.8.0-SNAPSHOT
### Dependency-Track Distribution
Executable WAR
### Database Server
H2
### Database Server Version
_No response_
### Browser
Google Chrome
### Checklist
- [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues)
- [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported | non_process | javax jdo jdouserexception field org dependencytrack model component repositorymeta is not marked as persistent so cannot be queried current behavior when opening the policy violations tab the policyviolationresource throws the following error and no violations are shown i think this bug was introduced somewhere last week javax jdo jdouserexception field org dependencytrack model component repositorymeta is not marked as persistent so cannot be queried at org datanucleus api jdo jdoadapter getjdoexceptionfornucleusexception jdoadapter java at org datanucleus api jdo jdoquery executeinternal jdoquery java at org datanucleus api jdo jdoquery execute jdoquery java at alpine persistence abstractalpinequerymanager execute abstractalpinequerymanager java at org dependencytrack persistence policyquerymanager getpolicyviolations policyquerymanager java at org dependencytrack persistence querymanager getpolicyviolations querymanager java at org dependencytrack resources policyviolationresource getviolationsbyproject policyviolationresource java steps to reproduce open the policy violations tab for a project in the frontend expected behavior no exception is thrown and violations are shown if any dependency track version snapshot dependency track distribution executable war database server database server version no response browser google chrome checklist i have read and understand the i have checked the for whether this defect was already reported | 0 |
47,346 | 5,887,959,365 | IssuesEvent | 2017-05-17 08:55:58 | EyeSeeTea/dhis2-android-datacapture | https://api.github.com/repos/EyeSeeTea/dhis2-android-datacapture | closed | Data elements on android do not appear in same order as web for datasets without sections | complexity - med (1-5hr) priority - medium testing type - bug | When entering data for a dataset WITHOUT sections, the data elements appear in a different order on the web version versus in the Android app.
Attached screenshots show the first part of play demo dataset "Clinical Monitoring Checklist" as it appears on the web and on the tablet.
This was tested on latest.mis-pis.org and play demo. | 1.0 | Data elements on android do not appear in same order as web for datasets without sections - When entering data for a dataset WITHOUT sections, the data elements appear in a different order on the web version versus in the Android app.
Attached screenshots show the first part of play demo dataset "Clinical Monitoring Checklist" as it appears on the web and on the tablet.
This was tested on latest.mis-pis.org and play demo. | non_process | data elements on android do not appear in same order as web for datasets without sections when entering data for a dataset without sections the data elements appear in a different order on the web version versus in the android app attached screenshots show the first part of play demo dataset clinical monitoring checklist as it appears on the web and on the tablet this was tested on latest mis pis org and play demo | 0 |
809,324 | 30,186,560,695 | IssuesEvent | 2023-07-04 12:31:39 | geneontology/go-annotation | https://api.github.com/repos/geneontology/go-annotation | closed | problem with PAINT family, or annotation transfer or both PANTHER:PTN000331677 | PAINT annotation high priority | * **PTHR ID & PTN node:**
see
https://www.ebi.ac.uk/QuickGO/annotations?geneProductId=P0AEH3
This is transferring
GO:0004343 glucosamine 6-phosphate N-acetyltransferase activity
GO:0006048 UDP-N-acetylglucosamine biosynthetic process
GO:0005793 endoplasmic reticulum-Golgi intermediate compartment
from eukaryotes to bacteria
* Add the label 'high priority' if needed. Generally high priority issues affect a lot of proteins, and annotations are incorrect (as opposed to just imprecise).
| 1.0 | problem with PAINT family, or annotation transfer or both PANTHER:PTN000331677 - * **PTHR ID & PTN node:**
see
https://www.ebi.ac.uk/QuickGO/annotations?geneProductId=P0AEH3
This is transferring
GO:0004343 glucosamine 6-phosphate N-acetyltransferase activity
GO:0006048 UDP-N-acetylglucosamine biosynthetic process
GO:0005793 endoplasmic reticulum-Golgi intermediate compartment
from eukaryotes to bacteria
* Add the label 'high priority' if needed. Generally high priority issues affect a lot of proteins, and annotations are incorrect (as opposed to just imprecise).
| non_process | problem with paint family or annotation transfer or both panther pthr id ptn node see this is transferring go glucosamine phosphate n acetyltransferase activity go udp n acetylglucosamine biosynthetic process go endoplasmic reticulum golgi intermediate compartment from eukaryotes to bacteria add the label high priority if needed generally high priority issues affect a lot of proteins and annotations are incorrect as opposed to just imprecise | 0 |
20,919 | 27,755,793,271 | IssuesEvent | 2023-03-16 02:17:25 | 0xPolygonMiden/miden-vm | https://api.github.com/repos/0xPolygonMiden/miden-vm | opened | Replace MerkleSets in the advice provider with MerkleStore | processor | Now that we have [MerkleStore](https://github.com/0xPolygonMiden/crypto/blob/next/src/merkle/store.rs) implemented in `miden-crypto`, we should use it for instead of a map of `MerkleSet`'s in [MemAdviceProvider](https://github.com/0xPolygonMiden/miden-vm/blob/main/processor/src/advice/mem_provider.rs). We should also probable rename the fields like this:
* `tape` -> `stack`
* `values` -> `map`
* `sets` -> store`
So, the `MemAdviceProvider` struct could look like this:
```Rust
pub struct MemAdviceProvider {
step: u32,
stack: Vec<Felt>,
map: BTreeMap<[u8; 32], Vec<Felt>>,
store: MerkleStore,
}
```
We might also want to consider renaming `tape` into `stack` in the [AdviceProvider](https://github.com/0xPolygonMiden/miden-vm/blob/main/processor/src/advice/mod.rs#L52) trait as well. | 1.0 | Replace MerkleSets in the advice provider with MerkleStore - Now that we have [MerkleStore](https://github.com/0xPolygonMiden/crypto/blob/next/src/merkle/store.rs) implemented in `miden-crypto`, we should use it for instead of a map of `MerkleSet`'s in [MemAdviceProvider](https://github.com/0xPolygonMiden/miden-vm/blob/main/processor/src/advice/mem_provider.rs). We should also probable rename the fields like this:
* `tape` -> `stack`
* `values` -> `map`
* `sets` -> store`
So, the `MemAdviceProvider` struct could look like this:
```Rust
pub struct MemAdviceProvider {
step: u32,
stack: Vec<Felt>,
map: BTreeMap<[u8; 32], Vec<Felt>>,
store: MerkleStore,
}
```
We might also want to consider renaming `tape` into `stack` in the [AdviceProvider](https://github.com/0xPolygonMiden/miden-vm/blob/main/processor/src/advice/mod.rs#L52) trait as well. | process | replace merklesets in the advice provider with merklestore now that we have implemented in miden crypto we should use it for instead of a map of merkleset s in we should also probable rename the fields like this tape stack values map sets store so the memadviceprovider struct could look like this rust pub struct memadviceprovider step stack vec map btreemap store merklestore we might also want to consider renaming tape into stack in the trait as well | 1 |
3,379 | 6,502,139,987 | IssuesEvent | 2017-08-23 12:37:24 | syndesisio/syndesis-ui | https://api.github.com/repos/syndesisio/syndesis-ui | opened | Automate Documentation Builds | dev process | At the moment you have to manually generate docs with `compodoc` or with the CompoDoc server `compodoc -s`. We should hook this up to our CircleCI GH hook to build on push. | 1.0 | Automate Documentation Builds - At the moment you have to manually generate docs with `compodoc` or with the CompoDoc server `compodoc -s`. We should hook this up to our CircleCI GH hook to build on push. | process | automate documentation builds at the moment you have to manually generate docs with compodoc or with the compodoc server compodoc s we should hook this up to our circleci gh hook to build on push | 1 |
361,303 | 10,707,058,880 | IssuesEvent | 2019-10-24 16:37:04 | smartdevicelink/sdl_core | https://api.github.com/repos/smartdevicelink/sdl_core | closed | Remove unused ShowCustomForm HMI RPC | best practice low priority | ### Bug Report
[UI.ShowCustomForm](https://github.com/smartdevicelink/sdl_core/blob/master/src/components/interfaces/HMI_API.xml#L4392) is an unimplemented RPC in the HMI API, it should be removed.
##### Reproduction Steps
N/A
##### OS & Version Information
* SDL Core Version: https://github.com/smartdevicelink/sdl_core/commit/d36316738785c96dab2ee892762ed08c059fffde
##### Test Case, Sample Code, and / or Example App
N/A | 1.0 | Remove unused ShowCustomForm HMI RPC - ### Bug Report
[UI.ShowCustomForm](https://github.com/smartdevicelink/sdl_core/blob/master/src/components/interfaces/HMI_API.xml#L4392) is an unimplemented RPC in the HMI API, it should be removed.
##### Reproduction Steps
N/A
##### OS & Version Information
* SDL Core Version: https://github.com/smartdevicelink/sdl_core/commit/d36316738785c96dab2ee892762ed08c059fffde
##### Test Case, Sample Code, and / or Example App
N/A | non_process | remove unused showcustomform hmi rpc bug report is an unimplemented rpc in the hmi api it should be removed reproduction steps n a os version information sdl core version test case sample code and or example app n a | 0 |
241,475 | 18,457,647,298 | IssuesEvent | 2021-10-15 18:48:12 | Ek-Z/course-aggregator | https://api.github.com/repos/Ek-Z/course-aggregator | closed | Используемые версии/ настроить в Openserver | documentation | PHP 7.4
MySQL 8.0
HTTP Apache_2.4-PHP_7.2-7.4
<img width="439" alt="Снимок экрана 2021-10-14 150849" src="https://user-images.githubusercontent.com/74703154/137315182-efa8a6ba-f86a-4f0b-8798-78ea409e9bb0.png">
| 1.0 | Используемые версии/ настроить в Openserver - PHP 7.4
MySQL 8.0
HTTP Apache_2.4-PHP_7.2-7.4
<img width="439" alt="Снимок экрана 2021-10-14 150849" src="https://user-images.githubusercontent.com/74703154/137315182-efa8a6ba-f86a-4f0b-8798-78ea409e9bb0.png">
| non_process | используемые версии настроить в openserver php mysql http apache php img width alt снимок экрана src | 0 |
6,253 | 9,214,291,861 | IssuesEvent | 2019-03-10 18:44:13 | chuminh712/BookStorage---Group-2 | https://api.github.com/repos/chuminh712/BookStorage---Group-2 | reopened | Detail Design | In Process | Design class diagram for Use Case Manage Goods Issue
Design class diagram for Use Case Manage Goods Issue Info | 1.0 | Detail Design - Design class diagram for Use Case Manage Goods Issue
Design class diagram for Use Case Manage Goods Issue Info | process | detail design design class diagram for use case manage goods issue design class diagram for use case manage goods issue info | 1 |
129,882 | 5,105,189,749 | IssuesEvent | 2017-01-05 05:55:05 | HuskieRobotics/roborioExpansion | https://api.github.com/repos/HuskieRobotics/roborioExpansion | closed | Don't clear entire buffer before each read | low-priority Propeller FW | ```bytefill(@serialBuffer, 0, 256)```
This really just isn't needed, we can just add a null byte after our last read byte, and everything should stay functional. This will speed up those routines considerably. | 1.0 | Don't clear entire buffer before each read - ```bytefill(@serialBuffer, 0, 256)```
This really just isn't needed, we can just add a null byte after our last read byte, and everything should stay functional. This will speed up those routines considerably. | non_process | don t clear entire buffer before each read bytefill serialbuffer this really just isn t needed we can just add a null byte after our last read byte and everything should stay functional this will speed up those routines considerably | 0 |
240,479 | 20,033,381,689 | IssuesEvent | 2022-02-02 09:17:26 | keycloak/keycloak | https://api.github.com/repos/keycloak/keycloak | opened | Test more recent versions of Spring Boot | area/testsuite kind/enhancement area/adapter/spring | ### Description
We should test Spring Boot 2.4 and 2.6 in our pipeline in order to cover the whole test domain for Spring Boot and ensure the adapters work with those versions.
Original issue: CIAM-1419
### Discussion
_No response_
### Motivation
_No response_
### Details
_No response_ | 1.0 | Test more recent versions of Spring Boot - ### Description
We should test Spring Boot 2.4 and 2.6 in our pipeline in order to cover the whole test domain for Spring Boot and ensure the adapters work with those versions.
Original issue: CIAM-1419
### Discussion
_No response_
### Motivation
_No response_
### Details
_No response_ | non_process | test more recent versions of spring boot description we should test spring boot and in our pipeline in order to cover the whole test domain for spring boot and ensure the adapters work with those versions original issue ciam discussion no response motivation no response details no response | 0 |
109,104 | 16,829,413,042 | IssuesEvent | 2021-06-18 00:42:18 | kevins01/CSPF-Founder | https://api.github.com/repos/kevins01/CSPF-Founder | opened | CVE-2019-14900 (Medium) detected in hibernate-core-4.0.1.Final.jar | security vulnerability | ## CVE-2019-14900 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hibernate-core-4.0.1.Final.jar</b></p></summary>
<p>A module of the Hibernate Core project</p>
<p>Library home page: <a href="http://hibernate.org">http://hibernate.org</a></p>
<p>Path to dependency file: CSPF-Founder/pom.xml</p>
<p>Path to vulnerable library: CSPF-Founder/target/JavaVulnerableLab/WEB-INF/lib/hibernate-core-4.0.1.Final.jar,canner/.m2/repository/org/hibernate/hibernate-core/4.0.1.Final/hibernate-core-4.0.1.Final.jar</p>
<p>
Dependency Hierarchy:
- :x: **hibernate-core-4.0.1.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kevins01/CSPF-Founder/commit/c1676893423d574bf87c62c26df4f3d097d8fd51">c1676893423d574bf87c62c26df4f3d097d8fd51</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in Hibernate ORM in versions before 5.3.18, 5.4.18 and 5.5.0.Beta1. A SQL injection in the implementation of the JPA Criteria API can permit unsanitized literals when a literal is used in the SELECT or GROUP BY parts of the query. This flaw could allow an attacker to access unauthorized information or possibly conduct further attacks.
<p>Publish Date: 2020-07-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14900>CVE-2019-14900</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14900">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14900</a></p>
<p>Release Date: 2020-07-06</p>
<p>Fix Resolution: org.hibernate:hibernate-core:5.4.18.Final</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-14900 (Medium) detected in hibernate-core-4.0.1.Final.jar - ## CVE-2019-14900 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hibernate-core-4.0.1.Final.jar</b></p></summary>
<p>A module of the Hibernate Core project</p>
<p>Library home page: <a href="http://hibernate.org">http://hibernate.org</a></p>
<p>Path to dependency file: CSPF-Founder/pom.xml</p>
<p>Path to vulnerable library: CSPF-Founder/target/JavaVulnerableLab/WEB-INF/lib/hibernate-core-4.0.1.Final.jar,canner/.m2/repository/org/hibernate/hibernate-core/4.0.1.Final/hibernate-core-4.0.1.Final.jar</p>
<p>
Dependency Hierarchy:
- :x: **hibernate-core-4.0.1.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kevins01/CSPF-Founder/commit/c1676893423d574bf87c62c26df4f3d097d8fd51">c1676893423d574bf87c62c26df4f3d097d8fd51</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in Hibernate ORM in versions before 5.3.18, 5.4.18 and 5.5.0.Beta1. A SQL injection in the implementation of the JPA Criteria API can permit unsanitized literals when a literal is used in the SELECT or GROUP BY parts of the query. This flaw could allow an attacker to access unauthorized information or possibly conduct further attacks.
<p>Publish Date: 2020-07-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14900>CVE-2019-14900</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14900">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14900</a></p>
<p>Release Date: 2020-07-06</p>
<p>Fix Resolution: org.hibernate:hibernate-core:5.4.18.Final</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve medium detected in hibernate core final jar cve medium severity vulnerability vulnerable library hibernate core final jar a module of the hibernate core project library home page a href path to dependency file cspf founder pom xml path to vulnerable library cspf founder target javavulnerablelab web inf lib hibernate core final jar canner repository org hibernate hibernate core final hibernate core final jar dependency hierarchy x hibernate core final jar vulnerable library found in head commit a href vulnerability details a flaw was found in hibernate orm in versions before and a sql injection in the implementation of the jpa criteria api can permit unsanitized literals when a literal is used in the select or group by parts of the query this flaw could allow an attacker to access unauthorized information or possibly conduct further attacks publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org hibernate hibernate core final step up your open source security game with whitesource | 0 |
20,467 | 27,129,039,233 | IssuesEvent | 2023-02-16 08:24:40 | bazelbuild/bazel | https://api.github.com/repos/bazelbuild/bazel | closed | `implies: ...` does not stack well with `static_link_msvcrt` | P4 type: support / not a bug (process) team-Rules-CPP stale | ### Description of the problem / feature request:
Hi. I can not enable `static_link_msvcrt` from another feature:
```
imply_static_link_msvcrt_feature = feature(
name = "imply_static_link_msvcrt",
implies = ["static_link_msvcrt"]
)
```
When I do:
```
cc_binary(
name = "hello-world",
srcs = ["hello-world.cpp"],
features = ["imply_static_link_msvcrt"],
linkopts = ["-DEFAULTLIB:libcmt.lib", "-WX"] # Ensuring static_link_msvcrt is enabled.
)
```
linker says that:
```
LINK : warning LNK4098: defaultlib 'msvcrt.lib' conflicts with use of other libs; use /NODEFAULTLIB:library
LINK : error LNK1218: warning treated as error; no output file generated
```
Second thing is that if I use non-existent feature, it is silently ignored by Bazel. I feel that that is not right.
See https://github.com/excitoon/bazel-issues/tree/master/imply-static-link-msvcrt
### What operating system are you running Bazel on?
Windows 10 x64
### What's the output of `bazel info release`?
Build label: 0.23.0
Build target: bazel-out/x64_windows-opt/bin/src/main/java/com/google/devtools/build/lib/bazel/BazelServer_deploy.jar
Build time: Tue Feb 26 17:40:30 2019 (1551202830)
Build timestamp: 1551202830
Build timestamp as int: 1551202830
| 1.0 | `implies: ...` does not stack well with `static_link_msvcrt` - ### Description of the problem / feature request:
Hi. I can not enable `static_link_msvcrt` from another feature:
```
imply_static_link_msvcrt_feature = feature(
name = "imply_static_link_msvcrt",
implies = ["static_link_msvcrt"]
)
```
When I do:
```
cc_binary(
name = "hello-world",
srcs = ["hello-world.cpp"],
features = ["imply_static_link_msvcrt"],
linkopts = ["-DEFAULTLIB:libcmt.lib", "-WX"] # Ensuring static_link_msvcrt is enabled.
)
```
linker says that:
```
LINK : warning LNK4098: defaultlib 'msvcrt.lib' conflicts with use of other libs; use /NODEFAULTLIB:library
LINK : error LNK1218: warning treated as error; no output file generated
```
Second thing is that if I use non-existent feature, it is silently ignored by Bazel. I feel that that is not right.
See https://github.com/excitoon/bazel-issues/tree/master/imply-static-link-msvcrt
### What operating system are you running Bazel on?
Windows 10 x64
### What's the output of `bazel info release`?
Build label: 0.23.0
Build target: bazel-out/x64_windows-opt/bin/src/main/java/com/google/devtools/build/lib/bazel/BazelServer_deploy.jar
Build time: Tue Feb 26 17:40:30 2019 (1551202830)
Build timestamp: 1551202830
Build timestamp as int: 1551202830
| process | implies does not stack well with static link msvcrt description of the problem feature request hi i can not enable static link msvcrt from another feature imply static link msvcrt feature feature name imply static link msvcrt implies when i do cc binary name hello world srcs features linkopts ensuring static link msvcrt is enabled linker says that link warning defaultlib msvcrt lib conflicts with use of other libs use nodefaultlib library link error warning treated as error no output file generated second thing is that if i use non existent feature it is silently ignored by bazel i feel that that is not right see what operating system are you running bazel on windows what s the output of bazel info release build label build target bazel out windows opt bin src main java com google devtools build lib bazel bazelserver deploy jar build time tue feb build timestamp build timestamp as int | 1 |
20,444 | 27,101,156,693 | IssuesEvent | 2023-02-15 08:46:30 | bazelbuild/bazel | https://api.github.com/repos/bazelbuild/bazel | closed | Clean up `@bazel_tools` | P3 type: process team-OSS stale | # Description of the problem / feature request:
This is an umbrella issue to track efforts to clean up `@bazel_tools`. We should remove unnecessary source code from this repo, and make it work for `bazel query @bazel_tools//...`.
## Details
This repo is built into Bazel. It's always available in every workspace. It's a place for commonly used tools like `tools/test/test-setup.sh`, which is the test wrapper used for every `bazel test` action.
But there is also source code in this repo, each file for their own reason. E.g. the runfiles libraries `@bazel_tools//tools/<lang>/runfiles` are included as source so they can be compiled into binaries.
Some source code's purpose in `@bazel_tools` is less obvious. Sometimes they cause problems, like https://github.com/bazelbuild/bazel/issues/8614 whose culprit is https://github.com/bazelbuild/bazel/pull/8742#issuecomment-506745688.
### What's the output of `bazel info release`?
`release 0.28.0` | 1.0 | Clean up `@bazel_tools` - # Description of the problem / feature request:
This is an umbrella issue to track efforts to clean up `@bazel_tools`. We should remove unnecessary source code from this repo, and make it work for `bazel query @bazel_tools//...`.
## Details
This repo is built into Bazel. It's always available in every workspace. It's a place for commonly used tools like `tools/test/test-setup.sh`, which is the test wrapper used for every `bazel test` action.
But there is also source code in this repo, each file for their own reason. E.g. the runfiles libraries `@bazel_tools//tools/<lang>/runfiles` are included as source so they can be compiled into binaries.
Some source code's purpose in `@bazel_tools` is less obvious. Sometimes they cause problems, like https://github.com/bazelbuild/bazel/issues/8614 whose culprit is https://github.com/bazelbuild/bazel/pull/8742#issuecomment-506745688.
### What's the output of `bazel info release`?
`release 0.28.0` | process | clean up bazel tools description of the problem feature request this is an umbrella issue to track efforts to clean up bazel tools we should remove unnecessary source code from this repo and make it work for bazel query bazel tools details this repo is built into bazel it s always available in every workspace it s a place for commonly used tools like tools test test setup sh which is the test wrapper used for every bazel test action but there is also source code in this repo each file for their own reason e g the runfiles libraries bazel tools tools runfiles are included as source so they can be compiled into binaries some source code s purpose in bazel tools is less obvious sometimes they cause problems like whose culprit is what s the output of bazel info release release | 1 |
356,564 | 25,176,214,188 | IssuesEvent | 2022-11-11 09:29:22 | peppapighs/pe | https://api.github.com/repos/peppapighs/pe | opened | Student class diagram should show `ModuleCode` class separately | severity.VeryLow type.DocumentationBug | Since the class `ModuleCode` is a custom class and is never mentioned before in the DG, the diagram should show this class separately and have the `Student` class assoicated with `ModuleCode` with `*` as the multiplicity.

<!--session: 1668153839735-43773ef4-6587-4127-9391-a44f44a8a6a7-->
<!--Version: Web v3.4.4--> | 1.0 | Student class diagram should show `ModuleCode` class separately - Since the class `ModuleCode` is a custom class and is never mentioned before in the DG, the diagram should show this class separately and have the `Student` class assoicated with `ModuleCode` with `*` as the multiplicity.

<!--session: 1668153839735-43773ef4-6587-4127-9391-a44f44a8a6a7-->
<!--Version: Web v3.4.4--> | non_process | student class diagram should show modulecode class separately since the class modulecode is a custom class and is never mentioned before in the dg the diagram should show this class separately and have the student class assoicated with modulecode with as the multiplicity | 0 |
142,738 | 11,491,618,729 | IssuesEvent | 2020-02-11 19:20:50 | GetTerminus/terminus-ui | https://api.github.com/repos/GetTerminus/terminus-ui | closed | Use new type guards from ngx-tools and remove all internal type guards | Focus: utility Goal: Library Stabilization Target: latest Type: chore | All existing UI library type guards will be added to ngx-tools by https://github.com/GetTerminus/ngx-tools/issues/292
- [ ] Use type guards from ngx-tools
- [ ] Remove all internal type guards
- [ ] Verify linting and tests still pass | 1.0 | Use new type guards from ngx-tools and remove all internal type guards - All existing UI library type guards will be added to ngx-tools by https://github.com/GetTerminus/ngx-tools/issues/292
- [ ] Use type guards from ngx-tools
- [ ] Remove all internal type guards
- [ ] Verify linting and tests still pass | non_process | use new type guards from ngx tools and remove all internal type guards all existing ui library type guards will be added to ngx tools by use type guards from ngx tools remove all internal type guards verify linting and tests still pass | 0 |
10,876 | 13,645,035,281 | IssuesEvent | 2020-09-25 20:00:36 | pystatgen/sgkit | https://api.github.com/repos/pystatgen/sgkit | closed | Apply to NumFOCUS by October 15 | process + tools | A reminder to myself from a previous [discussion](https://discourse.pystatgen.org/t/what-should-we-call-our-new-genetics-toolkit/52/11?u=hammer) in June.
I think we may want a logo too #30 | 1.0 | Apply to NumFOCUS by October 15 - A reminder to myself from a previous [discussion](https://discourse.pystatgen.org/t/what-should-we-call-our-new-genetics-toolkit/52/11?u=hammer) in June.
I think we may want a logo too #30 | process | apply to numfocus by october a reminder to myself from a previous in june i think we may want a logo too | 1 |
18,018 | 24,032,775,826 | IssuesEvent | 2022-09-15 16:18:34 | googleapis/java-beyondcorp-appconnectors | https://api.github.com/repos/googleapis/java-beyondcorp-appconnectors | opened | Your .repo-metadata.json file has a problem 🤒 | type: process repo-metadata: lint | You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* api_shortname 'beyondcorp-appconnectors' invalid in .repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions. | 1.0 | Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* api_shortname 'beyondcorp-appconnectors' invalid in .repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions. | process | your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 api shortname beyondcorp appconnectors invalid in repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions | 1 |
53,259 | 13,261,294,537 | IssuesEvent | 2020-08-20 19:37:57 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | [steamshovel] RangeSetting compile error (Trac #1059) | Migrated from Trac combo core defect | Fails to compile on SL6 cvmfs py2-v1:
```text
[ 80%] Building CXX object steamshovel/CMakeFiles/shovelart.dir/private/shovelart/artists/CoordinateSystem.cpp.o
In file included from
/scratch/dschultz/offline/src/steamshovel/private/shovelart/Artist.h:25,
from
/scratch/dschultz/offline/src/steamshovel/private/shovelart/artists/Coordin
ateSystem.h:4,
from /scratch/dschultz/offline/src/steamshovel/private/shovelart/artists/CoordinateSystem.cpp:8:
/scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h: In
constructor ‘RangeSetting::RangeSetting(float, float, int, double)’:
/scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h:40:
error: class ‘RangeSetting’ does not have any field named ‘
ComplexSetting’
/scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h:43:
error: No match for ‘ComplexSetting<float>::ComplexSetting()’
/scratch/dschultz/offline/src/steamshovel/private/shovelart/ComplexSetting.h:16:
note: candidates are: ComplexSetting<float>::ComplexSetting(float)
/scratch/dschultz/offline/src/steamshovel/private/shovelart/ComplexSetting.h:12:
note: ComplexSetting<float>::ComplexSetting(const ComplexSetting<float> &)
/scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h: In
constructor ‘RangeSetting::RangeSetting(float, float, int, float)’:
```
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1059">https://code.icecube.wisc.edu/projects/icecube/ticket/1059</a>, reported by david.schultzand owned by hdembinski</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-08-10T20:31:17",
"_ts": "1439238677802127",
"description": "Fails to compile on SL6 cvmfs py2-v1:\n\n{{{\n[ 80%] Building CXX object steamshovel/CMakeFiles/shovelart.dir/private/shovelart/artists/CoordinateSystem.cpp.o\nIn file included from\n /scratch/dschultz/offline/src/steamshovel/private/shovelart/Artist.h:25,\nfrom\n /scratch/dschultz/offline/src/steamshovel/private/shovelart/artists/Coordin\n ateSystem.h:4,\nfrom /scratch/dschultz/offline/src/steamshovel/private/shovelart/artists/CoordinateSystem.cpp:8:\n /scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h: In\n constructor \u2018RangeSetting::RangeSetting(float, float, int, double)\u2019:\n/scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h:40:\n error: class \u2018RangeSetting\u2019 does not have any field named \u2018\n ComplexSetting\u2019\n/scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h:43:\n error: No match for \u2018ComplexSetting<float>::ComplexSetting()\u2019\n/scratch/dschultz/offline/src/steamshovel/private/shovelart/ComplexSetting.h:16:\n note: candidates are: ComplexSetting<float>::ComplexSetting(float)\n/scratch/dschultz/offline/src/steamshovel/private/shovelart/ComplexSetting.h:12:\n note: ComplexSetting<float>::ComplexSetting(const ComplexSetting<float> &)\n/scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h: In\n constructor \u2018RangeSetting::RangeSetting(float, float, int, float)\u2019:\n}}}",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"time": "2015-07-20T17:26:52",
"component": "combo core",
"summary": "[steamshovel] RangeSetting compile error",
"priority": "blocker",
"keywords": "",
"milestone": "",
"owner": "hdembinski",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [steamshovel] RangeSetting compile error (Trac #1059) - Fails to compile on SL6 cvmfs py2-v1:
```text
[ 80%] Building CXX object steamshovel/CMakeFiles/shovelart.dir/private/shovelart/artists/CoordinateSystem.cpp.o
In file included from
/scratch/dschultz/offline/src/steamshovel/private/shovelart/Artist.h:25,
from
/scratch/dschultz/offline/src/steamshovel/private/shovelart/artists/Coordin
ateSystem.h:4,
from /scratch/dschultz/offline/src/steamshovel/private/shovelart/artists/CoordinateSystem.cpp:8:
/scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h: In
constructor ‘RangeSetting::RangeSetting(float, float, int, double)’:
/scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h:40:
error: class ‘RangeSetting’ does not have any field named ‘
ComplexSetting’
/scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h:43:
error: No match for ‘ComplexSetting<float>::ComplexSetting()’
/scratch/dschultz/offline/src/steamshovel/private/shovelart/ComplexSetting.h:16:
note: candidates are: ComplexSetting<float>::ComplexSetting(float)
/scratch/dschultz/offline/src/steamshovel/private/shovelart/ComplexSetting.h:12:
note: ComplexSetting<float>::ComplexSetting(const ComplexSetting<float> &)
/scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h: In
constructor ‘RangeSetting::RangeSetting(float, float, int, float)’:
```
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1059">https://code.icecube.wisc.edu/projects/icecube/ticket/1059</a>, reported by david.schultzand owned by hdembinski</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-08-10T20:31:17",
"_ts": "1439238677802127",
"description": "Fails to compile on SL6 cvmfs py2-v1:\n\n{{{\n[ 80%] Building CXX object steamshovel/CMakeFiles/shovelart.dir/private/shovelart/artists/CoordinateSystem.cpp.o\nIn file included from\n /scratch/dschultz/offline/src/steamshovel/private/shovelart/Artist.h:25,\nfrom\n /scratch/dschultz/offline/src/steamshovel/private/shovelart/artists/Coordin\n ateSystem.h:4,\nfrom /scratch/dschultz/offline/src/steamshovel/private/shovelart/artists/CoordinateSystem.cpp:8:\n /scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h: In\n constructor \u2018RangeSetting::RangeSetting(float, float, int, double)\u2019:\n/scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h:40:\n error: class \u2018RangeSetting\u2019 does not have any field named \u2018\n ComplexSetting\u2019\n/scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h:43:\n error: No match for \u2018ComplexSetting<float>::ComplexSetting()\u2019\n/scratch/dschultz/offline/src/steamshovel/private/shovelart/ComplexSetting.h:16:\n note: candidates are: ComplexSetting<float>::ComplexSetting(float)\n/scratch/dschultz/offline/src/steamshovel/private/shovelart/ComplexSetting.h:12:\n note: ComplexSetting<float>::ComplexSetting(const ComplexSetting<float> &)\n/scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h: In\n constructor \u2018RangeSetting::RangeSetting(float, float, int, float)\u2019:\n}}}",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"time": "2015-07-20T17:26:52",
"component": "combo core",
"summary": "[steamshovel] RangeSetting compile error",
"priority": "blocker",
"keywords": "",
"milestone": "",
"owner": "hdembinski",
"type": "defect"
}
```
</p>
</details>
| non_process | rangesetting compile error trac fails to compile on cvmfs text building cxx object steamshovel cmakefiles shovelart dir private shovelart artists coordinatesystem cpp o in file included from scratch dschultz offline src steamshovel private shovelart artist h from scratch dschultz offline src steamshovel private shovelart artists coordin atesystem h from scratch dschultz offline src steamshovel private shovelart artists coordinatesystem cpp scratch dschultz offline src steamshovel private shovelart rangesetting h in constructor ‘rangesetting rangesetting float float int double ’ scratch dschultz offline src steamshovel private shovelart rangesetting h error class ‘rangesetting’ does not have any field named ‘ complexsetting’ scratch dschultz offline src steamshovel private shovelart rangesetting h error no match for ‘complexsetting complexsetting ’ scratch dschultz offline src steamshovel private shovelart complexsetting h note candidates are complexsetting complexsetting float scratch dschultz offline src steamshovel private shovelart complexsetting h note complexsetting complexsetting const complexsetting scratch dschultz offline src steamshovel private shovelart rangesetting h in constructor ‘rangesetting rangesetting float float int float ’ migrated from json status closed changetime ts description fails to compile on cvmfs n n n building cxx object steamshovel cmakefiles shovelart dir private shovelart artists coordinatesystem cpp o nin file included from n scratch dschultz offline src steamshovel private shovelart artist h nfrom n scratch dschultz offline src steamshovel private shovelart artists coordin n atesystem h nfrom scratch dschultz offline src steamshovel private shovelart artists coordinatesystem cpp n scratch dschultz offline src steamshovel private shovelart rangesetting h in n constructor rangesetting float float int double n scratch dschultz offline src steamshovel private shovelart rangesetting h n error class does not have any field named n complexsetting n scratch dschultz offline src steamshovel private shovelart rangesetting h n error no match for complexsetting n scratch dschultz offline src steamshovel private shovelart complexsetting h n note candidates are complexsetting complexsetting float n scratch dschultz offline src steamshovel private shovelart complexsetting h n note complexsetting complexsetting const complexsetting n scratch dschultz offline src steamshovel private shovelart rangesetting h in n constructor rangesetting float float int float n reporter david schultz cc resolution fixed time component combo core summary rangesetting compile error priority blocker keywords milestone owner hdembinski type defect | 0 |
578,879 | 17,156,547,032 | IssuesEvent | 2021-07-14 07:41:19 | googleapis/java-bigtable-hbase | https://api.github.com/repos/googleapis/java-bigtable-hbase | closed | bigtable.hbase.TestBasicOps: testPutAlmostTooBigValueTenTimes failed | api: bigtable flakybot: issue priority: p1 type: bug | This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: a891335ce3179c45fade4f3683b7e09d38d0107a
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/64905140-8b04-4270-9ed2-a91e73d97e39), [Sponge](http://sponge2/64905140-8b04-4270-9ed2-a91e73d97e39)
status: failed
<details><summary>Test output</summary><br><pre>org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: Failed 1 action: UnauthenticatedException: 1 time, servers with issues: bigtable.googleapis.com
at com.google.cloud.bigtable.hbase.AbstractBigtableTable.createRetriesExhaustedWithDetailsException(AbstractBigtableTable.java:274)
at com.google.cloud.bigtable.hbase.AbstractBigtableTable.exists(AbstractBigtableTable.java:170)
at com.google.cloud.bigtable.hbase.TestBasicOps.testPutGetDelete(TestBasicOps.java:226)
at com.google.cloud.bigtable.hbase.TestBasicOps.testPutGetDeleteExists(TestBasicOps.java:190)
at com.google.cloud.bigtable.hbase.TestBasicOps.testPutAlmostTooBigValueTenTimes(TestBasicOps.java:166)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:410)
at org.apache.maven.surefire.junitcore.pc.InvokerStrategy.schedule(InvokerStrategy.java:54)
at org.apache.maven.surefire.junitcore.pc.Scheduler.schedule(Scheduler.java:367)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.junit.runners.Suite.runChild(Suite.java:128)
at org.junit.runners.Suite.runChild(Suite.java:27)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:410)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
</pre></details> | 1.0 | bigtable.hbase.TestBasicOps: testPutAlmostTooBigValueTenTimes failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: a891335ce3179c45fade4f3683b7e09d38d0107a
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/64905140-8b04-4270-9ed2-a91e73d97e39), [Sponge](http://sponge2/64905140-8b04-4270-9ed2-a91e73d97e39)
status: failed
<details><summary>Test output</summary><br><pre>org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: Failed 1 action: UnauthenticatedException: 1 time, servers with issues: bigtable.googleapis.com
at com.google.cloud.bigtable.hbase.AbstractBigtableTable.createRetriesExhaustedWithDetailsException(AbstractBigtableTable.java:274)
at com.google.cloud.bigtable.hbase.AbstractBigtableTable.exists(AbstractBigtableTable.java:170)
at com.google.cloud.bigtable.hbase.TestBasicOps.testPutGetDelete(TestBasicOps.java:226)
at com.google.cloud.bigtable.hbase.TestBasicOps.testPutGetDeleteExists(TestBasicOps.java:190)
at com.google.cloud.bigtable.hbase.TestBasicOps.testPutAlmostTooBigValueTenTimes(TestBasicOps.java:166)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:410)
at org.apache.maven.surefire.junitcore.pc.InvokerStrategy.schedule(InvokerStrategy.java:54)
at org.apache.maven.surefire.junitcore.pc.Scheduler.schedule(Scheduler.java:367)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.junit.runners.Suite.runChild(Suite.java:128)
at org.junit.runners.Suite.runChild(Suite.java:27)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:410)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
</pre></details> | non_process | bigtable hbase testbasicops testputalmosttoobigvaluetentimes failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output org apache hadoop hbase client retriesexhaustedwithdetailsexception failed action unauthenticatedexception time servers with issues bigtable googleapis com at com google cloud bigtable hbase abstractbigtabletable createretriesexhaustedwithdetailsexception abstractbigtabletable java at com google cloud bigtable hbase abstractbigtabletable exists abstractbigtabletable java at com google cloud bigtable hbase testbasicops testputgetdelete testbasicops java at com google cloud bigtable hbase testbasicops testputgetdeleteexists testbasicops java at com google cloud bigtable hbase testbasicops testputalmosttoobigvaluetentimes testbasicops java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org apache maven surefire junitcore pc scheduler run scheduler java at org apache maven surefire junitcore pc invokerstrategy schedule invokerstrategy java at org apache maven surefire junitcore pc scheduler schedule scheduler java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org junit runners suite runchild suite java at org junit runners suite runchild suite java at org junit runners parentrunner run parentrunner java at org apache maven surefire junitcore pc scheduler run scheduler java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java | 0 |
7,180 | 10,321,200,398 | IssuesEvent | 2019-08-30 23:56:42 | googleapis/google-api-java-client | https://api.github.com/repos/googleapis/google-api-java-client | opened | Publish SNAPSHOT releases nightly | type: process | This will help us detect downstream breakage before releasing the library
- [x] add build config in this repo
- [ ] add job config internally | 1.0 | Publish SNAPSHOT releases nightly - This will help us detect downstream breakage before releasing the library
- [x] add build config in this repo
- [ ] add job config internally | process | publish snapshot releases nightly this will help us detect downstream breakage before releasing the library add build config in this repo add job config internally | 1 |
16,365 | 21,052,204,938 | IssuesEvent | 2022-03-31 21:37:40 | nion-software/nionswift | https://api.github.com/repos/nion-software/nionswift | closed | Setting xdata property on should accept anything that can be converted to xdata | type - enhancement level - easy f - processing f - API | `target.xdata = numpy.zeros((3,3))` should work. | 1.0 | Setting xdata property on should accept anything that can be converted to xdata - `target.xdata = numpy.zeros((3,3))` should work. | process | setting xdata property on should accept anything that can be converted to xdata target xdata numpy zeros should work | 1 |
273,066 | 20,768,983,091 | IssuesEvent | 2022-03-16 00:58:51 | gth828r/brownify | https://api.github.com/repos/gth828r/brownify | closed | Generate and publish sphinx documentation | documentation | We should set up documentation in our library, given that we have already started doing a reasonable amount of docstring writing and type hinting.
- [x] Initialize the repository for sphinx documentation generation
- [x] Publish the documentation somewhere
- [x] Update the setup.cfg to point to where the documentation lives
- [x] Update the top level README to point to where documentation lives | 1.0 | Generate and publish sphinx documentation - We should set up documentation in our library, given that we have already started doing a reasonable amount of docstring writing and type hinting.
- [x] Initialize the repository for sphinx documentation generation
- [x] Publish the documentation somewhere
- [x] Update the setup.cfg to point to where the documentation lives
- [x] Update the top level README to point to where documentation lives | non_process | generate and publish sphinx documentation we should set up documentation in our library given that we have already started doing a reasonable amount of docstring writing and type hinting initialize the repository for sphinx documentation generation publish the documentation somewhere update the setup cfg to point to where the documentation lives update the top level readme to point to where documentation lives | 0 |
734 | 3,214,313,663 | IssuesEvent | 2015-10-07 00:44:43 | broadinstitute/hellbender | https://api.github.com/repos/broadinstitute/hellbender | closed | Dataflow BQSR Direct Runner fails with --knownSites | bug Dataflow DataflowPreprocessingPipeline | From @tomwhite:
I noticed that the test with "-knownSites" from BaseRecalibratorIntegrationTest (i.e. the non-dataflow version) fails with both the Direct and Spark runners. I had a look at the file output and there are a few discrepancies (see diff below).
```
diff /var/folders/d1/8f5_j4hx04z72w6wgqxkb2l40000gn/T/walktest.tmp_param.02172067147450353519.tmp src/test/resources/org/broadinstitute/hellbender/tools/BQSR/expected.NA12878.chr17_69k_70k.2inputs.txt
60c60
< 34 3051 34
---
> 34 3050 34
71c71
< 45 46942 45
---
> 45 46940 45
124,126c124,126
< 809R9ABXX101220.5 D 45.0000 45.0000 23471 0.00
< 809R9ABXX101220.5 I 45.0000 45.0000 23471 0.00
< 809R9ABXX101220.5 M 27.0000 27.0494 23471 49.13
---
> 809R9ABXX101220.5 D 45.0000 45.0000 23470 0.00
> 809R9ABXX101220.5 I 45.0000 45.0000 23470 0.00
> 809R9ABXX101220.5 M 27.0000 27.0493 23470 49.13
155c155
< 809R9ABXX101220.5 34 M 34.0000 3051 2.96
---
> 809R9ABXX101220.5 34 M 34.0000 3050 2.96
161,162c161,162
< 809R9ABXX101220.5 45 D 45.0000 23471 0.00
< 809R9ABXX101220.5 45 I 45.0000 23471 0.00
---
> 809R9ABXX101220.5 45 D 45.0000 23470 0.00
> 809R9ABXX101220.5 45 I 45.0000 23470 0.00
2714c2714
< 809R9ABXX101220.5 34 29 Cycle M 34.0000 20 0.00
---
> 809R9ABXX101220.5 34 29 Cycle M 34.0000 19 0.00
2773c2773
< 809R9ABXX101220.5 34 CA Context M 34.0000 506 0.00
---
> 809R9ABXX101220.5 34 CA Context M 34.0000 505 0.00
3464,3465c3464,3465
< 809R9ABXX101220.5 45 29 Cycle D 45.0000 180 0.00
< 809R9ABXX101220.5 45 29 Cycle I 45.0000 180 0.00
---
> 809R9ABXX101220.5 45 29 Cycle D 45.0000 179 0.00
> 809R9ABXX101220.5 45 29 Cycle I 45.0000 179 0.00
3634,3635c3634,3635
< 809R9ABXX101220.5 45 GCA Context D 45.0000 278 0.00
< 809R9ABXX101220.5 45 GCA Context I 45.0000 278 0.00
---
> 809R9ABXX101220.5 45 GCA Context D 45.0000 277 0.00
> 809R9ABXX101220.5 45 GCA Context I 45.0000 277 0.00
```
The relevant test is this one from `BaseRecalibratorDataflowIntegrationTest`
```
new BQSRTest(hg18Reference, HiSeqBam, dbSNPb37, "-knownSites " + moreSites, getResourceDir() + "expected.NA12878.chr17_69k_70k.2inputs.txt")
``` | 1.0 | Dataflow BQSR Direct Runner fails with --knownSites - From @tomwhite:
I noticed that the test with "-knownSites" from BaseRecalibratorIntegrationTest (i.e. the non-dataflow version) fails with both the Direct and Spark runners. I had a look at the file output and there are a few discrepancies (see diff below).
```
diff /var/folders/d1/8f5_j4hx04z72w6wgqxkb2l40000gn/T/walktest.tmp_param.02172067147450353519.tmp src/test/resources/org/broadinstitute/hellbender/tools/BQSR/expected.NA12878.chr17_69k_70k.2inputs.txt
60c60
< 34 3051 34
---
> 34 3050 34
71c71
< 45 46942 45
---
> 45 46940 45
124,126c124,126
< 809R9ABXX101220.5 D 45.0000 45.0000 23471 0.00
< 809R9ABXX101220.5 I 45.0000 45.0000 23471 0.00
< 809R9ABXX101220.5 M 27.0000 27.0494 23471 49.13
---
> 809R9ABXX101220.5 D 45.0000 45.0000 23470 0.00
> 809R9ABXX101220.5 I 45.0000 45.0000 23470 0.00
> 809R9ABXX101220.5 M 27.0000 27.0493 23470 49.13
155c155
< 809R9ABXX101220.5 34 M 34.0000 3051 2.96
---
> 809R9ABXX101220.5 34 M 34.0000 3050 2.96
161,162c161,162
< 809R9ABXX101220.5 45 D 45.0000 23471 0.00
< 809R9ABXX101220.5 45 I 45.0000 23471 0.00
---
> 809R9ABXX101220.5 45 D 45.0000 23470 0.00
> 809R9ABXX101220.5 45 I 45.0000 23470 0.00
2714c2714
< 809R9ABXX101220.5 34 29 Cycle M 34.0000 20 0.00
---
> 809R9ABXX101220.5 34 29 Cycle M 34.0000 19 0.00
2773c2773
< 809R9ABXX101220.5 34 CA Context M 34.0000 506 0.00
---
> 809R9ABXX101220.5 34 CA Context M 34.0000 505 0.00
3464,3465c3464,3465
< 809R9ABXX101220.5 45 29 Cycle D 45.0000 180 0.00
< 809R9ABXX101220.5 45 29 Cycle I 45.0000 180 0.00
---
> 809R9ABXX101220.5 45 29 Cycle D 45.0000 179 0.00
> 809R9ABXX101220.5 45 29 Cycle I 45.0000 179 0.00
3634,3635c3634,3635
< 809R9ABXX101220.5 45 GCA Context D 45.0000 278 0.00
< 809R9ABXX101220.5 45 GCA Context I 45.0000 278 0.00
---
> 809R9ABXX101220.5 45 GCA Context D 45.0000 277 0.00
> 809R9ABXX101220.5 45 GCA Context I 45.0000 277 0.00
```
The relevant test is this one from `BaseRecalibratorDataflowIntegrationTest`
```
new BQSRTest(hg18Reference, HiSeqBam, dbSNPb37, "-knownSites " + moreSites, getResourceDir() + "expected.NA12878.chr17_69k_70k.2inputs.txt")
``` | process | dataflow bqsr direct runner fails with knownsites from tomwhite i noticed that the test with knownsites from baserecalibratorintegrationtest i e the non dataflow version fails with both the direct and spark runners i had a look at the file output and there are a few discrepancies see diff below diff var folders t walktest tmp param tmp src test resources org broadinstitute hellbender tools bqsr expected txt d i m d i m m m d i d i cycle m cycle m ca context m ca context m cycle d cycle i cycle d cycle i gca context d gca context i gca context d gca context i the relevant test is this one from baserecalibratordataflowintegrationtest new bqsrtest hiseqbam knownsites moresites getresourcedir expected txt | 1 |
22,405 | 31,142,291,542 | IssuesEvent | 2023-08-16 01:44:42 | cypress-io/cypress | https://api.github.com/repos/cypress-io/cypress | closed | Flaky test: Error: done() called multiple times VI | OS: linux process: flaky test topic: flake ❄️ stage: flake "topic: done()" stale | ### Link to dashboard or CircleCI failure
https://app.circleci.com/pipelines/github/cypress-io/cypress/41102/workflows/219dd1e6-3a6a-44b0-a811-d5cfe222f25e/jobs/1699154
### Link to failing test in GitHub
https://github.com/cypress-io/cypress/blob/develop/packages/driver/cypress/e2e/commands/actions/select.cy.js#L176
### Analysis
<img width="812" alt="Screen Shot 2022-08-05 at 12 39 30 PM" src="https://user-images.githubusercontent.com/26726429/183149058-79aeca18-fbd1-4906-b270-0f8578628b91.png">
### Cypress Version
10.4.0
### Other
Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed | 1.0 | Flaky test: Error: done() called multiple times VI - ### Link to dashboard or CircleCI failure
https://app.circleci.com/pipelines/github/cypress-io/cypress/41102/workflows/219dd1e6-3a6a-44b0-a811-d5cfe222f25e/jobs/1699154
### Link to failing test in GitHub
https://github.com/cypress-io/cypress/blob/develop/packages/driver/cypress/e2e/commands/actions/select.cy.js#L176
### Analysis
<img width="812" alt="Screen Shot 2022-08-05 at 12 39 30 PM" src="https://user-images.githubusercontent.com/26726429/183149058-79aeca18-fbd1-4906-b270-0f8578628b91.png">
### Cypress Version
10.4.0
### Other
Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed | process | flaky test error done called multiple times vi link to dashboard or circleci failure link to failing test in github analysis img width alt screen shot at pm src cypress version other search for this issue number in the codebase to find the test s skipped until this issue is fixed | 1 |
9,482 | 12,477,761,345 | IssuesEvent | 2020-05-29 15:29:21 | googleapis/google-cloud-go | https://api.github.com/repos/googleapis/google-cloud-go | opened | all: add a note compatibility in README.md | type: process | Something along of the lines of https://github.com/golang/protobuf#compatibility, but tailored for our repository. | 1.0 | all: add a note compatibility in README.md - Something along of the lines of https://github.com/golang/protobuf#compatibility, but tailored for our repository. | process | all add a note compatibility in readme md something along of the lines of but tailored for our repository | 1 |
684,277 | 23,413,402,080 | IssuesEvent | 2022-08-12 20:19:35 | python/mypy | https://api.github.com/repos/python/mypy | closed | Support recursive types | feature needs discussion priority-1-normal meta topic-recursive-types | The following in particular would be useful:
```
Callback = Callable[[str], 'Callback']
Foo = Union[str, List['Foo']]
```
| 1.0 | Support recursive types - The following in particular would be useful:
```
Callback = Callable[[str], 'Callback']
Foo = Union[str, List['Foo']]
```
| non_process | support recursive types the following in particular would be useful callback callable callback foo union | 0 |
408,307 | 27,660,818,191 | IssuesEvent | 2023-03-12 13:57:43 | libusb/hidapi | https://api.github.com/repos/libusb/hidapi | closed | HIDAPI web site | documentation | Someone should work on a web site presenting the hidapi project and subprojects (like Python wrapper).
goals:
* make the new hidapi project more visible so that it is indexed by search engine
* document the history from https://github.com/signal11/hidapi
One technical solution is to use github pages https://pages.github.com/ and host the web site at https://hidapi.libusb.info/ (the name does not exist yet but I can create it) | 1.0 | HIDAPI web site - Someone should work on a web site presenting the hidapi project and subprojects (like Python wrapper).
goals:
* make the new hidapi project more visible so that it is indexed by search engine
* document the history from https://github.com/signal11/hidapi
One technical solution is to use github pages https://pages.github.com/ and host the web site at https://hidapi.libusb.info/ (the name does not exist yet but I can create it) | non_process | hidapi web site someone should work on a web site presenting the hidapi project and subprojects like python wrapper goals make the new hidapi project more visible so that it is indexed by search engine document the history from one technical solution is to use github pages and host the web site at the name does not exist yet but i can create it | 0 |
124,827 | 16,668,539,367 | IssuesEvent | 2021-06-07 08:04:30 | crev-dev/cargo-crev | https://api.github.com/repos/crev-dev/cargo-crev | closed | Thoughts about packaging | design enhancement | This is sort of related to #37, but probably warrants its own issue:
I'm involved with rust packaging in two major linux distros and looked into packaging crev. It seems there are both tagged releases and cargo uploads, which is very good, but I'm not sure which packages should be distributed using which package name.
It seems cargo-crev is used to tag the binary crates, should this ship only `cargo-crev` or also the other binary targets? Is rblake2sum required as a binary?
On a side note, specifically for debian it's important to make sure you're depending on the latest version of your dependencies. While we can upload outdated crates to debian it introduces quite a bit of complexity for us.
If you have some opinions on that topic please let me know. :) | 1.0 | Thoughts about packaging - This is sort of related to #37, but probably warrants its own issue:
I'm involved with rust packaging in two major linux distros and looked into packaging crev. It seems there are both tagged releases and cargo uploads, which is very good, but I'm not sure which packages should be distributed using which package name.
It seems cargo-crev is used to tag the binary crates, should this ship only `cargo-crev` or also the other binary targets? Is rblake2sum required as a binary?
On a side note, specifically for debian it's important to make sure you're depending on the latest version of your dependencies. While we can upload outdated crates to debian it introduces quite a bit of complexity for us.
If you have some opinions on that topic please let me know. :) | non_process | thoughts about packaging this is sort of related to but probably warrants its own issue i m involved with rust packaging in two major linux distros and looked into packaging crev it seems there are both tagged releases and cargo uploads which is very good but i m not sure which packages should be distributed using which package name it seems cargo crev is used to tag the binary crates should this ship only cargo crev or also the other binary targets is required as a binary on a side note specifically for debian it s important to make sure you re depending on the latest version of your dependencies while we can upload outdated crates to debian it introduces quite a bit of complexity for us if you have some opinions on that topic please let me know | 0 |
17,595 | 23,420,851,273 | IssuesEvent | 2022-08-13 17:18:01 | apache/arrow-rs | https://api.github.com/repos/apache/arrow-rs | closed | Release `object_store` `0.4.0` | development-process | As discussed, we will release a new version of object_store from the arrow-rs repo / under ASF process: https://github.com/apache/arrow-rs/issues/2180
* Planned Release Candidate: 2022-08-08
* Planned Release and Publish to crates.io: 2022-08-11
Items:
- [x] Update changelog and readme: https://github.com/apache/arrow-rs/pull/2392
- [x] Create release scripts: https://github.com/apache/arrow-rs/pull/2392
- [x] Create release candidate https://lists.apache.org/thread/xg8m8s12q2d0glp3gocmqt18v2gzn1ml
- [x] Release candidate approved : https://lists.apache.org/thread/8j9wfj5g0nzmvg5lz7cnlym84m2gw4g8
- [x] Release to crates.io
- [x] Create release scripts
- [x] Draft PR to update DataFusion to new object store: https://github.com/apache/arrow-datafusion/pull/3089
| 1.0 | Release `object_store` `0.4.0` - As discussed, we will release a new version of object_store from the arrow-rs repo / under ASF process: https://github.com/apache/arrow-rs/issues/2180
* Planned Release Candidate: 2022-08-08
* Planned Release and Publish to crates.io: 2022-08-11
Items:
- [x] Update changelog and readme: https://github.com/apache/arrow-rs/pull/2392
- [x] Create release scripts: https://github.com/apache/arrow-rs/pull/2392
- [x] Create release candidate https://lists.apache.org/thread/xg8m8s12q2d0glp3gocmqt18v2gzn1ml
- [x] Release candidate approved : https://lists.apache.org/thread/8j9wfj5g0nzmvg5lz7cnlym84m2gw4g8
- [x] Release to crates.io
- [x] Create release scripts
- [x] Draft PR to update DataFusion to new object store: https://github.com/apache/arrow-datafusion/pull/3089
| process | release object store as discussed we will release a new version of object store from the arrow rs repo under asf process planned release candidate planned release and publish to crates io items update changelog and readme create release scripts create release candidate release candidate approved release to crates io create release scripts draft pr to update datafusion to new object store | 1 |
12,857 | 15,241,852,641 | IssuesEvent | 2021-02-19 09:03:42 | prisma/prisma | https://api.github.com/repos/prisma/prisma | opened | Error `[migration-engine/connectors/sql-migration-connector/src/flavour/sqlite.rs:99:42] failed to truncate sqlite file: Os { code: 13, kind: PermissionDenied, message: "Permission denied" }` | bug/1-repro-available engines/migration engine kind/bug process/candidate team/migrations topic: migrate dev | <!-- If required, please update the title to be clear and descriptive -->
Command: `migrate dev --name init --preview-feature`
Version: `2.16.1`
Report: https://prisma-errors.netlify.app/report/12973
OS: `x64 darwin 20.2.0`
JS Stacktrace:
```
Error: Error in migration engine.
Reason: [migration-engine/connectors/sql-migration-connector/src/flavour/sqlite.rs:99:42] failed to truncate sqlite file: Os { code: 13, kind: PermissionDenied, message: "Permission denied" }
```
Rust Stacktrace:
```
[migration-engine/connectors/sql-migration-connector/src/flavour/sqlite.rs:99:42] failed to truncate sqlite file: Os { code: 13, kind: PermissionDenied, message: "Permission denied" }
```
Migrate Request
```
{
"id": 2,
"jsonrpc": "2.0",
"method": "reset",
"params": {}
}
``` | 1.0 | Error `[migration-engine/connectors/sql-migration-connector/src/flavour/sqlite.rs:99:42] failed to truncate sqlite file: Os { code: 13, kind: PermissionDenied, message: "Permission denied" }` - <!-- If required, please update the title to be clear and descriptive -->
Command: `migrate dev --name init --preview-feature`
Version: `2.16.1`
Report: https://prisma-errors.netlify.app/report/12973
OS: `x64 darwin 20.2.0`
JS Stacktrace:
```
Error: Error in migration engine.
Reason: [migration-engine/connectors/sql-migration-connector/src/flavour/sqlite.rs:99:42] failed to truncate sqlite file: Os { code: 13, kind: PermissionDenied, message: "Permission denied" }
```
Rust Stacktrace:
```
[migration-engine/connectors/sql-migration-connector/src/flavour/sqlite.rs:99:42] failed to truncate sqlite file: Os { code: 13, kind: PermissionDenied, message: "Permission denied" }
```
Migrate Request
```
{
"id": 2,
"jsonrpc": "2.0",
"method": "reset",
"params": {}
}
``` | process | error failed to truncate sqlite file os code kind permissiondenied message permission denied command migrate dev name init preview feature version report os darwin js stacktrace error error in migration engine reason failed to truncate sqlite file os code kind permissiondenied message permission denied rust stacktrace failed to truncate sqlite file os code kind permissiondenied message permission denied migrate request id jsonrpc method reset params | 1 |
226 | 2,652,505,362 | IssuesEvent | 2015-03-16 17:41:18 | AnalyticalGraphicsInc/cesium | https://api.github.com/repos/AnalyticalGraphicsInc/cesium | closed | Evaluate Jasmine 2.0 | dev process | [Jasmine 2.0](http://jasmine.github.io/2.0/introduction.html) was just released. We should take a closer look to see how hard upgrading is given all our custom changes. If it is a ton of work, we can hold off until we decide on Intern, #735 | 1.0 | Evaluate Jasmine 2.0 - [Jasmine 2.0](http://jasmine.github.io/2.0/introduction.html) was just released. We should take a closer look to see how hard upgrading is given all our custom changes. If it is a ton of work, we can hold off until we decide on Intern, #735 | process | evaluate jasmine was just released we should take a closer look to see how hard upgrading is given all our custom changes if it is a ton of work we can hold off until we decide on intern | 1 |
5,699 | 8,563,383,571 | IssuesEvent | 2018-11-09 13:53:31 | easy-software-ufal/annotations_repos | https://api.github.com/repos/easy-software-ufal/annotations_repos | opened | sjh37/EntityFramework-Reverse-POCO-Code-First-Generator Reverse Navigation Many To One | C# RMA wrong processing | Issue: `https://github.com/sjh37/EntityFramework-Reverse-POCO-Code-First-Generator/issues/385`
PR: `https://github.com/sjh37/EntityFramework-Reverse-POCO-Code-First-Generator/commit/4b1f3b493ca3899e831acdb91965c1aa48a41e75` | 1.0 | sjh37/EntityFramework-Reverse-POCO-Code-First-Generator Reverse Navigation Many To One - Issue: `https://github.com/sjh37/EntityFramework-Reverse-POCO-Code-First-Generator/issues/385`
PR: `https://github.com/sjh37/EntityFramework-Reverse-POCO-Code-First-Generator/commit/4b1f3b493ca3899e831acdb91965c1aa48a41e75` | process | entityframework reverse poco code first generator reverse navigation many to one issue pr | 1 |
17,341 | 23,163,547,562 | IssuesEvent | 2022-07-29 20:47:40 | cagov/design-system | https://api.github.com/repos/cagov/design-system | closed | Process: Consolidate Component audit lists to a central location | Process improvement PM | Work with team to pull various component audit lists for the following:
- Current State Web Template
- Current Design System
- Pending Components (in progress)
- Needed Components (not yet available) | 1.0 | Process: Consolidate Component audit lists to a central location - Work with team to pull various component audit lists for the following:
- Current State Web Template
- Current Design System
- Pending Components (in progress)
- Needed Components (not yet available) | process | process consolidate component audit lists to a central location work with team to pull various component audit lists for the following current state web template current design system pending components in progress needed components not yet available | 1 |
55,543 | 8,000,383,572 | IssuesEvent | 2018-07-22 15:19:32 | ichabod801/t_games | https://api.github.com/repos/ichabod801/t_games | closed | More help text. | documentation games | Look for more ways I can use the help command to provide information about playing games. | 1.0 | More help text. - Look for more ways I can use the help command to provide information about playing games. | non_process | more help text look for more ways i can use the help command to provide information about playing games | 0 |
279,222 | 21,134,318,851 | IssuesEvent | 2022-04-06 04:09:37 | AY2122S2-CS2103T-W14-2/tp | https://api.github.com/repos/AY2122S2-CS2103T-W14-2/tp | closed | [PE-D] Inconsistency in command format vs example | Documentation Bug | The following shows the example command in the quick start section:

which does not match the command format given later in the UG:

<!--session: 1648793114954-b1f7044c-6fa2-4da0-853f-3bda01691bfb-->
<!--Version: Web v3.4.2-->
-------------
Labels: `severity.Low` `type.DocumentationBug`
original: cpwill01/ped#1 | 1.0 | [PE-D] Inconsistency in command format vs example - The following shows the example command in the quick start section:

which does not match the command format given later in the UG:

<!--session: 1648793114954-b1f7044c-6fa2-4da0-853f-3bda01691bfb-->
<!--Version: Web v3.4.2-->
-------------
Labels: `severity.Low` `type.DocumentationBug`
original: cpwill01/ped#1 | non_process | inconsistency in command format vs example the following shows the example command in the quick start section which does not match the command format given later in the ug labels severity low type documentationbug original ped | 0 |
175,456 | 21,313,548,578 | IssuesEvent | 2022-04-16 00:08:07 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | [Security Solution] Session view icon is overlapping and not working under event renderer on alerts page | bug triage_needed impact:high Team: SecuritySolution v8.2.0 | **Describe the bug**
Session view icon is overlapping and not working under event renderer on alerts page
**Build Details:**
```
Version : 8.2.0 SNAPSHOT
Build : 51940
Commit : cbab89d8a43311d11dfaac3d9145a2be8841a857
```
**preconditions**
1. Alerts should be exist with session view
**Steps to Reproduce**
1. Navigate to Alerts page
2. Click on Event rendered view on alert table
3. Observe that Session view icon is overlapping and not working under event renderer on alerts page
**Actual Result**
Session view icon is overlapping and not working under event renderer on alerts page
**Expected Result**
Session view icon should not be overlapped and should be working under event renderer on alerts page
**Screen-Shot**

| True | [Security Solution] Session view icon is overlapping and not working under event renderer on alerts page - **Describe the bug**
Session view icon is overlapping and not working under event renderer on alerts page
**Build Details:**
```
Version : 8.2.0 SNAPSHOT
Build : 51940
Commit : cbab89d8a43311d11dfaac3d9145a2be8841a857
```
**preconditions**
1. Alerts should be exist with session view
**Steps to Reproduce**
1. Navigate to Alerts page
2. Click on Event rendered view on alert table
3. Observe that Session view icon is overlapping and not working under event renderer on alerts page
**Actual Result**
Session view icon is overlapping and not working under event renderer on alerts page
**Expected Result**
Session view icon should not be overlapped and should be working under event renderer on alerts page
**Screen-Shot**

| non_process | session view icon is overlapping and not working under event renderer on alerts page describe the bug session view icon is overlapping and not working under event renderer on alerts page build details version snapshot build commit preconditions alerts should be exist with session view steps to reproduce navigate to alerts page click on event rendered view on alert table observe that session view icon is overlapping and not working under event renderer on alerts page actual result session view icon is overlapping and not working under event renderer on alerts page expected result session view icon should not be overlapped and should be working under event renderer on alerts page screen shot | 0 |
226,389 | 18,015,527,881 | IssuesEvent | 2021-09-16 13:33:14 | apache/shardingsphere | https://api.github.com/repos/apache/shardingsphere | closed | GovernanceRepositoryAPIImplTest.assertWatch may be blocked forever if exception thrown in watch listener | type: bug in: test | ## Bug Report
### Which version of ShardingSphere did you use?
5.0.0-RC1-SNAPSHOT
### Which project did you use? ShardingSphere-JDBC or ShardingSphere-Proxy?
ShardingSphere-Scaling
### Expected behavior
Unit test not blocked.
### Actual behavior
`assertThat(event.getType(), is(DataChangedEvent.Type.ADDED));` fail randomly.
Unit test blocked.
From GitHub CI log:
```
[Curator-SafeNotifyService-0] o.a.c.f.l.MappingListenerManager - Listener (org.apache.curator.framework.recipes.cache.CuratorCacheListenerBuilderImpl$2@24ebbdf) threw an exception
java.lang.AssertionError:
Expected: is <ADDED>
but: was <UPDATED>
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
at org.junit.Assert.assertThat(Assert.java:956)
at org.junit.Assert.assertThat(Assert.java:923)
at org.apache.shardingsphere.scaling.core.api.impl.GovernanceRepositoryAPIImplTest.lambda$assertWatch$0(GovernanceRepositoryAPIImplTest.java:91)
```
### Reason analyze (If you can)
In `GovernanceRepositoryAPIImplTest.java`
```
@Test
public void assertWatch() throws InterruptedException {
CountDownLatch countDownLatch = new CountDownLatch(1);
String key = ScalingConstant.SCALING_ROOT + "/1";
governanceRepositoryAPI.watch(ScalingConstant.SCALING_ROOT, event -> {
if (event.getKey().equals(key)) {
assertThat(event.getType(), is(DataChangedEvent.Type.ADDED));
countDownLatch.countDown();
}
});
governanceRepositoryAPI.persist(key, "");
countDownLatch.await();
}
```
1, If `assertThat` throw exception, `countDownLatch.await();` will always be blocked.
2, Repository key is the same in unit test methods, and value is persisted. See test log:
```
[INFO ] 2021-09-16 12:07:34.817 [main] o.a.s.s.c.a.i.GovernanceRepositoryAPIImplTest - assertWatch, key=/scaling/1 value=assertGetChildrenKeys
[INFO ] 2021-09-16 12:07:34.919 [Curator-SafeNotifyService-0] o.a.s.s.c.a.i.GovernanceRepositoryAPIImplTest - event=DataChangedEvent(key=/scaling/1, value=assertWatch, type=ADDED)
```
In `CuratorZookeeperRepository.java`
```
private Type getChangedType(final TreeCacheEvent.Type type) {
switch (type) {
case NODE_ADDED:
return Type.ADDED;
case NODE_UPDATED:
return Type.UPDATED;
case NODE_REMOVED:
return Type.DELETED;
default:
return Type.IGNORED;
}
}
```
TreeCacheEvent type depends on whether path is locally cached or not, it could be `ADDED` or `UPDATED`.
### Steps to reproduce the behavior, such as: SQL to execute, sharding rule configuration, when exception occur etc.
Throw exception in `assertWatch`'s `watch` listener.
### Example codes for reproduce this issue (such as a github link).
| 1.0 | GovernanceRepositoryAPIImplTest.assertWatch may be blocked forever if exception thrown in watch listener - ## Bug Report
### Which version of ShardingSphere did you use?
5.0.0-RC1-SNAPSHOT
### Which project did you use? ShardingSphere-JDBC or ShardingSphere-Proxy?
ShardingSphere-Scaling
### Expected behavior
Unit test not blocked.
### Actual behavior
`assertThat(event.getType(), is(DataChangedEvent.Type.ADDED));` fail randomly.
Unit test blocked.
From GitHub CI log:
```
[Curator-SafeNotifyService-0] o.a.c.f.l.MappingListenerManager - Listener (org.apache.curator.framework.recipes.cache.CuratorCacheListenerBuilderImpl$2@24ebbdf) threw an exception
java.lang.AssertionError:
Expected: is <ADDED>
but: was <UPDATED>
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
at org.junit.Assert.assertThat(Assert.java:956)
at org.junit.Assert.assertThat(Assert.java:923)
at org.apache.shardingsphere.scaling.core.api.impl.GovernanceRepositoryAPIImplTest.lambda$assertWatch$0(GovernanceRepositoryAPIImplTest.java:91)
```
### Reason analyze (If you can)
In `GovernanceRepositoryAPIImplTest.java`
```
@Test
public void assertWatch() throws InterruptedException {
CountDownLatch countDownLatch = new CountDownLatch(1);
String key = ScalingConstant.SCALING_ROOT + "/1";
governanceRepositoryAPI.watch(ScalingConstant.SCALING_ROOT, event -> {
if (event.getKey().equals(key)) {
assertThat(event.getType(), is(DataChangedEvent.Type.ADDED));
countDownLatch.countDown();
}
});
governanceRepositoryAPI.persist(key, "");
countDownLatch.await();
}
```
1, If `assertThat` throw exception, `countDownLatch.await();` will always be blocked.
2, Repository key is the same in unit test methods, and value is persisted. See test log:
```
[INFO ] 2021-09-16 12:07:34.817 [main] o.a.s.s.c.a.i.GovernanceRepositoryAPIImplTest - assertWatch, key=/scaling/1 value=assertGetChildrenKeys
[INFO ] 2021-09-16 12:07:34.919 [Curator-SafeNotifyService-0] o.a.s.s.c.a.i.GovernanceRepositoryAPIImplTest - event=DataChangedEvent(key=/scaling/1, value=assertWatch, type=ADDED)
```
In `CuratorZookeeperRepository.java`
```
private Type getChangedType(final TreeCacheEvent.Type type) {
switch (type) {
case NODE_ADDED:
return Type.ADDED;
case NODE_UPDATED:
return Type.UPDATED;
case NODE_REMOVED:
return Type.DELETED;
default:
return Type.IGNORED;
}
}
```
TreeCacheEvent type depends on whether path is locally cached or not, it could be `ADDED` or `UPDATED`.
### Steps to reproduce the behavior, such as: SQL to execute, sharding rule configuration, when exception occur etc.
Throw exception in `assertWatch`'s `watch` listener.
### Example codes for reproduce this issue (such as a github link).
| non_process | governancerepositoryapiimpltest assertwatch may be blocked forever if exception thrown in watch listener bug report which version of shardingsphere did you use snapshot which project did you use shardingsphere jdbc or shardingsphere proxy shardingsphere scaling expected behavior unit test not blocked actual behavior assertthat event gettype is datachangedevent type added fail randomly unit test blocked from github ci log o a c f l mappinglistenermanager listener org apache curator framework recipes cache curatorcachelistenerbuilderimpl threw an exception java lang assertionerror expected is but was at org hamcrest matcherassert assertthat matcherassert java at org junit assert assertthat assert java at org junit assert assertthat assert java at org apache shardingsphere scaling core api impl governancerepositoryapiimpltest lambda assertwatch governancerepositoryapiimpltest java reason analyze if you can in governancerepositoryapiimpltest java test public void assertwatch throws interruptedexception countdownlatch countdownlatch new countdownlatch string key scalingconstant scaling root governancerepositoryapi watch scalingconstant scaling root event if event getkey equals key assertthat event gettype is datachangedevent type added countdownlatch countdown governancerepositoryapi persist key countdownlatch await if assertthat throw exception countdownlatch await will always be blocked repository key is the same in unit test methods and value is persisted see test log o a s s c a i governancerepositoryapiimpltest assertwatch key scaling value assertgetchildrenkeys o a s s c a i governancerepositoryapiimpltest event datachangedevent key scaling value assertwatch type added in curatorzookeeperrepository java private type getchangedtype final treecacheevent type type switch type case node added return type added case node updated return type updated case node removed return type deleted default return type ignored treecacheevent type depends on whether path is locally cached or not it could be added or updated steps to reproduce the behavior such as sql to execute sharding rule configuration when exception occur etc throw exception in assertwatch s watch listener example codes for reproduce this issue such as a github link | 0 |
688,960 | 23,601,855,164 | IssuesEvent | 2022-08-24 03:20:05 | ucb-rit/coldfront | https://api.github.com/repos/ucb-rit/coldfront | opened | Stand up REST API endpoints for serving Billing Activities | enhancement rest api high priority lrc-only | Currently, the primary key of the billing ID being used for Recharge for a particular `AllocationUser` may be fetched from `/api/allocation_users/{allocation_user_id}/attributes/?type=Billing Activity` (the `value` field).
Add an REST API endpoint (e.g., `/api/billing_activities/{billing_activity_id}/` for retrieving a particular `BillingActivity`, which includes a field for the fully-formed billing ID (i.e., `full_id`). | 1.0 | Stand up REST API endpoints for serving Billing Activities - Currently, the primary key of the billing ID being used for Recharge for a particular `AllocationUser` may be fetched from `/api/allocation_users/{allocation_user_id}/attributes/?type=Billing Activity` (the `value` field).
Add an REST API endpoint (e.g., `/api/billing_activities/{billing_activity_id}/` for retrieving a particular `BillingActivity`, which includes a field for the fully-formed billing ID (i.e., `full_id`). | non_process | stand up rest api endpoints for serving billing activities currently the primary key of the billing id being used for recharge for a particular allocationuser may be fetched from api allocation users allocation user id attributes type billing activity the value field add an rest api endpoint e g api billing activities billing activity id for retrieving a particular billingactivity which includes a field for the fully formed billing id i e full id | 0 |
3,154 | 6,205,297,603 | IssuesEvent | 2017-07-06 15:51:29 | pelias/model | https://api.github.com/repos/pelias/model | closed | Add ability to retrieve admin properties of a document | processed | There is currently no way to get admin properties of a document. One can only addParent. There are places where we've had to resort to checking the parent property of the document directly, which is less than ideal and should be changed as soon as this new getParent function is added. (see openstreetmap importer for example)
| 1.0 | Add ability to retrieve admin properties of a document - There is currently no way to get admin properties of a document. One can only addParent. There are places where we've had to resort to checking the parent property of the document directly, which is less than ideal and should be changed as soon as this new getParent function is added. (see openstreetmap importer for example)
| process | add ability to retrieve admin properties of a document there is currently no way to get admin properties of a document one can only addparent there are places where we ve had to resort to checking the parent property of the document directly which is less than ideal and should be changed as soon as this new getparent function is added see openstreetmap importer for example | 1 |
308,000 | 9,428,529,307 | IssuesEvent | 2019-04-12 01:37:16 | stuininga/Capstone_FoodBankApp | https://api.github.com/repos/stuininga/Capstone_FoodBankApp | closed | Create home page for Databases | top priority | Welcome user so and so, which database you're in, adds database navigation, | 1.0 | Create home page for Databases - Welcome user so and so, which database you're in, adds database navigation, | non_process | create home page for databases welcome user so and so which database you re in adds database navigation | 0 |
115,876 | 9,815,869,227 | IssuesEvent | 2019-06-13 13:34:29 | golang/go | https://api.github.com/repos/golang/go | reopened | Failing test for "net/http/cgi" on Ubuntu Xenial (16.04) | NeedsInvestigation Testing WaitingForInfo | Please answer these questions before submitting your issue. Thanks!
### What version of Go are you using (`go version`)?
`go version go1.11 linux/amd64`
### Does this issue reproduce with the latest release?
yes
### What operating system and processor architecture are you using (`go env`)?
```
GOARCH="amd64"
GOBIN=""
GOCACHE="/home/gvarouchas/.cache/go-build"
GOEXE=""
GOHOSTARCH="amd64"
GOHOSTOS="linux"
GOOS="linux"
GOPATH="/home/gvarouchas/golib"
GORACE=""
GOROOT="/usr/lib/go-1.10"
GOTMPDIR=""
GOTOOLDIR="/usr/lib/go-1.10/pkg/tool/linux_amd64"
GCCGO="gccgo"
CC="gcc"
CXX="g++"
CGO_ENABLED="1"
CGO_CFLAGS="-g -O2"
CGO_CPPFLAGS=""
CGO_CXXFLAGS="-g -O2"
CGO_FFLAGS="-g -O2"
CGO_LDFLAGS="-g -O2"
PKG_CONFIG="pkg-config"
GOGCCFLAGS="-fPIC -m64 -pthread -fmessage-length=0 -fdebug-prefix-map=/tmp/go-build117805775=/tmp/go-build -gno-record-gcc-switches"
```
Operating system is Ubuntu xenial 16.04.5
### What did you do?
From the go source directory, run the tests for "net/http/cgi" package
### What did you expect to see?
$ ../bin/go test net/http/cgi
ok net/http/cgi
### What did you see instead?
```
# "../bin/go" is the built go command :
$ ../bin/go version
go version go1.11 linux/amd64
$ ../bin/go test net/http/cgi
2018/09/21 10:20:51 cgi: copy error: write tcp 127.0.0.1:33176->127.0.0.1:47778: write: connection reset by peer
/usr/bin/env: 'bash': No such file or directory
2018/09/21 10:20:51 cgi: no headers
--- FAIL: TestEnvOverride (0.00s)
host_test.go:84: for key "env-REQUEST_URI" got ""; expected "/foo/bar"
host_test.go:84: for key "env-PATH" got ""; expected "/wibble"
host_test.go:84: for key "cwd" got ""; expected "/home/gvarouchas/go/src/net/http/cgi"
host_test.go:84: for key "env-SCRIPT_FILENAME" got ""; expected "/home/gvarouchas/go/src/net/http/cgi/testdata/test.cgi"
2018/09/21 10:20:51 cgi: copy error: past write limit
2018/09/21 10:20:51 cgi: no headers
2018/09/21 10:20:51 cgi: missing required Content-Type in headers
2018/09/21 10:20:51 cgi: no headers
FAIL
FAIL net/http/cgi 0.206s
```
I guess this issue is linked to the setup of my local machine (especially the "connection reset by peer" on a local connection), but I don't know where to start looking ? | 1.0 | Failing test for "net/http/cgi" on Ubuntu Xenial (16.04) - Please answer these questions before submitting your issue. Thanks!
### What version of Go are you using (`go version`)?
`go version go1.11 linux/amd64`
### Does this issue reproduce with the latest release?
yes
### What operating system and processor architecture are you using (`go env`)?
```
GOARCH="amd64"
GOBIN=""
GOCACHE="/home/gvarouchas/.cache/go-build"
GOEXE=""
GOHOSTARCH="amd64"
GOHOSTOS="linux"
GOOS="linux"
GOPATH="/home/gvarouchas/golib"
GORACE=""
GOROOT="/usr/lib/go-1.10"
GOTMPDIR=""
GOTOOLDIR="/usr/lib/go-1.10/pkg/tool/linux_amd64"
GCCGO="gccgo"
CC="gcc"
CXX="g++"
CGO_ENABLED="1"
CGO_CFLAGS="-g -O2"
CGO_CPPFLAGS=""
CGO_CXXFLAGS="-g -O2"
CGO_FFLAGS="-g -O2"
CGO_LDFLAGS="-g -O2"
PKG_CONFIG="pkg-config"
GOGCCFLAGS="-fPIC -m64 -pthread -fmessage-length=0 -fdebug-prefix-map=/tmp/go-build117805775=/tmp/go-build -gno-record-gcc-switches"
```
Operating system is Ubuntu xenial 16.04.5
### What did you do?
From the go source directory, run the tests for "net/http/cgi" package
### What did you expect to see?
$ ../bin/go test net/http/cgi
ok net/http/cgi
### What did you see instead?
```
# "../bin/go" is the built go command :
$ ../bin/go version
go version go1.11 linux/amd64
$ ../bin/go test net/http/cgi
2018/09/21 10:20:51 cgi: copy error: write tcp 127.0.0.1:33176->127.0.0.1:47778: write: connection reset by peer
/usr/bin/env: 'bash': No such file or directory
2018/09/21 10:20:51 cgi: no headers
--- FAIL: TestEnvOverride (0.00s)
host_test.go:84: for key "env-REQUEST_URI" got ""; expected "/foo/bar"
host_test.go:84: for key "env-PATH" got ""; expected "/wibble"
host_test.go:84: for key "cwd" got ""; expected "/home/gvarouchas/go/src/net/http/cgi"
host_test.go:84: for key "env-SCRIPT_FILENAME" got ""; expected "/home/gvarouchas/go/src/net/http/cgi/testdata/test.cgi"
2018/09/21 10:20:51 cgi: copy error: past write limit
2018/09/21 10:20:51 cgi: no headers
2018/09/21 10:20:51 cgi: missing required Content-Type in headers
2018/09/21 10:20:51 cgi: no headers
FAIL
FAIL net/http/cgi 0.206s
```
I guess this issue is linked to the setup of my local machine (especially the "connection reset by peer" on a local connection), but I don't know where to start looking ? | non_process | failing test for net http cgi on ubuntu xenial please answer these questions before submitting your issue thanks what version of go are you using go version go version linux does this issue reproduce with the latest release yes what operating system and processor architecture are you using go env goarch gobin gocache home gvarouchas cache go build goexe gohostarch gohostos linux goos linux gopath home gvarouchas golib gorace goroot usr lib go gotmpdir gotooldir usr lib go pkg tool linux gccgo gccgo cc gcc cxx g cgo enabled cgo cflags g cgo cppflags cgo cxxflags g cgo fflags g cgo ldflags g pkg config pkg config gogccflags fpic pthread fmessage length fdebug prefix map tmp go tmp go build gno record gcc switches operating system is ubuntu xenial what did you do from the go source directory run the tests for net http cgi package what did you expect to see bin go test net http cgi ok net http cgi what did you see instead bin go is the built go command bin go version go version linux bin go test net http cgi cgi copy error write tcp write connection reset by peer usr bin env bash no such file or directory cgi no headers fail testenvoverride host test go for key env request uri got expected foo bar host test go for key env path got expected wibble host test go for key cwd got expected home gvarouchas go src net http cgi host test go for key env script filename got expected home gvarouchas go src net http cgi testdata test cgi cgi copy error past write limit cgi no headers cgi missing required content type in headers cgi no headers fail fail net http cgi i guess this issue is linked to the setup of my local machine especially the connection reset by peer on a local connection but i don t know where to start looking | 0 |
299,900 | 25,936,248,157 | IssuesEvent | 2022-12-16 14:24:48 | DucTrann1310/FeedbackOnline | https://api.github.com/repos/DucTrann1310/FeedbackOnline | opened | [BugID_35]_FUNC_Quản lý topic_Tạo mới topic_Tạo mới thành công khi để trống trường Tên Topic | bug Open Integration Test Fun_Wrong Business logic Priority_Medium Severity_Medium | Precondition: Admin đang ở màn hình tạo mới Topic
Step:
1. Để trống [Tên Topic] textbox
2. Click [Hoàn Tất] button
Actual output: Topic vừa được tạo thành công, lưu trong DB table Topic và refresh grid
Expected output: Hiển thị inline message màu đỏ "Bạn chưa điền Tên Topic!" phía dưới trường Tên Topic và set focus tại đó
---------------
TestcaseID = 11
https://www.screencast.com/t/mnT50sTDhT | 1.0 | [BugID_35]_FUNC_Quản lý topic_Tạo mới topic_Tạo mới thành công khi để trống trường Tên Topic - Precondition: Admin đang ở màn hình tạo mới Topic
Step:
1. Để trống [Tên Topic] textbox
2. Click [Hoàn Tất] button
Actual output: Topic vừa được tạo thành công, lưu trong DB table Topic và refresh grid
Expected output: Hiển thị inline message màu đỏ "Bạn chưa điền Tên Topic!" phía dưới trường Tên Topic và set focus tại đó
---------------
TestcaseID = 11
https://www.screencast.com/t/mnT50sTDhT | non_process | func quản lý topic tạo mới topic tạo mới thành công khi để trống trường tên topic precondition admin đang ở màn hình tạo mới topic step để trống textbox click button actual output topic vừa được tạo thành công lưu trong db table topic và refresh grid expected output hiển thị inline message màu đỏ bạn chưa điền tên topic phía dưới trường tên topic và set focus tại đó testcaseid | 0 |
5,221 | 3,536,002,408 | IssuesEvent | 2016-01-16 23:14:25 | stedolan/jq | https://api.github.com/repos/stedolan/jq | closed | Cross-compile via Travis-CI too | build | It seems it's possible to cross-compile in Travis-CI builds.
A few links:
- http://stackoverflow.com/questions/28155308/how-to-cross-compile-with-mingw-on-linux-for-windows
This one has a great answer. We need grab the Mingw pkgs [from a third-party repo](https://docs.travis-ci.com/user/installing-dependencies/#installing-packages-from-a-custom-apt-repository) (which I guess we can trust as far as CI goes), then use them in the Travis build. Ideally Travis would just provide these...
- https://trac.osgeo.org/geos/ticket/659
- http://www.scriptscoop.net/t/d39c9eac38a2/how-to-cross-compile-with-mingw-on-linux-for-windows.html
- http://sourceforge.net/p/mingw-w64/mailman/message/31846195/ | 1.0 | Cross-compile via Travis-CI too - It seems it's possible to cross-compile in Travis-CI builds.
A few links:
- http://stackoverflow.com/questions/28155308/how-to-cross-compile-with-mingw-on-linux-for-windows
This one has a great answer. We need grab the Mingw pkgs [from a third-party repo](https://docs.travis-ci.com/user/installing-dependencies/#installing-packages-from-a-custom-apt-repository) (which I guess we can trust as far as CI goes), then use them in the Travis build. Ideally Travis would just provide these...
- https://trac.osgeo.org/geos/ticket/659
- http://www.scriptscoop.net/t/d39c9eac38a2/how-to-cross-compile-with-mingw-on-linux-for-windows.html
- http://sourceforge.net/p/mingw-w64/mailman/message/31846195/ | non_process | cross compile via travis ci too it seems it s possible to cross compile in travis ci builds a few links this one has a great answer we need grab the mingw pkgs which i guess we can trust as far as ci goes then use them in the travis build ideally travis would just provide these | 0 |
33,768 | 9,205,176,046 | IssuesEvent | 2019-03-08 09:50:22 | qissue-bot/QGIS | https://api.github.com/repos/qissue-bot/QGIS | closed | upgrade to titan but it won't start as I get a segmentation fault | Category: Build/Install Component: Affected QGIS version Component: Crashes QGIS or corrupts data Component: Easy fix? Component: Operating System Component: Pull Request or Patch supplied Component: Regression? Component: Resolution Priority: Low Project: QGIS Application Status: Closed Tracker: Bug report | ---
Author Name: **bshcm03-yahoo-co-uk -** (bshcm03-yahoo-co-uk -)
Original Redmine Issue: 573, https://issues.qgis.org/issues/573
Original Assignee: Magnus Homann
---
I have upgraded to titan but QGIS will no longer start - I get a segmentation fault on start up. I was able to use the previous version (with intermittent errors).
I have created (and attached) log files for configure, make and make install. I was getting a qt4 error but this was resolved after I chagned the qt4 installation directory from default to /usr/local/qt4
I ideally run qgis from within grass (grass 6.2.1) but cannot do this either.
I was getting some make install warnings,but not errors.
Do you have any thoughts?
If you need any further information etc, please do not hesitate to ask.
Thanks
---
- [config.log](https://issues.qgis.org/attachments/download/1862/config.log) (bshcm03-yahoo-co-uk -)
- [make.log](https://issues.qgis.org/attachments/download/1864/make.log) (anonymous -)
- [make.install.log](https://issues.qgis.org/attachments/download/1863/make.install.log) (bshcm03-yahoo-co-uk -)
- [config.2.log](https://issues.qgis.org/attachments/download/1859/config.2.log) (Magnus Homann)
- [config.3.log](https://issues.qgis.org/attachments/download/1860/config.3.log) (bshcm03 -)
- [bh.make.check.log](https://issues.qgis.org/attachments/download/1858/bh.make.check.log) (bshcm03 -)
- [config.4.log](https://issues.qgis.org/attachments/download/1861/config.4.log) (bshcm03 -)
- [qgis_error_8-2-07](https://issues.qgis.org/attachments/download/1865/qgis_error_8-2-07) (bshcm03 -) | 1.0 | upgrade to titan but it won't start as I get a segmentation fault - ---
Author Name: **bshcm03-yahoo-co-uk -** (bshcm03-yahoo-co-uk -)
Original Redmine Issue: 573, https://issues.qgis.org/issues/573
Original Assignee: Magnus Homann
---
I have upgraded to titan but QGIS will no longer start - I get a segmentation fault on start up. I was able to use the previous version (with intermittent errors).
I have created (and attached) log files for configure, make and make install. I was getting a qt4 error but this was resolved after I chagned the qt4 installation directory from default to /usr/local/qt4
I ideally run qgis from within grass (grass 6.2.1) but cannot do this either.
I was getting some make install warnings,but not errors.
Do you have any thoughts?
If you need any further information etc, please do not hesitate to ask.
Thanks
---
- [config.log](https://issues.qgis.org/attachments/download/1862/config.log) (bshcm03-yahoo-co-uk -)
- [make.log](https://issues.qgis.org/attachments/download/1864/make.log) (anonymous -)
- [make.install.log](https://issues.qgis.org/attachments/download/1863/make.install.log) (bshcm03-yahoo-co-uk -)
- [config.2.log](https://issues.qgis.org/attachments/download/1859/config.2.log) (Magnus Homann)
- [config.3.log](https://issues.qgis.org/attachments/download/1860/config.3.log) (bshcm03 -)
- [bh.make.check.log](https://issues.qgis.org/attachments/download/1858/bh.make.check.log) (bshcm03 -)
- [config.4.log](https://issues.qgis.org/attachments/download/1861/config.4.log) (bshcm03 -)
- [qgis_error_8-2-07](https://issues.qgis.org/attachments/download/1865/qgis_error_8-2-07) (bshcm03 -) | non_process | upgrade to titan but it won t start as i get a segmentation fault author name yahoo co uk yahoo co uk original redmine issue original assignee magnus homann i have upgraded to titan but qgis will no longer start i get a segmentation fault on start up i was able to use the previous version with intermittent errors i have created and attached log files for configure make and make install i was getting a error but this was resolved after i chagned the installation directory from default to usr local i ideally run qgis from within grass grass but cannot do this either i was getting some make install warnings but not errors do you have any thoughts if you need any further information etc please do not hesitate to ask thanks yahoo co uk anonymous yahoo co uk magnus homann | 0 |
49,416 | 6,025,491,915 | IssuesEvent | 2017-06-08 08:50:40 | healthlocker/healthlocker | https://api.github.com/repos/healthlocker/healthlocker | closed | Bug - Internal Error on Tracker Overview | bug please-test priority-2 | New account created (not connected to SLaM)
Added one entry to sleep tracker
Added one entry to problem tracker
Then clicked on tracker overview and got internal error page
Using safari on iPhone.
| 1.0 | Bug - Internal Error on Tracker Overview - New account created (not connected to SLaM)
Added one entry to sleep tracker
Added one entry to problem tracker
Then clicked on tracker overview and got internal error page
Using safari on iPhone.
| non_process | bug internal error on tracker overview new account created not connected to slam added one entry to sleep tracker added one entry to problem tracker then clicked on tracker overview and got internal error page using safari on iphone | 0 |
7,002 | 10,145,737,075 | IssuesEvent | 2019-08-05 05:53:12 | GroceriStar/food-static-files-generator | https://api.github.com/repos/GroceriStar/food-static-files-generator | closed | move readData into filesystem | enhancement good first issue help wanted in-process | **Is your feature request related to a problem? Please describe.**
we need to do it after #125 will be merged to master | 1.0 | move readData into filesystem - **Is your feature request related to a problem? Please describe.**
we need to do it after #125 will be merged to master | process | move readdata into filesystem is your feature request related to a problem please describe we need to do it after will be merged to master | 1 |
224,563 | 17,755,804,082 | IssuesEvent | 2021-08-28 18:33:00 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | ccl/benchccl/rttanalysisccl: BenchmarkAlterPrimaryRegion failed | C-test-failure O-robot branch-master | ccl/benchccl/rttanalysisccl.BenchmarkAlterPrimaryRegion [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3376361&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3376361&tab=artifacts#/) on master @ [93d95b97e53b11b9640a698535c563ca3ed5ba15](https://github.com/cockroachdb/cockroach/commits/93d95b97e53b11b9640a698535c563ca3ed5ba15):
```
BenchmarkAlterPrimaryRegion/alter_populated_database_alter_primary_region
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region651687629
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
BenchmarkAlterTableLocality/alter_from_global_to_rbr-16 1 3283756600 ns/op 22.00 roundtrips
BenchmarkAlterTableLocality/alter_from_regional_by_table_to_rbr
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr712218057
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterRegions_alter_populated_database_drop_region564512728
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
BenchmarkAlterSurvivalGoals/alter_populated_database_from_zone_to_region-16 1 359332916 ns/op 41.00 roundtrips
BenchmarkAlterSurvivalGoals/alter_populated_database_from_region_to_zone
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone141752125
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region085553020
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr356440840
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone849513004
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr215803079
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
rtt_analysis_bench.go:131: -- test log scope end --
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region053354155
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:54: -- test log scope end --
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone938059163
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr701030222
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region037082402
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone648673362
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
BenchmarkAlterRegions/alter_populated_database_drop_region-16 1 15275801616 ns/op 21.00 roundtrips
BenchmarkAlterRegions/alter_populated_database_add_region
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterRegions_alter_populated_database_add_region041894743
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
BenchmarkAlterPrimaryRegion/alter_populated_database_alter_primary_region-16 1 325849061 ns/op 21.00 roundtrips
```
<details><summary>Reproduce</summary>
<p>
To reproduce, try:
```bash
make stressrace TESTS=BenchmarkAlterPrimaryRegion PKG=./pkg/ccl/benchccl/rttanalysisccl TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
Parameters in this failure:
- GOFLAGS=-json
</p>
</details>
/cc @cockroachdb/sql-experience pawalt
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*BenchmarkAlterPrimaryRegion.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| 1.0 | ccl/benchccl/rttanalysisccl: BenchmarkAlterPrimaryRegion failed - ccl/benchccl/rttanalysisccl.BenchmarkAlterPrimaryRegion [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3376361&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3376361&tab=artifacts#/) on master @ [93d95b97e53b11b9640a698535c563ca3ed5ba15](https://github.com/cockroachdb/cockroach/commits/93d95b97e53b11b9640a698535c563ca3ed5ba15):
```
BenchmarkAlterPrimaryRegion/alter_populated_database_alter_primary_region
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region651687629
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
BenchmarkAlterTableLocality/alter_from_global_to_rbr-16 1 3283756600 ns/op 22.00 roundtrips
BenchmarkAlterTableLocality/alter_from_regional_by_table_to_rbr
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr712218057
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterRegions_alter_populated_database_drop_region564512728
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
BenchmarkAlterSurvivalGoals/alter_populated_database_from_zone_to_region-16 1 359332916 ns/op 41.00 roundtrips
BenchmarkAlterSurvivalGoals/alter_populated_database_from_region_to_zone
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone141752125
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region085553020
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr356440840
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone849513004
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr215803079
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
rtt_analysis_bench.go:131: -- test log scope end --
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region053354155
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:54: -- test log scope end --
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone938059163
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr701030222
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region037082402
test_log_scope.go:74: use -show-logs to present logs inline
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone648673362
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
BenchmarkAlterRegions/alter_populated_database_drop_region-16 1 15275801616 ns/op 21.00 roundtrips
BenchmarkAlterRegions/alter_populated_database_add_region
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterRegions_alter_populated_database_add_region041894743
test_log_scope.go:74: use -show-logs to present logs inline
rtt_analysis_bench.go:131: -- test log scope end --
rtt_analysis_bench.go:54: -- test log scope end --
BenchmarkAlterPrimaryRegion/alter_populated_database_alter_primary_region-16 1 325849061 ns/op 21.00 roundtrips
```
<details><summary>Reproduce</summary>
<p>
To reproduce, try:
```bash
make stressrace TESTS=BenchmarkAlterPrimaryRegion PKG=./pkg/ccl/benchccl/rttanalysisccl TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
Parameters in this failure:
- GOFLAGS=-json
</p>
</details>
/cc @cockroachdb/sql-experience pawalt
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*BenchmarkAlterPrimaryRegion.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| non_process | ccl benchccl rttanalysisccl benchmarkalterprimaryregion failed ccl benchccl rttanalysisccl benchmarkalterprimaryregion with on master benchmarkalterprimaryregion alter populated database alter primary region test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkalterprimaryregion alter populated database alter primary test log scope go use show logs to present logs inline rtt analysis bench go test log scope end rtt analysis bench go test log scope end benchmarkaltertablelocality alter from global to rbr ns op roundtrips benchmarkaltertablelocality alter from regional by table to rbr test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltertablelocality alter from regional by table to test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkalterregions alter populated database drop test log scope go use show logs to present logs inline rtt analysis bench go test log scope end rtt analysis bench go test log scope end benchmarkaltersurvivalgoals alter populated database from zone to region ns op roundtrips benchmarkaltersurvivalgoals alter populated database from region to zone test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltersurvivalgoals alter populated database from region to test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkalterprimaryregion alter populated database alter primary test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltertablelocality alter from regional by table to test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltersurvivalgoals alter populated database from region to test log scope go use show logs to present logs inline rtt analysis bench go test log scope end rtt analysis bench go test log scope end test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltertablelocality alter from regional by table to test log scope go use show logs to present logs inline rtt analysis bench go test log scope end rtt analysis bench go test log scope end rtt analysis bench go test log scope end test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkalterprimaryregion alter populated database alter primary test log scope go use show logs to present logs inline rtt analysis bench go test log scope end test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltersurvivalgoals alter populated database from region to test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltertablelocality alter from regional by table to test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkalterprimaryregion alter populated database alter primary test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltersurvivalgoals alter populated database from region to test log scope go use show logs to present logs inline rtt analysis bench go test log scope end rtt analysis bench go test log scope end benchmarkalterregions alter populated database drop region ns op roundtrips benchmarkalterregions alter populated database add region test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkalterregions alter populated database add test log scope go use show logs to present logs inline rtt analysis bench go test log scope end rtt analysis bench go test log scope end benchmarkalterprimaryregion alter populated database alter primary region ns op roundtrips reproduce to reproduce try bash make stressrace tests benchmarkalterprimaryregion pkg pkg ccl benchccl rttanalysisccl testtimeout stressflags timeout parameters in this failure goflags json cc cockroachdb sql experience pawalt | 0 |
77,742 | 14,915,092,671 | IssuesEvent | 2021-01-22 16:15:50 | ArctosDB/arctos | https://api.github.com/repos/ArctosDB/arctos | closed | csv download of validate results for Bulk Edit Container | Function-ContainerOrBarcode Priority-Normal | Issue Documentation is http://handbook.arctosdb.org/how_to/How-to-Use-Issues-in-Arctos.html
**Is your feature request related to a problem? Please describe.**
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]
I sometimes have to bulk edit a large container series (thousands of containers) & if validation fails I have had trouble finding the offending barcode in my web browser. can we get a csv download option of those validation results so we can more easily find the offending barcodes?
**Describe the solution you'd like**
A clear and concise description of what you want to happen.
csv download validation results after using the bulk edit container tool in object tracking
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
Fail to find the problem but can't proceed, email dusty...
**Additional context**
Add any other context or screenshots about the feature request here.
**Priority**
Please assign a priority-label.
| 1.0 | csv download of validate results for Bulk Edit Container - Issue Documentation is http://handbook.arctosdb.org/how_to/How-to-Use-Issues-in-Arctos.html
**Is your feature request related to a problem? Please describe.**
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]
I sometimes have to bulk edit a large container series (thousands of containers) & if validation fails I have had trouble finding the offending barcode in my web browser. can we get a csv download option of those validation results so we can more easily find the offending barcodes?
**Describe the solution you'd like**
A clear and concise description of what you want to happen.
csv download validation results after using the bulk edit container tool in object tracking
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
Fail to find the problem but can't proceed, email dusty...
**Additional context**
Add any other context or screenshots about the feature request here.
**Priority**
Please assign a priority-label.
| non_process | csv download of validate results for bulk edit container issue documentation is is your feature request related to a problem please describe a clear and concise description of what the problem is ex i m always frustrated when i sometimes have to bulk edit a large container series thousands of containers if validation fails i have had trouble finding the offending barcode in my web browser can we get a csv download option of those validation results so we can more easily find the offending barcodes describe the solution you d like a clear and concise description of what you want to happen csv download validation results after using the bulk edit container tool in object tracking describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered fail to find the problem but can t proceed email dusty additional context add any other context or screenshots about the feature request here priority please assign a priority label | 0 |
3,462 | 6,545,260,571 | IssuesEvent | 2017-09-04 03:07:14 | w3c/payment-request | https://api.github.com/repos/w3c/payment-request | closed | CR check list | Process aid | For all Transition Requests, to advance a specification to a new maturity level other than Note, the Working Group:
* [x] must [record the group's decision to request advancement](https://lists.w3.org/Archives/Public/public-payments-wg/2017Jul/0057.html).
* [ ] must obtain Director approval.
* [x] must provide [public documentation of all substantive changes to the technical report since the previous publication](https://github.com/w3c/browser-payment-api/pulls?q=is%3Apr+is%3Aclosed+label%3Acr-substantive).
* [ ] must [formally address all issues raised about the document since the previous maturity level](https://github.com/w3c/browser-payment-api/milestone/8).
* [x] must provide public documentation of any Formal Objections (none received).
* [x] should provide public documentation of [changes that are not substantive](https://github.com/w3c/browser-payment-api/commits/gh-pages).
* [x] should report which, if any, of the Working Group's requirements for this document have changed since the previous step.
* [x] should report any changes in dependencies with other groups.
* [x] should provide [information about implementations known to the Working Group](http://caniuse.com/#feat=payment-request).
To publish a Candidate recommendation, in addition to meeting the general requirements for advancement a Working Group:
* [x] must show that the specification has met all Working Group requirements, or explain why the requirements have changed or been deferred,
* [x] must document changes to dependencies during the development of the specification,
* [x] must document how adequate implementation experience will be demonstrated,
* [x] must specify the deadline for comments, which must be at least four weeks after publication, and should be longer for complex documents,
* [x] must show that the specification has received wide review, and
* [x] may identify features in the document as "at risk". These features may be removed before advancement to Proposed Recommendation without a requirement to publish a new Candidate Recommendation.
| 1.0 | CR check list - For all Transition Requests, to advance a specification to a new maturity level other than Note, the Working Group:
* [x] must [record the group's decision to request advancement](https://lists.w3.org/Archives/Public/public-payments-wg/2017Jul/0057.html).
* [ ] must obtain Director approval.
* [x] must provide [public documentation of all substantive changes to the technical report since the previous publication](https://github.com/w3c/browser-payment-api/pulls?q=is%3Apr+is%3Aclosed+label%3Acr-substantive).
* [ ] must [formally address all issues raised about the document since the previous maturity level](https://github.com/w3c/browser-payment-api/milestone/8).
* [x] must provide public documentation of any Formal Objections (none received).
* [x] should provide public documentation of [changes that are not substantive](https://github.com/w3c/browser-payment-api/commits/gh-pages).
* [x] should report which, if any, of the Working Group's requirements for this document have changed since the previous step.
* [x] should report any changes in dependencies with other groups.
* [x] should provide [information about implementations known to the Working Group](http://caniuse.com/#feat=payment-request).
To publish a Candidate recommendation, in addition to meeting the general requirements for advancement a Working Group:
* [x] must show that the specification has met all Working Group requirements, or explain why the requirements have changed or been deferred,
* [x] must document changes to dependencies during the development of the specification,
* [x] must document how adequate implementation experience will be demonstrated,
* [x] must specify the deadline for comments, which must be at least four weeks after publication, and should be longer for complex documents,
* [x] must show that the specification has received wide review, and
* [x] may identify features in the document as "at risk". These features may be removed before advancement to Proposed Recommendation without a requirement to publish a new Candidate Recommendation.
| process | cr check list for all transition requests to advance a specification to a new maturity level other than note the working group must must obtain director approval must provide must must provide public documentation of any formal objections none received should provide public documentation of should report which if any of the working group s requirements for this document have changed since the previous step should report any changes in dependencies with other groups should provide to publish a candidate recommendation in addition to meeting the general requirements for advancement a working group must show that the specification has met all working group requirements or explain why the requirements have changed or been deferred must document changes to dependencies during the development of the specification must document how adequate implementation experience will be demonstrated must specify the deadline for comments which must be at least four weeks after publication and should be longer for complex documents must show that the specification has received wide review and may identify features in the document as at risk these features may be removed before advancement to proposed recommendation without a requirement to publish a new candidate recommendation | 1 |
7,741 | 10,862,664,483 | IssuesEvent | 2019-11-14 13:45:14 | qgis/QGIS | https://api.github.com/repos/qgis/QGIS | closed | Processing: python error on startup | Bug Feedback High Priority Processing Regression | Author Name: **John Drewett** (John Drewett)
Original Redmine Issue: [21720](https://issues.qgis.org/issues/21720)
Affected QGIS version: 3.6.1
Redmine category:processing/saga
---
The following error occurs on start up. Any advice would be appreciated.
```
2019-04-01T13:25:00 WARNING Traceback (most recent call last):
File "C:/PROGRA~1/QGIS3~1.6/apps/qgis/./python/plugins\processing\algs\saga\SagaAlgorithmProvider.py", line 90, in loadAlgorithms
version = SagaUtils.getInstalledVersion(True)
File "C:/PROGRA~1/QGIS3~1.6/apps/qgis/./python/plugins\processing\algs\saga\SagaUtils.py", line 145, in getInstalledVersion
universal_newlines=True,
File "C:\PROGRA~1\QGIS3~1.6\apps\Python37\lib\subprocess.py", line 709, in __init__
errread, errwrite) = self._get_handles(stdin, stdout, stderr)
File "C:\PROGRA~1\QGIS3~1.6\apps\Python37\lib\subprocess.py", line 1015, in _get_handles
p2cread = msvcrt.get_osfhandle(self._get_devnull())
File "C:\PROGRA~1\QGIS3~1.6\apps\Python37\lib\subprocess.py", line 848, in _get_devnull
self._devnull = os.open(os.devnull, os.O_RDWR)
FileNotFoundError: [Errno 2] No such file or directory: 'nul'
```
| 1.0 | Processing: python error on startup - Author Name: **John Drewett** (John Drewett)
Original Redmine Issue: [21720](https://issues.qgis.org/issues/21720)
Affected QGIS version: 3.6.1
Redmine category:processing/saga
---
The following error occurs on start up. Any advice would be appreciated.
```
2019-04-01T13:25:00 WARNING Traceback (most recent call last):
File "C:/PROGRA~1/QGIS3~1.6/apps/qgis/./python/plugins\processing\algs\saga\SagaAlgorithmProvider.py", line 90, in loadAlgorithms
version = SagaUtils.getInstalledVersion(True)
File "C:/PROGRA~1/QGIS3~1.6/apps/qgis/./python/plugins\processing\algs\saga\SagaUtils.py", line 145, in getInstalledVersion
universal_newlines=True,
File "C:\PROGRA~1\QGIS3~1.6\apps\Python37\lib\subprocess.py", line 709, in __init__
errread, errwrite) = self._get_handles(stdin, stdout, stderr)
File "C:\PROGRA~1\QGIS3~1.6\apps\Python37\lib\subprocess.py", line 1015, in _get_handles
p2cread = msvcrt.get_osfhandle(self._get_devnull())
File "C:\PROGRA~1\QGIS3~1.6\apps\Python37\lib\subprocess.py", line 848, in _get_devnull
self._devnull = os.open(os.devnull, os.O_RDWR)
FileNotFoundError: [Errno 2] No such file or directory: 'nul'
```
| process | processing python error on startup author name john drewett john drewett original redmine issue affected qgis version redmine category processing saga the following error occurs on start up any advice would be appreciated warning traceback most recent call last file c progra apps qgis python plugins processing algs saga sagaalgorithmprovider py line in loadalgorithms version sagautils getinstalledversion true file c progra apps qgis python plugins processing algs saga sagautils py line in getinstalledversion universal newlines true file c progra apps lib subprocess py line in init errread errwrite self get handles stdin stdout stderr file c progra apps lib subprocess py line in get handles msvcrt get osfhandle self get devnull file c progra apps lib subprocess py line in get devnull self devnull os open os devnull os o rdwr filenotfounderror no such file or directory nul | 1 |
142,272 | 13,019,035,714 | IssuesEvent | 2020-07-26 20:21:18 | Vyctor661/king-typer | https://api.github.com/repos/Vyctor661/king-typer | closed | docs | documentation | ## Needed docs
#### Frontend
- [ ] home page
- [ ] start typing page
1. easy mode
2. normal mode
3. tutorials mode
4. multiplayer modes (easy normal custom)
- [ ] loging register forgot password page
- [ ] profile pages
- [ ] dashboard
- [ ] themes
- [ ] recent games/statistics
#### Backend
| 1.0 | docs - ## Needed docs
#### Frontend
- [ ] home page
- [ ] start typing page
1. easy mode
2. normal mode
3. tutorials mode
4. multiplayer modes (easy normal custom)
- [ ] loging register forgot password page
- [ ] profile pages
- [ ] dashboard
- [ ] themes
- [ ] recent games/statistics
#### Backend
| non_process | docs needed docs frontend home page start typing page easy mode normal mode tutorials mode multiplayer modes easy normal custom loging register forgot password page profile pages dashboard themes recent games statistics backend | 0 |
708,829 | 24,356,547,957 | IssuesEvent | 2022-10-03 08:00:54 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | weather.com - video or audio doesn't play | priority-important browser-focus-geckoview engine-gecko | <!-- @browser: Firefox Mobile 105.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 12; Mobile; rv:105.0) Gecko/105.0 Firefox/105.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/111743 -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://weather.com/news/weather/video/pattern-change-brings-cooler-air-to-northeast-warm-weather-to-west?par=samsung_widget
**Browser / Version**: Firefox Mobile 105.0
**Operating System**: Android 12
**Tested Another Browser**: Yes Chrome
**Problem type**: Video or audio doesn't play
**Description**: The video or audio does not play
**Steps to Reproduce**:
Video buttons unresponsive in firefox android, works on chrome mobile
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/10/77ebe704-6eef-4dd0-ae4e-c715fc92a506.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220915150737</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2022/10/ddba324c-ed42-41a6-98c3-255eb5e50dad)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | weather.com - video or audio doesn't play - <!-- @browser: Firefox Mobile 105.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 12; Mobile; rv:105.0) Gecko/105.0 Firefox/105.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/111743 -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://weather.com/news/weather/video/pattern-change-brings-cooler-air-to-northeast-warm-weather-to-west?par=samsung_widget
**Browser / Version**: Firefox Mobile 105.0
**Operating System**: Android 12
**Tested Another Browser**: Yes Chrome
**Problem type**: Video or audio doesn't play
**Description**: The video or audio does not play
**Steps to Reproduce**:
Video buttons unresponsive in firefox android, works on chrome mobile
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/10/77ebe704-6eef-4dd0-ae4e-c715fc92a506.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220915150737</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2022/10/ddba324c-ed42-41a6-98c3-255eb5e50dad)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_process | weather com video or audio doesn t play url browser version firefox mobile operating system android tested another browser yes chrome problem type video or audio doesn t play description the video or audio does not play steps to reproduce video buttons unresponsive in firefox android works on chrome mobile view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel release hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 0 |
113,241 | 9,633,378,778 | IssuesEvent | 2019-05-15 18:29:24 | pvlib/pvlib-python | https://api.github.com/repos/pvlib/pvlib-python | closed | requests package requirement not listed | bug installation testing | **Describe the bug**
`import pvlib` fails due to unspecified `requests` dependency:
```
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/Users/holmgren/miniconda3/envs/pvlibtest/lib/python3.7/site-packages/pvlib/__init__.py", line 9, in <module>
from pvlib import iotools
File "/Users/holmgren/miniconda3/envs/pvlibtest/lib/python3.7/site-packages/pvlib/iotools/__init__.py", line 13, in <module>
from pvlib.iotools.psm3 import get_psm3 # noqa: F401
File "/Users/holmgren/miniconda3/envs/pvlibtest/lib/python3.7/site-packages/pvlib/iotools/psm3.py", line 8, in <module>
import requests
ModuleNotFoundError: No module named 'requests'
```
**To Reproduce**
Steps to reproduce the behavior:
1. `conda create -n pvlibtest python=3.7`
2. `pip install pvlib`
3. `python -c import pvlib`
**Expected behavior**
`pip install pvlib` works from a clean environment.
**Versions:**
- ``pvlib.__version__``: 0.6.2
- ``pandas.__version__``: 0.24.2
- python: 3.7
**Additional context**
#694 added a dependency on requests. The CI environments pull in requests through some of their dependencies, so it wasn't caught there. Should make a new issue for adding CI tests for pip installs into a clean environment.
0.6.3 coming up...
| 1.0 | requests package requirement not listed - **Describe the bug**
`import pvlib` fails due to unspecified `requests` dependency:
```
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/Users/holmgren/miniconda3/envs/pvlibtest/lib/python3.7/site-packages/pvlib/__init__.py", line 9, in <module>
from pvlib import iotools
File "/Users/holmgren/miniconda3/envs/pvlibtest/lib/python3.7/site-packages/pvlib/iotools/__init__.py", line 13, in <module>
from pvlib.iotools.psm3 import get_psm3 # noqa: F401
File "/Users/holmgren/miniconda3/envs/pvlibtest/lib/python3.7/site-packages/pvlib/iotools/psm3.py", line 8, in <module>
import requests
ModuleNotFoundError: No module named 'requests'
```
**To Reproduce**
Steps to reproduce the behavior:
1. `conda create -n pvlibtest python=3.7`
2. `pip install pvlib`
3. `python -c import pvlib`
**Expected behavior**
`pip install pvlib` works from a clean environment.
**Versions:**
- ``pvlib.__version__``: 0.6.2
- ``pandas.__version__``: 0.24.2
- python: 3.7
**Additional context**
#694 added a dependency on requests. The CI environments pull in requests through some of their dependencies, so it wasn't caught there. Should make a new issue for adding CI tests for pip installs into a clean environment.
0.6.3 coming up...
| non_process | requests package requirement not listed describe the bug import pvlib fails due to unspecified requests dependency traceback most recent call last file line in file users holmgren envs pvlibtest lib site packages pvlib init py line in from pvlib import iotools file users holmgren envs pvlibtest lib site packages pvlib iotools init py line in from pvlib iotools import get noqa file users holmgren envs pvlibtest lib site packages pvlib iotools py line in import requests modulenotfounderror no module named requests to reproduce steps to reproduce the behavior conda create n pvlibtest python pip install pvlib python c import pvlib expected behavior pip install pvlib works from a clean environment versions pvlib version pandas version python additional context added a dependency on requests the ci environments pull in requests through some of their dependencies so it wasn t caught there should make a new issue for adding ci tests for pip installs into a clean environment coming up | 0 |
22,654 | 31,895,827,910 | IssuesEvent | 2023-09-18 01:31:59 | tdwg/dwc | https://api.github.com/repos/tdwg/dwc | closed | Change term - latestEraOrHighestErathem | Term - change Class - GeologicalContext normative Task Group - Material Sample Process - complete | ## Term change
* Submitter: [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/)
* Efficacy Justification (why is this change necessary?): Create consistency of terms for material in Darwin Core.
* Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/), which includes representatives of over 10 organizations.
* Stability Justification (what concerns are there that this might affect existing implementations?): None
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: No
Current Term definition: https://dwc.tdwg.org/list/#dwc_latestEraOrHighestErathem
Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~):
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes): latestEraOrHighestErathem
* Term label (English, not normative): Latest Era Or Highest Erathem
* * Organized in Class (e.g., Occurrence, Event, Location, Taxon): Geological Context
* Definition of the term (normative): The full name of the latest possible geochronologic era or highest chronostratigraphic erathem attributable to the stratigraphic horizon from which the ~~cataloged item~~**dwc:MaterialEntity** was collected.
* Usage comments (recommendations regarding content, etc., not normative):
* Examples (not normative): Cenozoic, Mesozoic
* Refines (identifier of the broader term this term refines; normative): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): not in ABCD
| 1.0 | Change term - latestEraOrHighestErathem - ## Term change
* Submitter: [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/)
* Efficacy Justification (why is this change necessary?): Create consistency of terms for material in Darwin Core.
* Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/), which includes representatives of over 10 organizations.
* Stability Justification (what concerns are there that this might affect existing implementations?): None
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: No
Current Term definition: https://dwc.tdwg.org/list/#dwc_latestEraOrHighestErathem
Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~):
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes): latestEraOrHighestErathem
* Term label (English, not normative): Latest Era Or Highest Erathem
* * Organized in Class (e.g., Occurrence, Event, Location, Taxon): Geological Context
* Definition of the term (normative): The full name of the latest possible geochronologic era or highest chronostratigraphic erathem attributable to the stratigraphic horizon from which the ~~cataloged item~~**dwc:MaterialEntity** was collected.
* Usage comments (recommendations regarding content, etc., not normative):
* Examples (not normative): Cenozoic, Mesozoic
* Refines (identifier of the broader term this term refines; normative): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): not in ABCD
| process | change term latesteraorhighesterathem term change submitter efficacy justification why is this change necessary create consistency of terms for material in darwin core demand justification if the change is semantic in nature name at least two organizations that independently need this term which includes representatives of over organizations stability justification what concerns are there that this might affect existing implementations none implications for dwciri namespace does this change affect a dwciri term version no current term definition proposed attributes of the new term version please put actual changes to be implemented in bold and strikethrough term name in lowercamelcase for properties uppercamelcase for classes latesteraorhighesterathem term label english not normative latest era or highest erathem organized in class e g occurrence event location taxon geological context definition of the term normative the full name of the latest possible geochronologic era or highest chronostratigraphic erathem attributable to the stratigraphic horizon from which the cataloged item dwc materialentity was collected usage comments recommendations regarding content etc not normative examples not normative cenozoic mesozoic refines identifier of the broader term this term refines normative none replaces identifier of the existing term that would be deprecated and replaced by this term normative none abcd xpath of the equivalent term in abcd or efg not normative not in abcd | 1 |
10,660 | 13,453,144,059 | IssuesEvent | 2020-09-09 00:04:24 | GoogleCloudPlatform/cloud-ops-sandbox | https://api.github.com/repos/GoogleCloudPlatform/cloud-ops-sandbox | closed | [telemetry] Add python subscriber script to report from project to Pub/Sub in GCP | lang: python priority: p2 type: process | In order to learn how frequently the installation script succeeds, and how often users run destroy script, we should add a subscriber script to report information to Pub / Sub in GCP. | 1.0 | [telemetry] Add python subscriber script to report from project to Pub/Sub in GCP - In order to learn how frequently the installation script succeeds, and how often users run destroy script, we should add a subscriber script to report information to Pub / Sub in GCP. | process | add python subscriber script to report from project to pub sub in gcp in order to learn how frequently the installation script succeeds and how often users run destroy script we should add a subscriber script to report information to pub sub in gcp | 1 |
8,281 | 11,439,587,808 | IssuesEvent | 2020-02-05 07:41:20 | cypress-io/cypress | https://api.github.com/repos/cypress-io/cypress | opened | Test against latest stable browser versions (open new PR to do so) | process: tests type: chore | ### Current behavior:
We recently implemented this update to our internal testing process to *always* test against the latest browser version: https://github.com/cypress-io/cypress/pull/6115
This has become problematic, as evidenced by Chrome 80 update that occurred Feb 4. We are trying to release 4.0 and in the midst of this release the tests are not passing due to some change in Chrome 80. This makes it difficult to isolate which test failures are coming from the current branch changes and which are coming from the new Chrome version changes.
Furthermore this issue will just compound as we add more browser support.
### Desired behavior:
We should do something similar to how our `renovatebot` works today for our repo.
Upon release of a new stable browser version:
- Create a new docker image with new Chrome version?
- Initiate a PR against `cypress` that runs all tests against the latest browser version.
Then we can take the time to isolate which tests failed due to the new stable release and track them down.
### Versions
Cypress 3.8.3
| 1.0 | Test against latest stable browser versions (open new PR to do so) - ### Current behavior:
We recently implemented this update to our internal testing process to *always* test against the latest browser version: https://github.com/cypress-io/cypress/pull/6115
This has become problematic, as evidenced by Chrome 80 update that occurred Feb 4. We are trying to release 4.0 and in the midst of this release the tests are not passing due to some change in Chrome 80. This makes it difficult to isolate which test failures are coming from the current branch changes and which are coming from the new Chrome version changes.
Furthermore this issue will just compound as we add more browser support.
### Desired behavior:
We should do something similar to how our `renovatebot` works today for our repo.
Upon release of a new stable browser version:
- Create a new docker image with new Chrome version?
- Initiate a PR against `cypress` that runs all tests against the latest browser version.
Then we can take the time to isolate which tests failed due to the new stable release and track them down.
### Versions
Cypress 3.8.3
| process | test against latest stable browser versions open new pr to do so current behavior we recently implemented this update to our internal testing process to always test against the latest browser version this has become problematic as evidenced by chrome update that occurred feb we are trying to release and in the midst of this release the tests are not passing due to some change in chrome this makes it difficult to isolate which test failures are coming from the current branch changes and which are coming from the new chrome version changes furthermore this issue will just compound as we add more browser support desired behavior we should do something similar to how our renovatebot works today for our repo upon release of a new stable browser version create a new docker image with new chrome version initiate a pr against cypress that runs all tests against the latest browser version then we can take the time to isolate which tests failed due to the new stable release and track them down versions cypress | 1 |
283,286 | 21,316,086,379 | IssuesEvent | 2022-04-16 09:49:51 | ElaineQT/pe | https://api.github.com/repos/ElaineQT/pe | opened | Text format in DG | type.DocumentationBug severity.VeryLow | Regarding DG, is it better to include all class names and methods etc. in " ` "?
E.g. This is done in `SearchCommandParser`, rather than

Just like in DG of AddressBook 3:

<!--session: 1650096021327-6753a3ca-01e0-4b98-8a9f-295e17840146-->
<!--Version: Web v3.4.2--> | 1.0 | Text format in DG - Regarding DG, is it better to include all class names and methods etc. in " ` "?
E.g. This is done in `SearchCommandParser`, rather than

Just like in DG of AddressBook 3:

<!--session: 1650096021327-6753a3ca-01e0-4b98-8a9f-295e17840146-->
<!--Version: Web v3.4.2--> | non_process | text format in dg regarding dg is it better to include all class names and methods etc in e g this is done in searchcommandparser rather than just like in dg of addressbook | 0 |
134 | 2,573,115,852 | IssuesEvent | 2015-02-11 06:05:38 | sysown/proxysql-0.2 | https://api.github.com/repos/sysown/proxysql-0.2 | opened | Implement global variable mysql_connect_timeout_server_error | CONNECTION POOL cxx_pa GLOBAL MYSQL PROTOCOL QUERY PROCESSOR | Related to global variable mysql_connect_timeout_server ( issue #202)
In case it is unable to connect to a backend to for mysql_connect_timeout_server , return the specified error to the client or disconnect it | 1.0 | Implement global variable mysql_connect_timeout_server_error - Related to global variable mysql_connect_timeout_server ( issue #202)
In case it is unable to connect to a backend to for mysql_connect_timeout_server , return the specified error to the client or disconnect it | process | implement global variable mysql connect timeout server error related to global variable mysql connect timeout server issue in case it is unable to connect to a backend to for mysql connect timeout server return the specified error to the client or disconnect it | 1 |
59,877 | 17,023,275,867 | IssuesEvent | 2021-07-03 01:10:52 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | aerialway=drag_lift is not on the mapnik map.. | Component: mapnik Priority: minor Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 9.07am, Tuesday, 22nd July 2008]**
The official map feature "aerialway=cable_car" isn't rendered by mapnik. But in the wiki is written, that it should be rendered. Could you please fix this. Here in Switzerland we have regions wich are only reachable over such aerialways..
link to the map features:
http://wiki.openstreetmap.org/index.php/Map_features#Aerialway
link to the problem
http://openstreetmap.org/?lat=46.89994&lon=8.50536&zoom=15&layers=B00FTF
thank you | 1.0 | aerialway=drag_lift is not on the mapnik map.. - **[Submitted to the original trac issue database at 9.07am, Tuesday, 22nd July 2008]**
The official map feature "aerialway=cable_car" isn't rendered by mapnik. But in the wiki is written, that it should be rendered. Could you please fix this. Here in Switzerland we have regions wich are only reachable over such aerialways..
link to the map features:
http://wiki.openstreetmap.org/index.php/Map_features#Aerialway
link to the problem
http://openstreetmap.org/?lat=46.89994&lon=8.50536&zoom=15&layers=B00FTF
thank you | non_process | aerialway drag lift is not on the mapnik map the official map feature aerialway cable car isn t rendered by mapnik but in the wiki is written that it should be rendered could you please fix this here in switzerland we have regions wich are only reachable over such aerialways link to the map features link to the problem thank you | 0 |
583 | 3,060,127,967 | IssuesEvent | 2015-08-14 18:50:41 | Microsoft/poshtools | https://api.github.com/repos/Microsoft/poshtools | closed | Port Support Does Not Always Work | Process Attaching | Using ports does not always seem to work. It appears that we are parsing out the port incorrectly. | 1.0 | Port Support Does Not Always Work - Using ports does not always seem to work. It appears that we are parsing out the port incorrectly. | process | port support does not always work using ports does not always seem to work it appears that we are parsing out the port incorrectly | 1 |
80,069 | 9,977,339,216 | IssuesEvent | 2019-07-09 17:01:07 | mozilla/foundation.mozilla.org | https://api.github.com/repos/mozilla/foundation.mozilla.org | closed | Add more context to homepage | ui design ux design | Concerns that our homepage doesn't give new users enough context about who we are and why they should care has surfaced in two different ways recently.
#### From @beccaklam's users testing https://github.com/mozilla/foundation.mozilla.org/issues/2995#issuecomment-491035740:
- We should consider users who don't know our issues/Mozilla well and think about how to onboard them on our homepage if we ever want to grow our audience
- We should be aware of two types of audience -- those who know our cause or are internet-savvy and those who don't know us or are not internet-savvy. We don't want to talk down to people who feel that they already know their stuff.
#### From Juan and Lotta around fundraising [(doc from meeting)](https://docs.google.com/document/d/1txC3guawnjoeXUvFbFABhR32kPNDtx2O0MKGP8OzTJs/edit#):
- Going to the site it’s hard to say “What Mozilla does”. This should frame the rest of the content about projects, news, etc.
- What is a short pithy way to describe what we do that is engaging and captivating? This should be a big piece of the landing page
- Everything we present should be about the “bigger plan”
- Needs to make donors be a part of the work we do so that they don’t feel like they’re just giving us money to do work but they’re giving us money to do work that they care about and are contributing to
- Landing page doesn’t talk about why they should participate - only how to participate (e.g. Apply to be a Fellow). We need to be able to tell our broader story better on this page.
- Where do we talk about what our plan is for a healthier internet? We should be selling that plan/vision. I.e. resonating with audiences with emotion, a story, rather than specific programs we run out of context. Answering “what is Mozilla about” should be the first touchpoint for people
- Juan just links people to the fo’mo website so the first touchpoint is the landing page.
- News should tell the story of our impact
- Projects need more context so that the broader picture is still described using it
#### We should:
- benchmark similar sites: a few examples but should gather more
- https://colorofchange.org/
- https://rebellion.earth/
- https://www.astraeafoundation.org/global-reach/
- http://www.freedomforall.org
- http://sfaf.org/
- https://www.fordfoundation.org/
- https://www.shuttleworthfoundation.org/
- write up a few recommendations (some quick fixes some long-term)
- one of those recommendations should be to add an intro sentence (who should write this?)
- mock-up a few options
- get feedback from staff
- test with different audiences
cc: @sabrinang @natalieworth @beccaklam @xmatthewx @anilkanji | 2.0 | Add more context to homepage - Concerns that our homepage doesn't give new users enough context about who we are and why they should care has surfaced in two different ways recently.
#### From @beccaklam's users testing https://github.com/mozilla/foundation.mozilla.org/issues/2995#issuecomment-491035740:
- We should consider users who don't know our issues/Mozilla well and think about how to onboard them on our homepage if we ever want to grow our audience
- We should be aware of two types of audience -- those who know our cause or are internet-savvy and those who don't know us or are not internet-savvy. We don't want to talk down to people who feel that they already know their stuff.
#### From Juan and Lotta around fundraising [(doc from meeting)](https://docs.google.com/document/d/1txC3guawnjoeXUvFbFABhR32kPNDtx2O0MKGP8OzTJs/edit#):
- Going to the site it’s hard to say “What Mozilla does”. This should frame the rest of the content about projects, news, etc.
- What is a short pithy way to describe what we do that is engaging and captivating? This should be a big piece of the landing page
- Everything we present should be about the “bigger plan”
- Needs to make donors be a part of the work we do so that they don’t feel like they’re just giving us money to do work but they’re giving us money to do work that they care about and are contributing to
- Landing page doesn’t talk about why they should participate - only how to participate (e.g. Apply to be a Fellow). We need to be able to tell our broader story better on this page.
- Where do we talk about what our plan is for a healthier internet? We should be selling that plan/vision. I.e. resonating with audiences with emotion, a story, rather than specific programs we run out of context. Answering “what is Mozilla about” should be the first touchpoint for people
- Juan just links people to the fo’mo website so the first touchpoint is the landing page.
- News should tell the story of our impact
- Projects need more context so that the broader picture is still described using it
#### We should:
- benchmark similar sites: a few examples but should gather more
- https://colorofchange.org/
- https://rebellion.earth/
- https://www.astraeafoundation.org/global-reach/
- http://www.freedomforall.org
- http://sfaf.org/
- https://www.fordfoundation.org/
- https://www.shuttleworthfoundation.org/
- write up a few recommendations (some quick fixes some long-term)
- one of those recommendations should be to add an intro sentence (who should write this?)
- mock-up a few options
- get feedback from staff
- test with different audiences
cc: @sabrinang @natalieworth @beccaklam @xmatthewx @anilkanji | non_process | add more context to homepage concerns that our homepage doesn t give new users enough context about who we are and why they should care has surfaced in two different ways recently from beccaklam s users testing we should consider users who don t know our issues mozilla well and think about how to onboard them on our homepage if we ever want to grow our audience we should be aware of two types of audience those who know our cause or are internet savvy and those who don t know us or are not internet savvy we don t want to talk down to people who feel that they already know their stuff from juan and lotta around fundraising going to the site it’s hard to say “what mozilla does” this should frame the rest of the content about projects news etc what is a short pithy way to describe what we do that is engaging and captivating this should be a big piece of the landing page everything we present should be about the “bigger plan” needs to make donors be a part of the work we do so that they don’t feel like they’re just giving us money to do work but they’re giving us money to do work that they care about and are contributing to landing page doesn’t talk about why they should participate only how to participate e g apply to be a fellow we need to be able to tell our broader story better on this page where do we talk about what our plan is for a healthier internet we should be selling that plan vision i e resonating with audiences with emotion a story rather than specific programs we run out of context answering “what is mozilla about” should be the first touchpoint for people juan just links people to the fo’mo website so the first touchpoint is the landing page news should tell the story of our impact projects need more context so that the broader picture is still described using it we should benchmark similar sites a few examples but should gather more write up a few recommendations some quick fixes some long term one of those recommendations should be to add an intro sentence who should write this mock up a few options get feedback from staff test with different audiences cc sabrinang natalieworth beccaklam xmatthewx anilkanji | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.