Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
1,725
4,381,729,581
IssuesEvent
2016-08-06 12:21:14
mlr-org/mlr
https://api.github.com/repos/mlr-org/mlr
closed
impute() does not handle non missing values in columns
bug enhancement preprocessing
impute should be able to do nothing if no missing values are present: ```r library(mlr) iris.m = iris imp = impute(data = iris.m, target = "Species", cols = list(Sepal.Width = imputeLearner(makeLearner("regr.rpart")))) #Error in impute(data = iris.m, target = "Species", cols = list(Sepal.Width = imputeLearner(makeLearner("regr.rpart")))) : # Assertion on 'newdata' failed: Must have at least 1 rows, but has 0 rows.
1.0
impute() does not handle non missing values in columns - impute should be able to do nothing if no missing values are present: ```r library(mlr) iris.m = iris imp = impute(data = iris.m, target = "Species", cols = list(Sepal.Width = imputeLearner(makeLearner("regr.rpart")))) #Error in impute(data = iris.m, target = "Species", cols = list(Sepal.Width = imputeLearner(makeLearner("regr.rpart")))) : # Assertion on 'newdata' failed: Must have at least 1 rows, but has 0 rows.
process
impute does not handle non missing values in columns impute should be able to do nothing if no missing values are present r library mlr iris m iris imp impute data iris m target species cols list sepal width imputelearner makelearner regr rpart error in impute data iris m target species cols list sepal width imputelearner makelearner regr rpart assertion on newdata failed must have at least rows but has rows
1
373,978
11,053,615,754
IssuesEvent
2019-12-10 11:46:17
wso2/product-is
https://api.github.com/repos/wso2/product-is
opened
Cannot set Access URL for a service provider through the management console
Priority/High Severity/Critical Type/Bug
Steps to reproduce. **Case I** 1. Create a service provider named as `test` 2. Set `Access URL` as `https://www.google.com/` in the next page. 3. Click on update. (Update seems to complete without any errors and directed to SP listing page) 4. Click on edit of the created `test` service provider. 5. `Access URL` shown as empty. **Case II** 1. Create a service provider named as `test` 2. Set `Access URL` as `https://www.google.com/` in the next page. 3. Click on `Discoverable Application` checkbox to make it enable. 4. Click on update. Update get failure with the following error message, `Error while updating Service Provider : Invalid application configuration for application: 'test' of tenantDomain: carbon.super. A valid accessURL needs to be defined if an application is marked as discoverable.`
1.0
Cannot set Access URL for a service provider through the management console - Steps to reproduce. **Case I** 1. Create a service provider named as `test` 2. Set `Access URL` as `https://www.google.com/` in the next page. 3. Click on update. (Update seems to complete without any errors and directed to SP listing page) 4. Click on edit of the created `test` service provider. 5. `Access URL` shown as empty. **Case II** 1. Create a service provider named as `test` 2. Set `Access URL` as `https://www.google.com/` in the next page. 3. Click on `Discoverable Application` checkbox to make it enable. 4. Click on update. Update get failure with the following error message, `Error while updating Service Provider : Invalid application configuration for application: 'test' of tenantDomain: carbon.super. A valid accessURL needs to be defined if an application is marked as discoverable.`
non_process
cannot set access url for a service provider through the management console steps to reproduce case i create a service provider named as test set access url as in the next page click on update update seems to complete without any errors and directed to sp listing page click on edit of the created test service provider access url shown as empty case ii create a service provider named as test set access url as in the next page click on discoverable application checkbox to make it enable click on update update get failure with the following error message error while updating service provider invalid application configuration for application test of tenantdomain carbon super a valid accessurl needs to be defined if an application is marked as discoverable
0
42,927
17,371,996,106
IssuesEvent
2021-07-30 15:08:34
Azure/azure-cli
https://api.github.com/repos/Azure/azure-cli
closed
az webapp deployment source config-zip: Deployment endpoint responded with status code 202. But fail to run?
Service Attention Web Apps
I'm running `az webapp deployment source config-zip --resource-group XX --name XX --src XX-package.zip` on a Linux Web Service instance. It's a node project. My zip contain a package.json with a npm start command that I expect Azure to run. ``` ... Getting scm site credentials for zip deployment Starting zip deployment. This operation can take a while to complete ... Deployment endpoint responded with status code 202 { "active": true, "author": "N/A", "author_email": "N/A", "complete": true, "deployer": "Push-Deployer", "end_time": "2021-07-30T09:41:17.9438556Z", "id": "a8063d7cc33943e9b852cd5a12a1184d", "is_readonly": true, "is_temp": false, "last_success_end_time": "2021-07-30T09:41:17.9438556Z", "log_url": "https:/xx.scm.azurewebsites.net/api/deployments/latest/log", "message": "Created via a push deployment", "progress": "", "received_time": "2021-07-30T09:41:13.0359297Z", "site_name": "xx", "start_time": "2021-07-30T09:41:14.2349318Z", "status": 4, "status_text": "", "url": "https://xx.scm.azurewebsites.net/api/deployments/latest" } ``` But then? The app do not seem to run.
1.0
az webapp deployment source config-zip: Deployment endpoint responded with status code 202. But fail to run? - I'm running `az webapp deployment source config-zip --resource-group XX --name XX --src XX-package.zip` on a Linux Web Service instance. It's a node project. My zip contain a package.json with a npm start command that I expect Azure to run. ``` ... Getting scm site credentials for zip deployment Starting zip deployment. This operation can take a while to complete ... Deployment endpoint responded with status code 202 { "active": true, "author": "N/A", "author_email": "N/A", "complete": true, "deployer": "Push-Deployer", "end_time": "2021-07-30T09:41:17.9438556Z", "id": "a8063d7cc33943e9b852cd5a12a1184d", "is_readonly": true, "is_temp": false, "last_success_end_time": "2021-07-30T09:41:17.9438556Z", "log_url": "https:/xx.scm.azurewebsites.net/api/deployments/latest/log", "message": "Created via a push deployment", "progress": "", "received_time": "2021-07-30T09:41:13.0359297Z", "site_name": "xx", "start_time": "2021-07-30T09:41:14.2349318Z", "status": 4, "status_text": "", "url": "https://xx.scm.azurewebsites.net/api/deployments/latest" } ``` But then? The app do not seem to run.
non_process
az webapp deployment source config zip deployment endpoint responded with status code but fail to run i m running az webapp deployment source config zip resource group xx name xx src xx package zip on a linux web service instance it s a node project my zip contain a package json with a npm start command that i expect azure to run getting scm site credentials for zip deployment starting zip deployment this operation can take a while to complete deployment endpoint responded with status code active true author n a author email n a complete true deployer push deployer end time id is readonly true is temp false last success end time log url https xx scm azurewebsites net api deployments latest log message created via a push deployment progress received time site name xx start time status status text url but then the app do not seem to run
0
2,253
5,088,654,207
IssuesEvent
2017-01-01 00:06:33
sw4j-org/tool-jpa-processor
https://api.github.com/repos/sw4j-org/tool-jpa-processor
opened
Handle @SecondaryTable Annotation
annotation processor task
Handle the `@SecondaryTable` annotation for an entity. See [JSR 338: Java Persistence API, Version 2.1](http://download.oracle.com/otn-pub/jcp/persistence-2_1-fr-eval-spec/JavaPersistence.pdf) - 11.1.46 SecondaryTable Annotation
1.0
Handle @SecondaryTable Annotation - Handle the `@SecondaryTable` annotation for an entity. See [JSR 338: Java Persistence API, Version 2.1](http://download.oracle.com/otn-pub/jcp/persistence-2_1-fr-eval-spec/JavaPersistence.pdf) - 11.1.46 SecondaryTable Annotation
process
handle secondarytable annotation handle the secondarytable annotation for an entity see secondarytable annotation
1
27,056
6,813,308,172
IssuesEvent
2017-11-06 08:45:09
BTDF/DeploymentFramework
https://api.github.com/repos/BTDF/DeploymentFramework
closed
Feature: More extensibilty on WIX targets
CodePlexMigrationInitiated enhancement Impact: Low MSI Creation and WiX Release 5.5
Are you considering to update the framework to provide extensibility to WIX targets? Here are few things I am looking at right now...   I am against to change the framework files for custom menu items as we have one build server for several applications. Changing framework files screws up other applications. Also, any future updates to framework will loose the customization. If you can provide more extensibility to WIX (i know right now this is in .js file) ie, moving wix xml-composition to targets with extensibility, we will have lot of flexibility in customizing the MSI.   Use of a property (something like InstallerFileName) for MSI name. [TFA -- available in BTDF v5.0]   Setting File/Product Versions for assemblies (Alternatively I am changing the each project file before compile the solution) [TFA -- this must be done before compilation, so it cannot be handled by the Deployment Framework]   Ability to set custom properties (like Version, Source, BuildServer) on MSI file. I don't know if that is possible with WIX. Right now I am using 'PackageComments' with concatenating all properties I need. #### This work item was migrated from CodePlex CodePlex work item ID: '8159' Assigned to: 'tfabraham' Vote count: '3'
1.0
Feature: More extensibilty on WIX targets - Are you considering to update the framework to provide extensibility to WIX targets? Here are few things I am looking at right now...   I am against to change the framework files for custom menu items as we have one build server for several applications. Changing framework files screws up other applications. Also, any future updates to framework will loose the customization. If you can provide more extensibility to WIX (i know right now this is in .js file) ie, moving wix xml-composition to targets with extensibility, we will have lot of flexibility in customizing the MSI.   Use of a property (something like InstallerFileName) for MSI name. [TFA -- available in BTDF v5.0]   Setting File/Product Versions for assemblies (Alternatively I am changing the each project file before compile the solution) [TFA -- this must be done before compilation, so it cannot be handled by the Deployment Framework]   Ability to set custom properties (like Version, Source, BuildServer) on MSI file. I don't know if that is possible with WIX. Right now I am using 'PackageComments' with concatenating all properties I need. #### This work item was migrated from CodePlex CodePlex work item ID: '8159' Assigned to: 'tfabraham' Vote count: '3'
non_process
feature more extensibilty on wix targets are you considering to update the framework to provide extensibility to wix targets here are few things i am looking at right now   i am against to change the framework files for custom menu items as we have one build server for several applications changing framework files screws up other applications also any future updates to framework will loose the customization if you can provide more extensibility to wix i know right now this is in js file ie moving wix xml composition to targets with extensibility we will have lot of flexibility in customizing the msi   use of a property something like installerfilename for msi name   setting file product versions for assemblies alternatively i am changing the each project file before compile the solution   ability to set custom properties like version source buildserver on msi file i don t know if that is possible with wix right now i am using packagecomments with concatenating all properties i need this work item was migrated from codeplex codeplex work item id assigned to tfabraham vote count
0
4,296
7,192,448,899
IssuesEvent
2018-02-03 03:41:35
Great-Hill-Corporation/quickBlocks
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
closed
ethslurp needs an Etherscan API key to work.
apps-ethslurp status-inprocess type-enhancement
We need to test the process of how a user is warned that they need an Etherscan API key and how to go about doing that. Referring to #248.
1.0
ethslurp needs an Etherscan API key to work. - We need to test the process of how a user is warned that they need an Etherscan API key and how to go about doing that. Referring to #248.
process
ethslurp needs an etherscan api key to work we need to test the process of how a user is warned that they need an etherscan api key and how to go about doing that referring to
1
17,553
23,367,332,901
IssuesEvent
2022-08-10 16:28:04
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
Process: maintainer involvement in triaging issues
Process
This is a re-hash of the initial discussion in #46759. Following some offline discussion with @nashif, we agreed that it would be worthwhile to start sending requests for action from maintainers to periodically triage open bugs and issues in parallel to the ongoing broader discussions about release criteria that turned into https://github.com/zephyrproject-rtos/zephyr/issues/38566. To resolve in particular: - Triaging of issues, especially low priority issues: what to use? "known issue" label? "milestone" metadata? - Any other triaging for initial effort?
1.0
Process: maintainer involvement in triaging issues - This is a re-hash of the initial discussion in #46759. Following some offline discussion with @nashif, we agreed that it would be worthwhile to start sending requests for action from maintainers to periodically triage open bugs and issues in parallel to the ongoing broader discussions about release criteria that turned into https://github.com/zephyrproject-rtos/zephyr/issues/38566. To resolve in particular: - Triaging of issues, especially low priority issues: what to use? "known issue" label? "milestone" metadata? - Any other triaging for initial effort?
process
process maintainer involvement in triaging issues this is a re hash of the initial discussion in following some offline discussion with nashif we agreed that it would be worthwhile to start sending requests for action from maintainers to periodically triage open bugs and issues in parallel to the ongoing broader discussions about release criteria that turned into to resolve in particular triaging of issues especially low priority issues what to use known issue label milestone metadata any other triaging for initial effort
1
145,445
5,575,654,671
IssuesEvent
2017-03-28 03:00:10
YaleSTC/vesta
https://api.github.com/repos/YaleSTC/vesta
opened
Improve enrollment errors / flashes
complexity: 3 priority: 4 type: bug
* [ ] Silent failures when users already exist (check for no flashes?) * [ ] Show successes/failures even after timeout
1.0
Improve enrollment errors / flashes - * [ ] Silent failures when users already exist (check for no flashes?) * [ ] Show successes/failures even after timeout
non_process
improve enrollment errors flashes silent failures when users already exist check for no flashes show successes failures even after timeout
0
9,671
12,676,767,184
IssuesEvent
2020-06-19 06:11:45
feast-dev/feast
https://api.github.com/repos/feast-dev/feast
closed
Link to BigQuery permissions for service account is broken
kind/process kind/question
The helm charts README mentioned that ``` In production, it is advised to give only the required permissions for the the service account, versus editor role which is very permissive.``` however the link to the required permissions is broken. Would be great to understand which permissions are required in production environment.
1.0
Link to BigQuery permissions for service account is broken - The helm charts README mentioned that ``` In production, it is advised to give only the required permissions for the the service account, versus editor role which is very permissive.``` however the link to the required permissions is broken. Would be great to understand which permissions are required in production environment.
process
link to bigquery permissions for service account is broken the helm charts readme mentioned that in production it is advised to give only the required permissions for the the service account versus editor role which is very permissive however the link to the required permissions is broken would be great to understand which permissions are required in production environment
1
197,894
14,948,419,258
IssuesEvent
2021-01-26 10:03:16
pints-team/functional-testing
https://api.github.com/repos/pints-team/functional-testing
opened
Add "simulation-based calibration" tests
new test
https://github.com/pints-team/pints/issues/843 https://arxiv.org/pdf/1804.06788.pdf Sean Talts, Michael Betancourt, Daniel Simpson, Aki Vehtari, Andrew Gelman (2020) Validating Bayesian Inference Algorithms with Simulation-Based Calibration
1.0
Add "simulation-based calibration" tests - https://github.com/pints-team/pints/issues/843 https://arxiv.org/pdf/1804.06788.pdf Sean Talts, Michael Betancourt, Daniel Simpson, Aki Vehtari, Andrew Gelman (2020) Validating Bayesian Inference Algorithms with Simulation-Based Calibration
non_process
add simulation based calibration tests sean talts michael betancourt daniel simpson aki vehtari andrew gelman validating bayesian inference algorithms with simulation based calibration
0
22,296
30,851,719,419
IssuesEvent
2023-08-02 17:15:38
googleapis/python-bigquery
https://api.github.com/repos/googleapis/python-bigquery
closed
Please add tests for the BigQuery v1 branch
api: bigquery status: will not fix type: process
There is an issue reported that `google-api-core` version 2.11 breaks compatibility with the `google-cloud-bigquery` v1 branch. See https://github.com/googleapis/python-api-core/issues/477 . Creating an issue to enable tests for the v1 branch if possible.
1.0
Please add tests for the BigQuery v1 branch - There is an issue reported that `google-api-core` version 2.11 breaks compatibility with the `google-cloud-bigquery` v1 branch. See https://github.com/googleapis/python-api-core/issues/477 . Creating an issue to enable tests for the v1 branch if possible.
process
please add tests for the bigquery branch there is an issue reported that google api core version breaks compatibility with the google cloud bigquery branch see creating an issue to enable tests for the branch if possible
1
19,559
25,883,173,720
IssuesEvent
2022-12-14 12:46:15
gradle/gradle
https://api.github.com/repos/gradle/gradle
closed
Module export rules needed with java17 defined via options.compilerArg are not applied to annotation processor
a:bug in:modular-java in:annotation-processing
<!--- Provide a brief summary of the issue in the title above --> ### Expected Behavior There should be a way to specify module export rules for those to apply to annotation processor ### Current Behavior only able to set module rules via global gradle JVM gradle.properies ### Context We have a happy codebase which works fine under java 11. JDK17 imposes additional module export requirments around javac internal API usages by our annotation processor, we are unable to set such rules locally just for specific compilation task. As a workaround we set export rules globally. Both problematic and workaround versions are attached. How we believe it export rules should be added does not get the job done see attached project ``` compileJava { options.compilerArgs += ["--add-exports=jdk.compiler/com.sun.tools.javac.api=ALL-UNNAMED", "--add-exports=jdk.compiler/com.sun.tools.javac.file=ALL-UNNAMED", "--add-exports=jdk.compiler/com.sun.tools.javac.main=ALL-UNNAMED", "--add-exports=jdk.compiler/com.sun.tools.javac.model=ALL-UNNAMED", "--add-exports=jdk.compiler/com.sun.tools.javac.parser=ALL-UNNAMED", "--add-exports=jdk.compiler/com.sun.tools.javac.processing=ALL-UNNAMED", "--add-exports=jdk.compiler/com.sun.tools.javac.tree=ALL-UNNAMED", "--add-exports=jdk.compiler/com.sun.tools.javac.util=ALL-UNNAMED", "--add-exports=jdk.compiler/com.sun.tools.javac.code=ALL-UNNAMED", "--add-exports=jdk.compiler/com.sun.tools.javac.comp=ALL-UNNAMED", "--add-exports=jdk.compiler/sun.nio.ch=ALL-UNNAMED"] } ``` how we work around the issue ``` org.gradle.jvmargs= --add-exports=jdk.compiler/com.sun.tools.javac.api=ALL-UNNAMED \ --add-exports=jdk.compiler/com.sun.tools.javac.file=ALL-UNNAMED \ --add-exports=jdk.compiler/com.sun.tools.javac.main=ALL-UNNAMED \ --add-exports=jdk.compiler/com.sun.tools.javac.model=ALL-UNNAMED \ --add-exports=jdk.compiler/com.sun.tools.javac.parser=ALL-UNNAMED \ --add-exports=jdk.compiler/com.sun.tools.javac.processing=ALL-UNNAMED \ --add-exports=jdk.compiler/com.sun.tools.javac.tree=ALL-UNNAMED \ --add-exports=jdk.compiler/com.sun.tools.javac.util=ALL-UNNAMED \ --add-exports=jdk.compiler/com.sun.tools.javac.code=ALL-UNNAMED \ --add-exports=jdk.compiler/com.sun.tools.javac.comp=ALL-UNNAMED \ --add-exports=jdk.compiler/sun.nio.ch=ALL-UNNAMED \ --add-exports=java.base/sun.nio.ch=ALL-UNNAMED ``` [devexperts-17-workaround.zip](https://github.com/gradle/gradle/files/9822041/devexperts-17-workaround.zip) [devexperts-17-issue.zip](https://github.com/gradle/gradle/files/9822042/devexperts-17-issue.zip) This adds steps-to-reproduce for https://github.com/gradle/gradle/issues/21310 This is a duplicate of https://github.com/gradle/gradle/issues/22065 where we have failed to express ourselves
1.0
Module export rules needed with java17 defined via options.compilerArg are not applied to annotation processor - <!--- Provide a brief summary of the issue in the title above --> ### Expected Behavior There should be a way to specify module export rules for those to apply to annotation processor ### Current Behavior only able to set module rules via global gradle JVM gradle.properies ### Context We have a happy codebase which works fine under java 11. JDK17 imposes additional module export requirments around javac internal API usages by our annotation processor, we are unable to set such rules locally just for specific compilation task. As a workaround we set export rules globally. Both problematic and workaround versions are attached. How we believe it export rules should be added does not get the job done see attached project ``` compileJava { options.compilerArgs += ["--add-exports=jdk.compiler/com.sun.tools.javac.api=ALL-UNNAMED", "--add-exports=jdk.compiler/com.sun.tools.javac.file=ALL-UNNAMED", "--add-exports=jdk.compiler/com.sun.tools.javac.main=ALL-UNNAMED", "--add-exports=jdk.compiler/com.sun.tools.javac.model=ALL-UNNAMED", "--add-exports=jdk.compiler/com.sun.tools.javac.parser=ALL-UNNAMED", "--add-exports=jdk.compiler/com.sun.tools.javac.processing=ALL-UNNAMED", "--add-exports=jdk.compiler/com.sun.tools.javac.tree=ALL-UNNAMED", "--add-exports=jdk.compiler/com.sun.tools.javac.util=ALL-UNNAMED", "--add-exports=jdk.compiler/com.sun.tools.javac.code=ALL-UNNAMED", "--add-exports=jdk.compiler/com.sun.tools.javac.comp=ALL-UNNAMED", "--add-exports=jdk.compiler/sun.nio.ch=ALL-UNNAMED"] } ``` how we work around the issue ``` org.gradle.jvmargs= --add-exports=jdk.compiler/com.sun.tools.javac.api=ALL-UNNAMED \ --add-exports=jdk.compiler/com.sun.tools.javac.file=ALL-UNNAMED \ --add-exports=jdk.compiler/com.sun.tools.javac.main=ALL-UNNAMED \ --add-exports=jdk.compiler/com.sun.tools.javac.model=ALL-UNNAMED \ --add-exports=jdk.compiler/com.sun.tools.javac.parser=ALL-UNNAMED \ --add-exports=jdk.compiler/com.sun.tools.javac.processing=ALL-UNNAMED \ --add-exports=jdk.compiler/com.sun.tools.javac.tree=ALL-UNNAMED \ --add-exports=jdk.compiler/com.sun.tools.javac.util=ALL-UNNAMED \ --add-exports=jdk.compiler/com.sun.tools.javac.code=ALL-UNNAMED \ --add-exports=jdk.compiler/com.sun.tools.javac.comp=ALL-UNNAMED \ --add-exports=jdk.compiler/sun.nio.ch=ALL-UNNAMED \ --add-exports=java.base/sun.nio.ch=ALL-UNNAMED ``` [devexperts-17-workaround.zip](https://github.com/gradle/gradle/files/9822041/devexperts-17-workaround.zip) [devexperts-17-issue.zip](https://github.com/gradle/gradle/files/9822042/devexperts-17-issue.zip) This adds steps-to-reproduce for https://github.com/gradle/gradle/issues/21310 This is a duplicate of https://github.com/gradle/gradle/issues/22065 where we have failed to express ourselves
process
module export rules needed with defined via options compilerarg are not applied to annotation processor expected behavior there should be a way to specify module export rules for those to apply to annotation processor current behavior only able to set module rules via global gradle jvm gradle properies context we have a happy codebase which works fine under java imposes additional module export requirments around javac internal api usages by our annotation processor we are unable to set such rules locally just for specific compilation task as a workaround we set export rules globally both problematic and workaround versions are attached how we believe it export rules should be added does not get the job done see attached project compilejava options compilerargs add exports jdk compiler com sun tools javac api all unnamed add exports jdk compiler com sun tools javac file all unnamed add exports jdk compiler com sun tools javac main all unnamed add exports jdk compiler com sun tools javac model all unnamed add exports jdk compiler com sun tools javac parser all unnamed add exports jdk compiler com sun tools javac processing all unnamed add exports jdk compiler com sun tools javac tree all unnamed add exports jdk compiler com sun tools javac util all unnamed add exports jdk compiler com sun tools javac code all unnamed add exports jdk compiler com sun tools javac comp all unnamed add exports jdk compiler sun nio ch all unnamed how we work around the issue org gradle jvmargs add exports jdk compiler com sun tools javac api all unnamed add exports jdk compiler com sun tools javac file all unnamed add exports jdk compiler com sun tools javac main all unnamed add exports jdk compiler com sun tools javac model all unnamed add exports jdk compiler com sun tools javac parser all unnamed add exports jdk compiler com sun tools javac processing all unnamed add exports jdk compiler com sun tools javac tree all unnamed add exports jdk compiler com sun tools javac util all unnamed add exports jdk compiler com sun tools javac code all unnamed add exports jdk compiler com sun tools javac comp all unnamed add exports jdk compiler sun nio ch all unnamed add exports java base sun nio ch all unnamed this adds steps to reproduce for this is a duplicate of where we have failed to express ourselves
1
694,935
23,836,788,050
IssuesEvent
2022-09-06 06:51:12
space-wizards/space-station-14
https://api.github.com/repos/space-wizards/space-station-14
closed
Traitor selection needs vibechecked
Issue: Bug Priority: 1-Urgent Difficulty: 2-Medium
## Description Seems to tend to give it to people who have antags all off.
1.0
Traitor selection needs vibechecked - ## Description Seems to tend to give it to people who have antags all off.
non_process
traitor selection needs vibechecked description seems to tend to give it to people who have antags all off
0
8,137
2,611,454,071
IssuesEvent
2015-02-27 05:01:13
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
closed
Unlimited Attacks: turn ends if no more enemy hedgehogs and when your hog is using rope, flying saucer, etc.
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Run fight with weapon set containing rope or flying saucer and game mode with Unlimited Attacks mode. 2. Kill all enemy hedgehog and use rope or flying saucer. What is the expected output? What do you see instead? The turn ends after few seconds. The situation might be annoying in some cases, for example when you have to escape from dynamite, and the only way is to use flying saucer (and the safe place is far away). What version of the product are you using? On what operating system? 0.9.14.1 on Windows XP SP2 Please provide any additional information below. ``` Original issue reported on code.google.com by `adibiaz...@gmail.com` on 22 Dec 2010 at 5:36 * Merged into: #129
1.0
Unlimited Attacks: turn ends if no more enemy hedgehogs and when your hog is using rope, flying saucer, etc. - ``` What steps will reproduce the problem? 1. Run fight with weapon set containing rope or flying saucer and game mode with Unlimited Attacks mode. 2. Kill all enemy hedgehog and use rope or flying saucer. What is the expected output? What do you see instead? The turn ends after few seconds. The situation might be annoying in some cases, for example when you have to escape from dynamite, and the only way is to use flying saucer (and the safe place is far away). What version of the product are you using? On what operating system? 0.9.14.1 on Windows XP SP2 Please provide any additional information below. ``` Original issue reported on code.google.com by `adibiaz...@gmail.com` on 22 Dec 2010 at 5:36 * Merged into: #129
non_process
unlimited attacks turn ends if no more enemy hedgehogs and when your hog is using rope flying saucer etc what steps will reproduce the problem run fight with weapon set containing rope or flying saucer and game mode with unlimited attacks mode kill all enemy hedgehog and use rope or flying saucer what is the expected output what do you see instead the turn ends after few seconds the situation might be annoying in some cases for example when you have to escape from dynamite and the only way is to use flying saucer and the safe place is far away what version of the product are you using on what operating system on windows xp please provide any additional information below original issue reported on code google com by adibiaz gmail com on dec at merged into
0
10,914
13,690,855,782
IssuesEvent
2020-09-30 14:52:14
prisma/prisma-engines
https://api.github.com/repos/prisma/prisma-engines
closed
Postgres: serial native type and `@default(autoincrement())` must not be possible at the same time
engines/data model parser process/candidate team/engines
The following schema should be invalid: ``` datasource test { provider = "postgresql" url = ... } model Blog { id Int @id serial Int @default(autoincrement()) @test.Serial } ``` The native types `serial` is a special type in Postgres that translates to an Integer column with an auto incrementing counter as a default. So it means exactly the same thing as `@default(autoincrement())`. In fact the migration engine translates `Int @default(autoincrement())` to the `serial` type underneath. Hence the parser should validate this and ask the user to remove one of the two attributes on Postgres.
1.0
Postgres: serial native type and `@default(autoincrement())` must not be possible at the same time - The following schema should be invalid: ``` datasource test { provider = "postgresql" url = ... } model Blog { id Int @id serial Int @default(autoincrement()) @test.Serial } ``` The native types `serial` is a special type in Postgres that translates to an Integer column with an auto incrementing counter as a default. So it means exactly the same thing as `@default(autoincrement())`. In fact the migration engine translates `Int @default(autoincrement())` to the `serial` type underneath. Hence the parser should validate this and ask the user to remove one of the two attributes on Postgres.
process
postgres serial native type and default autoincrement must not be possible at the same time the following schema should be invalid datasource test provider postgresql url model blog id int id serial int default autoincrement test serial the native types serial is a special type in postgres that translates to an integer column with an auto incrementing counter as a default so it means exactly the same thing as default autoincrement in fact the migration engine translates int default autoincrement to the serial type underneath hence the parser should validate this and ask the user to remove one of the two attributes on postgres
1
69,459
22,358,988,244
IssuesEvent
2022-06-15 18:26:06
hpi-swa-teaching/SVGMorph
https://api.github.com/repos/hpi-swa-teaching/SVGMorph
opened
Show elemets that do not have a color specified
defect User Story
As a viewer, I would be able to see all elements of an SVG that should be visible according to the SVG standard. This includes elements without a color specified, since the default color should be black, which is not implemented at the moment.
1.0
Show elemets that do not have a color specified - As a viewer, I would be able to see all elements of an SVG that should be visible according to the SVG standard. This includes elements without a color specified, since the default color should be black, which is not implemented at the moment.
non_process
show elemets that do not have a color specified as a viewer i would be able to see all elements of an svg that should be visible according to the svg standard this includes elements without a color specified since the default color should be black which is not implemented at the moment
0
347,208
10,426,551,865
IssuesEvent
2019-09-16 17:51:31
googleapis/google-cloud-python
https://api.github.com/repos/googleapis/google-cloud-python
closed
Vision: Operation returned from 'purge_product: raises TypeError in 'done'.
api: vision external priority: p1 type: bug
#### Environment details 1. Vision API, Product Search, Purge Product 2. OSX 3. 2.7.14 4. google-cloud-vision 0.39.0 #### Steps to reproduce 1. Run tests or either of the purge product samples from this PR: https://github.com/GoogleCloudPlatform/python-docs-samples/pull/2349 1. The LRO fails at `operation.done()` #### Code example see PR linked above. #### Stack trace ``` Traceback (most recent call last): File "/Users/yuhanliu/projects/dpe-github/python-docs-samples/vision/cloud-client/product_search/product_in_product_set_management_test.py", line 88, in test_purge_products_in_product_set PROJECT_ID, LOCATION, PRODUCT_SET_ID, force=True) File "/Users/yuhanliu/projects/dpe-github/python-docs-samples/vision/cloud-client/product_search/product_in_product_set_management.py", line 153, in purge_products_in_product_set if operation.done(): File "/Users/yuhanliu/.virtualenvs/vision-ga/lib/python2.7/site-packages/google/api_core/operation.py", line 162, in done self._refresh_and_update() File "/Users/yuhanliu/.virtualenvs/vision-ga/lib/python2.7/site-packages/google/api_core/operation.py", line 154, in _refresh_and_update self._set_result_from_operation() File "/Users/yuhanliu/.virtualenvs/vision-ga/lib/python2.7/site-packages/google/api_core/operation.py", line 131, in _set_result_from_operation self._result_type, self._operation.response File "/Users/yuhanliu/.virtualenvs/vision-ga/lib/python2.7/site-packages/google/api_core/protobuf_helpers.py", line 69, in from_any_pb any_pb.__class__.__name__, pb_type.__name__ TypeError: Could not convert Any to Empty ```
1.0
Vision: Operation returned from 'purge_product: raises TypeError in 'done'. - #### Environment details 1. Vision API, Product Search, Purge Product 2. OSX 3. 2.7.14 4. google-cloud-vision 0.39.0 #### Steps to reproduce 1. Run tests or either of the purge product samples from this PR: https://github.com/GoogleCloudPlatform/python-docs-samples/pull/2349 1. The LRO fails at `operation.done()` #### Code example see PR linked above. #### Stack trace ``` Traceback (most recent call last): File "/Users/yuhanliu/projects/dpe-github/python-docs-samples/vision/cloud-client/product_search/product_in_product_set_management_test.py", line 88, in test_purge_products_in_product_set PROJECT_ID, LOCATION, PRODUCT_SET_ID, force=True) File "/Users/yuhanliu/projects/dpe-github/python-docs-samples/vision/cloud-client/product_search/product_in_product_set_management.py", line 153, in purge_products_in_product_set if operation.done(): File "/Users/yuhanliu/.virtualenvs/vision-ga/lib/python2.7/site-packages/google/api_core/operation.py", line 162, in done self._refresh_and_update() File "/Users/yuhanliu/.virtualenvs/vision-ga/lib/python2.7/site-packages/google/api_core/operation.py", line 154, in _refresh_and_update self._set_result_from_operation() File "/Users/yuhanliu/.virtualenvs/vision-ga/lib/python2.7/site-packages/google/api_core/operation.py", line 131, in _set_result_from_operation self._result_type, self._operation.response File "/Users/yuhanliu/.virtualenvs/vision-ga/lib/python2.7/site-packages/google/api_core/protobuf_helpers.py", line 69, in from_any_pb any_pb.__class__.__name__, pb_type.__name__ TypeError: Could not convert Any to Empty ```
non_process
vision operation returned from purge product raises typeerror in done environment details vision api product search purge product osx google cloud vision steps to reproduce run tests or either of the purge product samples from this pr the lro fails at operation done code example see pr linked above stack trace traceback most recent call last file users yuhanliu projects dpe github python docs samples vision cloud client product search product in product set management test py line in test purge products in product set project id location product set id force true file users yuhanliu projects dpe github python docs samples vision cloud client product search product in product set management py line in purge products in product set if operation done file users yuhanliu virtualenvs vision ga lib site packages google api core operation py line in done self refresh and update file users yuhanliu virtualenvs vision ga lib site packages google api core operation py line in refresh and update self set result from operation file users yuhanliu virtualenvs vision ga lib site packages google api core operation py line in set result from operation self result type self operation response file users yuhanliu virtualenvs vision ga lib site packages google api core protobuf helpers py line in from any pb any pb class name pb type name typeerror could not convert any to empty
0
7,557
10,678,541,611
IssuesEvent
2019-10-21 17:29:06
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
opened
batch mode: cannot autofill "column" if if is an raster input
Bug High Priority Processing Regression
QGIS 3.4.12 on Ubuntu: Open a Processing tool (native or else) that uses an input raster layer. Fill the first input with a raster taken from the project or file system Add new lines to the batch mode. Try to fill the new lines by double clicking the input parameter name/column > do not work. Does not happen with vector inputs.
1.0
batch mode: cannot autofill "column" if if is an raster input - QGIS 3.4.12 on Ubuntu: Open a Processing tool (native or else) that uses an input raster layer. Fill the first input with a raster taken from the project or file system Add new lines to the batch mode. Try to fill the new lines by double clicking the input parameter name/column > do not work. Does not happen with vector inputs.
process
batch mode cannot autofill column if if is an raster input qgis on ubuntu open a processing tool native or else that uses an input raster layer fill the first input with a raster taken from the project or file system add new lines to the batch mode try to fill the new lines by double clicking the input parameter name column do not work does not happen with vector inputs
1
349,420
31,801,062,987
IssuesEvent
2023-09-13 11:08:31
wazuh/wazuh
https://api.github.com/repos/wazuh/wazuh
closed
Release 4.6.0 - Alpha 1 - E2E UX tests - Configuration assessment
type/test level/subtask
## End-to-End (E2E) Testing Guideline - **Documentation:** Always consult the development documentation for the current stage tag at [this link](https://documentation-dev.wazuh.com/v4.6.0-alpha1/index.html). Be careful because some of the description steps might refer to a current version in production, always navigate using the current development documention for the stage under test. - **Test Requirements:** Ensure your test comprehensively includes a full stack and agent/s deployment as per the Deployment requirements, detailing the machine OS, installed version, and revision. - **Deployment Options:** While deployments can be local (using VMs, Vagrant, or Docker) or on the aws-dev account, opt for local deployments when feasible. For AWS access, coordinate with the CICD team through [this link](https://github.com/wazuh/internal-devel-requests/issues/new?assignees=&labels=level%2Ftask%2C+request%2Foperational%2C+type%2Fchange&projects=&template=operational--request.md&title=%3CTitle%3E). - **External Accounts:** If tests require third-party accounts (e.g., GitHub, Azure, AWS, GCP), request the necessary access through the CICD team [here](https://github.com/wazuh/internal-devel-requests/issues/new?assignees=&labels=level%2Ftask%2C+request%2Foperational%2C+type%2Fchange&projects=&template=operational--request.md&title=%3CTitle%3E). - **Alerts:** Every test should generate a minimum of one end-to-end alert, from the agent to the dashboard, irrespective of test type. - **Multi-node Testing:** For multi-node wazuh-manager tests, ensure agents are connected to both workers and the master node. - **Package Verification:** Use the pre-release package that matches the current TAG you're testing. Confirm its version and revision. - **Filebeat Errors:** If you encounter errors with Filebeat during testing, refer to [this Slack discussion](https://wazuh-team.slack.com/archives/C03BDG0K6JC/p1672168163537809) for insights and resolutions. - **Known Issues:** Familiarize yourself with previously reported issues in the Known Issues section. This helps in identifying already recognized errors during testing. - **Reporting New Issues:** Any new errors discovered during testing that aren't listed under Known Issues should be reported. Communicate these to QA via the c-release Slack channel. - **Test Conduct:** It's imperative to be thorough in your testing, offering enough detail for reviewers. Incomplete tests might necessitate a redo. - **Documentation Feedback:** Encountering documentation gaps, unclear guidelines, or anything that disrupts the testing or UX? Open an issue, especially if it's not listed under Known Issues. - **Format:** If this is your first time doing this, refer to the format (but not necessarily the content, as it may vary) of previous E2E tests, here you have an example https://github.com/wazuh/wazuh/issues/13994. - **Status and completion:** Change the issue status within your team project accordingly. Once you finish testing and write the conclusions, move it to Pending review and notify the @wazuh/framework team via Slack using the [c-release channel](https://wazuh-team.slack.com/archives/C02A737S5MJ). Beware that the reviewers might request additional information or task repetitions. - **For reviewers:** Please move the issue to Pending final review and notify via Slack using the same thread if everything is ok, otherwise, perform an issue update with the requested changes and move it to On hold, increase the review_cycles in the team project by one and notify the issue assignee via Slack using the same thread. For the conclusions and the issue testing and updates, use the following legend: **Status legend** - 🟢 All checks passed - 🟡 Found a known issue - 🔴 Found a new error ## Deployment requirements | Component | Installation | Type | OS | |----------|--------------|------|----| | Indexer | [Installation assistant](https://documentation-dev.wazuh.com/v4.6.0-alpha1/installation-guide/wazuh-indexer/installation-assistant.html) | Multi node | Fedora 38 x86_64 | | Server | [Installation assistant](https://documentation-dev.wazuh.com/v4.6.0-alpha1/installation-guide/wazuh-indexer/installation-assistant.html) | Single node | Fedora 38 x86_64 | | Dashboard | [Installation assistant](https://documentation-dev.wazuh.com/v4.6.0-alpha1/installation-guide/wazuh-indexer/installation-assistant.html) | - | Fedora 38 x86_64 | | Agent | [Wazuh WUI one-liner deploy using FQDN](https://documentation-dev.wazuh.com/v4.6.0-alpha1/_images/deploy-new-agent-from-ui1.png) | - | Windows server 2016 x86_64, Debian 10 x86_64, Fedora 38 x86_64, macOS Ventura arm| ## Test description For the selected Wazuh Agent OS: - Check that their respective SCA policies are applied properly - Use and activate a custom policy for any of them - Disable an used policy and confirm it is not used anymore - Push SCA config through centralized config and check it applies properly ## Known issues - https://github.com/wazuh/wazuh-packages/issues/1422 - https://github.com/wazuh/wazuh/issues/12347 ## Conclusions Summarize the errors detected (Known Issues included). Illustrate using the table below, removing current examples: | **Status** | **Test** | **Failure type** | **Notes** | |----------------|-------------|---------------------|----------------| | 🟡 | macOS Ventura agent installation with Wazuh WUI one-liner deploy using FQDN | The `WAZUH_MANAGER` variable did not work on the installation | #18933 | | :yellow_circle: | Push SCA config through centralized config and check it applies properly with Fedora | Fedora operating system does not have an SCA policy | #18306 | | :yellow_circle: | Push SCA config through centralized config and check it applies properly with Windows Server 2016 | Failure when attempting to create an SCA policy with Windows Server 2016 | #18981 | ## Feedback We value your feedback. Please provide insights on your testing experience. - Was the testing guideline clear? Were there any ambiguities? Everything was clear and without any ambiguity - Did you face any challenges not covered by the guideline? No - Suggestions for improvement: ## Reviewers validation The criteria for completing this task is based on the validation of the conclusions and the test results by all reviewers. All the checkboxes below must be marked in order to close this issue. - [ ] @davidjiglesias - [ ] @wazuh/security-warriors
1.0
Release 4.6.0 - Alpha 1 - E2E UX tests - Configuration assessment - ## End-to-End (E2E) Testing Guideline - **Documentation:** Always consult the development documentation for the current stage tag at [this link](https://documentation-dev.wazuh.com/v4.6.0-alpha1/index.html). Be careful because some of the description steps might refer to a current version in production, always navigate using the current development documention for the stage under test. - **Test Requirements:** Ensure your test comprehensively includes a full stack and agent/s deployment as per the Deployment requirements, detailing the machine OS, installed version, and revision. - **Deployment Options:** While deployments can be local (using VMs, Vagrant, or Docker) or on the aws-dev account, opt for local deployments when feasible. For AWS access, coordinate with the CICD team through [this link](https://github.com/wazuh/internal-devel-requests/issues/new?assignees=&labels=level%2Ftask%2C+request%2Foperational%2C+type%2Fchange&projects=&template=operational--request.md&title=%3CTitle%3E). - **External Accounts:** If tests require third-party accounts (e.g., GitHub, Azure, AWS, GCP), request the necessary access through the CICD team [here](https://github.com/wazuh/internal-devel-requests/issues/new?assignees=&labels=level%2Ftask%2C+request%2Foperational%2C+type%2Fchange&projects=&template=operational--request.md&title=%3CTitle%3E). - **Alerts:** Every test should generate a minimum of one end-to-end alert, from the agent to the dashboard, irrespective of test type. - **Multi-node Testing:** For multi-node wazuh-manager tests, ensure agents are connected to both workers and the master node. - **Package Verification:** Use the pre-release package that matches the current TAG you're testing. Confirm its version and revision. - **Filebeat Errors:** If you encounter errors with Filebeat during testing, refer to [this Slack discussion](https://wazuh-team.slack.com/archives/C03BDG0K6JC/p1672168163537809) for insights and resolutions. - **Known Issues:** Familiarize yourself with previously reported issues in the Known Issues section. This helps in identifying already recognized errors during testing. - **Reporting New Issues:** Any new errors discovered during testing that aren't listed under Known Issues should be reported. Communicate these to QA via the c-release Slack channel. - **Test Conduct:** It's imperative to be thorough in your testing, offering enough detail for reviewers. Incomplete tests might necessitate a redo. - **Documentation Feedback:** Encountering documentation gaps, unclear guidelines, or anything that disrupts the testing or UX? Open an issue, especially if it's not listed under Known Issues. - **Format:** If this is your first time doing this, refer to the format (but not necessarily the content, as it may vary) of previous E2E tests, here you have an example https://github.com/wazuh/wazuh/issues/13994. - **Status and completion:** Change the issue status within your team project accordingly. Once you finish testing and write the conclusions, move it to Pending review and notify the @wazuh/framework team via Slack using the [c-release channel](https://wazuh-team.slack.com/archives/C02A737S5MJ). Beware that the reviewers might request additional information or task repetitions. - **For reviewers:** Please move the issue to Pending final review and notify via Slack using the same thread if everything is ok, otherwise, perform an issue update with the requested changes and move it to On hold, increase the review_cycles in the team project by one and notify the issue assignee via Slack using the same thread. For the conclusions and the issue testing and updates, use the following legend: **Status legend** - 🟢 All checks passed - 🟡 Found a known issue - 🔴 Found a new error ## Deployment requirements | Component | Installation | Type | OS | |----------|--------------|------|----| | Indexer | [Installation assistant](https://documentation-dev.wazuh.com/v4.6.0-alpha1/installation-guide/wazuh-indexer/installation-assistant.html) | Multi node | Fedora 38 x86_64 | | Server | [Installation assistant](https://documentation-dev.wazuh.com/v4.6.0-alpha1/installation-guide/wazuh-indexer/installation-assistant.html) | Single node | Fedora 38 x86_64 | | Dashboard | [Installation assistant](https://documentation-dev.wazuh.com/v4.6.0-alpha1/installation-guide/wazuh-indexer/installation-assistant.html) | - | Fedora 38 x86_64 | | Agent | [Wazuh WUI one-liner deploy using FQDN](https://documentation-dev.wazuh.com/v4.6.0-alpha1/_images/deploy-new-agent-from-ui1.png) | - | Windows server 2016 x86_64, Debian 10 x86_64, Fedora 38 x86_64, macOS Ventura arm| ## Test description For the selected Wazuh Agent OS: - Check that their respective SCA policies are applied properly - Use and activate a custom policy for any of them - Disable an used policy and confirm it is not used anymore - Push SCA config through centralized config and check it applies properly ## Known issues - https://github.com/wazuh/wazuh-packages/issues/1422 - https://github.com/wazuh/wazuh/issues/12347 ## Conclusions Summarize the errors detected (Known Issues included). Illustrate using the table below, removing current examples: | **Status** | **Test** | **Failure type** | **Notes** | |----------------|-------------|---------------------|----------------| | 🟡 | macOS Ventura agent installation with Wazuh WUI one-liner deploy using FQDN | The `WAZUH_MANAGER` variable did not work on the installation | #18933 | | :yellow_circle: | Push SCA config through centralized config and check it applies properly with Fedora | Fedora operating system does not have an SCA policy | #18306 | | :yellow_circle: | Push SCA config through centralized config and check it applies properly with Windows Server 2016 | Failure when attempting to create an SCA policy with Windows Server 2016 | #18981 | ## Feedback We value your feedback. Please provide insights on your testing experience. - Was the testing guideline clear? Were there any ambiguities? Everything was clear and without any ambiguity - Did you face any challenges not covered by the guideline? No - Suggestions for improvement: ## Reviewers validation The criteria for completing this task is based on the validation of the conclusions and the test results by all reviewers. All the checkboxes below must be marked in order to close this issue. - [ ] @davidjiglesias - [ ] @wazuh/security-warriors
non_process
release alpha ux tests configuration assessment end to end testing guideline documentation always consult the development documentation for the current stage tag at be careful because some of the description steps might refer to a current version in production always navigate using the current development documention for the stage under test test requirements ensure your test comprehensively includes a full stack and agent s deployment as per the deployment requirements detailing the machine os installed version and revision deployment options while deployments can be local using vms vagrant or docker or on the aws dev account opt for local deployments when feasible for aws access coordinate with the cicd team through external accounts if tests require third party accounts e g github azure aws gcp request the necessary access through the cicd team alerts every test should generate a minimum of one end to end alert from the agent to the dashboard irrespective of test type multi node testing for multi node wazuh manager tests ensure agents are connected to both workers and the master node package verification use the pre release package that matches the current tag you re testing confirm its version and revision filebeat errors if you encounter errors with filebeat during testing refer to for insights and resolutions known issues familiarize yourself with previously reported issues in the known issues section this helps in identifying already recognized errors during testing reporting new issues any new errors discovered during testing that aren t listed under known issues should be reported communicate these to qa via the c release slack channel test conduct it s imperative to be thorough in your testing offering enough detail for reviewers incomplete tests might necessitate a redo documentation feedback encountering documentation gaps unclear guidelines or anything that disrupts the testing or ux open an issue especially if it s not listed under known issues format if this is your first time doing this refer to the format but not necessarily the content as it may vary of previous tests here you have an example status and completion change the issue status within your team project accordingly once you finish testing and write the conclusions move it to pending review and notify the wazuh framework team via slack using the beware that the reviewers might request additional information or task repetitions for reviewers please move the issue to pending final review and notify via slack using the same thread if everything is ok otherwise perform an issue update with the requested changes and move it to on hold increase the review cycles in the team project by one and notify the issue assignee via slack using the same thread for the conclusions and the issue testing and updates use the following legend status legend 🟢 all checks passed 🟡 found a known issue 🔴 found a new error deployment requirements component installation type os indexer multi node fedora server single node fedora dashboard fedora agent windows server debian fedora macos ventura arm test description for the selected wazuh agent os check that their respective sca policies are applied properly use and activate a custom policy for any of them disable an used policy and confirm it is not used anymore push sca config through centralized config and check it applies properly known issues conclusions summarize the errors detected known issues included illustrate using the table below removing current examples status test failure type notes 🟡 macos ventura agent installation with wazuh wui one liner deploy using fqdn the wazuh manager variable did not work on the installation yellow circle push sca config through centralized config and check it applies properly with fedora fedora operating system does not have an sca policy yellow circle push sca config through centralized config and check it applies properly with windows server failure when attempting to create an sca policy with windows server feedback we value your feedback please provide insights on your testing experience was the testing guideline clear were there any ambiguities everything was clear and without any ambiguity did you face any challenges not covered by the guideline no suggestions for improvement reviewers validation the criteria for completing this task is based on the validation of the conclusions and the test results by all reviewers all the checkboxes below must be marked in order to close this issue davidjiglesias wazuh security warriors
0
3,466
6,546,017,393
IssuesEvent
2017-09-04 08:32:18
threefoldfoundation/app_backend
https://api.github.com/repos/threefoldfoundation/app_backend
closed
Make TFF app interface with IYO verification levels
priority_major process_duplicate type_feature
In IYO, there are different verification levels. Within TFF app, certain functionalities require different verification levels. TFF app should be able to check verification level of user and ask to initiate improved verification if needed. E.g., a hoster should be address verified (ID level 2).
1.0
Make TFF app interface with IYO verification levels - In IYO, there are different verification levels. Within TFF app, certain functionalities require different verification levels. TFF app should be able to check verification level of user and ask to initiate improved verification if needed. E.g., a hoster should be address verified (ID level 2).
process
make tff app interface with iyo verification levels in iyo there are different verification levels within tff app certain functionalities require different verification levels tff app should be able to check verification level of user and ask to initiate improved verification if needed e g a hoster should be address verified id level
1
79,157
9,843,015,484
IssuesEvent
2019-06-18 10:35:41
openSUSE/open-build-service
https://api.github.com/repos/openSUSE/open-build-service
closed
Workers part of /monitor suboptimal
Bootstrap :rocket: Design Frontend
I think the workers part of the monitor page deserves a little more love - the progress bars have rather dull colours and the font is too tiny to read (while it was enlarged everywhere else during bootstrap port). And it doesn't take all available space and shows a rather strange '#' for non building workers. I was quite confused what it's about before I switched back to old webui to verify. ![screencapture-build-opensuse-org-monitor-2019-06-12-18_28_01](https://user-images.githubusercontent.com/1067203/59369336-83ca3000-8d40-11e9-82ea-f7bf25c2f343.png) ![screencapture-build-opensuse-org-monitor-2019-06-12-18_29_48](https://user-images.githubusercontent.com/1067203/59369339-8462c680-8d40-11e9-8228-fc2254e31c92.png)
1.0
Workers part of /monitor suboptimal - I think the workers part of the monitor page deserves a little more love - the progress bars have rather dull colours and the font is too tiny to read (while it was enlarged everywhere else during bootstrap port). And it doesn't take all available space and shows a rather strange '#' for non building workers. I was quite confused what it's about before I switched back to old webui to verify. ![screencapture-build-opensuse-org-monitor-2019-06-12-18_28_01](https://user-images.githubusercontent.com/1067203/59369336-83ca3000-8d40-11e9-82ea-f7bf25c2f343.png) ![screencapture-build-opensuse-org-monitor-2019-06-12-18_29_48](https://user-images.githubusercontent.com/1067203/59369339-8462c680-8d40-11e9-8228-fc2254e31c92.png)
non_process
workers part of monitor suboptimal i think the workers part of the monitor page deserves a little more love the progress bars have rather dull colours and the font is too tiny to read while it was enlarged everywhere else during bootstrap port and it doesn t take all available space and shows a rather strange for non building workers i was quite confused what it s about before i switched back to old webui to verify
0
17,177
22,755,106,059
IssuesEvent
2022-07-07 15:54:08
hashgraph/hedera-json-rpc-relay
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
closed
Add acceptance test support for getLogs
enhancement limechain P2 process
### Problem #103 added support for getLogs. However , this preceded the acceptance tests so support was not added ### Solution Add a tests that calls `eth_getLogs` with coverage for a combination of the following params - fromBlock - toBlock - address - topics - blockhash ### Alternatives _No response_
1.0
Add acceptance test support for getLogs - ### Problem #103 added support for getLogs. However , this preceded the acceptance tests so support was not added ### Solution Add a tests that calls `eth_getLogs` with coverage for a combination of the following params - fromBlock - toBlock - address - topics - blockhash ### Alternatives _No response_
process
add acceptance test support for getlogs problem added support for getlogs however this preceded the acceptance tests so support was not added solution add a tests that calls eth getlogs with coverage for a combination of the following params fromblock toblock address topics blockhash alternatives no response
1
10,846
13,626,409,685
IssuesEvent
2020-09-24 10:59:15
eclipse/microprofile-config
https://api.github.com/repos/eclipse/microprofile-config
closed
Adopt Maven Formatter Plugin
process change :busts_in_silhouette:
We should probably use [Maven Formatter Plugin](https://github.com/revelc/formatter-maven-plugin) for consistent code styling and to avoid reformats on code changes.
1.0
Adopt Maven Formatter Plugin - We should probably use [Maven Formatter Plugin](https://github.com/revelc/formatter-maven-plugin) for consistent code styling and to avoid reformats on code changes.
process
adopt maven formatter plugin we should probably use for consistent code styling and to avoid reformats on code changes
1
287,763
21,672,886,658
IssuesEvent
2022-05-08 08:34:22
BastienFaivre/ExchangeHub
https://api.github.com/repos/BastienFaivre/ExchangeHub
opened
Write text about user feedback
documentation
As stated on the submission page: _Text files (gladly with pictures) on documented user feedback and how it was acted upon_
1.0
Write text about user feedback - As stated on the submission page: _Text files (gladly with pictures) on documented user feedback and how it was acted upon_
non_process
write text about user feedback as stated on the submission page text files gladly with pictures on documented user feedback and how it was acted upon
0
3,467
6,548,910,711
IssuesEvent
2017-09-05 02:49:32
Great-Hill-Corporation/quickBlocks
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
closed
ethprice initialization
apps-ethPrice status-inprocess type-bug
Remove ~/.quickBlocks/cache/prices folder (i.e. brand new user) Run test cases for ethprice (they all fail)
1.0
ethprice initialization - Remove ~/.quickBlocks/cache/prices folder (i.e. brand new user) Run test cases for ethprice (they all fail)
process
ethprice initialization remove quickblocks cache prices folder i e brand new user run test cases for ethprice they all fail
1
7,159
10,308,215,477
IssuesEvent
2019-08-29 10:24:07
natario1/CameraView
https://api.github.com/repos/natario1/CameraView
closed
E/Frame: Frame is dead! time: -1 lastTime: 1565808924463
about:frame processing is:bug status:has pr
### Describe the bug When i use frame processor in ML Kit Face detection.. in long time, always frame is dead.. is there a listener when frame is dead? - CameraView version: *2.0.0-rc2r* - Reproducible in official demo app: *yes/no* (i dont understand this) - Device / Android version: *Redmi note 4, Nougat* ### Expected behavior - I want a listener when frame is dead - i want the frame not dead. because, my tracking in ml kit is stoping for about 5 second and then works again. ### Logs E/Frame: Frame is dead! time: -1 lastTime: 1565811012759 W/CameraCallbacks: dispatchFrame: Error during processor implementation. Can happen when camera is closed while processors are running. java.lang.RuntimeException: You should not access a released frame. If this frame was passed to a FrameProcessor, you can only use its contents synchronously,for the duration of the process() method.
1.0
E/Frame: Frame is dead! time: -1 lastTime: 1565808924463 - ### Describe the bug When i use frame processor in ML Kit Face detection.. in long time, always frame is dead.. is there a listener when frame is dead? - CameraView version: *2.0.0-rc2r* - Reproducible in official demo app: *yes/no* (i dont understand this) - Device / Android version: *Redmi note 4, Nougat* ### Expected behavior - I want a listener when frame is dead - i want the frame not dead. because, my tracking in ml kit is stoping for about 5 second and then works again. ### Logs E/Frame: Frame is dead! time: -1 lastTime: 1565811012759 W/CameraCallbacks: dispatchFrame: Error during processor implementation. Can happen when camera is closed while processors are running. java.lang.RuntimeException: You should not access a released frame. If this frame was passed to a FrameProcessor, you can only use its contents synchronously,for the duration of the process() method.
process
e frame frame is dead time lasttime describe the bug when i use frame processor in ml kit face detection in long time always frame is dead is there a listener when frame is dead cameraview version reproducible in official demo app yes no i dont understand this device android version redmi note nougat expected behavior i want a listener when frame is dead i want the frame not dead because my tracking in ml kit is stoping for about second and then works again logs e frame frame is dead time lasttime w cameracallbacks dispatchframe error during processor implementation can happen when camera is closed while processors are running java lang runtimeexception you should not access a released frame if this frame was passed to a frameprocessor you can only use its contents synchronously for the duration of the process method
1
79,499
15,586,159,350
IssuesEvent
2021-03-18 01:18:26
jrshutske/unit-conversion-api
https://api.github.com/repos/jrshutske/unit-conversion-api
opened
CVE-2020-35491 (High) detected in jackson-databind-2.9.8.jar
security vulnerability
## CVE-2020-35491 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /unit-conversion-api/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-actuator-2.1.2.RELEASE.jar (Root Library) - spring-boot-actuator-autoconfigure-2.1.2.RELEASE.jar - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.datasources.SharedPoolDataSource. <p>Publish Date: 2020-12-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35491>CVE-2020-35491</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2986">https://github.com/FasterXML/jackson-databind/issues/2986</a></p> <p>Release Date: 2020-12-17</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-35491 (High) detected in jackson-databind-2.9.8.jar - ## CVE-2020-35491 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /unit-conversion-api/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-actuator-2.1.2.RELEASE.jar (Root Library) - spring-boot-actuator-autoconfigure-2.1.2.RELEASE.jar - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.datasources.SharedPoolDataSource. <p>Publish Date: 2020-12-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35491>CVE-2020-35491</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2986">https://github.com/FasterXML/jackson-databind/issues/2986</a></p> <p>Release Date: 2020-12-17</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file unit conversion api pom xml path to vulnerable library root repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter actuator release jar root library spring boot actuator autoconfigure release jar x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache commons datasources sharedpooldatasource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource
0
135,586
30,318,680,793
IssuesEvent
2023-07-10 17:27:15
bevyengine/bevy
https://api.github.com/repos/bevyengine/bevy
closed
Typo in bevy_reflect_derive
D-Good-First-Issue C-Code-Quality A-Reflection
`bevy_reflect_derive` has a little [typo](https://github.com/ClayenKitten/bevy/blob/1b51053f1958d750cdddffb99828bd0aacb61a78/crates/bevy_reflect/bevy_reflect_derive/src/type_path.rs#L76) in `NamedTypePathDef` enum variant. It is currently named `Primtive` and, obviously, should be changed to `Primitive`. Usage of the enum is tightly coupled with macro, where I don't have much experience, so I decided to leave it for more experienced contributor just in case.
1.0
Typo in bevy_reflect_derive - `bevy_reflect_derive` has a little [typo](https://github.com/ClayenKitten/bevy/blob/1b51053f1958d750cdddffb99828bd0aacb61a78/crates/bevy_reflect/bevy_reflect_derive/src/type_path.rs#L76) in `NamedTypePathDef` enum variant. It is currently named `Primtive` and, obviously, should be changed to `Primitive`. Usage of the enum is tightly coupled with macro, where I don't have much experience, so I decided to leave it for more experienced contributor just in case.
non_process
typo in bevy reflect derive bevy reflect derive has a little in namedtypepathdef enum variant it is currently named primtive and obviously should be changed to primitive usage of the enum is tightly coupled with macro where i don t have much experience so i decided to leave it for more experienced contributor just in case
0
4,625
7,468,800,826
IssuesEvent
2018-04-02 20:18:43
w3c/w3process
https://api.github.com/repos/w3c/w3process
closed
Can't update process as if it were a Rec
Process2019Candidate
There is no call for implementations, etc, so offering the Rec track as an alternative path makes no sense.
1.0
Can't update process as if it were a Rec - There is no call for implementations, etc, so offering the Rec track as an alternative path makes no sense.
process
can t update process as if it were a rec there is no call for implementations etc so offering the rec track as an alternative path makes no sense
1
15,365
19,538,432,749
IssuesEvent
2021-12-31 13:30:23
fadeoutsoftware/WASDI
https://api.github.com/repos/fadeoutsoftware/WASDI
closed
Automatic Autoban -> interface -> include agricolture areas
bug app / processor
There is an error using that flag. Check and solve.
1.0
Automatic Autoban -> interface -> include agricolture areas - There is an error using that flag. Check and solve.
process
automatic autoban interface include agricolture areas there is an error using that flag check and solve
1
80,616
15,586,301,994
IssuesEvent
2021-03-18 01:37:57
benchmarkdebricked/spring-boot
https://api.github.com/repos/benchmarkdebricked/spring-boot
opened
CVE-2020-27782 (High) detected in undertow-core-2.0.20.Final.jar
security vulnerability
## CVE-2020-27782 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>undertow-core-2.0.20.Final.jar</b></p></summary> <p>Undertow</p> <p>Library home page: <a href="http://www.jboss.org/">http://www.jboss.org/</a></p> <p>Path to dependency file: /spring-boot/spring-boot-project/spring-boot/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar,/root/.m2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar,/root/.m2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar,2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar,/root/.m2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar,/root/.m2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar</p> <p> Dependency Hierarchy: - undertow-servlet-2.0.20.Final.jar (Root Library) - :x: **undertow-core-2.0.20.Final.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in the Undertow AJP connector. Malicious requests and abrupt connection closes could be triggered by an attacker using query strings with non-RFC compliant characters resulting in a denial of service. The highest threat from this vulnerability is to system availability. This affects Undertow 2.1.5.SP1, 2.0.33.SP2, and 2.2.3.SP1. <p>Publish Date: 2021-02-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27782>CVE-2020-27782</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-27782">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-27782</a></p> <p>Release Date: 2021-02-23</p> <p>Fix Resolution: io.undertow:undertow-core:2.2.4.Final</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-27782 (High) detected in undertow-core-2.0.20.Final.jar - ## CVE-2020-27782 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>undertow-core-2.0.20.Final.jar</b></p></summary> <p>Undertow</p> <p>Library home page: <a href="http://www.jboss.org/">http://www.jboss.org/</a></p> <p>Path to dependency file: /spring-boot/spring-boot-project/spring-boot/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar,/root/.m2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar,/root/.m2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar,2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar,/root/.m2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar,/root/.m2/repository/io/undertow/undertow-core/2.0.20.Final/undertow-core-2.0.20.Final.jar</p> <p> Dependency Hierarchy: - undertow-servlet-2.0.20.Final.jar (Root Library) - :x: **undertow-core-2.0.20.Final.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in the Undertow AJP connector. Malicious requests and abrupt connection closes could be triggered by an attacker using query strings with non-RFC compliant characters resulting in a denial of service. The highest threat from this vulnerability is to system availability. This affects Undertow 2.1.5.SP1, 2.0.33.SP2, and 2.2.3.SP1. <p>Publish Date: 2021-02-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27782>CVE-2020-27782</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-27782">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-27782</a></p> <p>Release Date: 2021-02-23</p> <p>Fix Resolution: io.undertow:undertow-core:2.2.4.Final</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in undertow core final jar cve high severity vulnerability vulnerable library undertow core final jar undertow library home page a href path to dependency file spring boot spring boot project spring boot pom xml path to vulnerable library root repository io undertow undertow core final undertow core final jar root repository io undertow undertow core final undertow core final jar root repository io undertow undertow core final undertow core final jar repository io undertow undertow core final undertow core final jar root repository io undertow undertow core final undertow core final jar root repository io undertow undertow core final undertow core final jar dependency hierarchy undertow servlet final jar root library x undertow core final jar vulnerable library vulnerability details a flaw was found in the undertow ajp connector malicious requests and abrupt connection closes could be triggered by an attacker using query strings with non rfc compliant characters resulting in a denial of service the highest threat from this vulnerability is to system availability this affects undertow and publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io undertow undertow core final step up your open source security game with whitesource
0
12,424
14,924,138,177
IssuesEvent
2021-01-23 22:13:53
eddieantonio/predictive-text-studio
https://api.github.com/repos/eddieantonio/predictive-text-studio
closed
Remove headers from uploaded wordlists
data-backing data-processing good first issue 🔥 High priority
Remove headers such as "word" and "count" headers in the first row of all uploaded wordlists (either Google Sheets or Excel). This should be done automatically.
1.0
Remove headers from uploaded wordlists - Remove headers such as "word" and "count" headers in the first row of all uploaded wordlists (either Google Sheets or Excel). This should be done automatically.
process
remove headers from uploaded wordlists remove headers such as word and count headers in the first row of all uploaded wordlists either google sheets or excel this should be done automatically
1
74,823
14,346,736,069
IssuesEvent
2020-11-29 02:31:00
fps-game-jam-org/game-off-2020
https://api.github.com/repos/fps-game-jam-org/game-off-2020
closed
Make the SceneChanger make its GameObject be a singleton
code fix
Right now, if you change into a scene that has a Scene Changer GameObject, you'll keep your old Scene Changer and get the new one. This is not useful. There should be exactly 1 Scene Changer in the scene at any time.
1.0
Make the SceneChanger make its GameObject be a singleton - Right now, if you change into a scene that has a Scene Changer GameObject, you'll keep your old Scene Changer and get the new one. This is not useful. There should be exactly 1 Scene Changer in the scene at any time.
non_process
make the scenechanger make its gameobject be a singleton right now if you change into a scene that has a scene changer gameobject you ll keep your old scene changer and get the new one this is not useful there should be exactly scene changer in the scene at any time
0
21,001
27,868,901,112
IssuesEvent
2023-03-21 12:13:55
Living-with-machines/MapReader
https://api.github.com/repos/Living-with-machines/MapReader
closed
MapReader and Binder
maps post-processing
@ChristinaLast What do you think if we have more than one Binder link? Does it make sense to have: - [ ] Example for Maps - [x] Example for Plant images For the latter, we have a link now: https://mybinder.org/v2/gh/Living-with-machines/MapReader/main?labpath=examples%2Fquick_start%2Fquick_start.ipynb I just tested it, and it works, but I want to also add all the cells/codes of that notebook to CI. Related issue: https://github.com/Living-with-machines/MapReader/issues/28
1.0
MapReader and Binder - @ChristinaLast What do you think if we have more than one Binder link? Does it make sense to have: - [ ] Example for Maps - [x] Example for Plant images For the latter, we have a link now: https://mybinder.org/v2/gh/Living-with-machines/MapReader/main?labpath=examples%2Fquick_start%2Fquick_start.ipynb I just tested it, and it works, but I want to also add all the cells/codes of that notebook to CI. Related issue: https://github.com/Living-with-machines/MapReader/issues/28
process
mapreader and binder christinalast what do you think if we have more than one binder link does it make sense to have example for maps example for plant images for the latter we have a link now i just tested it and it works but i want to also add all the cells codes of that notebook to ci related issue
1
17,936
23,933,688,546
IssuesEvent
2022-09-10 23:19:59
GregTechCEu/gt-ideas
https://api.github.com/repos/GregTechCEu/gt-ideas
opened
Sea Water Purification
processing chain
## Details It should be no surprise that water obtained from the sea is full of impurities and should be purified before being used for industrial purposes. A process known as Multi-stage flash distillation is one of the most widely used methods of desalinating water (It is used to process 26% of today's desalinated water) ## Products Main Product: Water Side Product(s): Brine ## Steps Sea Water + Steam -> Water + Brine (Not sure how much of each liquid should be consumed or produced) ![image](https://user-images.githubusercontent.com/112270586/189504947-f6369cfd-3ddd-4bcf-bf1c-ee3f9e6472b6.png) ## Yield Idk ## Sources https://en.wikipedia.org/wiki/Multi-stage_flash_distillation
1.0
Sea Water Purification - ## Details It should be no surprise that water obtained from the sea is full of impurities and should be purified before being used for industrial purposes. A process known as Multi-stage flash distillation is one of the most widely used methods of desalinating water (It is used to process 26% of today's desalinated water) ## Products Main Product: Water Side Product(s): Brine ## Steps Sea Water + Steam -> Water + Brine (Not sure how much of each liquid should be consumed or produced) ![image](https://user-images.githubusercontent.com/112270586/189504947-f6369cfd-3ddd-4bcf-bf1c-ee3f9e6472b6.png) ## Yield Idk ## Sources https://en.wikipedia.org/wiki/Multi-stage_flash_distillation
process
sea water purification details it should be no surprise that water obtained from the sea is full of impurities and should be purified before being used for industrial purposes a process known as multi stage flash distillation is one of the most widely used methods of desalinating water it is used to process of today s desalinated water products main product water side product s brine steps sea water steam water brine not sure how much of each liquid should be consumed or produced yield idk sources
1
157,574
6,008,670,950
IssuesEvent
2017-06-06 08:31:55
GeekyAnts/NativeBase
https://api.github.com/repos/GeekyAnts/NativeBase
closed
How about the performance of nativebase?
1 high priority performance
The sideslip menu on the Android is obviously slow to switch, worrying about performance problems
1.0
How about the performance of nativebase? - The sideslip menu on the Android is obviously slow to switch, worrying about performance problems
non_process
how about the performance of nativebase the sideslip menu on the android is obviously slow to switch worrying about performance problems
0
11,449
14,270,367,545
IssuesEvent
2020-11-21 06:34:24
Altinn/altinn-studio
https://api.github.com/repos/Altinn/altinn-studio
opened
Implement functional signing support
area/process kind/user-story solution/app-backend solution/app-frontend ux
## Description Altinn 3 need to support a signing task. This signing task lets user "sign" parts of or all data submitted as part of an instance. ## Screenshots > Screenshots or links to Figma (make sure your sketch is public) ## Considerations ### Signing task ### Signing view ### Signature storage ### Signature API ## Acceptance criteria > Describe criteria here (i.e. What is allowed/not allowed (negative tesing), validations, error messages and warnings etc.) ## Specification tasks - [ ] Development tasks are defined - [ ] Test design / decide test need ## Development tasks > Add tasks here ## Definition of done Verify that this issue meets [DoD](https://confluence.brreg.no/display/T3KP/Definition+of+Done#DefinitionofDone-DoD%E2%80%93utvikling) (Only for project members) before closing. - [ ] Documentation is updated (if relevant) - [ ] Technical documentation (docs.altinn.studio) - [ ] User documentation (altinn.github.io/docs) - [ ] QA - [ ] Manual test is complete (if relevant) - [ ] Automated test is implemented (if relevant) - [ ] All tasks in this userstory are closed (i.e. remaining tasks are moved to other user stories or marked obsolete)
1.0
Implement functional signing support - ## Description Altinn 3 need to support a signing task. This signing task lets user "sign" parts of or all data submitted as part of an instance. ## Screenshots > Screenshots or links to Figma (make sure your sketch is public) ## Considerations ### Signing task ### Signing view ### Signature storage ### Signature API ## Acceptance criteria > Describe criteria here (i.e. What is allowed/not allowed (negative tesing), validations, error messages and warnings etc.) ## Specification tasks - [ ] Development tasks are defined - [ ] Test design / decide test need ## Development tasks > Add tasks here ## Definition of done Verify that this issue meets [DoD](https://confluence.brreg.no/display/T3KP/Definition+of+Done#DefinitionofDone-DoD%E2%80%93utvikling) (Only for project members) before closing. - [ ] Documentation is updated (if relevant) - [ ] Technical documentation (docs.altinn.studio) - [ ] User documentation (altinn.github.io/docs) - [ ] QA - [ ] Manual test is complete (if relevant) - [ ] Automated test is implemented (if relevant) - [ ] All tasks in this userstory are closed (i.e. remaining tasks are moved to other user stories or marked obsolete)
process
implement functional signing support description altinn need to support a signing task this signing task lets user sign parts of or all data submitted as part of an instance screenshots screenshots or links to figma make sure your sketch is public considerations signing task signing view signature storage signature api acceptance criteria describe criteria here i e what is allowed not allowed negative tesing validations error messages and warnings etc specification tasks development tasks are defined test design decide test need development tasks add tasks here definition of done verify that this issue meets only for project members before closing documentation is updated if relevant technical documentation docs altinn studio user documentation altinn github io docs qa manual test is complete if relevant automated test is implemented if relevant all tasks in this userstory are closed i e remaining tasks are moved to other user stories or marked obsolete
1
13,349
15,810,843,838
IssuesEvent
2021-04-05 00:12:59
darktable-org/darktable
https://api.github.com/repos/darktable-org/darktable
closed
Noise profiles are wrong
bug: wip difficulty: average no-issue-activity scope: image processing understood: clear
**Describe the bug** Noise profiles aim at having a constant variance of 1 after the generalized anscombe transform of denoiseprofiled, so that the denoising algorithms can work on gaussian noise of variance 1. Yet, if we check the variance obtained on a flat image at 100% zoom with the compute variance feature, I get variance values of about 14-15 for fujifilm xt3, and about 10 for lumix fz1000. The noiseprofiling tool assumes that noise is only fine grain, which is not true in practice. Also, it uses an input color profile of REC709, which (I guess) modifies the values, so the noiseprofiles are not made on the real data we get after demosaic. **To Reproduce** Steps to reproduce the behavior: 1. Shoot a defocused picture of a flat uniform image (e.g. https://artphotoheritage.files.wordpress.com/2011/06/gris-neutre.png ) at several ISO values. 2. In darktablerc, set plugins/darkroom/denoiseprofile/show_compute_variance_mode to "true" 3. Launch darktable 4. For each of the considered pictures: 4.1 Move denoiseprofile before colorin 4.2 Put denoiseprofile in "compute variance" mode 4.3 Zoom to 100% in the center of the image (to minimize the contribution of vignetting in the variance) 4.4 The displayed variance values are not close to 1... however, it looks like they are almost the same for all ISO values. **Expected behavior** At step 4.3, we should see variances close to 1. **Screenshots** ![Capture du 2019-03-30 14-57-34](https://user-images.githubusercontent.com/34063828/55277090-76360a00-52fc-11e9-8220-20d8364388c4.png) **How bad is it to have wrong profiles?** It is bad for making presets that works well on different cameras if the variance values we get differ from one camera to another. It does not change much in terms of denoising quality if the user uses the strength parameter to fit its need, as even with a wrong profile the generalized anscombe transform will succeed in stabilizing the variance. Let me go into further details: The generalized anscombe we do is (approximately) X -> 2\*sqrt(X\/a) The "a" number is supposed to be chosen such as a\*mean = variance V[ 2\*sqrt(X\/a) ] is approximately equal to: V [ 2\*sqrt(mean\/a) + 1\/(a\*sqrt(mean\/a)) \* (X-mean) ] (Taylor expansion of sqrt around the mean, see https://en.wikipedia.org/wiki/Taylor_expansions_for_the_moments_of_functions_of_random_variables ) which gives V[ 2\*sqrt(X\/a) ] is approximately equal to (a\/(a\*a\*mean))\*V[X] = 1\/(a\*mean)\*V[X] = 1 Thanks to the fact that a*mean = variance, we can get a variance of 1 after the transform. Now, what happens if we make mistakes and take a value which gives a\*mean=c\*variance? We can go through the same computations, but we will get V[X] = 1\/c at the end. So, the good new is that the variance remains constant! The question is, is the factor between the wrong profile and a true one (i.e. the number "c") always the same? Is our current noiseprofiling tool consistent in the errors it makes? If so, then it is not necessary to correct anything, we simply get profiles which allows to have a variance of 1\/c instead of 1, which is not problematic. However, if the number "c" changes from one camera to another, or from one ISO value to another, then we have an issue: the denoiseprofile will behave differently from one camera to another. From my first experiments, it _seems_ that the "c" does not change much from one ISO value to another, yet I get different values between my 2 cameras. What I would like to also check is if all bayer cameras have the same "c", and all xtrans cameras have the same "c". In that case, we would still be able to fix the issue without changing the noise profiling tool, and without having to redo all noise profiles. Otherwise, I guess the only option will be to do a new profiling tool, and to redo all noise profiles.
1.0
Noise profiles are wrong - **Describe the bug** Noise profiles aim at having a constant variance of 1 after the generalized anscombe transform of denoiseprofiled, so that the denoising algorithms can work on gaussian noise of variance 1. Yet, if we check the variance obtained on a flat image at 100% zoom with the compute variance feature, I get variance values of about 14-15 for fujifilm xt3, and about 10 for lumix fz1000. The noiseprofiling tool assumes that noise is only fine grain, which is not true in practice. Also, it uses an input color profile of REC709, which (I guess) modifies the values, so the noiseprofiles are not made on the real data we get after demosaic. **To Reproduce** Steps to reproduce the behavior: 1. Shoot a defocused picture of a flat uniform image (e.g. https://artphotoheritage.files.wordpress.com/2011/06/gris-neutre.png ) at several ISO values. 2. In darktablerc, set plugins/darkroom/denoiseprofile/show_compute_variance_mode to "true" 3. Launch darktable 4. For each of the considered pictures: 4.1 Move denoiseprofile before colorin 4.2 Put denoiseprofile in "compute variance" mode 4.3 Zoom to 100% in the center of the image (to minimize the contribution of vignetting in the variance) 4.4 The displayed variance values are not close to 1... however, it looks like they are almost the same for all ISO values. **Expected behavior** At step 4.3, we should see variances close to 1. **Screenshots** ![Capture du 2019-03-30 14-57-34](https://user-images.githubusercontent.com/34063828/55277090-76360a00-52fc-11e9-8220-20d8364388c4.png) **How bad is it to have wrong profiles?** It is bad for making presets that works well on different cameras if the variance values we get differ from one camera to another. It does not change much in terms of denoising quality if the user uses the strength parameter to fit its need, as even with a wrong profile the generalized anscombe transform will succeed in stabilizing the variance. Let me go into further details: The generalized anscombe we do is (approximately) X -> 2\*sqrt(X\/a) The "a" number is supposed to be chosen such as a\*mean = variance V[ 2\*sqrt(X\/a) ] is approximately equal to: V [ 2\*sqrt(mean\/a) + 1\/(a\*sqrt(mean\/a)) \* (X-mean) ] (Taylor expansion of sqrt around the mean, see https://en.wikipedia.org/wiki/Taylor_expansions_for_the_moments_of_functions_of_random_variables ) which gives V[ 2\*sqrt(X\/a) ] is approximately equal to (a\/(a\*a\*mean))\*V[X] = 1\/(a\*mean)\*V[X] = 1 Thanks to the fact that a*mean = variance, we can get a variance of 1 after the transform. Now, what happens if we make mistakes and take a value which gives a\*mean=c\*variance? We can go through the same computations, but we will get V[X] = 1\/c at the end. So, the good new is that the variance remains constant! The question is, is the factor between the wrong profile and a true one (i.e. the number "c") always the same? Is our current noiseprofiling tool consistent in the errors it makes? If so, then it is not necessary to correct anything, we simply get profiles which allows to have a variance of 1\/c instead of 1, which is not problematic. However, if the number "c" changes from one camera to another, or from one ISO value to another, then we have an issue: the denoiseprofile will behave differently from one camera to another. From my first experiments, it _seems_ that the "c" does not change much from one ISO value to another, yet I get different values between my 2 cameras. What I would like to also check is if all bayer cameras have the same "c", and all xtrans cameras have the same "c". In that case, we would still be able to fix the issue without changing the noise profiling tool, and without having to redo all noise profiles. Otherwise, I guess the only option will be to do a new profiling tool, and to redo all noise profiles.
process
noise profiles are wrong describe the bug noise profiles aim at having a constant variance of after the generalized anscombe transform of denoiseprofiled so that the denoising algorithms can work on gaussian noise of variance yet if we check the variance obtained on a flat image at zoom with the compute variance feature i get variance values of about for fujifilm and about for lumix the noiseprofiling tool assumes that noise is only fine grain which is not true in practice also it uses an input color profile of which i guess modifies the values so the noiseprofiles are not made on the real data we get after demosaic to reproduce steps to reproduce the behavior shoot a defocused picture of a flat uniform image e g at several iso values in darktablerc set plugins darkroom denoiseprofile show compute variance mode to true launch darktable for each of the considered pictures move denoiseprofile before colorin put denoiseprofile in compute variance mode zoom to in the center of the image to minimize the contribution of vignetting in the variance the displayed variance values are not close to however it looks like they are almost the same for all iso values expected behavior at step we should see variances close to screenshots how bad is it to have wrong profiles it is bad for making presets that works well on different cameras if the variance values we get differ from one camera to another it does not change much in terms of denoising quality if the user uses the strength parameter to fit its need as even with a wrong profile the generalized anscombe transform will succeed in stabilizing the variance let me go into further details the generalized anscombe we do is approximately x sqrt x a the a number is supposed to be chosen such as a mean variance v is approximately equal to v taylor expansion of sqrt around the mean see which gives v is approximately equal to a a a mean v a mean v thanks to the fact that a mean variance we can get a variance of after the transform now what happens if we make mistakes and take a value which gives a mean c variance we can go through the same computations but we will get v c at the end so the good new is that the variance remains constant the question is is the factor between the wrong profile and a true one i e the number c always the same is our current noiseprofiling tool consistent in the errors it makes if so then it is not necessary to correct anything we simply get profiles which allows to have a variance of c instead of which is not problematic however if the number c changes from one camera to another or from one iso value to another then we have an issue the denoiseprofile will behave differently from one camera to another from my first experiments it seems that the c does not change much from one iso value to another yet i get different values between my cameras what i would like to also check is if all bayer cameras have the same c and all xtrans cameras have the same c in that case we would still be able to fix the issue without changing the noise profiling tool and without having to redo all noise profiles otherwise i guess the only option will be to do a new profiling tool and to redo all noise profiles
1
15,048
18,762,689,258
IssuesEvent
2021-11-05 18:29:25
ORNL-AMO/AMO-Tools-Suite
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Suite
closed
Available Heat Algorithm
Needs Verification Process Heating
The calculations for calculating sensible heat of the fuel needs to be added to `GasCompositions::ProcessHeatPropertiesResults GasCompositions::getProcessHeatProperties` Don't need to use the separate method, but the math should be added from lines 154 - 161. replace `fuelTemp` with `ambientAirTempF`, do what you need to do to make it match the "style" of the rest of getProcessHeatProperties ``` double GasCompositions::calculateSensibleHeat(const double fuelTemp) { double specificHeatFuel = 0; for ( auto const & comp : gasses ) { specificHeatFuel += comp.second->compByWeight * (comp.second->specificHeat(520) / comp.second->molecularWeight); } return 1 * specificHeatFuel * (fuelTemp - 32); } ``` Whatever variable you use for sensible heat needs to be added to Line 131 ``` const double availableHeat = (100 * (WhateverSensibleHeat + heatValueFuel + preHeatedAirEff + preHeatedAirMoistureEff) - (H2OHeatContent + CO2HeatContent + N2HeatContent + O2HeatContent)) / (100 * heatValueFuel); ``` A few comments to make the code more understandable: I know I missed some that should be called out, so if there is anything you can think of to help make it more intuitive, please add. before ... line 47: temperature conversions line 54: calculate average Cp across temperature range of combustion air to flue gas line 84: something, not 100% sure what is right here, it is kind of used for several things line 111: calculate heat in combustion air moisture before whatever you add for sensible heat: calculate sensible heat before density: calculate final values
1.0
Available Heat Algorithm - The calculations for calculating sensible heat of the fuel needs to be added to `GasCompositions::ProcessHeatPropertiesResults GasCompositions::getProcessHeatProperties` Don't need to use the separate method, but the math should be added from lines 154 - 161. replace `fuelTemp` with `ambientAirTempF`, do what you need to do to make it match the "style" of the rest of getProcessHeatProperties ``` double GasCompositions::calculateSensibleHeat(const double fuelTemp) { double specificHeatFuel = 0; for ( auto const & comp : gasses ) { specificHeatFuel += comp.second->compByWeight * (comp.second->specificHeat(520) / comp.second->molecularWeight); } return 1 * specificHeatFuel * (fuelTemp - 32); } ``` Whatever variable you use for sensible heat needs to be added to Line 131 ``` const double availableHeat = (100 * (WhateverSensibleHeat + heatValueFuel + preHeatedAirEff + preHeatedAirMoistureEff) - (H2OHeatContent + CO2HeatContent + N2HeatContent + O2HeatContent)) / (100 * heatValueFuel); ``` A few comments to make the code more understandable: I know I missed some that should be called out, so if there is anything you can think of to help make it more intuitive, please add. before ... line 47: temperature conversions line 54: calculate average Cp across temperature range of combustion air to flue gas line 84: something, not 100% sure what is right here, it is kind of used for several things line 111: calculate heat in combustion air moisture before whatever you add for sensible heat: calculate sensible heat before density: calculate final values
process
available heat algorithm the calculations for calculating sensible heat of the fuel needs to be added to gascompositions processheatpropertiesresults gascompositions getprocessheatproperties don t need to use the separate method but the math should be added from lines replace fueltemp with ambientairtempf do what you need to do to make it match the style of the rest of getprocessheatproperties double gascompositions calculatesensibleheat const double fueltemp double specificheatfuel for auto const comp gasses specificheatfuel comp second compbyweight comp second specificheat comp second molecularweight return specificheatfuel fueltemp whatever variable you use for sensible heat needs to be added to line const double availableheat whateversensibleheat heatvaluefuel preheatedaireff preheatedairmoistureeff heatvaluefuel a few comments to make the code more understandable i know i missed some that should be called out so if there is anything you can think of to help make it more intuitive please add before line temperature conversions line calculate average cp across temperature range of combustion air to flue gas line something not sure what is right here it is kind of used for several things line calculate heat in combustion air moisture before whatever you add for sensible heat calculate sensible heat before density calculate final values
1
78,389
27,492,835,486
IssuesEvent
2023-03-04 20:49:08
DependencyTrack/dependency-track
https://api.github.com/repos/DependencyTrack/dependency-track
opened
javax.jdo.JDOUserException: Field org.dependencytrack.model.Component.repositoryMeta is not marked as persistent so cannot be queried
defect in triage
### Current Behavior When opening the Policy Violations tab, the PolicyViolationResource throws the following error and no violations are shown. I think this bug was introduced somewhere last week... `` javax.jdo.JDOUserException: Field org.dependencytrack.model.Component.repositoryMeta is not marked as persistent so cannot be queried at org.datanucleus.api.jdo.JDOAdapter.getJDOExceptionForNucleusException(JDOAdapter.java:698) at org.datanucleus.api.jdo.JDOQuery.executeInternal(JDOQuery.java:456) at org.datanucleus.api.jdo.JDOQuery.execute(JDOQuery.java:276) at alpine.persistence.AbstractAlpineQueryManager.execute(AbstractAlpineQueryManager.java:174) at org.dependencytrack.persistence.PolicyQueryManager.getPolicyViolations(PolicyQueryManager.java:280) at org.dependencytrack.persistence.QueryManager.getPolicyViolations(QueryManager.java:604) at org.dependencytrack.resources.v1.PolicyViolationResource.getViolationsByProject(PolicyViolationResource.java:102) ### Steps to Reproduce 1.Open the Policy Violations tab for a project in the frontend ### Expected Behavior No Exception is thrown and violations are shown (if any) ### Dependency-Track Version 4.8.0-SNAPSHOT ### Dependency-Track Distribution Executable WAR ### Database Server H2 ### Database Server Version _No response_ ### Browser Google Chrome ### Checklist - [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues) - [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported
1.0
javax.jdo.JDOUserException: Field org.dependencytrack.model.Component.repositoryMeta is not marked as persistent so cannot be queried - ### Current Behavior When opening the Policy Violations tab, the PolicyViolationResource throws the following error and no violations are shown. I think this bug was introduced somewhere last week... `` javax.jdo.JDOUserException: Field org.dependencytrack.model.Component.repositoryMeta is not marked as persistent so cannot be queried at org.datanucleus.api.jdo.JDOAdapter.getJDOExceptionForNucleusException(JDOAdapter.java:698) at org.datanucleus.api.jdo.JDOQuery.executeInternal(JDOQuery.java:456) at org.datanucleus.api.jdo.JDOQuery.execute(JDOQuery.java:276) at alpine.persistence.AbstractAlpineQueryManager.execute(AbstractAlpineQueryManager.java:174) at org.dependencytrack.persistence.PolicyQueryManager.getPolicyViolations(PolicyQueryManager.java:280) at org.dependencytrack.persistence.QueryManager.getPolicyViolations(QueryManager.java:604) at org.dependencytrack.resources.v1.PolicyViolationResource.getViolationsByProject(PolicyViolationResource.java:102) ### Steps to Reproduce 1.Open the Policy Violations tab for a project in the frontend ### Expected Behavior No Exception is thrown and violations are shown (if any) ### Dependency-Track Version 4.8.0-SNAPSHOT ### Dependency-Track Distribution Executable WAR ### Database Server H2 ### Database Server Version _No response_ ### Browser Google Chrome ### Checklist - [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues) - [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported
non_process
javax jdo jdouserexception field org dependencytrack model component repositorymeta is not marked as persistent so cannot be queried current behavior when opening the policy violations tab the policyviolationresource throws the following error and no violations are shown i think this bug was introduced somewhere last week javax jdo jdouserexception field org dependencytrack model component repositorymeta is not marked as persistent so cannot be queried at org datanucleus api jdo jdoadapter getjdoexceptionfornucleusexception jdoadapter java at org datanucleus api jdo jdoquery executeinternal jdoquery java at org datanucleus api jdo jdoquery execute jdoquery java at alpine persistence abstractalpinequerymanager execute abstractalpinequerymanager java at org dependencytrack persistence policyquerymanager getpolicyviolations policyquerymanager java at org dependencytrack persistence querymanager getpolicyviolations querymanager java at org dependencytrack resources policyviolationresource getviolationsbyproject policyviolationresource java steps to reproduce open the policy violations tab for a project in the frontend expected behavior no exception is thrown and violations are shown if any dependency track version snapshot dependency track distribution executable war database server database server version no response browser google chrome checklist i have read and understand the i have checked the for whether this defect was already reported
0
47,346
5,887,959,365
IssuesEvent
2017-05-17 08:55:58
EyeSeeTea/dhis2-android-datacapture
https://api.github.com/repos/EyeSeeTea/dhis2-android-datacapture
closed
Data elements on android do not appear in same order as web for datasets without sections
complexity - med (1-5hr) priority - medium testing type - bug
When entering data for a dataset WITHOUT sections, the data elements appear in a different order on the web version versus in the Android app. Attached screenshots show the first part of play demo dataset "Clinical Monitoring Checklist" as it appears on the web and on the tablet. This was tested on latest.mis-pis.org and play demo.
1.0
Data elements on android do not appear in same order as web for datasets without sections - When entering data for a dataset WITHOUT sections, the data elements appear in a different order on the web version versus in the Android app. Attached screenshots show the first part of play demo dataset "Clinical Monitoring Checklist" as it appears on the web and on the tablet. This was tested on latest.mis-pis.org and play demo.
non_process
data elements on android do not appear in same order as web for datasets without sections when entering data for a dataset without sections the data elements appear in a different order on the web version versus in the android app attached screenshots show the first part of play demo dataset clinical monitoring checklist as it appears on the web and on the tablet this was tested on latest mis pis org and play demo
0
809,324
30,186,560,695
IssuesEvent
2023-07-04 12:31:39
geneontology/go-annotation
https://api.github.com/repos/geneontology/go-annotation
closed
problem with PAINT family, or annotation transfer or both PANTHER:PTN000331677
PAINT annotation high priority
* **PTHR ID & PTN node:** see https://www.ebi.ac.uk/QuickGO/annotations?geneProductId=P0AEH3 This is transferring GO:0004343    glucosamine 6-phosphate N-acetyltransferase activity GO:0006048 UDP-N-acetylglucosamine biosynthetic process GO:0005793 endoplasmic reticulum-Golgi intermediate compartment from eukaryotes to bacteria * Add the label 'high priority' if needed. Generally high priority issues affect a lot of proteins, and annotations are incorrect (as opposed to just imprecise).
1.0
problem with PAINT family, or annotation transfer or both PANTHER:PTN000331677 - * **PTHR ID & PTN node:** see https://www.ebi.ac.uk/QuickGO/annotations?geneProductId=P0AEH3 This is transferring GO:0004343    glucosamine 6-phosphate N-acetyltransferase activity GO:0006048 UDP-N-acetylglucosamine biosynthetic process GO:0005793 endoplasmic reticulum-Golgi intermediate compartment from eukaryotes to bacteria * Add the label 'high priority' if needed. Generally high priority issues affect a lot of proteins, and annotations are incorrect (as opposed to just imprecise).
non_process
problem with paint family or annotation transfer or both panther pthr id ptn node see this is transferring go     glucosamine phosphate n acetyltransferase activity go udp n acetylglucosamine biosynthetic process go endoplasmic reticulum golgi intermediate compartment from eukaryotes to bacteria add the label high priority if needed generally high priority issues affect a lot of proteins and annotations are incorrect as opposed to just imprecise
0
20,919
27,755,793,271
IssuesEvent
2023-03-16 02:17:25
0xPolygonMiden/miden-vm
https://api.github.com/repos/0xPolygonMiden/miden-vm
opened
Replace MerkleSets in the advice provider with MerkleStore
processor
Now that we have [MerkleStore](https://github.com/0xPolygonMiden/crypto/blob/next/src/merkle/store.rs) implemented in `miden-crypto`, we should use it for instead of a map of `MerkleSet`'s in [MemAdviceProvider](https://github.com/0xPolygonMiden/miden-vm/blob/main/processor/src/advice/mem_provider.rs). We should also probable rename the fields like this: * `tape` -> `stack` * `values` -> `map` * `sets` -> store` So, the `MemAdviceProvider` struct could look like this: ```Rust pub struct MemAdviceProvider { step: u32, stack: Vec<Felt>, map: BTreeMap<[u8; 32], Vec<Felt>>, store: MerkleStore, } ``` We might also want to consider renaming `tape` into `stack` in the [AdviceProvider](https://github.com/0xPolygonMiden/miden-vm/blob/main/processor/src/advice/mod.rs#L52) trait as well.
1.0
Replace MerkleSets in the advice provider with MerkleStore - Now that we have [MerkleStore](https://github.com/0xPolygonMiden/crypto/blob/next/src/merkle/store.rs) implemented in `miden-crypto`, we should use it for instead of a map of `MerkleSet`'s in [MemAdviceProvider](https://github.com/0xPolygonMiden/miden-vm/blob/main/processor/src/advice/mem_provider.rs). We should also probable rename the fields like this: * `tape` -> `stack` * `values` -> `map` * `sets` -> store` So, the `MemAdviceProvider` struct could look like this: ```Rust pub struct MemAdviceProvider { step: u32, stack: Vec<Felt>, map: BTreeMap<[u8; 32], Vec<Felt>>, store: MerkleStore, } ``` We might also want to consider renaming `tape` into `stack` in the [AdviceProvider](https://github.com/0xPolygonMiden/miden-vm/blob/main/processor/src/advice/mod.rs#L52) trait as well.
process
replace merklesets in the advice provider with merklestore now that we have implemented in miden crypto we should use it for instead of a map of merkleset s in we should also probable rename the fields like this tape stack values map sets store so the memadviceprovider struct could look like this rust pub struct memadviceprovider step stack vec map btreemap store merklestore we might also want to consider renaming tape into stack in the trait as well
1
3,379
6,502,139,987
IssuesEvent
2017-08-23 12:37:24
syndesisio/syndesis-ui
https://api.github.com/repos/syndesisio/syndesis-ui
opened
Automate Documentation Builds
dev process
At the moment you have to manually generate docs with `compodoc` or with the CompoDoc server `compodoc -s`. We should hook this up to our CircleCI GH hook to build on push.
1.0
Automate Documentation Builds - At the moment you have to manually generate docs with `compodoc` or with the CompoDoc server `compodoc -s`. We should hook this up to our CircleCI GH hook to build on push.
process
automate documentation builds at the moment you have to manually generate docs with compodoc or with the compodoc server compodoc s we should hook this up to our circleci gh hook to build on push
1
361,303
10,707,058,880
IssuesEvent
2019-10-24 16:37:04
smartdevicelink/sdl_core
https://api.github.com/repos/smartdevicelink/sdl_core
closed
Remove unused ShowCustomForm HMI RPC
best practice low priority
### Bug Report [UI.ShowCustomForm](https://github.com/smartdevicelink/sdl_core/blob/master/src/components/interfaces/HMI_API.xml#L4392) is an unimplemented RPC in the HMI API, it should be removed. ##### Reproduction Steps N/A ##### OS & Version Information * SDL Core Version: https://github.com/smartdevicelink/sdl_core/commit/d36316738785c96dab2ee892762ed08c059fffde ##### Test Case, Sample Code, and / or Example App N/A
1.0
Remove unused ShowCustomForm HMI RPC - ### Bug Report [UI.ShowCustomForm](https://github.com/smartdevicelink/sdl_core/blob/master/src/components/interfaces/HMI_API.xml#L4392) is an unimplemented RPC in the HMI API, it should be removed. ##### Reproduction Steps N/A ##### OS & Version Information * SDL Core Version: https://github.com/smartdevicelink/sdl_core/commit/d36316738785c96dab2ee892762ed08c059fffde ##### Test Case, Sample Code, and / or Example App N/A
non_process
remove unused showcustomform hmi rpc bug report is an unimplemented rpc in the hmi api it should be removed reproduction steps n a os version information sdl core version test case sample code and or example app n a
0
241,475
18,457,647,298
IssuesEvent
2021-10-15 18:48:12
Ek-Z/course-aggregator
https://api.github.com/repos/Ek-Z/course-aggregator
closed
Используемые версии/ настроить в Openserver
documentation
PHP 7.4 MySQL 8.0 HTTP Apache_2.4-PHP_7.2-7.4 <img width="439" alt="Снимок экрана 2021-10-14 150849" src="https://user-images.githubusercontent.com/74703154/137315182-efa8a6ba-f86a-4f0b-8798-78ea409e9bb0.png">
1.0
Используемые версии/ настроить в Openserver - PHP 7.4 MySQL 8.0 HTTP Apache_2.4-PHP_7.2-7.4 <img width="439" alt="Снимок экрана 2021-10-14 150849" src="https://user-images.githubusercontent.com/74703154/137315182-efa8a6ba-f86a-4f0b-8798-78ea409e9bb0.png">
non_process
используемые версии настроить в openserver php mysql http apache php img width alt снимок экрана src
0
6,253
9,214,291,861
IssuesEvent
2019-03-10 18:44:13
chuminh712/BookStorage---Group-2
https://api.github.com/repos/chuminh712/BookStorage---Group-2
reopened
Detail Design
In Process
Design class diagram for Use Case Manage Goods Issue Design class diagram for Use Case Manage Goods Issue Info
1.0
Detail Design - Design class diagram for Use Case Manage Goods Issue Design class diagram for Use Case Manage Goods Issue Info
process
detail design design class diagram for use case manage goods issue design class diagram for use case manage goods issue info
1
129,882
5,105,189,749
IssuesEvent
2017-01-05 05:55:05
HuskieRobotics/roborioExpansion
https://api.github.com/repos/HuskieRobotics/roborioExpansion
closed
Don't clear entire buffer before each read
low-priority Propeller FW
```bytefill(@serialBuffer, 0, 256)``` This really just isn't needed, we can just add a null byte after our last read byte, and everything should stay functional. This will speed up those routines considerably.
1.0
Don't clear entire buffer before each read - ```bytefill(@serialBuffer, 0, 256)``` This really just isn't needed, we can just add a null byte after our last read byte, and everything should stay functional. This will speed up those routines considerably.
non_process
don t clear entire buffer before each read bytefill serialbuffer this really just isn t needed we can just add a null byte after our last read byte and everything should stay functional this will speed up those routines considerably
0
240,479
20,033,381,689
IssuesEvent
2022-02-02 09:17:26
keycloak/keycloak
https://api.github.com/repos/keycloak/keycloak
opened
Test more recent versions of Spring Boot
area/testsuite kind/enhancement area/adapter/spring
### Description We should test Spring Boot 2.4 and 2.6 in our pipeline in order to cover the whole test domain for Spring Boot and ensure the adapters work with those versions. Original issue: CIAM-1419 ### Discussion _No response_ ### Motivation _No response_ ### Details _No response_
1.0
Test more recent versions of Spring Boot - ### Description We should test Spring Boot 2.4 and 2.6 in our pipeline in order to cover the whole test domain for Spring Boot and ensure the adapters work with those versions. Original issue: CIAM-1419 ### Discussion _No response_ ### Motivation _No response_ ### Details _No response_
non_process
test more recent versions of spring boot description we should test spring boot and in our pipeline in order to cover the whole test domain for spring boot and ensure the adapters work with those versions original issue ciam discussion no response motivation no response details no response
0
109,104
16,829,413,042
IssuesEvent
2021-06-18 00:42:18
kevins01/CSPF-Founder
https://api.github.com/repos/kevins01/CSPF-Founder
opened
CVE-2019-14900 (Medium) detected in hibernate-core-4.0.1.Final.jar
security vulnerability
## CVE-2019-14900 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hibernate-core-4.0.1.Final.jar</b></p></summary> <p>A module of the Hibernate Core project</p> <p>Library home page: <a href="http://hibernate.org">http://hibernate.org</a></p> <p>Path to dependency file: CSPF-Founder/pom.xml</p> <p>Path to vulnerable library: CSPF-Founder/target/JavaVulnerableLab/WEB-INF/lib/hibernate-core-4.0.1.Final.jar,canner/.m2/repository/org/hibernate/hibernate-core/4.0.1.Final/hibernate-core-4.0.1.Final.jar</p> <p> Dependency Hierarchy: - :x: **hibernate-core-4.0.1.Final.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kevins01/CSPF-Founder/commit/c1676893423d574bf87c62c26df4f3d097d8fd51">c1676893423d574bf87c62c26df4f3d097d8fd51</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in Hibernate ORM in versions before 5.3.18, 5.4.18 and 5.5.0.Beta1. A SQL injection in the implementation of the JPA Criteria API can permit unsanitized literals when a literal is used in the SELECT or GROUP BY parts of the query. This flaw could allow an attacker to access unauthorized information or possibly conduct further attacks. <p>Publish Date: 2020-07-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14900>CVE-2019-14900</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14900">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14900</a></p> <p>Release Date: 2020-07-06</p> <p>Fix Resolution: org.hibernate:hibernate-core:5.4.18.Final</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-14900 (Medium) detected in hibernate-core-4.0.1.Final.jar - ## CVE-2019-14900 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hibernate-core-4.0.1.Final.jar</b></p></summary> <p>A module of the Hibernate Core project</p> <p>Library home page: <a href="http://hibernate.org">http://hibernate.org</a></p> <p>Path to dependency file: CSPF-Founder/pom.xml</p> <p>Path to vulnerable library: CSPF-Founder/target/JavaVulnerableLab/WEB-INF/lib/hibernate-core-4.0.1.Final.jar,canner/.m2/repository/org/hibernate/hibernate-core/4.0.1.Final/hibernate-core-4.0.1.Final.jar</p> <p> Dependency Hierarchy: - :x: **hibernate-core-4.0.1.Final.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kevins01/CSPF-Founder/commit/c1676893423d574bf87c62c26df4f3d097d8fd51">c1676893423d574bf87c62c26df4f3d097d8fd51</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in Hibernate ORM in versions before 5.3.18, 5.4.18 and 5.5.0.Beta1. A SQL injection in the implementation of the JPA Criteria API can permit unsanitized literals when a literal is used in the SELECT or GROUP BY parts of the query. This flaw could allow an attacker to access unauthorized information or possibly conduct further attacks. <p>Publish Date: 2020-07-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14900>CVE-2019-14900</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14900">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14900</a></p> <p>Release Date: 2020-07-06</p> <p>Fix Resolution: org.hibernate:hibernate-core:5.4.18.Final</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in hibernate core final jar cve medium severity vulnerability vulnerable library hibernate core final jar a module of the hibernate core project library home page a href path to dependency file cspf founder pom xml path to vulnerable library cspf founder target javavulnerablelab web inf lib hibernate core final jar canner repository org hibernate hibernate core final hibernate core final jar dependency hierarchy x hibernate core final jar vulnerable library found in head commit a href vulnerability details a flaw was found in hibernate orm in versions before and a sql injection in the implementation of the jpa criteria api can permit unsanitized literals when a literal is used in the select or group by parts of the query this flaw could allow an attacker to access unauthorized information or possibly conduct further attacks publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org hibernate hibernate core final step up your open source security game with whitesource
0
20,467
27,129,039,233
IssuesEvent
2023-02-16 08:24:40
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
`implies: ...` does not stack well with `static_link_msvcrt`
P4 type: support / not a bug (process) team-Rules-CPP stale
### Description of the problem / feature request: Hi. I can not enable `static_link_msvcrt` from another feature: ``` imply_static_link_msvcrt_feature = feature( name = "imply_static_link_msvcrt", implies = ["static_link_msvcrt"] ) ``` When I do: ``` cc_binary( name = "hello-world", srcs = ["hello-world.cpp"], features = ["imply_static_link_msvcrt"], linkopts = ["-DEFAULTLIB:libcmt.lib", "-WX"] # Ensuring static_link_msvcrt is enabled. ) ``` linker says that: ``` LINK : warning LNK4098: defaultlib 'msvcrt.lib' conflicts with use of other libs; use /NODEFAULTLIB:library LINK : error LNK1218: warning treated as error; no output file generated ``` Second thing is that if I use non-existent feature, it is silently ignored by Bazel. I feel that that is not right. See https://github.com/excitoon/bazel-issues/tree/master/imply-static-link-msvcrt ### What operating system are you running Bazel on? Windows 10 x64 ### What's the output of `bazel info release`? Build label: 0.23.0 Build target: bazel-out/x64_windows-opt/bin/src/main/java/com/google/devtools/build/lib/bazel/BazelServer_deploy.jar Build time: Tue Feb 26 17:40:30 2019 (1551202830) Build timestamp: 1551202830 Build timestamp as int: 1551202830
1.0
`implies: ...` does not stack well with `static_link_msvcrt` - ### Description of the problem / feature request: Hi. I can not enable `static_link_msvcrt` from another feature: ``` imply_static_link_msvcrt_feature = feature( name = "imply_static_link_msvcrt", implies = ["static_link_msvcrt"] ) ``` When I do: ``` cc_binary( name = "hello-world", srcs = ["hello-world.cpp"], features = ["imply_static_link_msvcrt"], linkopts = ["-DEFAULTLIB:libcmt.lib", "-WX"] # Ensuring static_link_msvcrt is enabled. ) ``` linker says that: ``` LINK : warning LNK4098: defaultlib 'msvcrt.lib' conflicts with use of other libs; use /NODEFAULTLIB:library LINK : error LNK1218: warning treated as error; no output file generated ``` Second thing is that if I use non-existent feature, it is silently ignored by Bazel. I feel that that is not right. See https://github.com/excitoon/bazel-issues/tree/master/imply-static-link-msvcrt ### What operating system are you running Bazel on? Windows 10 x64 ### What's the output of `bazel info release`? Build label: 0.23.0 Build target: bazel-out/x64_windows-opt/bin/src/main/java/com/google/devtools/build/lib/bazel/BazelServer_deploy.jar Build time: Tue Feb 26 17:40:30 2019 (1551202830) Build timestamp: 1551202830 Build timestamp as int: 1551202830
process
implies does not stack well with static link msvcrt description of the problem feature request hi i can not enable static link msvcrt from another feature imply static link msvcrt feature feature name imply static link msvcrt implies when i do cc binary name hello world srcs features linkopts ensuring static link msvcrt is enabled linker says that link warning defaultlib msvcrt lib conflicts with use of other libs use nodefaultlib library link error warning treated as error no output file generated second thing is that if i use non existent feature it is silently ignored by bazel i feel that that is not right see what operating system are you running bazel on windows what s the output of bazel info release build label build target bazel out windows opt bin src main java com google devtools build lib bazel bazelserver deploy jar build time tue feb build timestamp build timestamp as int
1
20,444
27,101,156,693
IssuesEvent
2023-02-15 08:46:30
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
Clean up `@bazel_tools`
P3 type: process team-OSS stale
# Description of the problem / feature request: This is an umbrella issue to track efforts to clean up `@bazel_tools`. We should remove unnecessary source code from this repo, and make it work for `bazel query @bazel_tools//...`. ## Details This repo is built into Bazel. It's always available in every workspace. It's a place for commonly used tools like `tools/test/test-setup.sh`, which is the test wrapper used for every `bazel test` action. But there is also source code in this repo, each file for their own reason. E.g. the runfiles libraries `@bazel_tools//tools/<lang>/runfiles` are included as source so they can be compiled into binaries. Some source code's purpose in `@bazel_tools` is less obvious. Sometimes they cause problems, like https://github.com/bazelbuild/bazel/issues/8614 whose culprit is https://github.com/bazelbuild/bazel/pull/8742#issuecomment-506745688. ### What's the output of `bazel info release`? `release 0.28.0`
1.0
Clean up `@bazel_tools` - # Description of the problem / feature request: This is an umbrella issue to track efforts to clean up `@bazel_tools`. We should remove unnecessary source code from this repo, and make it work for `bazel query @bazel_tools//...`. ## Details This repo is built into Bazel. It's always available in every workspace. It's a place for commonly used tools like `tools/test/test-setup.sh`, which is the test wrapper used for every `bazel test` action. But there is also source code in this repo, each file for their own reason. E.g. the runfiles libraries `@bazel_tools//tools/<lang>/runfiles` are included as source so they can be compiled into binaries. Some source code's purpose in `@bazel_tools` is less obvious. Sometimes they cause problems, like https://github.com/bazelbuild/bazel/issues/8614 whose culprit is https://github.com/bazelbuild/bazel/pull/8742#issuecomment-506745688. ### What's the output of `bazel info release`? `release 0.28.0`
process
clean up bazel tools description of the problem feature request this is an umbrella issue to track efforts to clean up bazel tools we should remove unnecessary source code from this repo and make it work for bazel query bazel tools details this repo is built into bazel it s always available in every workspace it s a place for commonly used tools like tools test test setup sh which is the test wrapper used for every bazel test action but there is also source code in this repo each file for their own reason e g the runfiles libraries bazel tools tools runfiles are included as source so they can be compiled into binaries some source code s purpose in bazel tools is less obvious sometimes they cause problems like whose culprit is what s the output of bazel info release release
1
356,564
25,176,214,188
IssuesEvent
2022-11-11 09:29:22
peppapighs/pe
https://api.github.com/repos/peppapighs/pe
opened
Student class diagram should show `ModuleCode` class separately
severity.VeryLow type.DocumentationBug
Since the class `ModuleCode` is a custom class and is never mentioned before in the DG, the diagram should show this class separately and have the `Student` class assoicated with `ModuleCode` with `*` as the multiplicity. ![image.png](https://raw.githubusercontent.com/peppapighs/pe/main/files/b3650d59-9091-460c-b2c2-2d6d0a32e762.png) <!--session: 1668153839735-43773ef4-6587-4127-9391-a44f44a8a6a7--> <!--Version: Web v3.4.4-->
1.0
Student class diagram should show `ModuleCode` class separately - Since the class `ModuleCode` is a custom class and is never mentioned before in the DG, the diagram should show this class separately and have the `Student` class assoicated with `ModuleCode` with `*` as the multiplicity. ![image.png](https://raw.githubusercontent.com/peppapighs/pe/main/files/b3650d59-9091-460c-b2c2-2d6d0a32e762.png) <!--session: 1668153839735-43773ef4-6587-4127-9391-a44f44a8a6a7--> <!--Version: Web v3.4.4-->
non_process
student class diagram should show modulecode class separately since the class modulecode is a custom class and is never mentioned before in the dg the diagram should show this class separately and have the student class assoicated with modulecode with as the multiplicity
0
142,738
11,491,618,729
IssuesEvent
2020-02-11 19:20:50
GetTerminus/terminus-ui
https://api.github.com/repos/GetTerminus/terminus-ui
closed
Use new type guards from ngx-tools and remove all internal type guards
Focus: utility Goal: Library Stabilization Target: latest Type: chore
All existing UI library type guards will be added to ngx-tools by https://github.com/GetTerminus/ngx-tools/issues/292 - [ ] Use type guards from ngx-tools - [ ] Remove all internal type guards - [ ] Verify linting and tests still pass
1.0
Use new type guards from ngx-tools and remove all internal type guards - All existing UI library type guards will be added to ngx-tools by https://github.com/GetTerminus/ngx-tools/issues/292 - [ ] Use type guards from ngx-tools - [ ] Remove all internal type guards - [ ] Verify linting and tests still pass
non_process
use new type guards from ngx tools and remove all internal type guards all existing ui library type guards will be added to ngx tools by use type guards from ngx tools remove all internal type guards verify linting and tests still pass
0
10,876
13,645,035,281
IssuesEvent
2020-09-25 20:00:36
pystatgen/sgkit
https://api.github.com/repos/pystatgen/sgkit
closed
Apply to NumFOCUS by October 15
process + tools
A reminder to myself from a previous [discussion](https://discourse.pystatgen.org/t/what-should-we-call-our-new-genetics-toolkit/52/11?u=hammer) in June. I think we may want a logo too #30
1.0
Apply to NumFOCUS by October 15 - A reminder to myself from a previous [discussion](https://discourse.pystatgen.org/t/what-should-we-call-our-new-genetics-toolkit/52/11?u=hammer) in June. I think we may want a logo too #30
process
apply to numfocus by october a reminder to myself from a previous in june i think we may want a logo too
1
18,018
24,032,775,826
IssuesEvent
2022-09-15 16:18:34
googleapis/java-beyondcorp-appconnectors
https://api.github.com/repos/googleapis/java-beyondcorp-appconnectors
opened
Your .repo-metadata.json file has a problem 🤒
type: process repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan 📈: * api_shortname 'beyondcorp-appconnectors' invalid in .repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file: Result of scan 📈: * api_shortname 'beyondcorp-appconnectors' invalid in .repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 api shortname beyondcorp appconnectors invalid in repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
1
53,259
13,261,294,537
IssuesEvent
2020-08-20 19:37:57
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
[steamshovel] RangeSetting compile error (Trac #1059)
Migrated from Trac combo core defect
Fails to compile on SL6 cvmfs py2-v1: ```text [ 80%] Building CXX object steamshovel/CMakeFiles/shovelart.dir/private/shovelart/artists/CoordinateSystem.cpp.o In file included from /scratch/dschultz/offline/src/steamshovel/private/shovelart/Artist.h:25, from /scratch/dschultz/offline/src/steamshovel/private/shovelart/artists/Coordin ateSystem.h:4, from /scratch/dschultz/offline/src/steamshovel/private/shovelart/artists/CoordinateSystem.cpp:8: /scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h: In constructor ‘RangeSetting::RangeSetting(float, float, int, double)’: /scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h:40: error: class ‘RangeSetting’ does not have any field named ‘ ComplexSetting’ /scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h:43: error: No match for ‘ComplexSetting<float>::ComplexSetting()’ /scratch/dschultz/offline/src/steamshovel/private/shovelart/ComplexSetting.h:16: note: candidates are: ComplexSetting<float>::ComplexSetting(float) /scratch/dschultz/offline/src/steamshovel/private/shovelart/ComplexSetting.h:12: note: ComplexSetting<float>::ComplexSetting(const ComplexSetting<float> &) /scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h: In constructor ‘RangeSetting::RangeSetting(float, float, int, float)’: ``` <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1059">https://code.icecube.wisc.edu/projects/icecube/ticket/1059</a>, reported by david.schultzand owned by hdembinski</em></summary> <p> ```json { "status": "closed", "changetime": "2015-08-10T20:31:17", "_ts": "1439238677802127", "description": "Fails to compile on SL6 cvmfs py2-v1:\n\n{{{\n[ 80%] Building CXX object steamshovel/CMakeFiles/shovelart.dir/private/shovelart/artists/CoordinateSystem.cpp.o\nIn file included from\n /scratch/dschultz/offline/src/steamshovel/private/shovelart/Artist.h:25,\nfrom\n /scratch/dschultz/offline/src/steamshovel/private/shovelart/artists/Coordin\n ateSystem.h:4,\nfrom /scratch/dschultz/offline/src/steamshovel/private/shovelart/artists/CoordinateSystem.cpp:8:\n /scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h: In\n constructor \u2018RangeSetting::RangeSetting(float, float, int, double)\u2019:\n/scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h:40:\n error: class \u2018RangeSetting\u2019 does not have any field named \u2018\n ComplexSetting\u2019\n/scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h:43:\n error: No match for \u2018ComplexSetting<float>::ComplexSetting()\u2019\n/scratch/dschultz/offline/src/steamshovel/private/shovelart/ComplexSetting.h:16:\n note: candidates are: ComplexSetting<float>::ComplexSetting(float)\n/scratch/dschultz/offline/src/steamshovel/private/shovelart/ComplexSetting.h:12:\n note: ComplexSetting<float>::ComplexSetting(const ComplexSetting<float> &)\n/scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h: In\n constructor \u2018RangeSetting::RangeSetting(float, float, int, float)\u2019:\n}}}", "reporter": "david.schultz", "cc": "", "resolution": "fixed", "time": "2015-07-20T17:26:52", "component": "combo core", "summary": "[steamshovel] RangeSetting compile error", "priority": "blocker", "keywords": "", "milestone": "", "owner": "hdembinski", "type": "defect" } ``` </p> </details>
1.0
[steamshovel] RangeSetting compile error (Trac #1059) - Fails to compile on SL6 cvmfs py2-v1: ```text [ 80%] Building CXX object steamshovel/CMakeFiles/shovelart.dir/private/shovelart/artists/CoordinateSystem.cpp.o In file included from /scratch/dschultz/offline/src/steamshovel/private/shovelart/Artist.h:25, from /scratch/dschultz/offline/src/steamshovel/private/shovelart/artists/Coordin ateSystem.h:4, from /scratch/dschultz/offline/src/steamshovel/private/shovelart/artists/CoordinateSystem.cpp:8: /scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h: In constructor ‘RangeSetting::RangeSetting(float, float, int, double)’: /scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h:40: error: class ‘RangeSetting’ does not have any field named ‘ ComplexSetting’ /scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h:43: error: No match for ‘ComplexSetting<float>::ComplexSetting()’ /scratch/dschultz/offline/src/steamshovel/private/shovelart/ComplexSetting.h:16: note: candidates are: ComplexSetting<float>::ComplexSetting(float) /scratch/dschultz/offline/src/steamshovel/private/shovelart/ComplexSetting.h:12: note: ComplexSetting<float>::ComplexSetting(const ComplexSetting<float> &) /scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h: In constructor ‘RangeSetting::RangeSetting(float, float, int, float)’: ``` <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1059">https://code.icecube.wisc.edu/projects/icecube/ticket/1059</a>, reported by david.schultzand owned by hdembinski</em></summary> <p> ```json { "status": "closed", "changetime": "2015-08-10T20:31:17", "_ts": "1439238677802127", "description": "Fails to compile on SL6 cvmfs py2-v1:\n\n{{{\n[ 80%] Building CXX object steamshovel/CMakeFiles/shovelart.dir/private/shovelart/artists/CoordinateSystem.cpp.o\nIn file included from\n /scratch/dschultz/offline/src/steamshovel/private/shovelart/Artist.h:25,\nfrom\n /scratch/dschultz/offline/src/steamshovel/private/shovelart/artists/Coordin\n ateSystem.h:4,\nfrom /scratch/dschultz/offline/src/steamshovel/private/shovelart/artists/CoordinateSystem.cpp:8:\n /scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h: In\n constructor \u2018RangeSetting::RangeSetting(float, float, int, double)\u2019:\n/scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h:40:\n error: class \u2018RangeSetting\u2019 does not have any field named \u2018\n ComplexSetting\u2019\n/scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h:43:\n error: No match for \u2018ComplexSetting<float>::ComplexSetting()\u2019\n/scratch/dschultz/offline/src/steamshovel/private/shovelart/ComplexSetting.h:16:\n note: candidates are: ComplexSetting<float>::ComplexSetting(float)\n/scratch/dschultz/offline/src/steamshovel/private/shovelart/ComplexSetting.h:12:\n note: ComplexSetting<float>::ComplexSetting(const ComplexSetting<float> &)\n/scratch/dschultz/offline/src/steamshovel/private/shovelart/RangeSetting.h: In\n constructor \u2018RangeSetting::RangeSetting(float, float, int, float)\u2019:\n}}}", "reporter": "david.schultz", "cc": "", "resolution": "fixed", "time": "2015-07-20T17:26:52", "component": "combo core", "summary": "[steamshovel] RangeSetting compile error", "priority": "blocker", "keywords": "", "milestone": "", "owner": "hdembinski", "type": "defect" } ``` </p> </details>
non_process
rangesetting compile error trac fails to compile on cvmfs text building cxx object steamshovel cmakefiles shovelart dir private shovelart artists coordinatesystem cpp o in file included from scratch dschultz offline src steamshovel private shovelart artist h from scratch dschultz offline src steamshovel private shovelart artists coordin atesystem h from scratch dschultz offline src steamshovel private shovelart artists coordinatesystem cpp scratch dschultz offline src steamshovel private shovelart rangesetting h in constructor ‘rangesetting rangesetting float float int double ’ scratch dschultz offline src steamshovel private shovelart rangesetting h error class ‘rangesetting’ does not have any field named ‘ complexsetting’ scratch dschultz offline src steamshovel private shovelart rangesetting h error no match for ‘complexsetting complexsetting ’ scratch dschultz offline src steamshovel private shovelart complexsetting h note candidates are complexsetting complexsetting float scratch dschultz offline src steamshovel private shovelart complexsetting h note complexsetting complexsetting const complexsetting scratch dschultz offline src steamshovel private shovelart rangesetting h in constructor ‘rangesetting rangesetting float float int float ’ migrated from json status closed changetime ts description fails to compile on cvmfs n n n building cxx object steamshovel cmakefiles shovelart dir private shovelart artists coordinatesystem cpp o nin file included from n scratch dschultz offline src steamshovel private shovelart artist h nfrom n scratch dschultz offline src steamshovel private shovelart artists coordin n atesystem h nfrom scratch dschultz offline src steamshovel private shovelart artists coordinatesystem cpp n scratch dschultz offline src steamshovel private shovelart rangesetting h in n constructor rangesetting float float int double n scratch dschultz offline src steamshovel private shovelart rangesetting h n error class does not have any field named n complexsetting n scratch dschultz offline src steamshovel private shovelart rangesetting h n error no match for complexsetting n scratch dschultz offline src steamshovel private shovelart complexsetting h n note candidates are complexsetting complexsetting float n scratch dschultz offline src steamshovel private shovelart complexsetting h n note complexsetting complexsetting const complexsetting n scratch dschultz offline src steamshovel private shovelart rangesetting h in n constructor rangesetting float float int float n reporter david schultz cc resolution fixed time component combo core summary rangesetting compile error priority blocker keywords milestone owner hdembinski type defect
0
578,879
17,156,547,032
IssuesEvent
2021-07-14 07:41:19
googleapis/java-bigtable-hbase
https://api.github.com/repos/googleapis/java-bigtable-hbase
closed
bigtable.hbase.TestBasicOps: testPutAlmostTooBigValueTenTimes failed
api: bigtable flakybot: issue priority: p1 type: bug
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: a891335ce3179c45fade4f3683b7e09d38d0107a buildURL: [Build Status](https://source.cloud.google.com/results/invocations/64905140-8b04-4270-9ed2-a91e73d97e39), [Sponge](http://sponge2/64905140-8b04-4270-9ed2-a91e73d97e39) status: failed <details><summary>Test output</summary><br><pre>org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: Failed 1 action: UnauthenticatedException: 1 time, servers with issues: bigtable.googleapis.com at com.google.cloud.bigtable.hbase.AbstractBigtableTable.createRetriesExhaustedWithDetailsException(AbstractBigtableTable.java:274) at com.google.cloud.bigtable.hbase.AbstractBigtableTable.exists(AbstractBigtableTable.java:170) at com.google.cloud.bigtable.hbase.TestBasicOps.testPutGetDelete(TestBasicOps.java:226) at com.google.cloud.bigtable.hbase.TestBasicOps.testPutGetDeleteExists(TestBasicOps.java:190) at com.google.cloud.bigtable.hbase.TestBasicOps.testPutAlmostTooBigValueTenTimes(TestBasicOps.java:166) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:410) at org.apache.maven.surefire.junitcore.pc.InvokerStrategy.schedule(InvokerStrategy.java:54) at org.apache.maven.surefire.junitcore.pc.Scheduler.schedule(Scheduler.java:367) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:410) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) </pre></details>
1.0
bigtable.hbase.TestBasicOps: testPutAlmostTooBigValueTenTimes failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: a891335ce3179c45fade4f3683b7e09d38d0107a buildURL: [Build Status](https://source.cloud.google.com/results/invocations/64905140-8b04-4270-9ed2-a91e73d97e39), [Sponge](http://sponge2/64905140-8b04-4270-9ed2-a91e73d97e39) status: failed <details><summary>Test output</summary><br><pre>org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: Failed 1 action: UnauthenticatedException: 1 time, servers with issues: bigtable.googleapis.com at com.google.cloud.bigtable.hbase.AbstractBigtableTable.createRetriesExhaustedWithDetailsException(AbstractBigtableTable.java:274) at com.google.cloud.bigtable.hbase.AbstractBigtableTable.exists(AbstractBigtableTable.java:170) at com.google.cloud.bigtable.hbase.TestBasicOps.testPutGetDelete(TestBasicOps.java:226) at com.google.cloud.bigtable.hbase.TestBasicOps.testPutGetDeleteExists(TestBasicOps.java:190) at com.google.cloud.bigtable.hbase.TestBasicOps.testPutAlmostTooBigValueTenTimes(TestBasicOps.java:166) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:410) at org.apache.maven.surefire.junitcore.pc.InvokerStrategy.schedule(InvokerStrategy.java:54) at org.apache.maven.surefire.junitcore.pc.Scheduler.schedule(Scheduler.java:367) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:410) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) </pre></details>
non_process
bigtable hbase testbasicops testputalmosttoobigvaluetentimes failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output org apache hadoop hbase client retriesexhaustedwithdetailsexception failed action unauthenticatedexception time servers with issues bigtable googleapis com at com google cloud bigtable hbase abstractbigtabletable createretriesexhaustedwithdetailsexception abstractbigtabletable java at com google cloud bigtable hbase abstractbigtabletable exists abstractbigtabletable java at com google cloud bigtable hbase testbasicops testputgetdelete testbasicops java at com google cloud bigtable hbase testbasicops testputgetdeleteexists testbasicops java at com google cloud bigtable hbase testbasicops testputalmosttoobigvaluetentimes testbasicops java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org apache maven surefire junitcore pc scheduler run scheduler java at org apache maven surefire junitcore pc invokerstrategy schedule invokerstrategy java at org apache maven surefire junitcore pc scheduler schedule scheduler java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org junit runners suite runchild suite java at org junit runners suite runchild suite java at org junit runners parentrunner run parentrunner java at org apache maven surefire junitcore pc scheduler run scheduler java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java
0
7,180
10,321,200,398
IssuesEvent
2019-08-30 23:56:42
googleapis/google-api-java-client
https://api.github.com/repos/googleapis/google-api-java-client
opened
Publish SNAPSHOT releases nightly
type: process
This will help us detect downstream breakage before releasing the library - [x] add build config in this repo - [ ] add job config internally
1.0
Publish SNAPSHOT releases nightly - This will help us detect downstream breakage before releasing the library - [x] add build config in this repo - [ ] add job config internally
process
publish snapshot releases nightly this will help us detect downstream breakage before releasing the library add build config in this repo add job config internally
1
16,365
21,052,204,938
IssuesEvent
2022-03-31 21:37:40
nion-software/nionswift
https://api.github.com/repos/nion-software/nionswift
closed
Setting xdata property on should accept anything that can be converted to xdata
type - enhancement level - easy f - processing f - API
`target.xdata = numpy.zeros((3,3))` should work.
1.0
Setting xdata property on should accept anything that can be converted to xdata - `target.xdata = numpy.zeros((3,3))` should work.
process
setting xdata property on should accept anything that can be converted to xdata target xdata numpy zeros should work
1
273,066
20,768,983,091
IssuesEvent
2022-03-16 00:58:51
gth828r/brownify
https://api.github.com/repos/gth828r/brownify
closed
Generate and publish sphinx documentation
documentation
We should set up documentation in our library, given that we have already started doing a reasonable amount of docstring writing and type hinting. - [x] Initialize the repository for sphinx documentation generation - [x] Publish the documentation somewhere - [x] Update the setup.cfg to point to where the documentation lives - [x] Update the top level README to point to where documentation lives
1.0
Generate and publish sphinx documentation - We should set up documentation in our library, given that we have already started doing a reasonable amount of docstring writing and type hinting. - [x] Initialize the repository for sphinx documentation generation - [x] Publish the documentation somewhere - [x] Update the setup.cfg to point to where the documentation lives - [x] Update the top level README to point to where documentation lives
non_process
generate and publish sphinx documentation we should set up documentation in our library given that we have already started doing a reasonable amount of docstring writing and type hinting initialize the repository for sphinx documentation generation publish the documentation somewhere update the setup cfg to point to where the documentation lives update the top level readme to point to where documentation lives
0
734
3,214,313,663
IssuesEvent
2015-10-07 00:44:43
broadinstitute/hellbender
https://api.github.com/repos/broadinstitute/hellbender
closed
Dataflow BQSR Direct Runner fails with --knownSites
bug Dataflow DataflowPreprocessingPipeline
From @tomwhite: I noticed that the test with "-knownSites" from BaseRecalibratorIntegrationTest (i.e. the non-dataflow version) fails with both the Direct and Spark runners. I had a look at the file output and there are a few discrepancies (see diff below). ``` diff /var/folders/d1/8f5_j4hx04z72w6wgqxkb2l40000gn/T/walktest.tmp_param.02172067147450353519.tmp src/test/resources/org/broadinstitute/hellbender/tools/BQSR/expected.NA12878.chr17_69k_70k.2inputs.txt 60c60 < 34 3051 34 --- > 34 3050 34 71c71 < 45 46942 45 --- > 45 46940 45 124,126c124,126 < 809R9ABXX101220.5 D 45.0000 45.0000 23471 0.00 < 809R9ABXX101220.5 I 45.0000 45.0000 23471 0.00 < 809R9ABXX101220.5 M 27.0000 27.0494 23471 49.13 --- > 809R9ABXX101220.5 D 45.0000 45.0000 23470 0.00 > 809R9ABXX101220.5 I 45.0000 45.0000 23470 0.00 > 809R9ABXX101220.5 M 27.0000 27.0493 23470 49.13 155c155 < 809R9ABXX101220.5 34 M 34.0000 3051 2.96 --- > 809R9ABXX101220.5 34 M 34.0000 3050 2.96 161,162c161,162 < 809R9ABXX101220.5 45 D 45.0000 23471 0.00 < 809R9ABXX101220.5 45 I 45.0000 23471 0.00 --- > 809R9ABXX101220.5 45 D 45.0000 23470 0.00 > 809R9ABXX101220.5 45 I 45.0000 23470 0.00 2714c2714 < 809R9ABXX101220.5 34 29 Cycle M 34.0000 20 0.00 --- > 809R9ABXX101220.5 34 29 Cycle M 34.0000 19 0.00 2773c2773 < 809R9ABXX101220.5 34 CA Context M 34.0000 506 0.00 --- > 809R9ABXX101220.5 34 CA Context M 34.0000 505 0.00 3464,3465c3464,3465 < 809R9ABXX101220.5 45 29 Cycle D 45.0000 180 0.00 < 809R9ABXX101220.5 45 29 Cycle I 45.0000 180 0.00 --- > 809R9ABXX101220.5 45 29 Cycle D 45.0000 179 0.00 > 809R9ABXX101220.5 45 29 Cycle I 45.0000 179 0.00 3634,3635c3634,3635 < 809R9ABXX101220.5 45 GCA Context D 45.0000 278 0.00 < 809R9ABXX101220.5 45 GCA Context I 45.0000 278 0.00 --- > 809R9ABXX101220.5 45 GCA Context D 45.0000 277 0.00 > 809R9ABXX101220.5 45 GCA Context I 45.0000 277 0.00 ``` The relevant test is this one from `BaseRecalibratorDataflowIntegrationTest` ``` new BQSRTest(hg18Reference, HiSeqBam, dbSNPb37, "-knownSites " + moreSites, getResourceDir() + "expected.NA12878.chr17_69k_70k.2inputs.txt") ```
1.0
Dataflow BQSR Direct Runner fails with --knownSites - From @tomwhite: I noticed that the test with "-knownSites" from BaseRecalibratorIntegrationTest (i.e. the non-dataflow version) fails with both the Direct and Spark runners. I had a look at the file output and there are a few discrepancies (see diff below). ``` diff /var/folders/d1/8f5_j4hx04z72w6wgqxkb2l40000gn/T/walktest.tmp_param.02172067147450353519.tmp src/test/resources/org/broadinstitute/hellbender/tools/BQSR/expected.NA12878.chr17_69k_70k.2inputs.txt 60c60 < 34 3051 34 --- > 34 3050 34 71c71 < 45 46942 45 --- > 45 46940 45 124,126c124,126 < 809R9ABXX101220.5 D 45.0000 45.0000 23471 0.00 < 809R9ABXX101220.5 I 45.0000 45.0000 23471 0.00 < 809R9ABXX101220.5 M 27.0000 27.0494 23471 49.13 --- > 809R9ABXX101220.5 D 45.0000 45.0000 23470 0.00 > 809R9ABXX101220.5 I 45.0000 45.0000 23470 0.00 > 809R9ABXX101220.5 M 27.0000 27.0493 23470 49.13 155c155 < 809R9ABXX101220.5 34 M 34.0000 3051 2.96 --- > 809R9ABXX101220.5 34 M 34.0000 3050 2.96 161,162c161,162 < 809R9ABXX101220.5 45 D 45.0000 23471 0.00 < 809R9ABXX101220.5 45 I 45.0000 23471 0.00 --- > 809R9ABXX101220.5 45 D 45.0000 23470 0.00 > 809R9ABXX101220.5 45 I 45.0000 23470 0.00 2714c2714 < 809R9ABXX101220.5 34 29 Cycle M 34.0000 20 0.00 --- > 809R9ABXX101220.5 34 29 Cycle M 34.0000 19 0.00 2773c2773 < 809R9ABXX101220.5 34 CA Context M 34.0000 506 0.00 --- > 809R9ABXX101220.5 34 CA Context M 34.0000 505 0.00 3464,3465c3464,3465 < 809R9ABXX101220.5 45 29 Cycle D 45.0000 180 0.00 < 809R9ABXX101220.5 45 29 Cycle I 45.0000 180 0.00 --- > 809R9ABXX101220.5 45 29 Cycle D 45.0000 179 0.00 > 809R9ABXX101220.5 45 29 Cycle I 45.0000 179 0.00 3634,3635c3634,3635 < 809R9ABXX101220.5 45 GCA Context D 45.0000 278 0.00 < 809R9ABXX101220.5 45 GCA Context I 45.0000 278 0.00 --- > 809R9ABXX101220.5 45 GCA Context D 45.0000 277 0.00 > 809R9ABXX101220.5 45 GCA Context I 45.0000 277 0.00 ``` The relevant test is this one from `BaseRecalibratorDataflowIntegrationTest` ``` new BQSRTest(hg18Reference, HiSeqBam, dbSNPb37, "-knownSites " + moreSites, getResourceDir() + "expected.NA12878.chr17_69k_70k.2inputs.txt") ```
process
dataflow bqsr direct runner fails with knownsites from tomwhite i noticed that the test with knownsites from baserecalibratorintegrationtest i e the non dataflow version fails with both the direct and spark runners i had a look at the file output and there are a few discrepancies see diff below diff var folders t walktest tmp param tmp src test resources org broadinstitute hellbender tools bqsr expected txt d i m d i m m m d i d i cycle m cycle m ca context m ca context m cycle d cycle i cycle d cycle i gca context d gca context i gca context d gca context i the relevant test is this one from baserecalibratordataflowintegrationtest new bqsrtest hiseqbam knownsites moresites getresourcedir expected txt
1
22,405
31,142,291,542
IssuesEvent
2023-08-16 01:44:42
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
Flaky test: Error: done() called multiple times VI
OS: linux process: flaky test topic: flake ❄️ stage: flake "topic: done()" stale
### Link to dashboard or CircleCI failure https://app.circleci.com/pipelines/github/cypress-io/cypress/41102/workflows/219dd1e6-3a6a-44b0-a811-d5cfe222f25e/jobs/1699154 ### Link to failing test in GitHub https://github.com/cypress-io/cypress/blob/develop/packages/driver/cypress/e2e/commands/actions/select.cy.js#L176 ### Analysis <img width="812" alt="Screen Shot 2022-08-05 at 12 39 30 PM" src="https://user-images.githubusercontent.com/26726429/183149058-79aeca18-fbd1-4906-b270-0f8578628b91.png"> ### Cypress Version 10.4.0 ### Other Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed
1.0
Flaky test: Error: done() called multiple times VI - ### Link to dashboard or CircleCI failure https://app.circleci.com/pipelines/github/cypress-io/cypress/41102/workflows/219dd1e6-3a6a-44b0-a811-d5cfe222f25e/jobs/1699154 ### Link to failing test in GitHub https://github.com/cypress-io/cypress/blob/develop/packages/driver/cypress/e2e/commands/actions/select.cy.js#L176 ### Analysis <img width="812" alt="Screen Shot 2022-08-05 at 12 39 30 PM" src="https://user-images.githubusercontent.com/26726429/183149058-79aeca18-fbd1-4906-b270-0f8578628b91.png"> ### Cypress Version 10.4.0 ### Other Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed
process
flaky test error done called multiple times vi link to dashboard or circleci failure link to failing test in github analysis img width alt screen shot at pm src cypress version other search for this issue number in the codebase to find the test s skipped until this issue is fixed
1
9,482
12,477,761,345
IssuesEvent
2020-05-29 15:29:21
googleapis/google-cloud-go
https://api.github.com/repos/googleapis/google-cloud-go
opened
all: add a note compatibility in README.md
type: process
Something along of the lines of https://github.com/golang/protobuf#compatibility, but tailored for our repository.
1.0
all: add a note compatibility in README.md - Something along of the lines of https://github.com/golang/protobuf#compatibility, but tailored for our repository.
process
all add a note compatibility in readme md something along of the lines of but tailored for our repository
1
684,277
23,413,402,080
IssuesEvent
2022-08-12 20:19:35
python/mypy
https://api.github.com/repos/python/mypy
closed
Support recursive types
feature needs discussion priority-1-normal meta topic-recursive-types
The following in particular would be useful: ``` Callback = Callable[[str], 'Callback'] Foo = Union[str, List['Foo']] ```
1.0
Support recursive types - The following in particular would be useful: ``` Callback = Callable[[str], 'Callback'] Foo = Union[str, List['Foo']] ```
non_process
support recursive types the following in particular would be useful callback callable callback foo union
0
408,307
27,660,818,191
IssuesEvent
2023-03-12 13:57:43
libusb/hidapi
https://api.github.com/repos/libusb/hidapi
closed
HIDAPI web site
documentation
Someone should work on a web site presenting the hidapi project and subprojects (like Python wrapper). goals: * make the new hidapi project more visible so that it is indexed by search engine * document the history from https://github.com/signal11/hidapi One technical solution is to use github pages https://pages.github.com/ and host the web site at https://hidapi.libusb.info/ (the name does not exist yet but I can create it)
1.0
HIDAPI web site - Someone should work on a web site presenting the hidapi project and subprojects (like Python wrapper). goals: * make the new hidapi project more visible so that it is indexed by search engine * document the history from https://github.com/signal11/hidapi One technical solution is to use github pages https://pages.github.com/ and host the web site at https://hidapi.libusb.info/ (the name does not exist yet but I can create it)
non_process
hidapi web site someone should work on a web site presenting the hidapi project and subprojects like python wrapper goals make the new hidapi project more visible so that it is indexed by search engine document the history from one technical solution is to use github pages and host the web site at the name does not exist yet but i can create it
0
124,827
16,668,539,367
IssuesEvent
2021-06-07 08:04:30
crev-dev/cargo-crev
https://api.github.com/repos/crev-dev/cargo-crev
closed
Thoughts about packaging
design enhancement
This is sort of related to #37, but probably warrants its own issue: I'm involved with rust packaging in two major linux distros and looked into packaging crev. It seems there are both tagged releases and cargo uploads, which is very good, but I'm not sure which packages should be distributed using which package name. It seems cargo-crev is used to tag the binary crates, should this ship only `cargo-crev` or also the other binary targets? Is rblake2sum required as a binary? On a side note, specifically for debian it's important to make sure you're depending on the latest version of your dependencies. While we can upload outdated crates to debian it introduces quite a bit of complexity for us. If you have some opinions on that topic please let me know. :)
1.0
Thoughts about packaging - This is sort of related to #37, but probably warrants its own issue: I'm involved with rust packaging in two major linux distros and looked into packaging crev. It seems there are both tagged releases and cargo uploads, which is very good, but I'm not sure which packages should be distributed using which package name. It seems cargo-crev is used to tag the binary crates, should this ship only `cargo-crev` or also the other binary targets? Is rblake2sum required as a binary? On a side note, specifically for debian it's important to make sure you're depending on the latest version of your dependencies. While we can upload outdated crates to debian it introduces quite a bit of complexity for us. If you have some opinions on that topic please let me know. :)
non_process
thoughts about packaging this is sort of related to but probably warrants its own issue i m involved with rust packaging in two major linux distros and looked into packaging crev it seems there are both tagged releases and cargo uploads which is very good but i m not sure which packages should be distributed using which package name it seems cargo crev is used to tag the binary crates should this ship only cargo crev or also the other binary targets is required as a binary on a side note specifically for debian it s important to make sure you re depending on the latest version of your dependencies while we can upload outdated crates to debian it introduces quite a bit of complexity for us if you have some opinions on that topic please let me know
0
17,595
23,420,851,273
IssuesEvent
2022-08-13 17:18:01
apache/arrow-rs
https://api.github.com/repos/apache/arrow-rs
closed
Release `object_store` `0.4.0`
development-process
As discussed, we will release a new version of object_store from the arrow-rs repo / under ASF process: https://github.com/apache/arrow-rs/issues/2180 * Planned Release Candidate: 2022-08-08 * Planned Release and Publish to crates.io: 2022-08-11 Items: - [x] Update changelog and readme: https://github.com/apache/arrow-rs/pull/2392 - [x] Create release scripts: https://github.com/apache/arrow-rs/pull/2392 - [x] Create release candidate https://lists.apache.org/thread/xg8m8s12q2d0glp3gocmqt18v2gzn1ml - [x] Release candidate approved : https://lists.apache.org/thread/8j9wfj5g0nzmvg5lz7cnlym84m2gw4g8 - [x] Release to crates.io - [x] Create release scripts - [x] Draft PR to update DataFusion to new object store: https://github.com/apache/arrow-datafusion/pull/3089
1.0
Release `object_store` `0.4.0` - As discussed, we will release a new version of object_store from the arrow-rs repo / under ASF process: https://github.com/apache/arrow-rs/issues/2180 * Planned Release Candidate: 2022-08-08 * Planned Release and Publish to crates.io: 2022-08-11 Items: - [x] Update changelog and readme: https://github.com/apache/arrow-rs/pull/2392 - [x] Create release scripts: https://github.com/apache/arrow-rs/pull/2392 - [x] Create release candidate https://lists.apache.org/thread/xg8m8s12q2d0glp3gocmqt18v2gzn1ml - [x] Release candidate approved : https://lists.apache.org/thread/8j9wfj5g0nzmvg5lz7cnlym84m2gw4g8 - [x] Release to crates.io - [x] Create release scripts - [x] Draft PR to update DataFusion to new object store: https://github.com/apache/arrow-datafusion/pull/3089
process
release object store as discussed we will release a new version of object store from the arrow rs repo under asf process planned release candidate planned release and publish to crates io items update changelog and readme create release scripts create release candidate release candidate approved release to crates io create release scripts draft pr to update datafusion to new object store
1
12,857
15,241,852,641
IssuesEvent
2021-02-19 09:03:42
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
Error `[migration-engine/connectors/sql-migration-connector/src/flavour/sqlite.rs:99:42] failed to truncate sqlite file: Os { code: 13, kind: PermissionDenied, message: "Permission denied" }`
bug/1-repro-available engines/migration engine kind/bug process/candidate team/migrations topic: migrate dev
<!-- If required, please update the title to be clear and descriptive --> Command: `migrate dev --name init --preview-feature` Version: `2.16.1` Report: https://prisma-errors.netlify.app/report/12973 OS: `x64 darwin 20.2.0` JS Stacktrace: ``` Error: Error in migration engine. Reason: [migration-engine/connectors/sql-migration-connector/src/flavour/sqlite.rs:99:42] failed to truncate sqlite file: Os { code: 13, kind: PermissionDenied, message: "Permission denied" } ``` Rust Stacktrace: ``` [migration-engine/connectors/sql-migration-connector/src/flavour/sqlite.rs:99:42] failed to truncate sqlite file: Os { code: 13, kind: PermissionDenied, message: "Permission denied" } ``` Migrate Request ``` { "id": 2, "jsonrpc": "2.0", "method": "reset", "params": {} } ```
1.0
Error `[migration-engine/connectors/sql-migration-connector/src/flavour/sqlite.rs:99:42] failed to truncate sqlite file: Os { code: 13, kind: PermissionDenied, message: "Permission denied" }` - <!-- If required, please update the title to be clear and descriptive --> Command: `migrate dev --name init --preview-feature` Version: `2.16.1` Report: https://prisma-errors.netlify.app/report/12973 OS: `x64 darwin 20.2.0` JS Stacktrace: ``` Error: Error in migration engine. Reason: [migration-engine/connectors/sql-migration-connector/src/flavour/sqlite.rs:99:42] failed to truncate sqlite file: Os { code: 13, kind: PermissionDenied, message: "Permission denied" } ``` Rust Stacktrace: ``` [migration-engine/connectors/sql-migration-connector/src/flavour/sqlite.rs:99:42] failed to truncate sqlite file: Os { code: 13, kind: PermissionDenied, message: "Permission denied" } ``` Migrate Request ``` { "id": 2, "jsonrpc": "2.0", "method": "reset", "params": {} } ```
process
error failed to truncate sqlite file os code kind permissiondenied message permission denied command migrate dev name init preview feature version report os darwin js stacktrace error error in migration engine reason failed to truncate sqlite file os code kind permissiondenied message permission denied rust stacktrace failed to truncate sqlite file os code kind permissiondenied message permission denied migrate request id jsonrpc method reset params
1
226
2,652,505,362
IssuesEvent
2015-03-16 17:41:18
AnalyticalGraphicsInc/cesium
https://api.github.com/repos/AnalyticalGraphicsInc/cesium
closed
Evaluate Jasmine 2.0
dev process
[Jasmine 2.0](http://jasmine.github.io/2.0/introduction.html) was just released. We should take a closer look to see how hard upgrading is given all our custom changes. If it is a ton of work, we can hold off until we decide on Intern, #735
1.0
Evaluate Jasmine 2.0 - [Jasmine 2.0](http://jasmine.github.io/2.0/introduction.html) was just released. We should take a closer look to see how hard upgrading is given all our custom changes. If it is a ton of work, we can hold off until we decide on Intern, #735
process
evaluate jasmine was just released we should take a closer look to see how hard upgrading is given all our custom changes if it is a ton of work we can hold off until we decide on intern
1
5,699
8,563,383,571
IssuesEvent
2018-11-09 13:53:31
easy-software-ufal/annotations_repos
https://api.github.com/repos/easy-software-ufal/annotations_repos
opened
sjh37/EntityFramework-Reverse-POCO-Code-First-Generator Reverse Navigation Many To One
C# RMA wrong processing
Issue: `https://github.com/sjh37/EntityFramework-Reverse-POCO-Code-First-Generator/issues/385` PR: `https://github.com/sjh37/EntityFramework-Reverse-POCO-Code-First-Generator/commit/4b1f3b493ca3899e831acdb91965c1aa48a41e75`
1.0
sjh37/EntityFramework-Reverse-POCO-Code-First-Generator Reverse Navigation Many To One - Issue: `https://github.com/sjh37/EntityFramework-Reverse-POCO-Code-First-Generator/issues/385` PR: `https://github.com/sjh37/EntityFramework-Reverse-POCO-Code-First-Generator/commit/4b1f3b493ca3899e831acdb91965c1aa48a41e75`
process
entityframework reverse poco code first generator reverse navigation many to one issue pr
1
17,341
23,163,547,562
IssuesEvent
2022-07-29 20:47:40
cagov/design-system
https://api.github.com/repos/cagov/design-system
closed
Process: Consolidate Component audit lists to a central location
Process improvement PM
Work with team to pull various component audit lists for the following: - Current State Web Template - Current Design System - Pending Components (in progress) - Needed Components (not yet available)
1.0
Process: Consolidate Component audit lists to a central location - Work with team to pull various component audit lists for the following: - Current State Web Template - Current Design System - Pending Components (in progress) - Needed Components (not yet available)
process
process consolidate component audit lists to a central location work with team to pull various component audit lists for the following current state web template current design system pending components in progress needed components not yet available
1
55,543
8,000,383,572
IssuesEvent
2018-07-22 15:19:32
ichabod801/t_games
https://api.github.com/repos/ichabod801/t_games
closed
More help text.
documentation games
Look for more ways I can use the help command to provide information about playing games.
1.0
More help text. - Look for more ways I can use the help command to provide information about playing games.
non_process
more help text look for more ways i can use the help command to provide information about playing games
0
279,222
21,134,318,851
IssuesEvent
2022-04-06 04:09:37
AY2122S2-CS2103T-W14-2/tp
https://api.github.com/repos/AY2122S2-CS2103T-W14-2/tp
closed
[PE-D] Inconsistency in command format vs example
Documentation Bug
The following shows the example command in the quick start section: ![image.png](https://raw.githubusercontent.com/cpwill01/ped/main/files/e850d358-10f9-4f6f-b46f-7c8daee448a5.png) which does not match the command format given later in the UG: ![image.png](https://raw.githubusercontent.com/cpwill01/ped/main/files/dfd490b2-f5fd-4f9e-95b7-e0e2056abbef.png) <!--session: 1648793114954-b1f7044c-6fa2-4da0-853f-3bda01691bfb--> <!--Version: Web v3.4.2--> ------------- Labels: `severity.Low` `type.DocumentationBug` original: cpwill01/ped#1
1.0
[PE-D] Inconsistency in command format vs example - The following shows the example command in the quick start section: ![image.png](https://raw.githubusercontent.com/cpwill01/ped/main/files/e850d358-10f9-4f6f-b46f-7c8daee448a5.png) which does not match the command format given later in the UG: ![image.png](https://raw.githubusercontent.com/cpwill01/ped/main/files/dfd490b2-f5fd-4f9e-95b7-e0e2056abbef.png) <!--session: 1648793114954-b1f7044c-6fa2-4da0-853f-3bda01691bfb--> <!--Version: Web v3.4.2--> ------------- Labels: `severity.Low` `type.DocumentationBug` original: cpwill01/ped#1
non_process
inconsistency in command format vs example the following shows the example command in the quick start section which does not match the command format given later in the ug labels severity low type documentationbug original ped
0
175,456
21,313,548,578
IssuesEvent
2022-04-16 00:08:07
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
[Security Solution] Session view icon is overlapping and not working under event renderer on alerts page
bug triage_needed impact:high Team: SecuritySolution v8.2.0
**Describe the bug** Session view icon is overlapping and not working under event renderer on alerts page **Build Details:** ``` Version : 8.2.0 SNAPSHOT Build : 51940 Commit : cbab89d8a43311d11dfaac3d9145a2be8841a857 ``` **preconditions** 1. Alerts should be exist with session view **Steps to Reproduce** 1. Navigate to Alerts page 2. Click on Event rendered view on alert table 3. Observe that Session view icon is overlapping and not working under event renderer on alerts page **Actual Result** Session view icon is overlapping and not working under event renderer on alerts page **Expected Result** Session view icon should not be overlapped and should be working under event renderer on alerts page **Screen-Shot** ![image](https://user-images.githubusercontent.com/61860752/163563790-75bb789f-b398-47b2-a645-dd1c7cffa03c.png)
True
[Security Solution] Session view icon is overlapping and not working under event renderer on alerts page - **Describe the bug** Session view icon is overlapping and not working under event renderer on alerts page **Build Details:** ``` Version : 8.2.0 SNAPSHOT Build : 51940 Commit : cbab89d8a43311d11dfaac3d9145a2be8841a857 ``` **preconditions** 1. Alerts should be exist with session view **Steps to Reproduce** 1. Navigate to Alerts page 2. Click on Event rendered view on alert table 3. Observe that Session view icon is overlapping and not working under event renderer on alerts page **Actual Result** Session view icon is overlapping and not working under event renderer on alerts page **Expected Result** Session view icon should not be overlapped and should be working under event renderer on alerts page **Screen-Shot** ![image](https://user-images.githubusercontent.com/61860752/163563790-75bb789f-b398-47b2-a645-dd1c7cffa03c.png)
non_process
session view icon is overlapping and not working under event renderer on alerts page describe the bug session view icon is overlapping and not working under event renderer on alerts page build details version snapshot build commit preconditions alerts should be exist with session view steps to reproduce navigate to alerts page click on event rendered view on alert table observe that session view icon is overlapping and not working under event renderer on alerts page actual result session view icon is overlapping and not working under event renderer on alerts page expected result session view icon should not be overlapped and should be working under event renderer on alerts page screen shot
0
226,389
18,015,527,881
IssuesEvent
2021-09-16 13:33:14
apache/shardingsphere
https://api.github.com/repos/apache/shardingsphere
closed
GovernanceRepositoryAPIImplTest.assertWatch may be blocked forever if exception thrown in watch listener
type: bug in: test
## Bug Report ### Which version of ShardingSphere did you use? 5.0.0-RC1-SNAPSHOT ### Which project did you use? ShardingSphere-JDBC or ShardingSphere-Proxy? ShardingSphere-Scaling ### Expected behavior Unit test not blocked. ### Actual behavior `assertThat(event.getType(), is(DataChangedEvent.Type.ADDED));` fail randomly. Unit test blocked. From GitHub CI log: ``` [Curator-SafeNotifyService-0] o.a.c.f.l.MappingListenerManager - Listener (org.apache.curator.framework.recipes.cache.CuratorCacheListenerBuilderImpl$2@24ebbdf) threw an exception java.lang.AssertionError: Expected: is <ADDED> but: was <UPDATED> at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) at org.junit.Assert.assertThat(Assert.java:956) at org.junit.Assert.assertThat(Assert.java:923) at org.apache.shardingsphere.scaling.core.api.impl.GovernanceRepositoryAPIImplTest.lambda$assertWatch$0(GovernanceRepositoryAPIImplTest.java:91) ``` ### Reason analyze (If you can) In `GovernanceRepositoryAPIImplTest.java` ``` @Test public void assertWatch() throws InterruptedException { CountDownLatch countDownLatch = new CountDownLatch(1); String key = ScalingConstant.SCALING_ROOT + "/1"; governanceRepositoryAPI.watch(ScalingConstant.SCALING_ROOT, event -> { if (event.getKey().equals(key)) { assertThat(event.getType(), is(DataChangedEvent.Type.ADDED)); countDownLatch.countDown(); } }); governanceRepositoryAPI.persist(key, ""); countDownLatch.await(); } ``` 1, If `assertThat` throw exception, `countDownLatch.await();` will always be blocked. 2, Repository key is the same in unit test methods, and value is persisted. See test log: ``` [INFO ] 2021-09-16 12:07:34.817 [main] o.a.s.s.c.a.i.GovernanceRepositoryAPIImplTest - assertWatch, key=/scaling/1 value=assertGetChildrenKeys [INFO ] 2021-09-16 12:07:34.919 [Curator-SafeNotifyService-0] o.a.s.s.c.a.i.GovernanceRepositoryAPIImplTest - event=DataChangedEvent(key=/scaling/1, value=assertWatch, type=ADDED) ``` In `CuratorZookeeperRepository.java` ``` private Type getChangedType(final TreeCacheEvent.Type type) { switch (type) { case NODE_ADDED: return Type.ADDED; case NODE_UPDATED: return Type.UPDATED; case NODE_REMOVED: return Type.DELETED; default: return Type.IGNORED; } } ``` TreeCacheEvent type depends on whether path is locally cached or not, it could be `ADDED` or `UPDATED`. ### Steps to reproduce the behavior, such as: SQL to execute, sharding rule configuration, when exception occur etc. Throw exception in `assertWatch`'s `watch` listener. ### Example codes for reproduce this issue (such as a github link).
1.0
GovernanceRepositoryAPIImplTest.assertWatch may be blocked forever if exception thrown in watch listener - ## Bug Report ### Which version of ShardingSphere did you use? 5.0.0-RC1-SNAPSHOT ### Which project did you use? ShardingSphere-JDBC or ShardingSphere-Proxy? ShardingSphere-Scaling ### Expected behavior Unit test not blocked. ### Actual behavior `assertThat(event.getType(), is(DataChangedEvent.Type.ADDED));` fail randomly. Unit test blocked. From GitHub CI log: ``` [Curator-SafeNotifyService-0] o.a.c.f.l.MappingListenerManager - Listener (org.apache.curator.framework.recipes.cache.CuratorCacheListenerBuilderImpl$2@24ebbdf) threw an exception java.lang.AssertionError: Expected: is <ADDED> but: was <UPDATED> at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) at org.junit.Assert.assertThat(Assert.java:956) at org.junit.Assert.assertThat(Assert.java:923) at org.apache.shardingsphere.scaling.core.api.impl.GovernanceRepositoryAPIImplTest.lambda$assertWatch$0(GovernanceRepositoryAPIImplTest.java:91) ``` ### Reason analyze (If you can) In `GovernanceRepositoryAPIImplTest.java` ``` @Test public void assertWatch() throws InterruptedException { CountDownLatch countDownLatch = new CountDownLatch(1); String key = ScalingConstant.SCALING_ROOT + "/1"; governanceRepositoryAPI.watch(ScalingConstant.SCALING_ROOT, event -> { if (event.getKey().equals(key)) { assertThat(event.getType(), is(DataChangedEvent.Type.ADDED)); countDownLatch.countDown(); } }); governanceRepositoryAPI.persist(key, ""); countDownLatch.await(); } ``` 1, If `assertThat` throw exception, `countDownLatch.await();` will always be blocked. 2, Repository key is the same in unit test methods, and value is persisted. See test log: ``` [INFO ] 2021-09-16 12:07:34.817 [main] o.a.s.s.c.a.i.GovernanceRepositoryAPIImplTest - assertWatch, key=/scaling/1 value=assertGetChildrenKeys [INFO ] 2021-09-16 12:07:34.919 [Curator-SafeNotifyService-0] o.a.s.s.c.a.i.GovernanceRepositoryAPIImplTest - event=DataChangedEvent(key=/scaling/1, value=assertWatch, type=ADDED) ``` In `CuratorZookeeperRepository.java` ``` private Type getChangedType(final TreeCacheEvent.Type type) { switch (type) { case NODE_ADDED: return Type.ADDED; case NODE_UPDATED: return Type.UPDATED; case NODE_REMOVED: return Type.DELETED; default: return Type.IGNORED; } } ``` TreeCacheEvent type depends on whether path is locally cached or not, it could be `ADDED` or `UPDATED`. ### Steps to reproduce the behavior, such as: SQL to execute, sharding rule configuration, when exception occur etc. Throw exception in `assertWatch`'s `watch` listener. ### Example codes for reproduce this issue (such as a github link).
non_process
governancerepositoryapiimpltest assertwatch may be blocked forever if exception thrown in watch listener bug report which version of shardingsphere did you use snapshot which project did you use shardingsphere jdbc or shardingsphere proxy shardingsphere scaling expected behavior unit test not blocked actual behavior assertthat event gettype is datachangedevent type added fail randomly unit test blocked from github ci log o a c f l mappinglistenermanager listener org apache curator framework recipes cache curatorcachelistenerbuilderimpl threw an exception java lang assertionerror expected is but was at org hamcrest matcherassert assertthat matcherassert java at org junit assert assertthat assert java at org junit assert assertthat assert java at org apache shardingsphere scaling core api impl governancerepositoryapiimpltest lambda assertwatch governancerepositoryapiimpltest java reason analyze if you can in governancerepositoryapiimpltest java test public void assertwatch throws interruptedexception countdownlatch countdownlatch new countdownlatch string key scalingconstant scaling root governancerepositoryapi watch scalingconstant scaling root event if event getkey equals key assertthat event gettype is datachangedevent type added countdownlatch countdown governancerepositoryapi persist key countdownlatch await if assertthat throw exception countdownlatch await will always be blocked repository key is the same in unit test methods and value is persisted see test log o a s s c a i governancerepositoryapiimpltest assertwatch key scaling value assertgetchildrenkeys o a s s c a i governancerepositoryapiimpltest event datachangedevent key scaling value assertwatch type added in curatorzookeeperrepository java private type getchangedtype final treecacheevent type type switch type case node added return type added case node updated return type updated case node removed return type deleted default return type ignored treecacheevent type depends on whether path is locally cached or not it could be added or updated steps to reproduce the behavior such as sql to execute sharding rule configuration when exception occur etc throw exception in assertwatch s watch listener example codes for reproduce this issue such as a github link
0
688,960
23,601,855,164
IssuesEvent
2022-08-24 03:20:05
ucb-rit/coldfront
https://api.github.com/repos/ucb-rit/coldfront
opened
Stand up REST API endpoints for serving Billing Activities
enhancement rest api high priority lrc-only
Currently, the primary key of the billing ID being used for Recharge for a particular `AllocationUser` may be fetched from `/api/allocation_users/{allocation_user_id}/attributes/?type=Billing Activity` (the `value` field). Add an REST API endpoint (e.g., `/api/billing_activities/{billing_activity_id}/` for retrieving a particular `BillingActivity`, which includes a field for the fully-formed billing ID (i.e., `full_id`).
1.0
Stand up REST API endpoints for serving Billing Activities - Currently, the primary key of the billing ID being used for Recharge for a particular `AllocationUser` may be fetched from `/api/allocation_users/{allocation_user_id}/attributes/?type=Billing Activity` (the `value` field). Add an REST API endpoint (e.g., `/api/billing_activities/{billing_activity_id}/` for retrieving a particular `BillingActivity`, which includes a field for the fully-formed billing ID (i.e., `full_id`).
non_process
stand up rest api endpoints for serving billing activities currently the primary key of the billing id being used for recharge for a particular allocationuser may be fetched from api allocation users allocation user id attributes type billing activity the value field add an rest api endpoint e g api billing activities billing activity id for retrieving a particular billingactivity which includes a field for the fully formed billing id i e full id
0
3,154
6,205,297,603
IssuesEvent
2017-07-06 15:51:29
pelias/model
https://api.github.com/repos/pelias/model
closed
Add ability to retrieve admin properties of a document
processed
There is currently no way to get admin properties of a document. One can only addParent. There are places where we've had to resort to checking the parent property of the document directly, which is less than ideal and should be changed as soon as this new getParent function is added. (see openstreetmap importer for example)
1.0
Add ability to retrieve admin properties of a document - There is currently no way to get admin properties of a document. One can only addParent. There are places where we've had to resort to checking the parent property of the document directly, which is less than ideal and should be changed as soon as this new getParent function is added. (see openstreetmap importer for example)
process
add ability to retrieve admin properties of a document there is currently no way to get admin properties of a document one can only addparent there are places where we ve had to resort to checking the parent property of the document directly which is less than ideal and should be changed as soon as this new getparent function is added see openstreetmap importer for example
1
308,000
9,428,529,307
IssuesEvent
2019-04-12 01:37:16
stuininga/Capstone_FoodBankApp
https://api.github.com/repos/stuininga/Capstone_FoodBankApp
closed
Create home page for Databases
top priority
Welcome user so and so, which database you're in, adds database navigation,
1.0
Create home page for Databases - Welcome user so and so, which database you're in, adds database navigation,
non_process
create home page for databases welcome user so and so which database you re in adds database navigation
0
115,876
9,815,869,227
IssuesEvent
2019-06-13 13:34:29
golang/go
https://api.github.com/repos/golang/go
reopened
Failing test for "net/http/cgi" on Ubuntu Xenial (16.04)
NeedsInvestigation Testing WaitingForInfo
Please answer these questions before submitting your issue. Thanks! ### What version of Go are you using (`go version`)? `go version go1.11 linux/amd64` ### Does this issue reproduce with the latest release? yes ### What operating system and processor architecture are you using (`go env`)? ``` GOARCH="amd64" GOBIN="" GOCACHE="/home/gvarouchas/.cache/go-build" GOEXE="" GOHOSTARCH="amd64" GOHOSTOS="linux" GOOS="linux" GOPATH="/home/gvarouchas/golib" GORACE="" GOROOT="/usr/lib/go-1.10" GOTMPDIR="" GOTOOLDIR="/usr/lib/go-1.10/pkg/tool/linux_amd64" GCCGO="gccgo" CC="gcc" CXX="g++" CGO_ENABLED="1" CGO_CFLAGS="-g -O2" CGO_CPPFLAGS="" CGO_CXXFLAGS="-g -O2" CGO_FFLAGS="-g -O2" CGO_LDFLAGS="-g -O2" PKG_CONFIG="pkg-config" GOGCCFLAGS="-fPIC -m64 -pthread -fmessage-length=0 -fdebug-prefix-map=/tmp/go-build117805775=/tmp/go-build -gno-record-gcc-switches" ``` Operating system is Ubuntu xenial 16.04.5 ### What did you do? From the go source directory, run the tests for "net/http/cgi" package ### What did you expect to see? $ ../bin/go test net/http/cgi ok net/http/cgi ### What did you see instead? ``` # "../bin/go" is the built go command : $ ../bin/go version go version go1.11 linux/amd64 $ ../bin/go test net/http/cgi 2018/09/21 10:20:51 cgi: copy error: write tcp 127.0.0.1:33176->127.0.0.1:47778: write: connection reset by peer /usr/bin/env: 'bash': No such file or directory 2018/09/21 10:20:51 cgi: no headers --- FAIL: TestEnvOverride (0.00s) host_test.go:84: for key "env-REQUEST_URI" got ""; expected "/foo/bar" host_test.go:84: for key "env-PATH" got ""; expected "/wibble" host_test.go:84: for key "cwd" got ""; expected "/home/gvarouchas/go/src/net/http/cgi" host_test.go:84: for key "env-SCRIPT_FILENAME" got ""; expected "/home/gvarouchas/go/src/net/http/cgi/testdata/test.cgi" 2018/09/21 10:20:51 cgi: copy error: past write limit 2018/09/21 10:20:51 cgi: no headers 2018/09/21 10:20:51 cgi: missing required Content-Type in headers 2018/09/21 10:20:51 cgi: no headers FAIL FAIL net/http/cgi 0.206s ``` I guess this issue is linked to the setup of my local machine (especially the "connection reset by peer" on a local connection), but I don't know where to start looking ?
1.0
Failing test for "net/http/cgi" on Ubuntu Xenial (16.04) - Please answer these questions before submitting your issue. Thanks! ### What version of Go are you using (`go version`)? `go version go1.11 linux/amd64` ### Does this issue reproduce with the latest release? yes ### What operating system and processor architecture are you using (`go env`)? ``` GOARCH="amd64" GOBIN="" GOCACHE="/home/gvarouchas/.cache/go-build" GOEXE="" GOHOSTARCH="amd64" GOHOSTOS="linux" GOOS="linux" GOPATH="/home/gvarouchas/golib" GORACE="" GOROOT="/usr/lib/go-1.10" GOTMPDIR="" GOTOOLDIR="/usr/lib/go-1.10/pkg/tool/linux_amd64" GCCGO="gccgo" CC="gcc" CXX="g++" CGO_ENABLED="1" CGO_CFLAGS="-g -O2" CGO_CPPFLAGS="" CGO_CXXFLAGS="-g -O2" CGO_FFLAGS="-g -O2" CGO_LDFLAGS="-g -O2" PKG_CONFIG="pkg-config" GOGCCFLAGS="-fPIC -m64 -pthread -fmessage-length=0 -fdebug-prefix-map=/tmp/go-build117805775=/tmp/go-build -gno-record-gcc-switches" ``` Operating system is Ubuntu xenial 16.04.5 ### What did you do? From the go source directory, run the tests for "net/http/cgi" package ### What did you expect to see? $ ../bin/go test net/http/cgi ok net/http/cgi ### What did you see instead? ``` # "../bin/go" is the built go command : $ ../bin/go version go version go1.11 linux/amd64 $ ../bin/go test net/http/cgi 2018/09/21 10:20:51 cgi: copy error: write tcp 127.0.0.1:33176->127.0.0.1:47778: write: connection reset by peer /usr/bin/env: 'bash': No such file or directory 2018/09/21 10:20:51 cgi: no headers --- FAIL: TestEnvOverride (0.00s) host_test.go:84: for key "env-REQUEST_URI" got ""; expected "/foo/bar" host_test.go:84: for key "env-PATH" got ""; expected "/wibble" host_test.go:84: for key "cwd" got ""; expected "/home/gvarouchas/go/src/net/http/cgi" host_test.go:84: for key "env-SCRIPT_FILENAME" got ""; expected "/home/gvarouchas/go/src/net/http/cgi/testdata/test.cgi" 2018/09/21 10:20:51 cgi: copy error: past write limit 2018/09/21 10:20:51 cgi: no headers 2018/09/21 10:20:51 cgi: missing required Content-Type in headers 2018/09/21 10:20:51 cgi: no headers FAIL FAIL net/http/cgi 0.206s ``` I guess this issue is linked to the setup of my local machine (especially the "connection reset by peer" on a local connection), but I don't know where to start looking ?
non_process
failing test for net http cgi on ubuntu xenial please answer these questions before submitting your issue thanks what version of go are you using go version go version linux does this issue reproduce with the latest release yes what operating system and processor architecture are you using go env goarch gobin gocache home gvarouchas cache go build goexe gohostarch gohostos linux goos linux gopath home gvarouchas golib gorace goroot usr lib go gotmpdir gotooldir usr lib go pkg tool linux gccgo gccgo cc gcc cxx g cgo enabled cgo cflags g cgo cppflags cgo cxxflags g cgo fflags g cgo ldflags g pkg config pkg config gogccflags fpic pthread fmessage length fdebug prefix map tmp go tmp go build gno record gcc switches operating system is ubuntu xenial what did you do from the go source directory run the tests for net http cgi package what did you expect to see bin go test net http cgi ok net http cgi what did you see instead bin go is the built go command bin go version go version linux bin go test net http cgi cgi copy error write tcp write connection reset by peer usr bin env bash no such file or directory cgi no headers fail testenvoverride host test go for key env request uri got expected foo bar host test go for key env path got expected wibble host test go for key cwd got expected home gvarouchas go src net http cgi host test go for key env script filename got expected home gvarouchas go src net http cgi testdata test cgi cgi copy error past write limit cgi no headers cgi missing required content type in headers cgi no headers fail fail net http cgi i guess this issue is linked to the setup of my local machine especially the connection reset by peer on a local connection but i don t know where to start looking
0
299,900
25,936,248,157
IssuesEvent
2022-12-16 14:24:48
DucTrann1310/FeedbackOnline
https://api.github.com/repos/DucTrann1310/FeedbackOnline
opened
[BugID_35]_FUNC_Quản lý topic_Tạo mới topic_Tạo mới thành công khi để trống trường Tên Topic
bug Open Integration Test Fun_Wrong Business logic Priority_Medium Severity_Medium
Precondition: Admin đang ở màn hình tạo mới Topic Step: 1. Để trống [Tên Topic] textbox 2. Click [Hoàn Tất] button Actual output: Topic vừa được tạo thành công, lưu trong DB table Topic và refresh grid Expected output: Hiển thị inline message màu đỏ "Bạn chưa điền Tên Topic!" phía dưới trường Tên Topic và set focus tại đó --------------- TestcaseID = 11 https://www.screencast.com/t/mnT50sTDhT
1.0
[BugID_35]_FUNC_Quản lý topic_Tạo mới topic_Tạo mới thành công khi để trống trường Tên Topic - Precondition: Admin đang ở màn hình tạo mới Topic Step: 1. Để trống [Tên Topic] textbox 2. Click [Hoàn Tất] button Actual output: Topic vừa được tạo thành công, lưu trong DB table Topic và refresh grid Expected output: Hiển thị inline message màu đỏ "Bạn chưa điền Tên Topic!" phía dưới trường Tên Topic và set focus tại đó --------------- TestcaseID = 11 https://www.screencast.com/t/mnT50sTDhT
non_process
func quản lý topic tạo mới topic tạo mới thành công khi để trống trường tên topic precondition admin đang ở màn hình tạo mới topic step để trống textbox click button actual output topic vừa được tạo thành công lưu trong db table topic và refresh grid expected output hiển thị inline message màu đỏ bạn chưa điền tên topic phía dưới trường tên topic và set focus tại đó testcaseid
0
5,221
3,536,002,408
IssuesEvent
2016-01-16 23:14:25
stedolan/jq
https://api.github.com/repos/stedolan/jq
closed
Cross-compile via Travis-CI too
build
It seems it's possible to cross-compile in Travis-CI builds. A few links: - http://stackoverflow.com/questions/28155308/how-to-cross-compile-with-mingw-on-linux-for-windows This one has a great answer. We need grab the Mingw pkgs [from a third-party repo](https://docs.travis-ci.com/user/installing-dependencies/#installing-packages-from-a-custom-apt-repository) (which I guess we can trust as far as CI goes), then use them in the Travis build. Ideally Travis would just provide these... - https://trac.osgeo.org/geos/ticket/659 - http://www.scriptscoop.net/t/d39c9eac38a2/how-to-cross-compile-with-mingw-on-linux-for-windows.html - http://sourceforge.net/p/mingw-w64/mailman/message/31846195/
1.0
Cross-compile via Travis-CI too - It seems it's possible to cross-compile in Travis-CI builds. A few links: - http://stackoverflow.com/questions/28155308/how-to-cross-compile-with-mingw-on-linux-for-windows This one has a great answer. We need grab the Mingw pkgs [from a third-party repo](https://docs.travis-ci.com/user/installing-dependencies/#installing-packages-from-a-custom-apt-repository) (which I guess we can trust as far as CI goes), then use them in the Travis build. Ideally Travis would just provide these... - https://trac.osgeo.org/geos/ticket/659 - http://www.scriptscoop.net/t/d39c9eac38a2/how-to-cross-compile-with-mingw-on-linux-for-windows.html - http://sourceforge.net/p/mingw-w64/mailman/message/31846195/
non_process
cross compile via travis ci too it seems it s possible to cross compile in travis ci builds a few links this one has a great answer we need grab the mingw pkgs which i guess we can trust as far as ci goes then use them in the travis build ideally travis would just provide these
0
33,768
9,205,176,046
IssuesEvent
2019-03-08 09:50:22
qissue-bot/QGIS
https://api.github.com/repos/qissue-bot/QGIS
closed
upgrade to titan but it won't start as I get a segmentation fault
Category: Build/Install Component: Affected QGIS version Component: Crashes QGIS or corrupts data Component: Easy fix? Component: Operating System Component: Pull Request or Patch supplied Component: Regression? Component: Resolution Priority: Low Project: QGIS Application Status: Closed Tracker: Bug report
--- Author Name: **bshcm03-yahoo-co-uk -** (bshcm03-yahoo-co-uk -) Original Redmine Issue: 573, https://issues.qgis.org/issues/573 Original Assignee: Magnus Homann --- I have upgraded to titan but QGIS will no longer start - I get a segmentation fault on start up. I was able to use the previous version (with intermittent errors). I have created (and attached) log files for configure, make and make install. I was getting a qt4 error but this was resolved after I chagned the qt4 installation directory from default to /usr/local/qt4 I ideally run qgis from within grass (grass 6.2.1) but cannot do this either. I was getting some make install warnings,but not errors. Do you have any thoughts? If you need any further information etc, please do not hesitate to ask. Thanks --- - [config.log](https://issues.qgis.org/attachments/download/1862/config.log) (bshcm03-yahoo-co-uk -) - [make.log](https://issues.qgis.org/attachments/download/1864/make.log) (anonymous -) - [make.install.log](https://issues.qgis.org/attachments/download/1863/make.install.log) (bshcm03-yahoo-co-uk -) - [config.2.log](https://issues.qgis.org/attachments/download/1859/config.2.log) (Magnus Homann) - [config.3.log](https://issues.qgis.org/attachments/download/1860/config.3.log) (bshcm03 -) - [bh.make.check.log](https://issues.qgis.org/attachments/download/1858/bh.make.check.log) (bshcm03 -) - [config.4.log](https://issues.qgis.org/attachments/download/1861/config.4.log) (bshcm03 -) - [qgis_error_8-2-07](https://issues.qgis.org/attachments/download/1865/qgis_error_8-2-07) (bshcm03 -)
1.0
upgrade to titan but it won't start as I get a segmentation fault - --- Author Name: **bshcm03-yahoo-co-uk -** (bshcm03-yahoo-co-uk -) Original Redmine Issue: 573, https://issues.qgis.org/issues/573 Original Assignee: Magnus Homann --- I have upgraded to titan but QGIS will no longer start - I get a segmentation fault on start up. I was able to use the previous version (with intermittent errors). I have created (and attached) log files for configure, make and make install. I was getting a qt4 error but this was resolved after I chagned the qt4 installation directory from default to /usr/local/qt4 I ideally run qgis from within grass (grass 6.2.1) but cannot do this either. I was getting some make install warnings,but not errors. Do you have any thoughts? If you need any further information etc, please do not hesitate to ask. Thanks --- - [config.log](https://issues.qgis.org/attachments/download/1862/config.log) (bshcm03-yahoo-co-uk -) - [make.log](https://issues.qgis.org/attachments/download/1864/make.log) (anonymous -) - [make.install.log](https://issues.qgis.org/attachments/download/1863/make.install.log) (bshcm03-yahoo-co-uk -) - [config.2.log](https://issues.qgis.org/attachments/download/1859/config.2.log) (Magnus Homann) - [config.3.log](https://issues.qgis.org/attachments/download/1860/config.3.log) (bshcm03 -) - [bh.make.check.log](https://issues.qgis.org/attachments/download/1858/bh.make.check.log) (bshcm03 -) - [config.4.log](https://issues.qgis.org/attachments/download/1861/config.4.log) (bshcm03 -) - [qgis_error_8-2-07](https://issues.qgis.org/attachments/download/1865/qgis_error_8-2-07) (bshcm03 -)
non_process
upgrade to titan but it won t start as i get a segmentation fault author name yahoo co uk yahoo co uk original redmine issue original assignee magnus homann i have upgraded to titan but qgis will no longer start i get a segmentation fault on start up i was able to use the previous version with intermittent errors i have created and attached log files for configure make and make install i was getting a error but this was resolved after i chagned the installation directory from default to usr local i ideally run qgis from within grass grass but cannot do this either i was getting some make install warnings but not errors do you have any thoughts if you need any further information etc please do not hesitate to ask thanks yahoo co uk anonymous yahoo co uk magnus homann
0
49,416
6,025,491,915
IssuesEvent
2017-06-08 08:50:40
healthlocker/healthlocker
https://api.github.com/repos/healthlocker/healthlocker
closed
Bug - Internal Error on Tracker Overview
bug please-test priority-2
New account created (not connected to SLaM) Added one entry to sleep tracker Added one entry to problem tracker Then clicked on tracker overview and got internal error page Using safari on iPhone.
1.0
Bug - Internal Error on Tracker Overview - New account created (not connected to SLaM) Added one entry to sleep tracker Added one entry to problem tracker Then clicked on tracker overview and got internal error page Using safari on iPhone.
non_process
bug internal error on tracker overview new account created not connected to slam added one entry to sleep tracker added one entry to problem tracker then clicked on tracker overview and got internal error page using safari on iphone
0
7,002
10,145,737,075
IssuesEvent
2019-08-05 05:53:12
GroceriStar/food-static-files-generator
https://api.github.com/repos/GroceriStar/food-static-files-generator
closed
move readData into filesystem
enhancement good first issue help wanted in-process
**Is your feature request related to a problem? Please describe.** we need to do it after #125 will be merged to master
1.0
move readData into filesystem - **Is your feature request related to a problem? Please describe.** we need to do it after #125 will be merged to master
process
move readdata into filesystem is your feature request related to a problem please describe we need to do it after will be merged to master
1
224,563
17,755,804,082
IssuesEvent
2021-08-28 18:33:00
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
ccl/benchccl/rttanalysisccl: BenchmarkAlterPrimaryRegion failed
C-test-failure O-robot branch-master
ccl/benchccl/rttanalysisccl.BenchmarkAlterPrimaryRegion [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3376361&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3376361&tab=artifacts#/) on master @ [93d95b97e53b11b9640a698535c563ca3ed5ba15](https://github.com/cockroachdb/cockroach/commits/93d95b97e53b11b9640a698535c563ca3ed5ba15): ``` BenchmarkAlterPrimaryRegion/alter_populated_database_alter_primary_region test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region651687629 test_log_scope.go:74: use -show-logs to present logs inline rtt_analysis_bench.go:131: -- test log scope end -- rtt_analysis_bench.go:54: -- test log scope end -- BenchmarkAlterTableLocality/alter_from_global_to_rbr-16 1 3283756600 ns/op 22.00 roundtrips BenchmarkAlterTableLocality/alter_from_regional_by_table_to_rbr test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr712218057 test_log_scope.go:74: use -show-logs to present logs inline test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterRegions_alter_populated_database_drop_region564512728 test_log_scope.go:74: use -show-logs to present logs inline rtt_analysis_bench.go:131: -- test log scope end -- rtt_analysis_bench.go:54: -- test log scope end -- BenchmarkAlterSurvivalGoals/alter_populated_database_from_zone_to_region-16 1 359332916 ns/op 41.00 roundtrips BenchmarkAlterSurvivalGoals/alter_populated_database_from_region_to_zone test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone141752125 test_log_scope.go:74: use -show-logs to present logs inline test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region085553020 test_log_scope.go:74: use -show-logs to present logs inline test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr356440840 test_log_scope.go:74: use -show-logs to present logs inline test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone849513004 test_log_scope.go:74: use -show-logs to present logs inline rtt_analysis_bench.go:131: -- test log scope end -- rtt_analysis_bench.go:54: -- test log scope end -- test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr215803079 test_log_scope.go:74: use -show-logs to present logs inline rtt_analysis_bench.go:131: -- test log scope end -- rtt_analysis_bench.go:54: -- test log scope end -- rtt_analysis_bench.go:131: -- test log scope end -- test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region053354155 test_log_scope.go:74: use -show-logs to present logs inline rtt_analysis_bench.go:54: -- test log scope end -- test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone938059163 test_log_scope.go:74: use -show-logs to present logs inline test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr701030222 test_log_scope.go:74: use -show-logs to present logs inline test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region037082402 test_log_scope.go:74: use -show-logs to present logs inline test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone648673362 test_log_scope.go:74: use -show-logs to present logs inline rtt_analysis_bench.go:131: -- test log scope end -- rtt_analysis_bench.go:54: -- test log scope end -- BenchmarkAlterRegions/alter_populated_database_drop_region-16 1 15275801616 ns/op 21.00 roundtrips BenchmarkAlterRegions/alter_populated_database_add_region test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterRegions_alter_populated_database_add_region041894743 test_log_scope.go:74: use -show-logs to present logs inline rtt_analysis_bench.go:131: -- test log scope end -- rtt_analysis_bench.go:54: -- test log scope end -- BenchmarkAlterPrimaryRegion/alter_populated_database_alter_primary_region-16 1 325849061 ns/op 21.00 roundtrips ``` <details><summary>Reproduce</summary> <p> To reproduce, try: ```bash make stressrace TESTS=BenchmarkAlterPrimaryRegion PKG=./pkg/ccl/benchccl/rttanalysisccl TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1 ``` Parameters in this failure: - GOFLAGS=-json </p> </details> /cc @cockroachdb/sql-experience pawalt <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*BenchmarkAlterPrimaryRegion.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
1.0
ccl/benchccl/rttanalysisccl: BenchmarkAlterPrimaryRegion failed - ccl/benchccl/rttanalysisccl.BenchmarkAlterPrimaryRegion [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3376361&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3376361&tab=artifacts#/) on master @ [93d95b97e53b11b9640a698535c563ca3ed5ba15](https://github.com/cockroachdb/cockroach/commits/93d95b97e53b11b9640a698535c563ca3ed5ba15): ``` BenchmarkAlterPrimaryRegion/alter_populated_database_alter_primary_region test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region651687629 test_log_scope.go:74: use -show-logs to present logs inline rtt_analysis_bench.go:131: -- test log scope end -- rtt_analysis_bench.go:54: -- test log scope end -- BenchmarkAlterTableLocality/alter_from_global_to_rbr-16 1 3283756600 ns/op 22.00 roundtrips BenchmarkAlterTableLocality/alter_from_regional_by_table_to_rbr test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr712218057 test_log_scope.go:74: use -show-logs to present logs inline test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterRegions_alter_populated_database_drop_region564512728 test_log_scope.go:74: use -show-logs to present logs inline rtt_analysis_bench.go:131: -- test log scope end -- rtt_analysis_bench.go:54: -- test log scope end -- BenchmarkAlterSurvivalGoals/alter_populated_database_from_zone_to_region-16 1 359332916 ns/op 41.00 roundtrips BenchmarkAlterSurvivalGoals/alter_populated_database_from_region_to_zone test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone141752125 test_log_scope.go:74: use -show-logs to present logs inline test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region085553020 test_log_scope.go:74: use -show-logs to present logs inline test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr356440840 test_log_scope.go:74: use -show-logs to present logs inline test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone849513004 test_log_scope.go:74: use -show-logs to present logs inline rtt_analysis_bench.go:131: -- test log scope end -- rtt_analysis_bench.go:54: -- test log scope end -- test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr215803079 test_log_scope.go:74: use -show-logs to present logs inline rtt_analysis_bench.go:131: -- test log scope end -- rtt_analysis_bench.go:54: -- test log scope end -- rtt_analysis_bench.go:131: -- test log scope end -- test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region053354155 test_log_scope.go:74: use -show-logs to present logs inline rtt_analysis_bench.go:54: -- test log scope end -- test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone938059163 test_log_scope.go:74: use -show-logs to present logs inline test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterTableLocality_alter_from_regional_by_table_to_rbr701030222 test_log_scope.go:74: use -show-logs to present logs inline test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterPrimaryRegion_alter_populated_database_alter_primary_region037082402 test_log_scope.go:74: use -show-logs to present logs inline test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterSurvivalGoals_alter_populated_database_from_region_to_zone648673362 test_log_scope.go:74: use -show-logs to present logs inline rtt_analysis_bench.go:131: -- test log scope end -- rtt_analysis_bench.go:54: -- test log scope end -- BenchmarkAlterRegions/alter_populated_database_drop_region-16 1 15275801616 ns/op 21.00 roundtrips BenchmarkAlterRegions/alter_populated_database_add_region test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logBenchmarkAlterRegions_alter_populated_database_add_region041894743 test_log_scope.go:74: use -show-logs to present logs inline rtt_analysis_bench.go:131: -- test log scope end -- rtt_analysis_bench.go:54: -- test log scope end -- BenchmarkAlterPrimaryRegion/alter_populated_database_alter_primary_region-16 1 325849061 ns/op 21.00 roundtrips ``` <details><summary>Reproduce</summary> <p> To reproduce, try: ```bash make stressrace TESTS=BenchmarkAlterPrimaryRegion PKG=./pkg/ccl/benchccl/rttanalysisccl TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1 ``` Parameters in this failure: - GOFLAGS=-json </p> </details> /cc @cockroachdb/sql-experience pawalt <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*BenchmarkAlterPrimaryRegion.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
non_process
ccl benchccl rttanalysisccl benchmarkalterprimaryregion failed ccl benchccl rttanalysisccl benchmarkalterprimaryregion with on master benchmarkalterprimaryregion alter populated database alter primary region test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkalterprimaryregion alter populated database alter primary test log scope go use show logs to present logs inline rtt analysis bench go test log scope end rtt analysis bench go test log scope end benchmarkaltertablelocality alter from global to rbr ns op roundtrips benchmarkaltertablelocality alter from regional by table to rbr test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltertablelocality alter from regional by table to test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkalterregions alter populated database drop test log scope go use show logs to present logs inline rtt analysis bench go test log scope end rtt analysis bench go test log scope end benchmarkaltersurvivalgoals alter populated database from zone to region ns op roundtrips benchmarkaltersurvivalgoals alter populated database from region to zone test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltersurvivalgoals alter populated database from region to test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkalterprimaryregion alter populated database alter primary test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltertablelocality alter from regional by table to test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltersurvivalgoals alter populated database from region to test log scope go use show logs to present logs inline rtt analysis bench go test log scope end rtt analysis bench go test log scope end test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltertablelocality alter from regional by table to test log scope go use show logs to present logs inline rtt analysis bench go test log scope end rtt analysis bench go test log scope end rtt analysis bench go test log scope end test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkalterprimaryregion alter populated database alter primary test log scope go use show logs to present logs inline rtt analysis bench go test log scope end test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltersurvivalgoals alter populated database from region to test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltertablelocality alter from regional by table to test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkalterprimaryregion alter populated database alter primary test log scope go use show logs to present logs inline test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkaltersurvivalgoals alter populated database from region to test log scope go use show logs to present logs inline rtt analysis bench go test log scope end rtt analysis bench go test log scope end benchmarkalterregions alter populated database drop region ns op roundtrips benchmarkalterregions alter populated database add region test log scope go test logs captured to go src github com cockroachdb cockroach artifacts logbenchmarkalterregions alter populated database add test log scope go use show logs to present logs inline rtt analysis bench go test log scope end rtt analysis bench go test log scope end benchmarkalterprimaryregion alter populated database alter primary region ns op roundtrips reproduce to reproduce try bash make stressrace tests benchmarkalterprimaryregion pkg pkg ccl benchccl rttanalysisccl testtimeout stressflags timeout parameters in this failure goflags json cc cockroachdb sql experience pawalt
0
77,742
14,915,092,671
IssuesEvent
2021-01-22 16:15:50
ArctosDB/arctos
https://api.github.com/repos/ArctosDB/arctos
closed
csv download of validate results for Bulk Edit Container
Function-ContainerOrBarcode Priority-Normal
Issue Documentation is http://handbook.arctosdb.org/how_to/How-to-Use-Issues-in-Arctos.html **Is your feature request related to a problem? Please describe.** A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] I sometimes have to bulk edit a large container series (thousands of containers) & if validation fails I have had trouble finding the offending barcode in my web browser. can we get a csv download option of those validation results so we can more easily find the offending barcodes? **Describe the solution you'd like** A clear and concise description of what you want to happen. csv download validation results after using the bulk edit container tool in object tracking **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. Fail to find the problem but can't proceed, email dusty... **Additional context** Add any other context or screenshots about the feature request here. **Priority** Please assign a priority-label.
1.0
csv download of validate results for Bulk Edit Container - Issue Documentation is http://handbook.arctosdb.org/how_to/How-to-Use-Issues-in-Arctos.html **Is your feature request related to a problem? Please describe.** A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] I sometimes have to bulk edit a large container series (thousands of containers) & if validation fails I have had trouble finding the offending barcode in my web browser. can we get a csv download option of those validation results so we can more easily find the offending barcodes? **Describe the solution you'd like** A clear and concise description of what you want to happen. csv download validation results after using the bulk edit container tool in object tracking **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. Fail to find the problem but can't proceed, email dusty... **Additional context** Add any other context or screenshots about the feature request here. **Priority** Please assign a priority-label.
non_process
csv download of validate results for bulk edit container issue documentation is is your feature request related to a problem please describe a clear and concise description of what the problem is ex i m always frustrated when i sometimes have to bulk edit a large container series thousands of containers if validation fails i have had trouble finding the offending barcode in my web browser can we get a csv download option of those validation results so we can more easily find the offending barcodes describe the solution you d like a clear and concise description of what you want to happen csv download validation results after using the bulk edit container tool in object tracking describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered fail to find the problem but can t proceed email dusty additional context add any other context or screenshots about the feature request here priority please assign a priority label
0
3,462
6,545,260,571
IssuesEvent
2017-09-04 03:07:14
w3c/payment-request
https://api.github.com/repos/w3c/payment-request
closed
CR check list
Process aid
For all Transition Requests, to advance a specification to a new maturity level other than Note, the Working Group: * [x] must [record the group's decision to request advancement](https://lists.w3.org/Archives/Public/public-payments-wg/2017Jul/0057.html). * [ ] must obtain Director approval. * [x] must provide [public documentation of all substantive changes to the technical report since the previous publication](https://github.com/w3c/browser-payment-api/pulls?q=is%3Apr+is%3Aclosed+label%3Acr-substantive). * [ ] must [formally address all issues raised about the document since the previous maturity level](https://github.com/w3c/browser-payment-api/milestone/8). * [x] must provide public documentation of any Formal Objections (none received). * [x] should provide public documentation of [changes that are not substantive](https://github.com/w3c/browser-payment-api/commits/gh-pages). * [x] should report which, if any, of the Working Group's requirements for this document have changed since the previous step. * [x] should report any changes in dependencies with other groups. * [x] should provide [information about implementations known to the Working Group](http://caniuse.com/#feat=payment-request). To publish a Candidate recommendation, in addition to meeting the general requirements for advancement a Working Group: * [x] must show that the specification has met all Working Group requirements, or explain why the requirements have changed or been deferred, * [x] must document changes to dependencies during the development of the specification, * [x] must document how adequate implementation experience will be demonstrated, * [x] must specify the deadline for comments, which must be at least four weeks after publication, and should be longer for complex documents, * [x] must show that the specification has received wide review, and * [x] may identify features in the document as "at risk". These features may be removed before advancement to Proposed Recommendation without a requirement to publish a new Candidate Recommendation.
1.0
CR check list - For all Transition Requests, to advance a specification to a new maturity level other than Note, the Working Group: * [x] must [record the group's decision to request advancement](https://lists.w3.org/Archives/Public/public-payments-wg/2017Jul/0057.html). * [ ] must obtain Director approval. * [x] must provide [public documentation of all substantive changes to the technical report since the previous publication](https://github.com/w3c/browser-payment-api/pulls?q=is%3Apr+is%3Aclosed+label%3Acr-substantive). * [ ] must [formally address all issues raised about the document since the previous maturity level](https://github.com/w3c/browser-payment-api/milestone/8). * [x] must provide public documentation of any Formal Objections (none received). * [x] should provide public documentation of [changes that are not substantive](https://github.com/w3c/browser-payment-api/commits/gh-pages). * [x] should report which, if any, of the Working Group's requirements for this document have changed since the previous step. * [x] should report any changes in dependencies with other groups. * [x] should provide [information about implementations known to the Working Group](http://caniuse.com/#feat=payment-request). To publish a Candidate recommendation, in addition to meeting the general requirements for advancement a Working Group: * [x] must show that the specification has met all Working Group requirements, or explain why the requirements have changed or been deferred, * [x] must document changes to dependencies during the development of the specification, * [x] must document how adequate implementation experience will be demonstrated, * [x] must specify the deadline for comments, which must be at least four weeks after publication, and should be longer for complex documents, * [x] must show that the specification has received wide review, and * [x] may identify features in the document as "at risk". These features may be removed before advancement to Proposed Recommendation without a requirement to publish a new Candidate Recommendation.
process
cr check list for all transition requests to advance a specification to a new maturity level other than note the working group must must obtain director approval must provide must must provide public documentation of any formal objections none received should provide public documentation of should report which if any of the working group s requirements for this document have changed since the previous step should report any changes in dependencies with other groups should provide to publish a candidate recommendation in addition to meeting the general requirements for advancement a working group must show that the specification has met all working group requirements or explain why the requirements have changed or been deferred must document changes to dependencies during the development of the specification must document how adequate implementation experience will be demonstrated must specify the deadline for comments which must be at least four weeks after publication and should be longer for complex documents must show that the specification has received wide review and may identify features in the document as at risk these features may be removed before advancement to proposed recommendation without a requirement to publish a new candidate recommendation
1
7,741
10,862,664,483
IssuesEvent
2019-11-14 13:45:14
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Processing: python error on startup
Bug Feedback High Priority Processing Regression
Author Name: **John Drewett** (John Drewett) Original Redmine Issue: [21720](https://issues.qgis.org/issues/21720) Affected QGIS version: 3.6.1 Redmine category:processing/saga --- The following error occurs on start up. Any advice would be appreciated. ``` 2019-04-01T13:25:00 WARNING Traceback (most recent call last): File "C:/PROGRA~1/QGIS3~1.6/apps/qgis/./python/plugins\processing\algs\saga\SagaAlgorithmProvider.py", line 90, in loadAlgorithms version = SagaUtils.getInstalledVersion(True) File "C:/PROGRA~1/QGIS3~1.6/apps/qgis/./python/plugins\processing\algs\saga\SagaUtils.py", line 145, in getInstalledVersion universal_newlines=True, File "C:\PROGRA~1\QGIS3~1.6\apps\Python37\lib\subprocess.py", line 709, in __init__ errread, errwrite) = self._get_handles(stdin, stdout, stderr) File "C:\PROGRA~1\QGIS3~1.6\apps\Python37\lib\subprocess.py", line 1015, in _get_handles p2cread = msvcrt.get_osfhandle(self._get_devnull()) File "C:\PROGRA~1\QGIS3~1.6\apps\Python37\lib\subprocess.py", line 848, in _get_devnull self._devnull = os.open(os.devnull, os.O_RDWR) FileNotFoundError: [Errno 2] No such file or directory: 'nul' ```
1.0
Processing: python error on startup - Author Name: **John Drewett** (John Drewett) Original Redmine Issue: [21720](https://issues.qgis.org/issues/21720) Affected QGIS version: 3.6.1 Redmine category:processing/saga --- The following error occurs on start up. Any advice would be appreciated. ``` 2019-04-01T13:25:00 WARNING Traceback (most recent call last): File "C:/PROGRA~1/QGIS3~1.6/apps/qgis/./python/plugins\processing\algs\saga\SagaAlgorithmProvider.py", line 90, in loadAlgorithms version = SagaUtils.getInstalledVersion(True) File "C:/PROGRA~1/QGIS3~1.6/apps/qgis/./python/plugins\processing\algs\saga\SagaUtils.py", line 145, in getInstalledVersion universal_newlines=True, File "C:\PROGRA~1\QGIS3~1.6\apps\Python37\lib\subprocess.py", line 709, in __init__ errread, errwrite) = self._get_handles(stdin, stdout, stderr) File "C:\PROGRA~1\QGIS3~1.6\apps\Python37\lib\subprocess.py", line 1015, in _get_handles p2cread = msvcrt.get_osfhandle(self._get_devnull()) File "C:\PROGRA~1\QGIS3~1.6\apps\Python37\lib\subprocess.py", line 848, in _get_devnull self._devnull = os.open(os.devnull, os.O_RDWR) FileNotFoundError: [Errno 2] No such file or directory: 'nul' ```
process
processing python error on startup author name john drewett john drewett original redmine issue affected qgis version redmine category processing saga the following error occurs on start up any advice would be appreciated warning traceback most recent call last file c progra apps qgis python plugins processing algs saga sagaalgorithmprovider py line in loadalgorithms version sagautils getinstalledversion true file c progra apps qgis python plugins processing algs saga sagautils py line in getinstalledversion universal newlines true file c progra apps lib subprocess py line in init errread errwrite self get handles stdin stdout stderr file c progra apps lib subprocess py line in get handles msvcrt get osfhandle self get devnull file c progra apps lib subprocess py line in get devnull self devnull os open os devnull os o rdwr filenotfounderror no such file or directory nul
1
142,272
13,019,035,714
IssuesEvent
2020-07-26 20:21:18
Vyctor661/king-typer
https://api.github.com/repos/Vyctor661/king-typer
closed
docs
documentation
## Needed docs #### Frontend - [ ] home page - [ ] start typing page 1. easy mode 2. normal mode 3. tutorials mode 4. multiplayer modes (easy normal custom) - [ ] loging register forgot password page - [ ] profile pages - [ ] dashboard - [ ] themes - [ ] recent games/statistics #### Backend
1.0
docs - ## Needed docs #### Frontend - [ ] home page - [ ] start typing page 1. easy mode 2. normal mode 3. tutorials mode 4. multiplayer modes (easy normal custom) - [ ] loging register forgot password page - [ ] profile pages - [ ] dashboard - [ ] themes - [ ] recent games/statistics #### Backend
non_process
docs needed docs frontend home page start typing page easy mode normal mode tutorials mode multiplayer modes easy normal custom loging register forgot password page profile pages dashboard themes recent games statistics backend
0
708,829
24,356,547,957
IssuesEvent
2022-10-03 08:00:54
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
weather.com - video or audio doesn't play
priority-important browser-focus-geckoview engine-gecko
<!-- @browser: Firefox Mobile 105.0 --> <!-- @ua_header: Mozilla/5.0 (Android 12; Mobile; rv:105.0) Gecko/105.0 Firefox/105.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/111743 --> <!-- @extra_labels: browser-focus-geckoview --> **URL**: https://weather.com/news/weather/video/pattern-change-brings-cooler-air-to-northeast-warm-weather-to-west?par=samsung_widget **Browser / Version**: Firefox Mobile 105.0 **Operating System**: Android 12 **Tested Another Browser**: Yes Chrome **Problem type**: Video or audio doesn't play **Description**: The video or audio does not play **Steps to Reproduce**: Video buttons unresponsive in firefox android, works on chrome mobile <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2022/10/77ebe704-6eef-4dd0-ae4e-c715fc92a506.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220915150737</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2022/10/ddba324c-ed42-41a6-98c3-255eb5e50dad) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
weather.com - video or audio doesn't play - <!-- @browser: Firefox Mobile 105.0 --> <!-- @ua_header: Mozilla/5.0 (Android 12; Mobile; rv:105.0) Gecko/105.0 Firefox/105.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/111743 --> <!-- @extra_labels: browser-focus-geckoview --> **URL**: https://weather.com/news/weather/video/pattern-change-brings-cooler-air-to-northeast-warm-weather-to-west?par=samsung_widget **Browser / Version**: Firefox Mobile 105.0 **Operating System**: Android 12 **Tested Another Browser**: Yes Chrome **Problem type**: Video or audio doesn't play **Description**: The video or audio does not play **Steps to Reproduce**: Video buttons unresponsive in firefox android, works on chrome mobile <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2022/10/77ebe704-6eef-4dd0-ae4e-c715fc92a506.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220915150737</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2022/10/ddba324c-ed42-41a6-98c3-255eb5e50dad) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
weather com video or audio doesn t play url browser version firefox mobile operating system android tested another browser yes chrome problem type video or audio doesn t play description the video or audio does not play steps to reproduce video buttons unresponsive in firefox android works on chrome mobile view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel release hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
113,241
9,633,378,778
IssuesEvent
2019-05-15 18:29:24
pvlib/pvlib-python
https://api.github.com/repos/pvlib/pvlib-python
closed
requests package requirement not listed
bug installation testing
**Describe the bug** `import pvlib` fails due to unspecified `requests` dependency: ``` Traceback (most recent call last): File "<string>", line 1, in <module> File "/Users/holmgren/miniconda3/envs/pvlibtest/lib/python3.7/site-packages/pvlib/__init__.py", line 9, in <module> from pvlib import iotools File "/Users/holmgren/miniconda3/envs/pvlibtest/lib/python3.7/site-packages/pvlib/iotools/__init__.py", line 13, in <module> from pvlib.iotools.psm3 import get_psm3 # noqa: F401 File "/Users/holmgren/miniconda3/envs/pvlibtest/lib/python3.7/site-packages/pvlib/iotools/psm3.py", line 8, in <module> import requests ModuleNotFoundError: No module named 'requests' ``` **To Reproduce** Steps to reproduce the behavior: 1. `conda create -n pvlibtest python=3.7` 2. `pip install pvlib` 3. `python -c import pvlib` **Expected behavior** `pip install pvlib` works from a clean environment. **Versions:** - ``pvlib.__version__``: 0.6.2 - ``pandas.__version__``: 0.24.2 - python: 3.7 **Additional context** #694 added a dependency on requests. The CI environments pull in requests through some of their dependencies, so it wasn't caught there. Should make a new issue for adding CI tests for pip installs into a clean environment. 0.6.3 coming up...
1.0
requests package requirement not listed - **Describe the bug** `import pvlib` fails due to unspecified `requests` dependency: ``` Traceback (most recent call last): File "<string>", line 1, in <module> File "/Users/holmgren/miniconda3/envs/pvlibtest/lib/python3.7/site-packages/pvlib/__init__.py", line 9, in <module> from pvlib import iotools File "/Users/holmgren/miniconda3/envs/pvlibtest/lib/python3.7/site-packages/pvlib/iotools/__init__.py", line 13, in <module> from pvlib.iotools.psm3 import get_psm3 # noqa: F401 File "/Users/holmgren/miniconda3/envs/pvlibtest/lib/python3.7/site-packages/pvlib/iotools/psm3.py", line 8, in <module> import requests ModuleNotFoundError: No module named 'requests' ``` **To Reproduce** Steps to reproduce the behavior: 1. `conda create -n pvlibtest python=3.7` 2. `pip install pvlib` 3. `python -c import pvlib` **Expected behavior** `pip install pvlib` works from a clean environment. **Versions:** - ``pvlib.__version__``: 0.6.2 - ``pandas.__version__``: 0.24.2 - python: 3.7 **Additional context** #694 added a dependency on requests. The CI environments pull in requests through some of their dependencies, so it wasn't caught there. Should make a new issue for adding CI tests for pip installs into a clean environment. 0.6.3 coming up...
non_process
requests package requirement not listed describe the bug import pvlib fails due to unspecified requests dependency traceback most recent call last file line in file users holmgren envs pvlibtest lib site packages pvlib init py line in from pvlib import iotools file users holmgren envs pvlibtest lib site packages pvlib iotools init py line in from pvlib iotools import get noqa file users holmgren envs pvlibtest lib site packages pvlib iotools py line in import requests modulenotfounderror no module named requests to reproduce steps to reproduce the behavior conda create n pvlibtest python pip install pvlib python c import pvlib expected behavior pip install pvlib works from a clean environment versions pvlib version pandas version python additional context added a dependency on requests the ci environments pull in requests through some of their dependencies so it wasn t caught there should make a new issue for adding ci tests for pip installs into a clean environment coming up
0
22,654
31,895,827,910
IssuesEvent
2023-09-18 01:31:59
tdwg/dwc
https://api.github.com/repos/tdwg/dwc
closed
Change term - latestEraOrHighestErathem
Term - change Class - GeologicalContext normative Task Group - Material Sample Process - complete
## Term change * Submitter: [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/) * Efficacy Justification (why is this change necessary?): Create consistency of terms for material in Darwin Core. * Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/), which includes representatives of over 10 organizations. * Stability Justification (what concerns are there that this might affect existing implementations?): None * Implications for dwciri: namespace (does this change affect a dwciri term version)?: No Current Term definition: https://dwc.tdwg.org/list/#dwc_latestEraOrHighestErathem Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~): * Term name (in lowerCamelCase for properties, UpperCamelCase for classes): latestEraOrHighestErathem * Term label (English, not normative): Latest Era Or Highest Erathem * * Organized in Class (e.g., Occurrence, Event, Location, Taxon): Geological Context * Definition of the term (normative): The full name of the latest possible geochronologic era or highest chronostratigraphic erathem attributable to the stratigraphic horizon from which the ~~cataloged item~~**dwc:MaterialEntity** was collected. * Usage comments (recommendations regarding content, etc., not normative): * Examples (not normative): Cenozoic, Mesozoic * Refines (identifier of the broader term this term refines; normative): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): not in ABCD
1.0
Change term - latestEraOrHighestErathem - ## Term change * Submitter: [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/) * Efficacy Justification (why is this change necessary?): Create consistency of terms for material in Darwin Core. * Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/), which includes representatives of over 10 organizations. * Stability Justification (what concerns are there that this might affect existing implementations?): None * Implications for dwciri: namespace (does this change affect a dwciri term version)?: No Current Term definition: https://dwc.tdwg.org/list/#dwc_latestEraOrHighestErathem Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~): * Term name (in lowerCamelCase for properties, UpperCamelCase for classes): latestEraOrHighestErathem * Term label (English, not normative): Latest Era Or Highest Erathem * * Organized in Class (e.g., Occurrence, Event, Location, Taxon): Geological Context * Definition of the term (normative): The full name of the latest possible geochronologic era or highest chronostratigraphic erathem attributable to the stratigraphic horizon from which the ~~cataloged item~~**dwc:MaterialEntity** was collected. * Usage comments (recommendations regarding content, etc., not normative): * Examples (not normative): Cenozoic, Mesozoic * Refines (identifier of the broader term this term refines; normative): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): not in ABCD
process
change term latesteraorhighesterathem term change submitter efficacy justification why is this change necessary create consistency of terms for material in darwin core demand justification if the change is semantic in nature name at least two organizations that independently need this term which includes representatives of over organizations stability justification what concerns are there that this might affect existing implementations none implications for dwciri namespace does this change affect a dwciri term version no current term definition proposed attributes of the new term version please put actual changes to be implemented in bold and strikethrough term name in lowercamelcase for properties uppercamelcase for classes latesteraorhighesterathem term label english not normative latest era or highest erathem organized in class e g occurrence event location taxon geological context definition of the term normative the full name of the latest possible geochronologic era or highest chronostratigraphic erathem attributable to the stratigraphic horizon from which the cataloged item dwc materialentity was collected usage comments recommendations regarding content etc not normative examples not normative cenozoic mesozoic refines identifier of the broader term this term refines normative none replaces identifier of the existing term that would be deprecated and replaced by this term normative none abcd xpath of the equivalent term in abcd or efg not normative not in abcd
1
10,660
13,453,144,059
IssuesEvent
2020-09-09 00:04:24
GoogleCloudPlatform/cloud-ops-sandbox
https://api.github.com/repos/GoogleCloudPlatform/cloud-ops-sandbox
closed
[telemetry] Add python subscriber script to report from project to Pub/Sub in GCP
lang: python priority: p2 type: process
In order to learn how frequently the installation script succeeds, and how often users run destroy script, we should add a subscriber script to report information to Pub / Sub in GCP.
1.0
[telemetry] Add python subscriber script to report from project to Pub/Sub in GCP - In order to learn how frequently the installation script succeeds, and how often users run destroy script, we should add a subscriber script to report information to Pub / Sub in GCP.
process
add python subscriber script to report from project to pub sub in gcp in order to learn how frequently the installation script succeeds and how often users run destroy script we should add a subscriber script to report information to pub sub in gcp
1
8,281
11,439,587,808
IssuesEvent
2020-02-05 07:41:20
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
opened
Test against latest stable browser versions (open new PR to do so)
process: tests type: chore
### Current behavior: We recently implemented this update to our internal testing process to *always* test against the latest browser version: https://github.com/cypress-io/cypress/pull/6115 This has become problematic, as evidenced by Chrome 80 update that occurred Feb 4. We are trying to release 4.0 and in the midst of this release the tests are not passing due to some change in Chrome 80. This makes it difficult to isolate which test failures are coming from the current branch changes and which are coming from the new Chrome version changes. Furthermore this issue will just compound as we add more browser support. ### Desired behavior: We should do something similar to how our `renovatebot` works today for our repo. Upon release of a new stable browser version: - Create a new docker image with new Chrome version? - Initiate a PR against `cypress` that runs all tests against the latest browser version. Then we can take the time to isolate which tests failed due to the new stable release and track them down. ### Versions Cypress 3.8.3
1.0
Test against latest stable browser versions (open new PR to do so) - ### Current behavior: We recently implemented this update to our internal testing process to *always* test against the latest browser version: https://github.com/cypress-io/cypress/pull/6115 This has become problematic, as evidenced by Chrome 80 update that occurred Feb 4. We are trying to release 4.0 and in the midst of this release the tests are not passing due to some change in Chrome 80. This makes it difficult to isolate which test failures are coming from the current branch changes and which are coming from the new Chrome version changes. Furthermore this issue will just compound as we add more browser support. ### Desired behavior: We should do something similar to how our `renovatebot` works today for our repo. Upon release of a new stable browser version: - Create a new docker image with new Chrome version? - Initiate a PR against `cypress` that runs all tests against the latest browser version. Then we can take the time to isolate which tests failed due to the new stable release and track them down. ### Versions Cypress 3.8.3
process
test against latest stable browser versions open new pr to do so current behavior we recently implemented this update to our internal testing process to always test against the latest browser version this has become problematic as evidenced by chrome update that occurred feb we are trying to release and in the midst of this release the tests are not passing due to some change in chrome this makes it difficult to isolate which test failures are coming from the current branch changes and which are coming from the new chrome version changes furthermore this issue will just compound as we add more browser support desired behavior we should do something similar to how our renovatebot works today for our repo upon release of a new stable browser version create a new docker image with new chrome version initiate a pr against cypress that runs all tests against the latest browser version then we can take the time to isolate which tests failed due to the new stable release and track them down versions cypress
1
283,286
21,316,086,379
IssuesEvent
2022-04-16 09:49:51
ElaineQT/pe
https://api.github.com/repos/ElaineQT/pe
opened
Text format in DG
type.DocumentationBug severity.VeryLow
Regarding DG, is it better to include all class names and methods etc. in " ` "? E.g. This is done in `SearchCommandParser`, rather than ![image.png](https://raw.githubusercontent.com/ElaineQT/pe/main/files/1bf59f12-3a4c-408c-9795-a2f9be1ccf2c.png) Just like in DG of AddressBook 3: ![image.png](https://raw.githubusercontent.com/ElaineQT/pe/main/files/c087d5b9-7e7f-4312-b759-a886b584d4f8.png) <!--session: 1650096021327-6753a3ca-01e0-4b98-8a9f-295e17840146--> <!--Version: Web v3.4.2-->
1.0
Text format in DG - Regarding DG, is it better to include all class names and methods etc. in " ` "? E.g. This is done in `SearchCommandParser`, rather than ![image.png](https://raw.githubusercontent.com/ElaineQT/pe/main/files/1bf59f12-3a4c-408c-9795-a2f9be1ccf2c.png) Just like in DG of AddressBook 3: ![image.png](https://raw.githubusercontent.com/ElaineQT/pe/main/files/c087d5b9-7e7f-4312-b759-a886b584d4f8.png) <!--session: 1650096021327-6753a3ca-01e0-4b98-8a9f-295e17840146--> <!--Version: Web v3.4.2-->
non_process
text format in dg regarding dg is it better to include all class names and methods etc in e g this is done in searchcommandparser rather than just like in dg of addressbook
0
134
2,573,115,852
IssuesEvent
2015-02-11 06:05:38
sysown/proxysql-0.2
https://api.github.com/repos/sysown/proxysql-0.2
opened
Implement global variable mysql_connect_timeout_server_error
CONNECTION POOL cxx_pa GLOBAL MYSQL PROTOCOL QUERY PROCESSOR
Related to global variable mysql_connect_timeout_server ( issue #202) In case it is unable to connect to a backend to for mysql_connect_timeout_server , return the specified error to the client or disconnect it
1.0
Implement global variable mysql_connect_timeout_server_error - Related to global variable mysql_connect_timeout_server ( issue #202) In case it is unable to connect to a backend to for mysql_connect_timeout_server , return the specified error to the client or disconnect it
process
implement global variable mysql connect timeout server error related to global variable mysql connect timeout server issue in case it is unable to connect to a backend to for mysql connect timeout server return the specified error to the client or disconnect it
1
59,877
17,023,275,867
IssuesEvent
2021-07-03 01:10:52
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
aerialway=drag_lift is not on the mapnik map..
Component: mapnik Priority: minor Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 9.07am, Tuesday, 22nd July 2008]** The official map feature "aerialway=cable_car" isn't rendered by mapnik. But in the wiki is written, that it should be rendered. Could you please fix this. Here in Switzerland we have regions wich are only reachable over such aerialways.. link to the map features: http://wiki.openstreetmap.org/index.php/Map_features#Aerialway link to the problem http://openstreetmap.org/?lat=46.89994&lon=8.50536&zoom=15&layers=B00FTF thank you
1.0
aerialway=drag_lift is not on the mapnik map.. - **[Submitted to the original trac issue database at 9.07am, Tuesday, 22nd July 2008]** The official map feature "aerialway=cable_car" isn't rendered by mapnik. But in the wiki is written, that it should be rendered. Could you please fix this. Here in Switzerland we have regions wich are only reachable over such aerialways.. link to the map features: http://wiki.openstreetmap.org/index.php/Map_features#Aerialway link to the problem http://openstreetmap.org/?lat=46.89994&lon=8.50536&zoom=15&layers=B00FTF thank you
non_process
aerialway drag lift is not on the mapnik map the official map feature aerialway cable car isn t rendered by mapnik but in the wiki is written that it should be rendered could you please fix this here in switzerland we have regions wich are only reachable over such aerialways link to the map features link to the problem thank you
0
583
3,060,127,967
IssuesEvent
2015-08-14 18:50:41
Microsoft/poshtools
https://api.github.com/repos/Microsoft/poshtools
closed
Port Support Does Not Always Work
Process Attaching
Using ports does not always seem to work. It appears that we are parsing out the port incorrectly.
1.0
Port Support Does Not Always Work - Using ports does not always seem to work. It appears that we are parsing out the port incorrectly.
process
port support does not always work using ports does not always seem to work it appears that we are parsing out the port incorrectly
1
80,069
9,977,339,216
IssuesEvent
2019-07-09 17:01:07
mozilla/foundation.mozilla.org
https://api.github.com/repos/mozilla/foundation.mozilla.org
closed
Add more context to homepage
ui design ux design
Concerns that our homepage doesn't give new users enough context about who we are and why they should care has surfaced in two different ways recently. #### From @beccaklam's users testing https://github.com/mozilla/foundation.mozilla.org/issues/2995#issuecomment-491035740: - We should consider users who don't know our issues/Mozilla well and think about how to onboard them on our homepage if we ever want to grow our audience - We should be aware of two types of audience -- those who know our cause or are internet-savvy and those who don't know us or are not internet-savvy. We don't want to talk down to people who feel that they already know their stuff. #### From Juan and Lotta around fundraising [(doc from meeting)](https://docs.google.com/document/d/1txC3guawnjoeXUvFbFABhR32kPNDtx2O0MKGP8OzTJs/edit#): - Going to the site it’s hard to say “What Mozilla does”. This should frame the rest of the content about projects, news, etc. - What is a short pithy way to describe what we do that is engaging and captivating? This should be a big piece of the landing page - Everything we present should be about the “bigger plan” - Needs to make donors be a part of the work we do so that they don’t feel like they’re just giving us money to do work but they’re giving us money to do work that they care about and are contributing to - Landing page doesn’t talk about why they should participate - only how to participate (e.g. Apply to be a Fellow). We need to be able to tell our broader story better on this page. - Where do we talk about what our plan is for a healthier internet? We should be selling that plan/vision. I.e. resonating with audiences with emotion, a story, rather than specific programs we run out of context. Answering “what is Mozilla about” should be the first touchpoint for people - Juan just links people to the fo’mo website so the first touchpoint is the landing page. - News should tell the story of our impact - Projects need more context so that the broader picture is still described using it #### We should: - benchmark similar sites: a few examples but should gather more - https://colorofchange.org/ - https://rebellion.earth/ - https://www.astraeafoundation.org/global-reach/ - http://www.freedomforall.org - http://sfaf.org/ - https://www.fordfoundation.org/ - https://www.shuttleworthfoundation.org/ - write up a few recommendations (some quick fixes some long-term) - one of those recommendations should be to add an intro sentence (who should write this?) - mock-up a few options - get feedback from staff - test with different audiences cc: @sabrinang @natalieworth @beccaklam @xmatthewx @anilkanji
2.0
Add more context to homepage - Concerns that our homepage doesn't give new users enough context about who we are and why they should care has surfaced in two different ways recently. #### From @beccaklam's users testing https://github.com/mozilla/foundation.mozilla.org/issues/2995#issuecomment-491035740: - We should consider users who don't know our issues/Mozilla well and think about how to onboard them on our homepage if we ever want to grow our audience - We should be aware of two types of audience -- those who know our cause or are internet-savvy and those who don't know us or are not internet-savvy. We don't want to talk down to people who feel that they already know their stuff. #### From Juan and Lotta around fundraising [(doc from meeting)](https://docs.google.com/document/d/1txC3guawnjoeXUvFbFABhR32kPNDtx2O0MKGP8OzTJs/edit#): - Going to the site it’s hard to say “What Mozilla does”. This should frame the rest of the content about projects, news, etc. - What is a short pithy way to describe what we do that is engaging and captivating? This should be a big piece of the landing page - Everything we present should be about the “bigger plan” - Needs to make donors be a part of the work we do so that they don’t feel like they’re just giving us money to do work but they’re giving us money to do work that they care about and are contributing to - Landing page doesn’t talk about why they should participate - only how to participate (e.g. Apply to be a Fellow). We need to be able to tell our broader story better on this page. - Where do we talk about what our plan is for a healthier internet? We should be selling that plan/vision. I.e. resonating with audiences with emotion, a story, rather than specific programs we run out of context. Answering “what is Mozilla about” should be the first touchpoint for people - Juan just links people to the fo’mo website so the first touchpoint is the landing page. - News should tell the story of our impact - Projects need more context so that the broader picture is still described using it #### We should: - benchmark similar sites: a few examples but should gather more - https://colorofchange.org/ - https://rebellion.earth/ - https://www.astraeafoundation.org/global-reach/ - http://www.freedomforall.org - http://sfaf.org/ - https://www.fordfoundation.org/ - https://www.shuttleworthfoundation.org/ - write up a few recommendations (some quick fixes some long-term) - one of those recommendations should be to add an intro sentence (who should write this?) - mock-up a few options - get feedback from staff - test with different audiences cc: @sabrinang @natalieworth @beccaklam @xmatthewx @anilkanji
non_process
add more context to homepage concerns that our homepage doesn t give new users enough context about who we are and why they should care has surfaced in two different ways recently from beccaklam s users testing we should consider users who don t know our issues mozilla well and think about how to onboard them on our homepage if we ever want to grow our audience we should be aware of two types of audience those who know our cause or are internet savvy and those who don t know us or are not internet savvy we don t want to talk down to people who feel that they already know their stuff from juan and lotta around fundraising going to the site it’s hard to say “what mozilla does” this should frame the rest of the content about projects news etc what is a short pithy way to describe what we do that is engaging and captivating this should be a big piece of the landing page everything we present should be about the “bigger plan” needs to make donors be a part of the work we do so that they don’t feel like they’re just giving us money to do work but they’re giving us money to do work that they care about and are contributing to landing page doesn’t talk about why they should participate only how to participate e g apply to be a fellow we need to be able to tell our broader story better on this page where do we talk about what our plan is for a healthier internet we should be selling that plan vision i e resonating with audiences with emotion a story rather than specific programs we run out of context answering “what is mozilla about” should be the first touchpoint for people juan just links people to the fo’mo website so the first touchpoint is the landing page news should tell the story of our impact projects need more context so that the broader picture is still described using it we should benchmark similar sites a few examples but should gather more write up a few recommendations some quick fixes some long term one of those recommendations should be to add an intro sentence who should write this mock up a few options get feedback from staff test with different audiences cc sabrinang natalieworth beccaklam xmatthewx anilkanji
0