Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
98,162
12,300,036,481
IssuesEvent
2020-05-11 13:23:35
Opentrons/opentrons
https://api.github.com/repos/Opentrons/opentrons
closed
[PD] Thermocycler Graph of Profile
WIP design protocol designer
# User Story As a user, I should be able to reference a standard graph/visualization of the profile I’ve created which shows the profile’s temperature, time spans, and cycles. # UI Design _In progress_ # Acceptance Criteria - [ ] The graph is made up of equal width/height blocks. These blocks contain: - [ ] A horizontal line, the height of which represents temperature - [ ] A temperature - [ ] A duration - [ ] Blocks that are together in a cycle have an arrow underneath with the cycle # in it. These blocks also have a different fill than individual blocks. - [ ] The horizontal lines in each block are connected to their neighbors by a line that spans the gutter between blocks
2.0
[PD] Thermocycler Graph of Profile - # User Story As a user, I should be able to reference a standard graph/visualization of the profile I’ve created which shows the profile’s temperature, time spans, and cycles. # UI Design _In progress_ # Acceptance Criteria - [ ] The graph is made up of equal width/height blocks. These blocks contain: - [ ] A horizontal line, the height of which represents temperature - [ ] A temperature - [ ] A duration - [ ] Blocks that are together in a cycle have an arrow underneath with the cycle # in it. These blocks also have a different fill than individual blocks. - [ ] The horizontal lines in each block are connected to their neighbors by a line that spans the gutter between blocks
non_test
thermocycler graph of profile user story as a user i should be able to reference a standard graph visualization of the profile i’ve created which shows the profile’s temperature time spans and cycles ui design in progress acceptance criteria the graph is made up of equal width height blocks these blocks contain a horizontal line the height of which represents temperature a temperature a duration blocks that are together in a cycle have an arrow underneath with the cycle in it these blocks also have a different fill than individual blocks the horizontal lines in each block are connected to their neighbors by a line that spans the gutter between blocks
0
346,012
10,383,016,236
IssuesEvent
2019-09-10 08:46:54
red-hat-storage/ocs-ci
https://api.github.com/repos/red-hat-storage/ocs-ci
opened
ocs_operator_storage_cluster_cr points to nonexisting url
High Priority bug
Installation of upstream version via operator is failing on following error: ``` E AssertionError: Couldn't load URL: https://raw.githubusercontent.com/openshift/ocs-operator/master/deploy/crds/ocs_v1alpha1_storagecluster_cr.yaml content! Status: 404. ``` The URL is set in default config [ocs_ci/framework/conf/default_config.yaml](https://github.com/red-hat-storage/ocs-ci/blob/master/ocs_ci/framework/conf/default_config.yaml#L44). The original file does not exists any more, but it seems to be just renamed to `ocs_v1_storagecluster_cr.yaml`: https://github.com/openshift/ocs-operator/tree/master/deploy/crds
1.0
ocs_operator_storage_cluster_cr points to nonexisting url - Installation of upstream version via operator is failing on following error: ``` E AssertionError: Couldn't load URL: https://raw.githubusercontent.com/openshift/ocs-operator/master/deploy/crds/ocs_v1alpha1_storagecluster_cr.yaml content! Status: 404. ``` The URL is set in default config [ocs_ci/framework/conf/default_config.yaml](https://github.com/red-hat-storage/ocs-ci/blob/master/ocs_ci/framework/conf/default_config.yaml#L44). The original file does not exists any more, but it seems to be just renamed to `ocs_v1_storagecluster_cr.yaml`: https://github.com/openshift/ocs-operator/tree/master/deploy/crds
non_test
ocs operator storage cluster cr points to nonexisting url installation of upstream version via operator is failing on following error e assertionerror couldn t load url content status the url is set in default config the original file does not exists any more but it seems to be just renamed to ocs storagecluster cr yaml
0
30,983
6,385,767,674
IssuesEvent
2017-08-03 09:23:27
bridgedotnet/Bridge
https://api.github.com/repos/bridgedotnet/Bridge
closed
Cannot cast null to nullable
defect
Cannot cast null to nullable. ### Steps To Reproduce https://dev.deck.net/16c8630e00f0554844f810648e73f9e9/third https://dotnetfiddle.net/FeYHFH ```c# public class Program { public static void Main() { Console.WriteLine((((object)null) as Int64?).HasValue ? "Failed" : "Passed"); } } ```
1.0
Cannot cast null to nullable - Cannot cast null to nullable. ### Steps To Reproduce https://dev.deck.net/16c8630e00f0554844f810648e73f9e9/third https://dotnetfiddle.net/FeYHFH ```c# public class Program { public static void Main() { Console.WriteLine((((object)null) as Int64?).HasValue ? "Failed" : "Passed"); } } ```
non_test
cannot cast null to nullable cannot cast null to nullable steps to reproduce c public class program public static void main console writeline object null as hasvalue failed passed
0
821,359
30,819,127,865
IssuesEvent
2023-08-01 15:11:40
opendatahub-io/data-science-pipelines-operator
https://api.github.com/repos/opendatahub-io/data-science-pipelines-operator
opened
Use openshift-goimports to sort go importants
triage/accepted priority/normal
### Feature description We should follow openshift best practices when organizing our go imports, we should utilize [openshift go imports for this](https://github.com/openshift-eng/openshift-goimports). Acceptance criteria: * Figure out how best to include this as part of development workflow, (can this be part of pre-commit?) * add a gh action to verify that PRs adhere to the imports via openshift go imports. * Also add a documentation to dspo readme on how to develop/run go imports when developing. ### Describe alternatives you've considered _No response_ ### Anything else? _No response_
1.0
Use openshift-goimports to sort go importants - ### Feature description We should follow openshift best practices when organizing our go imports, we should utilize [openshift go imports for this](https://github.com/openshift-eng/openshift-goimports). Acceptance criteria: * Figure out how best to include this as part of development workflow, (can this be part of pre-commit?) * add a gh action to verify that PRs adhere to the imports via openshift go imports. * Also add a documentation to dspo readme on how to develop/run go imports when developing. ### Describe alternatives you've considered _No response_ ### Anything else? _No response_
non_test
use openshift goimports to sort go importants feature description we should follow openshift best practices when organizing our go imports we should utilize acceptance criteria figure out how best to include this as part of development workflow can this be part of pre commit add a gh action to verify that prs adhere to the imports via openshift go imports also add a documentation to dspo readme on how to develop run go imports when developing describe alternatives you ve considered no response anything else no response
0
304,486
9,332,750,125
IssuesEvent
2019-03-28 12:59:57
mlibrary/heliotrope
https://api.github.com/repos/mlibrary/heliotrope
closed
Vanilla Fulcrum
EPIC low priority refactor systems
Separate Fulcrum from Heliotrope such that following the setup instructions in the README.md creates a generic 'Vanilla Fulcrum' application.
1.0
Vanilla Fulcrum - Separate Fulcrum from Heliotrope such that following the setup instructions in the README.md creates a generic 'Vanilla Fulcrum' application.
non_test
vanilla fulcrum separate fulcrum from heliotrope such that following the setup instructions in the readme md creates a generic vanilla fulcrum application
0
68,508
3,288,906,608
IssuesEvent
2015-10-29 16:51:42
INN/Largo
https://api.github.com/repos/INN/Largo
closed
Pulling .rst function documentation out of .php
priority: low type: question
#### What's currently working. Using __[`doxphp`](https://github.com/avalanche123/doxphp)__ (a php phar), it's possible to generate a `.json` representation of documentation directly from our `.php` files. From there, we can generate `*.rst` using the __`doxphp2sphinx`__ renderer supplied in the same phar. $ doxphp < functions.php | doxphp2sphinx > functions.rst As a proof of concept, this is currently in our sphinx Makefile for a few select files [here](https://github.com/INN/Largo/blob/develop/docs/Makefile#L179-L185). With `make php`, It will generate an `.rst` files for each specified `.php` file. With `make html`, those files will compile to `.html` [like this](http://largo.readthedocs.org/api/inc/helpers.html). It requires you to have **`doxphp`** installed. **tl;dr:** `.php > .json > .rst > .html` We could modify this pipeline, writing either our own `.php > .json` parser (seems unnecessary) or sphinx extension to render the included `.json > .rst` in a different way. In theory, this shouldn't be too hard. #### Questions. 1. Is organizing function documentation like this the best practice? (i.e is the way we organize our code is the best way to organize our function documentation?) 2. Not all files have functions that need documentation. I'd assume most of what should be included resides in the `./inc/` folder, but is there anything else that should be included? Is there anything in `./inc/` that shouldn't be included? #### Others. What others are doing * [WordPress Codex](http://codex.wordpress.org/): Organizes function reference one function per page. I tend to like this format and think our users would be familiar with it. It also would allow us to add examples and longer form documentation for those functions that need it. * [WooThemes](http://docs.woothemes.com/): Seems to have their documentation all over the place. - Their **WooCommerce** plugin uses [APIgen](http://www.apigen.org/) to generate documentation for their WooCommerce plugin. and keeps it separate from - **WooCodex** seems to mirror the structure of WordPress for some of their more [commonly used](http://docs.woothemes.com/documentation/woocodex/) functions ([like this](http://docs.woothemes.com/document/woocommerce_breadcrumb/)). * Others?
1.0
Pulling .rst function documentation out of .php - #### What's currently working. Using __[`doxphp`](https://github.com/avalanche123/doxphp)__ (a php phar), it's possible to generate a `.json` representation of documentation directly from our `.php` files. From there, we can generate `*.rst` using the __`doxphp2sphinx`__ renderer supplied in the same phar. $ doxphp < functions.php | doxphp2sphinx > functions.rst As a proof of concept, this is currently in our sphinx Makefile for a few select files [here](https://github.com/INN/Largo/blob/develop/docs/Makefile#L179-L185). With `make php`, It will generate an `.rst` files for each specified `.php` file. With `make html`, those files will compile to `.html` [like this](http://largo.readthedocs.org/api/inc/helpers.html). It requires you to have **`doxphp`** installed. **tl;dr:** `.php > .json > .rst > .html` We could modify this pipeline, writing either our own `.php > .json` parser (seems unnecessary) or sphinx extension to render the included `.json > .rst` in a different way. In theory, this shouldn't be too hard. #### Questions. 1. Is organizing function documentation like this the best practice? (i.e is the way we organize our code is the best way to organize our function documentation?) 2. Not all files have functions that need documentation. I'd assume most of what should be included resides in the `./inc/` folder, but is there anything else that should be included? Is there anything in `./inc/` that shouldn't be included? #### Others. What others are doing * [WordPress Codex](http://codex.wordpress.org/): Organizes function reference one function per page. I tend to like this format and think our users would be familiar with it. It also would allow us to add examples and longer form documentation for those functions that need it. * [WooThemes](http://docs.woothemes.com/): Seems to have their documentation all over the place. - Their **WooCommerce** plugin uses [APIgen](http://www.apigen.org/) to generate documentation for their WooCommerce plugin. and keeps it separate from - **WooCodex** seems to mirror the structure of WordPress for some of their more [commonly used](http://docs.woothemes.com/documentation/woocodex/) functions ([like this](http://docs.woothemes.com/document/woocommerce_breadcrumb/)). * Others?
non_test
pulling rst function documentation out of php what s currently working using a php phar it s possible to generate a json representation of documentation directly from our php files from there we can generate rst using the renderer supplied in the same phar doxphp functions rst as a proof of concept this is currently in our sphinx makefile for a few select files with make php it will generate an rst files for each specified php file with make html those files will compile to html it requires you to have doxphp installed tl dr php json rst html we could modify this pipeline writing either our own php json parser seems unnecessary or sphinx extension to render the included json rst in a different way in theory this shouldn t be too hard questions is organizing function documentation like this the best practice i e is the way we organize our code is the best way to organize our function documentation not all files have functions that need documentation i d assume most of what should be included resides in the inc folder but is there anything else that should be included is there anything in inc that shouldn t be included others what others are doing organizes function reference one function per page i tend to like this format and think our users would be familiar with it it also would allow us to add examples and longer form documentation for those functions that need it seems to have their documentation all over the place their woocommerce plugin uses to generate documentation for their woocommerce plugin and keeps it separate from woocodex seems to mirror the structure of wordpress for some of their more functions others
0
170,442
6,444,659,752
IssuesEvent
2017-08-12 15:16:17
drckf/paysage
https://api.github.com/repos/drckf/paysage
closed
Dropout RBMs
Priority: Medium
Dropout RBMs are discussed in the original dropout paper with results superior to normal RBMs.
1.0
Dropout RBMs - Dropout RBMs are discussed in the original dropout paper with results superior to normal RBMs.
non_test
dropout rbms dropout rbms are discussed in the original dropout paper with results superior to normal rbms
0
329,055
28,146,631,717
IssuesEvent
2023-04-02 15:03:16
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
closed
Fix elementwise.test_count_nonzero
Sub Task Ivy API Experimental Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4588728136/jobs/8103151468" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4588728136/jobs/8103151468" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4588728136/jobs/8103151468" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4588728136/jobs/8103151468" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix elementwise.test_count_nonzero - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4588728136/jobs/8103151468" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4588728136/jobs/8103151468" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4588728136/jobs/8103151468" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4588728136/jobs/8103151468" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
test
fix elementwise test count nonzero tensorflow img src torch img src numpy img src jax img src
1
613,759
19,097,793,857
IssuesEvent
2021-11-29 18:36:20
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
Able to push a package with a local repo dependancy to central
Type/Bug Priority/High Status/Blocked Team/DevTools SwanLakeDump Area/ProjectAPI
**Description:** Even though a similar package name does or does not exist in central, should not be able to push a package with a local dependency to the central. Currently, it's allowed to push for both scenarios. **Steps to reproduce:** **Affected Versions:** **OS, DB, other environment details and versions:** **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> **Suggested Labels (optional):** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees (optional):** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
1.0
Able to push a package with a local repo dependancy to central - **Description:** Even though a similar package name does or does not exist in central, should not be able to push a package with a local dependency to the central. Currently, it's allowed to push for both scenarios. **Steps to reproduce:** **Affected Versions:** **OS, DB, other environment details and versions:** **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> **Suggested Labels (optional):** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees (optional):** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
non_test
able to push a package with a local repo dependancy to central description even though a similar package name does or does not exist in central should not be able to push a package with a local dependency to the central currently it s allowed to push for both scenarios steps to reproduce affected versions os db other environment details and versions related issues optional suggested labels optional suggested assignees optional
0
516,420
14,981,892,391
IssuesEvent
2021-01-28 15:21:49
jetstack/cert-manager
https://api.github.com/repos/jetstack/cert-manager
closed
backup instructions in the docs are incorrect
kind/bug kind/feature priority/important-longterm
**Describe the bug**: This page has instruction for backup up the certs and restoring them: https://docs.cert-manager.io/en/latest/tasks/backup-restore-crds.html The instruction don't work as is. The backup part work, but the restore won't work for many reasons. ``` status.conditions.lastTransitionTime in body must be of type string: "null" ``` a quick hack here is to replace `lastTransitionTime: null` with some valid date, but filtering the status part would be better. That fixed I get ``` Error from server (Conflict): Operation cannot be fulfilled on issuers.certmanager.k8s.io "letsencrypt-prod": the object has been modified; please apply your changes to the latest version and try again Error from server (Conflict): Operation cannot be fulfilled on certificates.certmanager.k8s.io "cloud-master": the object has been modified; please apply your changes to the latest version and try again ``` **Expected behaviour**: Restore should be tested and working **Environment details:**: I was upgrading from 0.5.2 to 0.6.7 (in order to upgrade further). /kind bug
1.0
backup instructions in the docs are incorrect - **Describe the bug**: This page has instruction for backup up the certs and restoring them: https://docs.cert-manager.io/en/latest/tasks/backup-restore-crds.html The instruction don't work as is. The backup part work, but the restore won't work for many reasons. ``` status.conditions.lastTransitionTime in body must be of type string: "null" ``` a quick hack here is to replace `lastTransitionTime: null` with some valid date, but filtering the status part would be better. That fixed I get ``` Error from server (Conflict): Operation cannot be fulfilled on issuers.certmanager.k8s.io "letsencrypt-prod": the object has been modified; please apply your changes to the latest version and try again Error from server (Conflict): Operation cannot be fulfilled on certificates.certmanager.k8s.io "cloud-master": the object has been modified; please apply your changes to the latest version and try again ``` **Expected behaviour**: Restore should be tested and working **Environment details:**: I was upgrading from 0.5.2 to 0.6.7 (in order to upgrade further). /kind bug
non_test
backup instructions in the docs are incorrect describe the bug this page has instruction for backup up the certs and restoring them the instruction don t work as is the backup part work but the restore won t work for many reasons status conditions lasttransitiontime in body must be of type string null a quick hack here is to replace lasttransitiontime null with some valid date but filtering the status part would be better that fixed i get error from server conflict operation cannot be fulfilled on issuers certmanager io letsencrypt prod the object has been modified please apply your changes to the latest version and try again error from server conflict operation cannot be fulfilled on certificates certmanager io cloud master the object has been modified please apply your changes to the latest version and try again expected behaviour restore should be tested and working environment details i was upgrading from to in order to upgrade further kind bug
0
199,125
15,024,780,032
IssuesEvent
2021-02-01 20:07:48
CARTAvis/carta-backend
https://api.github.com/repos/CARTAvis/carta-backend
closed
extra channel info derived from header with ra-dec-stokes
awaiting testing bug
If the image has 3 axes as ra-dec-stokes, there is an extra channel info in the file info tab. ![6ABDF640-88D2-4589-AA6B-628EF059DD05](https://user-images.githubusercontent.com/20819712/106127763-c94ba300-6199-11eb-89e3-540e74fea672.jpeg) ![86F8DD20-F515-4546-AEF9-0625FA491BD9](https://user-images.githubusercontent.com/20819712/106127779-cbadfd00-6199-11eb-97de-3744ba9ec4eb.jpeg) image from @zarda
1.0
extra channel info derived from header with ra-dec-stokes - If the image has 3 axes as ra-dec-stokes, there is an extra channel info in the file info tab. ![6ABDF640-88D2-4589-AA6B-628EF059DD05](https://user-images.githubusercontent.com/20819712/106127763-c94ba300-6199-11eb-89e3-540e74fea672.jpeg) ![86F8DD20-F515-4546-AEF9-0625FA491BD9](https://user-images.githubusercontent.com/20819712/106127779-cbadfd00-6199-11eb-97de-3744ba9ec4eb.jpeg) image from @zarda
test
extra channel info derived from header with ra dec stokes if the image has axes as ra dec stokes there is an extra channel info in the file info tab image from zarda
1
147,582
13,210,679,832
IssuesEvent
2020-08-15 18:14:54
Gizra/og
https://api.github.com/repos/Gizra/og
closed
og_ungroup() documentation incorrectly informs about @return value.
Documentation Drupal 7
The current documentation of the og_ungroup() function states that an entity is returned but it's not currently the case: ``` /** * Delete an association (e.g. unsubscribe) of an entity to a group. * * @param $group_type * The entity type (e.g. "node"). * @param $gid * The group entity object or ID, to ungroup. * @param $entity_type * (optional) The entity type (e.g. "node" or "user"). * @param $etid * (optional) The entity object or ID, to ungroup. * * @return * The entity with the fields updated. */ function og_ungroup($group_type, $gid, $entity_type = 'user', $etid = NULL) { if (is_object($gid)) { list($gid) = entity_extract_ids($group_type, $gid); } if ($entity_type == 'user' && empty($etid)) { global $user; $etid = $user->uid; } elseif (is_object($etid)) { list($etid) = entity_extract_ids($entity_type, $etid); } if ($og_membership = og_get_membership($group_type, $gid, $entity_type, $etid)) { $og_membership->delete(); } } ``` Not providing a patch / PR because I don't know if what needs to be corrected is the documentation to follow the code, or to change the code to return something (perhaps more usefull...)
1.0
og_ungroup() documentation incorrectly informs about @return value. - The current documentation of the og_ungroup() function states that an entity is returned but it's not currently the case: ``` /** * Delete an association (e.g. unsubscribe) of an entity to a group. * * @param $group_type * The entity type (e.g. "node"). * @param $gid * The group entity object or ID, to ungroup. * @param $entity_type * (optional) The entity type (e.g. "node" or "user"). * @param $etid * (optional) The entity object or ID, to ungroup. * * @return * The entity with the fields updated. */ function og_ungroup($group_type, $gid, $entity_type = 'user', $etid = NULL) { if (is_object($gid)) { list($gid) = entity_extract_ids($group_type, $gid); } if ($entity_type == 'user' && empty($etid)) { global $user; $etid = $user->uid; } elseif (is_object($etid)) { list($etid) = entity_extract_ids($entity_type, $etid); } if ($og_membership = og_get_membership($group_type, $gid, $entity_type, $etid)) { $og_membership->delete(); } } ``` Not providing a patch / PR because I don't know if what needs to be corrected is the documentation to follow the code, or to change the code to return something (perhaps more usefull...)
non_test
og ungroup documentation incorrectly informs about return value the current documentation of the og ungroup function states that an entity is returned but it s not currently the case delete an association e g unsubscribe of an entity to a group param group type the entity type e g node param gid the group entity object or id to ungroup param entity type optional the entity type e g node or user param etid optional the entity object or id to ungroup return the entity with the fields updated function og ungroup group type gid entity type user etid null if is object gid list gid entity extract ids group type gid if entity type user empty etid global user etid user uid elseif is object etid list etid entity extract ids entity type etid if og membership og get membership group type gid entity type etid og membership delete not providing a patch pr because i don t know if what needs to be corrected is the documentation to follow the code or to change the code to return something perhaps more usefull
0
351,773
32,025,684,178
IssuesEvent
2023-09-22 08:40:25
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
closed
Fix solving_equations_and_inverting_matrices.test_numpy_tensorinv
NumPy Frontend Sub Task Failing Test
| | | |---|---| |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6271730752"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6265043606"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/6265043606"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6265043606"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/6265043606"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix solving_equations_and_inverting_matrices.test_numpy_tensorinv - | | | |---|---| |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6271730752"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6265043606"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/6265043606"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6265043606"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/6265043606"><img src=https://img.shields.io/badge/-success-success></a>
test
fix solving equations and inverting matrices test numpy tensorinv paddle a href src numpy a href src jax a href src tensorflow a href src torch a href src
1
13,294
2,750,721,889
IssuesEvent
2015-04-24 01:41:59
micheldumontier/semanticscience
https://api.github.com/repos/micheldumontier/semanticscience
closed
"is source of" is both property and class
auto-migrated Priority-Medium Type-Defect
``` http://semanticscience.org/resource/SIO_000219 'is source of' in sio-bio.owl seems to be both a property and (probably incorrect) a class. 'drug regulatory authority' is a subclass of this entity. Viewed in TopBraid Composer Free Edition without inferencing. ``` Original issue reported on code.google.com by `matthias...@gmail.com` on 29 Jun 2012 at 9:22
1.0
"is source of" is both property and class - ``` http://semanticscience.org/resource/SIO_000219 'is source of' in sio-bio.owl seems to be both a property and (probably incorrect) a class. 'drug regulatory authority' is a subclass of this entity. Viewed in TopBraid Composer Free Edition without inferencing. ``` Original issue reported on code.google.com by `matthias...@gmail.com` on 29 Jun 2012 at 9:22
non_test
is source of is both property and class is source of in sio bio owl seems to be both a property and probably incorrect a class drug regulatory authority is a subclass of this entity viewed in topbraid composer free edition without inferencing original issue reported on code google com by matthias gmail com on jun at
0
268,009
23,339,202,079
IssuesEvent
2022-08-09 12:44:53
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Failing ES Promotion: Chrome X-Pack UI Functional Tests - ML anomaly_detection.x-pack/test/functional/apps/ml/anomaly_detection/anomaly_explorer·ts - machine learning - anomaly detection anomaly explorer with farequote based multi metric job renders View By swim lane
blocker :ml skipped-test failed-es-promotion Team:ML v7.17.6
**Chrome X-Pack UI Functional Tests - ML anomaly_detection** **x-pack/test/functional/apps/ml/anomaly_detection/anomaly_explorer.ts** **machine learning - anomaly detection anomaly explorer with farequote based multi metric job renders View By swim lane** This failure is preventing the promotion of the current Elasticsearch nightly snapshot. For more information on the Elasticsearch snapshot promotion process including how to reproduce using the unverified nightly ES build: https://www.elastic.co/guide/en/kibana/master/development-es-snapshots.html * [Failed promotion job](https://buildkite.com/elastic/kibana-elasticsearch-snapshot-verify/builds/1539#01827dcc-e0f1-4a59-a15c-acc245fdaa48) * [Test Failure](https://buildkite.com/organizations/elastic/pipelines/kibana-elasticsearch-snapshot-verify/builds/1539/jobs/01827dcc-e0f1-4a59-a15c-acc245fdaa48/artifacts/01827df9-0925-4dfb-8d8a-f996e6024ebf) ``` Error: Expected swim lane y labels to be AAL,VRD,EGF,SWR,AMX,JZA,TRS,ACA,BAW,ASA, got AAL,EGF,VRD,SWR,JZA,AMX,TRS,ACA,BAW,ASA at Assertion.assert (node_modules/@kbn/expect/expect.js:100:11) at Assertion.eql (node_modules/@kbn/expect/expect.js:244:8) at Object.assertAxisLabels (x-pack/test/functional/services/ml/swim_lane.ts:88:31) at runMicrotasks (<anonymous>) at processTicksAndRejections (node:internal/process/task_queues:96:5) at Context.<anonymous> (x-pack/test/functional/apps/ml/anomaly_detection/anomaly_explorer.ts:167:11) at Object.apply (node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) { actual: '[\n' + ' "AAL"\n' + ' "EGF"\n' + ' "VRD"\n' + ' "SWR"\n' + ' "JZA"\n' + ' "AMX"\n' + ' "TRS"\n' + ' "ACA"\n' + ' "BAW"\n' + ' "ASA"\n' + ']', expected: '[\n' + ' "AAL"\n' + ' "VRD"\n' + ' "EGF"\n' + ' "SWR"\n' + ' "AMX"\n' + ' "JZA"\n' + ' "TRS"\n' + ' "ACA"\n' + ' "BAW"\n' + ' "ASA"\n' + ']', showDiff: true } ```
1.0
Failing ES Promotion: Chrome X-Pack UI Functional Tests - ML anomaly_detection.x-pack/test/functional/apps/ml/anomaly_detection/anomaly_explorer·ts - machine learning - anomaly detection anomaly explorer with farequote based multi metric job renders View By swim lane - **Chrome X-Pack UI Functional Tests - ML anomaly_detection** **x-pack/test/functional/apps/ml/anomaly_detection/anomaly_explorer.ts** **machine learning - anomaly detection anomaly explorer with farequote based multi metric job renders View By swim lane** This failure is preventing the promotion of the current Elasticsearch nightly snapshot. For more information on the Elasticsearch snapshot promotion process including how to reproduce using the unverified nightly ES build: https://www.elastic.co/guide/en/kibana/master/development-es-snapshots.html * [Failed promotion job](https://buildkite.com/elastic/kibana-elasticsearch-snapshot-verify/builds/1539#01827dcc-e0f1-4a59-a15c-acc245fdaa48) * [Test Failure](https://buildkite.com/organizations/elastic/pipelines/kibana-elasticsearch-snapshot-verify/builds/1539/jobs/01827dcc-e0f1-4a59-a15c-acc245fdaa48/artifacts/01827df9-0925-4dfb-8d8a-f996e6024ebf) ``` Error: Expected swim lane y labels to be AAL,VRD,EGF,SWR,AMX,JZA,TRS,ACA,BAW,ASA, got AAL,EGF,VRD,SWR,JZA,AMX,TRS,ACA,BAW,ASA at Assertion.assert (node_modules/@kbn/expect/expect.js:100:11) at Assertion.eql (node_modules/@kbn/expect/expect.js:244:8) at Object.assertAxisLabels (x-pack/test/functional/services/ml/swim_lane.ts:88:31) at runMicrotasks (<anonymous>) at processTicksAndRejections (node:internal/process/task_queues:96:5) at Context.<anonymous> (x-pack/test/functional/apps/ml/anomaly_detection/anomaly_explorer.ts:167:11) at Object.apply (node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) { actual: '[\n' + ' "AAL"\n' + ' "EGF"\n' + ' "VRD"\n' + ' "SWR"\n' + ' "JZA"\n' + ' "AMX"\n' + ' "TRS"\n' + ' "ACA"\n' + ' "BAW"\n' + ' "ASA"\n' + ']', expected: '[\n' + ' "AAL"\n' + ' "VRD"\n' + ' "EGF"\n' + ' "SWR"\n' + ' "AMX"\n' + ' "JZA"\n' + ' "TRS"\n' + ' "ACA"\n' + ' "BAW"\n' + ' "ASA"\n' + ']', showDiff: true } ```
test
failing es promotion chrome x pack ui functional tests ml anomaly detection x pack test functional apps ml anomaly detection anomaly explorer·ts machine learning anomaly detection anomaly explorer with farequote based multi metric job renders view by swim lane chrome x pack ui functional tests ml anomaly detection x pack test functional apps ml anomaly detection anomaly explorer ts machine learning anomaly detection anomaly explorer with farequote based multi metric job renders view by swim lane this failure is preventing the promotion of the current elasticsearch nightly snapshot for more information on the elasticsearch snapshot promotion process including how to reproduce using the unverified nightly es build error expected swim lane y labels to be aal vrd egf swr amx jza trs aca baw asa got aal egf vrd swr jza amx trs aca baw asa at assertion assert node modules kbn expect expect js at assertion eql node modules kbn expect expect js at object assertaxislabels x pack test functional services ml swim lane ts at runmicrotasks at processticksandrejections node internal process task queues at context x pack test functional apps ml anomaly detection anomaly explorer ts at object apply node modules kbn test target node functional test runner lib mocha wrap function js actual n aal n egf n vrd n swr n jza n amx n trs n aca n baw n asa n expected n aal n vrd n egf n swr n amx n jza n trs n aca n baw n asa n showdiff true
1
121,884
10,197,017,424
IssuesEvent
2019-08-12 22:32:15
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
teamcity: failed test: gossip/restart
C-test-failure O-robot
The following tests appear to have failed on master (roachtest): acceptance/gossip/restart You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+gossip/restart). [#1436098](https://teamcity.cockroachdb.com/viewLog.html?buildId=1436098): ``` acceptance/gossip/restart --- FAIL: roachtest/acceptance/gossip/restart (26.519s) test artifacts and logs in: artifacts/acceptance/gossip/restart/run_1 gossip.go:226,gossip.go:286,acceptance.go:69,test_runner.go:691: dial tcp 127.0.0.1:26261: connect: connection refused test artifacts and logs in: artifacts/acceptance/gossip/restart/run_1 gossip.go:226,gossip.go:286,acceptance.go:69,test_runner.go:691: dial tcp 127.0.0.1:26261: connect: connection refused ``` Please assign, take a look and update the issue accordingly.
1.0
teamcity: failed test: gossip/restart - The following tests appear to have failed on master (roachtest): acceptance/gossip/restart You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+gossip/restart). [#1436098](https://teamcity.cockroachdb.com/viewLog.html?buildId=1436098): ``` acceptance/gossip/restart --- FAIL: roachtest/acceptance/gossip/restart (26.519s) test artifacts and logs in: artifacts/acceptance/gossip/restart/run_1 gossip.go:226,gossip.go:286,acceptance.go:69,test_runner.go:691: dial tcp 127.0.0.1:26261: connect: connection refused test artifacts and logs in: artifacts/acceptance/gossip/restart/run_1 gossip.go:226,gossip.go:286,acceptance.go:69,test_runner.go:691: dial tcp 127.0.0.1:26261: connect: connection refused ``` Please assign, take a look and update the issue accordingly.
test
teamcity failed test gossip restart the following tests appear to have failed on master roachtest acceptance gossip restart you may want to check acceptance gossip restart fail roachtest acceptance gossip restart test artifacts and logs in artifacts acceptance gossip restart run gossip go gossip go acceptance go test runner go dial tcp connect connection refused test artifacts and logs in artifacts acceptance gossip restart run gossip go gossip go acceptance go test runner go dial tcp connect connection refused please assign take a look and update the issue accordingly
1
90,532
15,856,201,240
IssuesEvent
2021-04-08 01:46:26
AnhaaD/hacknightvol4
https://api.github.com/repos/AnhaaD/hacknightvol4
opened
CVE-2019-20149 (High) detected in kind-of-6.0.2.tgz
security vulnerability
## CVE-2019-20149 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>kind-of-6.0.2.tgz</b></p></summary> <p>Get the native type of a value.</p> <p>Library home page: <a href="https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz">https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz</a></p> <p>Path to dependency file: /hacknightvol4/package.json</p> <p>Path to vulnerable library: hacknightvol4/node_modules/snapdragon-node/node_modules/kind-of/package.json</p> <p> Dependency Hierarchy: - nodemon-1.17.5.tgz (Root Library) - chokidar-2.1.6.tgz - anymatch-2.0.0.tgz - micromatch-3.1.10.tgz - :x: **kind-of-6.0.2.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ctorName in index.js in kind-of v6.0.2 allows external user input to overwrite certain internal attributes via a conflicting name, as demonstrated by 'constructor': {'name':'Symbol'}. Hence, a crafted payload can overwrite this builtin attribute to manipulate the type detection result. <p>Publish Date: 2019-12-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20149>CVE-2019-20149</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-20149">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-20149</a></p> <p>Release Date: 2019-12-30</p> <p>Fix Resolution: 6.0.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-20149 (High) detected in kind-of-6.0.2.tgz - ## CVE-2019-20149 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>kind-of-6.0.2.tgz</b></p></summary> <p>Get the native type of a value.</p> <p>Library home page: <a href="https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz">https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz</a></p> <p>Path to dependency file: /hacknightvol4/package.json</p> <p>Path to vulnerable library: hacknightvol4/node_modules/snapdragon-node/node_modules/kind-of/package.json</p> <p> Dependency Hierarchy: - nodemon-1.17.5.tgz (Root Library) - chokidar-2.1.6.tgz - anymatch-2.0.0.tgz - micromatch-3.1.10.tgz - :x: **kind-of-6.0.2.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ctorName in index.js in kind-of v6.0.2 allows external user input to overwrite certain internal attributes via a conflicting name, as demonstrated by 'constructor': {'name':'Symbol'}. Hence, a crafted payload can overwrite this builtin attribute to manipulate the type detection result. <p>Publish Date: 2019-12-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20149>CVE-2019-20149</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-20149">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-20149</a></p> <p>Release Date: 2019-12-30</p> <p>Fix Resolution: 6.0.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in kind of tgz cve high severity vulnerability vulnerable library kind of tgz get the native type of a value library home page a href path to dependency file package json path to vulnerable library node modules snapdragon node node modules kind of package json dependency hierarchy nodemon tgz root library chokidar tgz anymatch tgz micromatch tgz x kind of tgz vulnerable library vulnerability details ctorname in index js in kind of allows external user input to overwrite certain internal attributes via a conflicting name as demonstrated by constructor name symbol hence a crafted payload can overwrite this builtin attribute to manipulate the type detection result publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
264,340
8,307,891,840
IssuesEvent
2018-09-23 14:51:36
publiclab/plots2
https://api.github.com/repos/publiclab/plots2
opened
Digest email bug
bug priority
The digest emails are currently not being sent. `ArgumentError: wrong number of arguments (given 0, expected 1)` is shown on Sidekiq dashboard (https://publiclab.org/sidekiq/retries).
1.0
Digest email bug - The digest emails are currently not being sent. `ArgumentError: wrong number of arguments (given 0, expected 1)` is shown on Sidekiq dashboard (https://publiclab.org/sidekiq/retries).
non_test
digest email bug the digest emails are currently not being sent argumenterror wrong number of arguments given expected is shown on sidekiq dashboard
0
412,971
12,058,928,646
IssuesEvent
2020-04-15 18:20:39
zulip/zulip
https://api.github.com/repos/zulip/zulip
closed
Optimize rate_limiter performance for get_events queries
area: production in progress priority: high
See https://chat.zulip.org/#narrow/stream/3-backend/topic/profiling.20get_events/near/816860 for profiling details, but basically, currently a get_events request spends 1.4ms/request talking to redis for our rate limiter, which is somewhere between 15% and 50% of the total request runtime (my measurement technique is susceptible to issues like the first request on a code path being extra expensive). Since get_events is our most scalability-critical endpoint, this is a big deal. We should do some rethinking of the redis internals for our rate limiter. I have a few ideas: * Writing an alternative rate-limiter implementation for `get_events `specifically that's entirely in-process and would be basically instant. Since the Tornado system has a relatively strong constraint that a given user always connect to the same server, this might be fairly cheap to implement and would bring that 1.4ms to probably 50us or less. (And gate it on `RUNNING_INSIDE_TORNADO`). * Look at rewriting our redis transactions to be more efficient for the highest-traffic cases (E.g. user is not close to limit, or user is way over limit). E.g. maybe `incr_rateimit` should automatically return the `api_calls_left` result rather than requiring 2 transactions. * Looking at https://github.com/popravich/python-redis-benchmark, there may be some alternative async IO redis clients we could consider migrating to, and possibly some that are just faster. Given how little code we have interacting with redis directly, this might be an easy port to do; I'm not sure whether or not it would help. (And unlike the in-process hack approach, this would have side benefits to non-Tornado endpoints).
1.0
Optimize rate_limiter performance for get_events queries - See https://chat.zulip.org/#narrow/stream/3-backend/topic/profiling.20get_events/near/816860 for profiling details, but basically, currently a get_events request spends 1.4ms/request talking to redis for our rate limiter, which is somewhere between 15% and 50% of the total request runtime (my measurement technique is susceptible to issues like the first request on a code path being extra expensive). Since get_events is our most scalability-critical endpoint, this is a big deal. We should do some rethinking of the redis internals for our rate limiter. I have a few ideas: * Writing an alternative rate-limiter implementation for `get_events `specifically that's entirely in-process and would be basically instant. Since the Tornado system has a relatively strong constraint that a given user always connect to the same server, this might be fairly cheap to implement and would bring that 1.4ms to probably 50us or less. (And gate it on `RUNNING_INSIDE_TORNADO`). * Look at rewriting our redis transactions to be more efficient for the highest-traffic cases (E.g. user is not close to limit, or user is way over limit). E.g. maybe `incr_rateimit` should automatically return the `api_calls_left` result rather than requiring 2 transactions. * Looking at https://github.com/popravich/python-redis-benchmark, there may be some alternative async IO redis clients we could consider migrating to, and possibly some that are just faster. Given how little code we have interacting with redis directly, this might be an easy port to do; I'm not sure whether or not it would help. (And unlike the in-process hack approach, this would have side benefits to non-Tornado endpoints).
non_test
optimize rate limiter performance for get events queries see for profiling details but basically currently a get events request spends request talking to redis for our rate limiter which is somewhere between and of the total request runtime my measurement technique is susceptible to issues like the first request on a code path being extra expensive since get events is our most scalability critical endpoint this is a big deal we should do some rethinking of the redis internals for our rate limiter i have a few ideas writing an alternative rate limiter implementation for get events specifically that s entirely in process and would be basically instant since the tornado system has a relatively strong constraint that a given user always connect to the same server this might be fairly cheap to implement and would bring that to probably or less and gate it on running inside tornado look at rewriting our redis transactions to be more efficient for the highest traffic cases e g user is not close to limit or user is way over limit e g maybe incr rateimit should automatically return the api calls left result rather than requiring transactions looking at there may be some alternative async io redis clients we could consider migrating to and possibly some that are just faster given how little code we have interacting with redis directly this might be an easy port to do i m not sure whether or not it would help and unlike the in process hack approach this would have side benefits to non tornado endpoints
0
4,195
4,968,932,741
IssuesEvent
2016-12-05 11:33:47
core-wg/oscoap
https://api.github.com/repos/core-wg/oscoap
closed
Replay window is an input parameter?
core-object-security-00
Section 3.2 - Why is the replay window an input - this seems to be odd as you would not pre-fill part of the window. I assume this is really just the Replay Window Size.
True
Replay window is an input parameter? - Section 3.2 - Why is the replay window an input - this seems to be odd as you would not pre-fill part of the window. I assume this is really just the Replay Window Size.
non_test
replay window is an input parameter section why is the replay window an input this seems to be odd as you would not pre fill part of the window i assume this is really just the replay window size
0
138,488
11,202,492,452
IssuesEvent
2020-01-04 12:59:02
searchkit/searchkit
https://api.github.com/repos/searchkit/searchkit
closed
rangeFormatter function not used in filter
2.3.0-9 Ready For Testing stale
<img width="286" alt="screenshot 2017-06-13 15 59 49" src="https://user-images.githubusercontent.com/7115982/27072441-f190638e-5052-11e7-9b15-1a5bb8c7018d.png"> <img width="433" alt="screenshot 2017-06-13 15 59 53" src="https://user-images.githubusercontent.com/7115982/27072443-f29c5710-5052-11e7-837b-43106a0e66c8.png"> rangeFormatter function used in rangeFilter but not used in top filter
1.0
rangeFormatter function not used in filter - <img width="286" alt="screenshot 2017-06-13 15 59 49" src="https://user-images.githubusercontent.com/7115982/27072441-f190638e-5052-11e7-9b15-1a5bb8c7018d.png"> <img width="433" alt="screenshot 2017-06-13 15 59 53" src="https://user-images.githubusercontent.com/7115982/27072443-f29c5710-5052-11e7-837b-43106a0e66c8.png"> rangeFormatter function used in rangeFilter but not used in top filter
test
rangeformatter function not used in filter img width alt screenshot src img width alt screenshot src rangeformatter function used in rangefilter but not used in top filter
1
263,587
28,047,514,574
IssuesEvent
2023-03-29 01:03:58
tabacws-sandbox/juice-shop-checkPR
https://api.github.com/repos/tabacws-sandbox/juice-shop-checkPR
closed
check-dependencies-1.1.0.tgz: 1 vulnerabilities (highest severity is: 9.8) - autoclosed
Mend: dependency security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>check-dependencies-1.1.0.tgz</b></p></summary> <p></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/set-value/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/tabacws-sandbox/juice-shop-checkPR/commit/898e55dce59f24513206f629f1dd595ca468b56f">898e55dce59f24513206f629f1dd595ca468b56f</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (check-dependencies version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2021-23440](https://www.mend.io/vulnerability-database/CVE-2021-23440) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> Critical | 9.8 | set-value-2.0.1.tgz | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the "Details" section below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> CVE-2021-23440</summary> ### Vulnerable Library - <b>set-value-2.0.1.tgz</b></p> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-2.0.1.tgz">https://registry.npmjs.org/set-value/-/set-value-2.0.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/set-value/package.json</p> <p> Dependency Hierarchy: - check-dependencies-1.1.0.tgz (Root Library) - findup-sync-2.0.0.tgz - micromatch-3.1.10.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - :x: **set-value-2.0.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/tabacws-sandbox/juice-shop-checkPR/commit/898e55dce59f24513206f629f1dd595ca468b56f">898e55dce59f24513206f629f1dd595ca468b56f</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> This affects the package set-value before <2.0.1, >=3.0.0 <4.0.1. A type confusion vulnerability can lead to a bypass of CVE-2019-10747 when the user-provided keys used in the path parameter are arrays. Mend Note: After conducting further research, Mend has determined that all versions of set-value up to version 4.0.0 are vulnerable to CVE-2021-23440. <p>Publish Date: 2021-09-12 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23440>CVE-2021-23440</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2021-09-12</p> <p>Fix Resolution: set-value - 4.0.1 </p> </p> <p></p> </details>
True
check-dependencies-1.1.0.tgz: 1 vulnerabilities (highest severity is: 9.8) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>check-dependencies-1.1.0.tgz</b></p></summary> <p></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/set-value/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/tabacws-sandbox/juice-shop-checkPR/commit/898e55dce59f24513206f629f1dd595ca468b56f">898e55dce59f24513206f629f1dd595ca468b56f</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (check-dependencies version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2021-23440](https://www.mend.io/vulnerability-database/CVE-2021-23440) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> Critical | 9.8 | set-value-2.0.1.tgz | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the "Details" section below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> CVE-2021-23440</summary> ### Vulnerable Library - <b>set-value-2.0.1.tgz</b></p> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-2.0.1.tgz">https://registry.npmjs.org/set-value/-/set-value-2.0.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/set-value/package.json</p> <p> Dependency Hierarchy: - check-dependencies-1.1.0.tgz (Root Library) - findup-sync-2.0.0.tgz - micromatch-3.1.10.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - :x: **set-value-2.0.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/tabacws-sandbox/juice-shop-checkPR/commit/898e55dce59f24513206f629f1dd595ca468b56f">898e55dce59f24513206f629f1dd595ca468b56f</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> This affects the package set-value before <2.0.1, >=3.0.0 <4.0.1. A type confusion vulnerability can lead to a bypass of CVE-2019-10747 when the user-provided keys used in the path parameter are arrays. Mend Note: After conducting further research, Mend has determined that all versions of set-value up to version 4.0.0 are vulnerable to CVE-2021-23440. <p>Publish Date: 2021-09-12 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23440>CVE-2021-23440</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2021-09-12</p> <p>Fix Resolution: set-value - 4.0.1 </p> </p> <p></p> </details>
non_test
check dependencies tgz vulnerabilities highest severity is autoclosed vulnerable library check dependencies tgz path to dependency file package json path to vulnerable library node modules set value package json found in head commit a href vulnerabilities cve severity cvss dependency type fixed in check dependencies version remediation available critical set value tgz transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the details section below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library set value tgz create nested values and any intermediaries using dot notation a b c paths library home page a href path to dependency file package json path to vulnerable library node modules set value package json dependency hierarchy check dependencies tgz root library findup sync tgz micromatch tgz snapdragon tgz base tgz cache base tgz x set value tgz vulnerable library found in head commit a href found in base branch master vulnerability details this affects the package set value before a type confusion vulnerability can lead to a bypass of cve when the user provided keys used in the path parameter are arrays mend note after conducting further research mend has determined that all versions of set value up to version are vulnerable to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution set value
0
38,842
8,554,684,441
IssuesEvent
2018-11-08 07:32:00
openshiftio/openshift.io
https://api.github.com/repos/openshiftio/openshift.io
closed
Support Quick fixes with codeAction in case of CVEs flagged from LSP
area/analytics env/vs-code status/in-progress team/analytics type/user-story
### Planner Link : https://openshift.io/openshiftio/Openshift_io/plan/detail/948 Currently in VSCode Extension, we flag CVEs for any core dependencies, proposed would be to enable quick fixes to those with enabling codeAction from Language Server. ### Tasks: - [x] Update language server and client version - [x] Bind codeAction supportin Language Server - [x] Register command in client to listen and apply edits
1.0
Support Quick fixes with codeAction in case of CVEs flagged from LSP - ### Planner Link : https://openshift.io/openshiftio/Openshift_io/plan/detail/948 Currently in VSCode Extension, we flag CVEs for any core dependencies, proposed would be to enable quick fixes to those with enabling codeAction from Language Server. ### Tasks: - [x] Update language server and client version - [x] Bind codeAction supportin Language Server - [x] Register command in client to listen and apply edits
non_test
support quick fixes with codeaction in case of cves flagged from lsp planner link currently in vscode extension we flag cves for any core dependencies proposed would be to enable quick fixes to those with enabling codeaction from language server tasks update language server and client version bind codeaction supportin language server register command in client to listen and apply edits
0
245,192
20,751,961,035
IssuesEvent
2022-03-15 08:34:34
openvinotoolkit/openvino
https://api.github.com/repos/openvinotoolkit/openvino
closed
[Bug] (minor) test namespace ?
bug category: IE Tests support_request PSE
##### System information (version) <!-- Example - OpenVINO => 2020.4 - Operating System / Platform => Windows 64 Bit - Compiler => Visual Studio 2017 - Problem classification: Model Conversion - Framework: TensorFlow (if applicable) - Model name: ResNet50 (if applicable) --> - OpenVINO=> :grey_question: - Operating System / Platform => :grey_question: - Compiler => :grey_question: - Problem classification => :grey_question: ##### Detailed description <!-- your description --> I suppose this namespace was meant to be "LayerTest**s**Definitions", instead of "LayerTestDefinitions", to be aligned with the rest of the filters. I was getting an unexpected compile error and I barely noticed the missing 's': https://github.com/openvinotoolkit/openvino/blob/e8d5cf43d0e153f4f52c7be133f71f65c9eb4512/src/tests/functional/shared_test_classes/include/shared_test_classes/single_layer/prior_box.hpp#L29 ##### Steps to reproduce <!-- Describe your problem and steps you've done before you got to this point. to add code example fence it with triple backticks and optional file extension ```.cpp // C++ code example ``` or attach as .txt or .zip file --> ##### Issue submission checklist - [x] I report the issue, it's not a question <!-- OpenVINO team works with support forum, Stack Overflow and other communities to discuss problems. Tickets with question without real issue statement will be closed. --> - [ ] I checked the problem with documentation, FAQ, open issues, Stack Overflow, etc and have not found solution <!-- Places to check: * OpenVINO documentation: https://docs.openvinotoolkit.org/ * OpenVINO forum: https://community.intel.com/t5/Intel-Distribution-of-OpenVINO/bd-p/distribution-openvino-toolkit * OpenVINO issue tracker: https://github.com/openvinotoolkit/openvino/issues?q=is%3Aissue * Stack Overflow branch: https://stackoverflow.com/questions/tagged/openvino --> - [ ] There is reproducer code and related data files: images, videos, models, etc. <!-- The best reproducer -- test case for OpenVINO that we can add to the library. -->
1.0
[Bug] (minor) test namespace ? - ##### System information (version) <!-- Example - OpenVINO => 2020.4 - Operating System / Platform => Windows 64 Bit - Compiler => Visual Studio 2017 - Problem classification: Model Conversion - Framework: TensorFlow (if applicable) - Model name: ResNet50 (if applicable) --> - OpenVINO=> :grey_question: - Operating System / Platform => :grey_question: - Compiler => :grey_question: - Problem classification => :grey_question: ##### Detailed description <!-- your description --> I suppose this namespace was meant to be "LayerTest**s**Definitions", instead of "LayerTestDefinitions", to be aligned with the rest of the filters. I was getting an unexpected compile error and I barely noticed the missing 's': https://github.com/openvinotoolkit/openvino/blob/e8d5cf43d0e153f4f52c7be133f71f65c9eb4512/src/tests/functional/shared_test_classes/include/shared_test_classes/single_layer/prior_box.hpp#L29 ##### Steps to reproduce <!-- Describe your problem and steps you've done before you got to this point. to add code example fence it with triple backticks and optional file extension ```.cpp // C++ code example ``` or attach as .txt or .zip file --> ##### Issue submission checklist - [x] I report the issue, it's not a question <!-- OpenVINO team works with support forum, Stack Overflow and other communities to discuss problems. Tickets with question without real issue statement will be closed. --> - [ ] I checked the problem with documentation, FAQ, open issues, Stack Overflow, etc and have not found solution <!-- Places to check: * OpenVINO documentation: https://docs.openvinotoolkit.org/ * OpenVINO forum: https://community.intel.com/t5/Intel-Distribution-of-OpenVINO/bd-p/distribution-openvino-toolkit * OpenVINO issue tracker: https://github.com/openvinotoolkit/openvino/issues?q=is%3Aissue * Stack Overflow branch: https://stackoverflow.com/questions/tagged/openvino --> - [ ] There is reproducer code and related data files: images, videos, models, etc. <!-- The best reproducer -- test case for OpenVINO that we can add to the library. -->
test
minor test namespace system information version example openvino operating system platform windows bit compiler visual studio problem classification model conversion framework tensorflow if applicable model name if applicable openvino grey question operating system platform grey question compiler grey question problem classification grey question detailed description i suppose this namespace was meant to be layertest s definitions instead of layertestdefinitions to be aligned with the rest of the filters i was getting an unexpected compile error and i barely noticed the missing s steps to reproduce describe your problem and steps you ve done before you got to this point to add code example fence it with triple backticks and optional file extension cpp c code example or attach as txt or zip file issue submission checklist i report the issue it s not a question openvino team works with support forum stack overflow and other communities to discuss problems tickets with question without real issue statement will be closed i checked the problem with documentation faq open issues stack overflow etc and have not found solution places to check openvino documentation openvino forum openvino issue tracker stack overflow branch there is reproducer code and related data files images videos models etc the best reproducer test case for openvino that we can add to the library
1
654,339
21,648,352,441
IssuesEvent
2022-05-06 06:27:33
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
video.gazzetta.it - video or audio doesn't play
priority-normal browser-focus-geckoview engine-gecko
<!-- @browser: Firefox Mobile 100.0 --> <!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:100.0) Gecko/100.0 Firefox/100.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/103955 --> <!-- @extra_labels: browser-focus-geckoview --> **URL**: https://video.gazzetta.it/video-reral-madrid-finale-ancelotti-abbraccia-florentino-perez/5460ed74-cc92-11ec-9cc1-9243b00d5567?vclk=videobar **Browser / Version**: Firefox Mobile 100.0 **Operating System**: Android 11 **Tested Another Browser**: No **Problem type**: Video or audio doesn't play **Description**: The video or audio does not play **Steps to Reproduce**: Video does not play after clicking the play button <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2022/5/b4dbf693-5e8e-4678-b9ea-89544e9bb28a.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220428192727</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2022/5/d917ecb7-c99a-4fa5-8776-91389cf22c19) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
video.gazzetta.it - video or audio doesn't play - <!-- @browser: Firefox Mobile 100.0 --> <!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:100.0) Gecko/100.0 Firefox/100.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/103955 --> <!-- @extra_labels: browser-focus-geckoview --> **URL**: https://video.gazzetta.it/video-reral-madrid-finale-ancelotti-abbraccia-florentino-perez/5460ed74-cc92-11ec-9cc1-9243b00d5567?vclk=videobar **Browser / Version**: Firefox Mobile 100.0 **Operating System**: Android 11 **Tested Another Browser**: No **Problem type**: Video or audio doesn't play **Description**: The video or audio does not play **Steps to Reproduce**: Video does not play after clicking the play button <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2022/5/b4dbf693-5e8e-4678-b9ea-89544e9bb28a.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220428192727</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2022/5/d917ecb7-c99a-4fa5-8776-91389cf22c19) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_test
video gazzetta it video or audio doesn t play url browser version firefox mobile operating system android tested another browser no problem type video or audio doesn t play description the video or audio does not play steps to reproduce video does not play after clicking the play button view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel release hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
12,450
14,696,699,262
IssuesEvent
2021-01-04 00:27:12
Electroblob77/Wizardry
https://api.github.com/repos/Electroblob77/Wizardry
closed
MOD CONFLICTS
bug compatibility crash workaround exists
Minecraft version: 1.12.2 Wizardry version: 4.3.1 Environment: Singleplayer/LAN game/Server Issue details: I was creating a modpack for minecraft. Obviosly it had to have my favorite mod: electroblob's wizardry, but the modpack contained morph and when i tried to use morph something went wrong and said that electroblob's wizardry was responsible. Can you look and see what is the problem please? I really want to create a modpack with that mod before christmas Other mods involved: Morph (maybe ichunutil) [crash-2020-12-18_18.45.35-client.txt](https://github.com/Electroblob77/Wizardry/files/5717584/crash-2020-12-18_18.45.35-client.txt)
True
MOD CONFLICTS - Minecraft version: 1.12.2 Wizardry version: 4.3.1 Environment: Singleplayer/LAN game/Server Issue details: I was creating a modpack for minecraft. Obviosly it had to have my favorite mod: electroblob's wizardry, but the modpack contained morph and when i tried to use morph something went wrong and said that electroblob's wizardry was responsible. Can you look and see what is the problem please? I really want to create a modpack with that mod before christmas Other mods involved: Morph (maybe ichunutil) [crash-2020-12-18_18.45.35-client.txt](https://github.com/Electroblob77/Wizardry/files/5717584/crash-2020-12-18_18.45.35-client.txt)
non_test
mod conflicts minecraft version wizardry version environment singleplayer lan game server issue details i was creating a modpack for minecraft obviosly it had to have my favorite mod electroblob s wizardry but the modpack contained morph and when i tried to use morph something went wrong and said that electroblob s wizardry was responsible can you look and see what is the problem please i really want to create a modpack with that mod before christmas other mods involved morph maybe ichunutil
0
65,196
7,861,027,508
IssuesEvent
2018-06-21 22:09:08
Opentrons/opentrons
https://api.github.com/repos/Opentrons/opentrons
opened
TempDeck: Real Time Status During Run
app design feature
As a Run App user, I would like to be able to see my TempDeck's status during a run. ## Acceptance Criteria - [ ] Display current temperature in Run sidebar - [ ] Display target temperature in Run sidebar - [ ] Display status (string) in Run sidebar ## Implementation Notes - Dependent on 'Get Modules / 'Info' endpoint, #1653, including current and target temps - Will just poll this endpoint
1.0
TempDeck: Real Time Status During Run - As a Run App user, I would like to be able to see my TempDeck's status during a run. ## Acceptance Criteria - [ ] Display current temperature in Run sidebar - [ ] Display target temperature in Run sidebar - [ ] Display status (string) in Run sidebar ## Implementation Notes - Dependent on 'Get Modules / 'Info' endpoint, #1653, including current and target temps - Will just poll this endpoint
non_test
tempdeck real time status during run as a run app user i would like to be able to see my tempdeck s status during a run acceptance criteria display current temperature in run sidebar display target temperature in run sidebar display status string in run sidebar implementation notes dependent on get modules info endpoint including current and target temps will just poll this endpoint
0
214,612
16,601,168,924
IssuesEvent
2021-06-01 19:40:41
FEniCS/dolfinx
https://api.github.com/repos/FEniCS/dolfinx
closed
Bump gmsh version
testing
Currently, we are not using the latest release of gmsh due to the Eigen aligment issues, #1277 . Once we have sorted this out, we should get on to the latest version of gmsh. With this, we should remove the download of gmsh binaries, and rather use `pip3`, as only the following is required to make gmsh installable with pip3 (tested using ubuntu:20.04 as base): ``` apt-get update apt-get install libglu1 libxcursor-dev libxinerama1libxft2 python3-pip pip3 install gmsh --user export PATH=$PATH:/root/.local/bin python3 -c "import gmsh; print(gmsh.__version__)" ```
1.0
Bump gmsh version - Currently, we are not using the latest release of gmsh due to the Eigen aligment issues, #1277 . Once we have sorted this out, we should get on to the latest version of gmsh. With this, we should remove the download of gmsh binaries, and rather use `pip3`, as only the following is required to make gmsh installable with pip3 (tested using ubuntu:20.04 as base): ``` apt-get update apt-get install libglu1 libxcursor-dev libxinerama1libxft2 python3-pip pip3 install gmsh --user export PATH=$PATH:/root/.local/bin python3 -c "import gmsh; print(gmsh.__version__)" ```
test
bump gmsh version currently we are not using the latest release of gmsh due to the eigen aligment issues once we have sorted this out we should get on to the latest version of gmsh with this we should remove the download of gmsh binaries and rather use as only the following is required to make gmsh installable with tested using ubuntu as base apt get update apt get install libxcursor dev pip install gmsh user export path path root local bin c import gmsh print gmsh version
1
207,284
15,802,467,440
IssuesEvent
2021-04-03 09:52:51
Level/abstract-leveldown
https://api.github.com/repos/Level/abstract-leveldown
closed
Remove isTypedArray from tests
test
Continuing from https://github.com/Level/abstract-leveldown/issues/121#issuecomment-343722989, in our tests we have various instances of: https://github.com/Level/abstract-leveldown/blob/f9143c8bf5e9a4f42b0028ed612d8ede0d1fe513/test/get-test.js#L67-L69 Investigate if we still need this, because all implementations are now able to use `Buffer`, including in browsers.
1.0
Remove isTypedArray from tests - Continuing from https://github.com/Level/abstract-leveldown/issues/121#issuecomment-343722989, in our tests we have various instances of: https://github.com/Level/abstract-leveldown/blob/f9143c8bf5e9a4f42b0028ed612d8ede0d1fe513/test/get-test.js#L67-L69 Investigate if we still need this, because all implementations are now able to use `Buffer`, including in browsers.
test
remove istypedarray from tests continuing from in our tests we have various instances of investigate if we still need this because all implementations are now able to use buffer including in browsers
1
286,117
31,239,893,296
IssuesEvent
2023-08-20 18:33:29
MValle21/lamby_site
https://api.github.com/repos/MValle21/lamby_site
opened
CVE-2020-23064 (Medium) detected in jquery-3.3.1.js
Mend: dependency security vulnerability
## CVE-2020-23064 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-3.3.1.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.3.1/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.3.1/jquery.js</a></p> <p>Path to vulnerable library: /vendor/assets/javascripts/jquery-3.3.1.js</p> <p> Dependency Hierarchy: - :x: **jquery-3.3.1.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/MValle21/lamby_site/commit/58d2ba7cfe9616216cb126c69803c5ccd10d32b9">58d2ba7cfe9616216cb126c69803c5ccd10d32b9</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Cross Site Scripting vulnerability in jQuery 2.2.0 through 3.x before 3.5.0 allows a remote attacker to execute arbitrary code via the <options> element. <p>Publish Date: 2023-06-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-23064>CVE-2020-23064</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2023-06-26</p> <p>Fix Resolution: jquery - 3.5.0</p> </p> </details> <p></p>
True
CVE-2020-23064 (Medium) detected in jquery-3.3.1.js - ## CVE-2020-23064 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-3.3.1.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.3.1/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.3.1/jquery.js</a></p> <p>Path to vulnerable library: /vendor/assets/javascripts/jquery-3.3.1.js</p> <p> Dependency Hierarchy: - :x: **jquery-3.3.1.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/MValle21/lamby_site/commit/58d2ba7cfe9616216cb126c69803c5ccd10d32b9">58d2ba7cfe9616216cb126c69803c5ccd10d32b9</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Cross Site Scripting vulnerability in jQuery 2.2.0 through 3.x before 3.5.0 allows a remote attacker to execute arbitrary code via the <options> element. <p>Publish Date: 2023-06-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-23064>CVE-2020-23064</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2023-06-26</p> <p>Fix Resolution: jquery - 3.5.0</p> </p> </details> <p></p>
non_test
cve medium detected in jquery js cve medium severity vulnerability vulnerable library jquery js javascript library for dom operations library home page a href path to vulnerable library vendor assets javascripts jquery js dependency hierarchy x jquery js vulnerable library found in head commit a href found in base branch master vulnerability details cross site scripting vulnerability in jquery through x before allows a remote attacker to execute arbitrary code via the element publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery
0
53,087
6,299,649,498
IssuesEvent
2017-07-21 00:00:05
intel-analytics/BigDL
https://api.github.com/repos/intel-analytics/BigDL
opened
python textclassifier doc needs update for pip install
0.2 release test document python
https://github.com/intel-analytics/BigDL/tree/master/pyspark/bigdl/models/textclassifier only provide instructions for the manually installation, pip install users will not know where to find the bigdl home and jar files.
1.0
python textclassifier doc needs update for pip install - https://github.com/intel-analytics/BigDL/tree/master/pyspark/bigdl/models/textclassifier only provide instructions for the manually installation, pip install users will not know where to find the bigdl home and jar files.
test
python textclassifier doc needs update for pip install only provide instructions for the manually installation pip install users will not know where to find the bigdl home and jar files
1
141,074
11,392,230,592
IssuesEvent
2020-01-30 01:44:49
ray-project/ray
https://api.github.com/repos/ray-project/ray
opened
test_gather_benchmark in async_test.py is flaky
tests failing
For example, see https://travis-ci.com/ray-project/ray/jobs/281702565. The error is the following. ``` ―――――――――――――――――――――――――――― test_gather_benchmark ――――――――――――――――――――――――――――― init = None def test_gather_benchmark(init): @ray.remote def f(n): time.sleep(0.001 * n) return 42 async def test_async(): sum_time = 0. for _ in range(50): tasks = [f.remote(n) for n in range(20)] start = time.time() futures = [async_api.as_future(obj_id) for obj_id in tasks] await asyncio.gather(*futures) sum_time += time.time() - start return sum_time def baseline(): sum_time = 0. for _ in range(50): tasks = [f.remote(n) for n in range(20)] start = time.time() ray.get(tasks) sum_time += time.time() - start return sum_time # warm up baseline() # async get sum_time_1 = asyncio.get_event_loop().run_until_complete(test_async()) # get sum_time_2 = baseline() # Ensure the new implementation is not too slow. > assert sum_time_2 * 1.2 > sum_time_1 E assert (11.759512901306152 * 1.2) > 14.426274538040161 ``` Also, a test that takes 14 seconds seems too slow. Can we shorten this test and also not test anything that depends on timing to avoid flakiness? cc @sven1977 @suquark @simon-mo
1.0
test_gather_benchmark in async_test.py is flaky - For example, see https://travis-ci.com/ray-project/ray/jobs/281702565. The error is the following. ``` ―――――――――――――――――――――――――――― test_gather_benchmark ――――――――――――――――――――――――――――― init = None def test_gather_benchmark(init): @ray.remote def f(n): time.sleep(0.001 * n) return 42 async def test_async(): sum_time = 0. for _ in range(50): tasks = [f.remote(n) for n in range(20)] start = time.time() futures = [async_api.as_future(obj_id) for obj_id in tasks] await asyncio.gather(*futures) sum_time += time.time() - start return sum_time def baseline(): sum_time = 0. for _ in range(50): tasks = [f.remote(n) for n in range(20)] start = time.time() ray.get(tasks) sum_time += time.time() - start return sum_time # warm up baseline() # async get sum_time_1 = asyncio.get_event_loop().run_until_complete(test_async()) # get sum_time_2 = baseline() # Ensure the new implementation is not too slow. > assert sum_time_2 * 1.2 > sum_time_1 E assert (11.759512901306152 * 1.2) > 14.426274538040161 ``` Also, a test that takes 14 seconds seems too slow. Can we shorten this test and also not test anything that depends on timing to avoid flakiness? cc @sven1977 @suquark @simon-mo
test
test gather benchmark in async test py is flaky for example see the error is the following ―――――――――――――――――――――――――――― test gather benchmark ――――――――――――――――――――――――――――― init none def test gather benchmark init ray remote def f n time sleep n return async def test async sum time for in range tasks start time time futures await asyncio gather futures sum time time time start return sum time def baseline sum time for in range tasks start time time ray get tasks sum time time time start return sum time warm up baseline async get sum time asyncio get event loop run until complete test async get sum time baseline ensure the new implementation is not too slow assert sum time sum time e assert also a test that takes seconds seems too slow can we shorten this test and also not test anything that depends on timing to avoid flakiness cc suquark simon mo
1
30,433
7,197,030,240
IssuesEvent
2018-02-05 07:13:21
GSA/code-gov-api
https://api.github.com/repos/GSA/code-gov-api
closed
Add sort functionality
[code.gov] [help wanted] [issue-type] enhancement
## Context Sorting has not been implemented. The frontend would benefit from receiving sorted responses from the API. There seems to be an initial attempt to implement this. Take a look at `services/searcher/index.js`
1.0
Add sort functionality - ## Context Sorting has not been implemented. The frontend would benefit from receiving sorted responses from the API. There seems to be an initial attempt to implement this. Take a look at `services/searcher/index.js`
non_test
add sort functionality context sorting has not been implemented the frontend would benefit from receiving sorted responses from the api there seems to be an initial attempt to implement this take a look at services searcher index js
0
221,159
17,295,127,443
IssuesEvent
2021-07-25 15:13:05
Chatterino/chatterino2
https://api.github.com/repos/Chatterino/chatterino2
closed
PubSub creates only 1 connection, strictly limitting amount of topics Chatterino can listen to
bug help wanted needs testing pubsub
As defined here, these are the limits of how many connections and listen requests can PubSub make: https://github.com/Chatterino/chatterino2/blob/5ead70a314ebee8b93541e6cd6855ad48c0f42b9/src/providers/twitch/PubsubClient.hpp#L47-L48 While those do comply with [twitch docs](https://dev.twitch.tv/docs/pubsub#api-limits), we don't scale connections to 10 at max (`MAX_PUBSUB_CONNECTIONS` isn't used anywhere), making it sometimes impossible to listen to certain topics. As a proof to my previous statement, I've added a some code ([git diff](https://haste.zneix.eu/raw/hypakuguhu)) to see how many times we actually try to listen to a PubSub topic: ![debug popup](https://cdn.zneix.eu/sPOBcKj.png) Note: 100 listen attempts instead of expected 50 might be caused due to following found by @Bun ![thanks cbdg](https://cdn.zneix.eu/IF6mofx.png)
1.0
PubSub creates only 1 connection, strictly limitting amount of topics Chatterino can listen to - As defined here, these are the limits of how many connections and listen requests can PubSub make: https://github.com/Chatterino/chatterino2/blob/5ead70a314ebee8b93541e6cd6855ad48c0f42b9/src/providers/twitch/PubsubClient.hpp#L47-L48 While those do comply with [twitch docs](https://dev.twitch.tv/docs/pubsub#api-limits), we don't scale connections to 10 at max (`MAX_PUBSUB_CONNECTIONS` isn't used anywhere), making it sometimes impossible to listen to certain topics. As a proof to my previous statement, I've added a some code ([git diff](https://haste.zneix.eu/raw/hypakuguhu)) to see how many times we actually try to listen to a PubSub topic: ![debug popup](https://cdn.zneix.eu/sPOBcKj.png) Note: 100 listen attempts instead of expected 50 might be caused due to following found by @Bun ![thanks cbdg](https://cdn.zneix.eu/IF6mofx.png)
test
pubsub creates only connection strictly limitting amount of topics chatterino can listen to as defined here these are the limits of how many connections and listen requests can pubsub make while those do comply with we don t scale connections to at max max pubsub connections isn t used anywhere making it sometimes impossible to listen to certain topics as a proof to my previous statement i ve added a some code to see how many times we actually try to listen to a pubsub topic note listen attempts instead of expected might be caused due to following found by bun
1
143,234
11,525,120,669
IssuesEvent
2020-02-15 05:48:50
mgba-emu/mgba
https://api.github.com/repos/mgba-emu/mgba
closed
Stuck using Qt Multimedia
category:needs retest severity:bug
Change the Audio Driver to Qt Multimedia and try to load any rom, it will stuck with a white screen. In the second time you try to load a rom, you'll see this message: ``` using null output device, none available ``` Using qtmultimedia5-dev 5.2.1, xubuntu 14.04 amd64.
1.0
Stuck using Qt Multimedia - Change the Audio Driver to Qt Multimedia and try to load any rom, it will stuck with a white screen. In the second time you try to load a rom, you'll see this message: ``` using null output device, none available ``` Using qtmultimedia5-dev 5.2.1, xubuntu 14.04 amd64.
test
stuck using qt multimedia change the audio driver to qt multimedia and try to load any rom it will stuck with a white screen in the second time you try to load a rom you ll see this message using null output device none available using dev xubuntu
1
148,869
11,869,362,500
IssuesEvent
2020-03-26 10:49:21
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
hashi_vault should report secret when an error occurs
affects_2.10 feature needs_triage support:community support:core test
<!--- Verify first that your feature was not already discussed on GitHub --> <!--- Complete *all* sections as described, this form is processed automatically --> ##### SUMMARY <!--- Describe the new feature/improvement briefly below --> Currently, `hashi_vault` only reports a very basic exception when it can't access a secret. On large playbooks, where many secrets are accessed, this gives no indication which secret was failing. ##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME <!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure --> hashi_vault ##### ADDITIONAL INFORMATION <!--- Describe how the feature would be used, why it is needed and what it would solve --> The error reported by `hashi_vault` currently looks like this: <!--- Paste example playbooks or commands between quotes below --> ``` failed: [localhost] (item=example) => { "changed": false, "item": "example", "msg": "AnsibleError: An unhandled exception occurred while running the lookup plugin 'hashi_vault'. Error was a <class 'hvac.exceptions.Forbidden'>, original message: 1 error occurred:\n\t* permission denied\n\n" } ``` This should be more like: ``` failed: [localhost] (item=example) => { "changed": false, "item": "example", "msg": "AnsibleError: An unhandled exception occurred while running the lookup plugin 'hashi_vault'. Error was a <class 'hvac.exceptions.Forbidden'>, original message: 1 error occurred:\n\t* permission denied on secret secret=secret/example:secret\n\n" } ``` Or, the hvac.exceptions.Forbidden exception should be caught by the module and handled with a proper error message. <!--- HINT: You can also paste gist.github.com links for larger files -->
1.0
hashi_vault should report secret when an error occurs - <!--- Verify first that your feature was not already discussed on GitHub --> <!--- Complete *all* sections as described, this form is processed automatically --> ##### SUMMARY <!--- Describe the new feature/improvement briefly below --> Currently, `hashi_vault` only reports a very basic exception when it can't access a secret. On large playbooks, where many secrets are accessed, this gives no indication which secret was failing. ##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME <!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure --> hashi_vault ##### ADDITIONAL INFORMATION <!--- Describe how the feature would be used, why it is needed and what it would solve --> The error reported by `hashi_vault` currently looks like this: <!--- Paste example playbooks or commands between quotes below --> ``` failed: [localhost] (item=example) => { "changed": false, "item": "example", "msg": "AnsibleError: An unhandled exception occurred while running the lookup plugin 'hashi_vault'. Error was a <class 'hvac.exceptions.Forbidden'>, original message: 1 error occurred:\n\t* permission denied\n\n" } ``` This should be more like: ``` failed: [localhost] (item=example) => { "changed": false, "item": "example", "msg": "AnsibleError: An unhandled exception occurred while running the lookup plugin 'hashi_vault'. Error was a <class 'hvac.exceptions.Forbidden'>, original message: 1 error occurred:\n\t* permission denied on secret secret=secret/example:secret\n\n" } ``` Or, the hvac.exceptions.Forbidden exception should be caught by the module and handled with a proper error message. <!--- HINT: You can also paste gist.github.com links for larger files -->
test
hashi vault should report secret when an error occurs summary currently hashi vault only reports a very basic exception when it can t access a secret on large playbooks where many secrets are accessed this gives no indication which secret was failing issue type feature idea component name hashi vault additional information the error reported by hashi vault currently looks like this failed item example changed false item example msg ansibleerror an unhandled exception occurred while running the lookup plugin hashi vault error was a original message error occurred n t permission denied n n this should be more like failed item example changed false item example msg ansibleerror an unhandled exception occurred while running the lookup plugin hashi vault error was a original message error occurred n t permission denied on secret secret secret example secret n n or the hvac exceptions forbidden exception should be caught by the module and handled with a proper error message
1
262,381
22,837,578,219
IssuesEvent
2022-07-12 18:13:49
metaplex-foundation/metaplex
https://api.github.com/repos/metaplex-foundation/metaplex
closed
[Bug]: CandyMachineUI InvalidRentPayingAccount
needs tests bug
### Which package is this bug report for? candy machine ui ### Issue description Deployed a collection of 1 NFT to mainnet and it works as it should. I deployed the full collection of 1200 NFTs and when I try to mint I get an error `InvalidRentPayingAccount` when awaitConfirmationSignature is called. Seems to fail on the Set and Verify Collection instruction. Anyone run into this before ? ### Command _No response_ ### Relevant log output ```shell From the console of candymachineui (running latest) REST null result for jnt4qwRgHRRntsJKiwJD8dueztKgowGjoT5wfiMTZgccPXoux3EMEUWC4W9TpLLqV9avNezEWdVuNxpNFkvkqLN null connection.tsx:496 Rejected via websocket InvalidRentPayingAccount index.js:1 Timeout Error caught {err: 'InvalidRentPayingAccount', slot: 139694285, confirmations: 0} Caught failure: Error: Transaction failed: Instruction: Set and Verify Collection ``` ``` ### Operating system Linux ### Priority this issue should have Medium (should be fixed soon) ### Check the Docs First - [X] I have checked the docs and it didn't solve my issue
1.0
[Bug]: CandyMachineUI InvalidRentPayingAccount - ### Which package is this bug report for? candy machine ui ### Issue description Deployed a collection of 1 NFT to mainnet and it works as it should. I deployed the full collection of 1200 NFTs and when I try to mint I get an error `InvalidRentPayingAccount` when awaitConfirmationSignature is called. Seems to fail on the Set and Verify Collection instruction. Anyone run into this before ? ### Command _No response_ ### Relevant log output ```shell From the console of candymachineui (running latest) REST null result for jnt4qwRgHRRntsJKiwJD8dueztKgowGjoT5wfiMTZgccPXoux3EMEUWC4W9TpLLqV9avNezEWdVuNxpNFkvkqLN null connection.tsx:496 Rejected via websocket InvalidRentPayingAccount index.js:1 Timeout Error caught {err: 'InvalidRentPayingAccount', slot: 139694285, confirmations: 0} Caught failure: Error: Transaction failed: Instruction: Set and Verify Collection ``` ``` ### Operating system Linux ### Priority this issue should have Medium (should be fixed soon) ### Check the Docs First - [X] I have checked the docs and it didn't solve my issue
test
candymachineui invalidrentpayingaccount which package is this bug report for candy machine ui issue description deployed a collection of nft to mainnet and it works as it should i deployed the full collection of nfts and when i try to mint i get an error invalidrentpayingaccount when awaitconfirmationsignature is called seems to fail on the set and verify collection instruction anyone run into this before command no response relevant log output shell from the console of candymachineui running latest rest null result for null connection tsx rejected via websocket invalidrentpayingaccount index js timeout error caught err invalidrentpayingaccount slot confirmations caught failure error transaction failed instruction set and verify collection operating system linux priority this issue should have medium should be fixed soon check the docs first i have checked the docs and it didn t solve my issue
1
240,684
20,070,485,120
IssuesEvent
2022-02-04 05:48:40
StollLab/EasySpin
https://api.github.com/repos/StollLab/EasySpin
closed
Add tests that compare DStrain etc to explicit averaging
test
The goal is to assert that `Sys.DStrain` (and other strain parameters) do exactly what they advertise.
1.0
Add tests that compare DStrain etc to explicit averaging - The goal is to assert that `Sys.DStrain` (and other strain parameters) do exactly what they advertise.
test
add tests that compare dstrain etc to explicit averaging the goal is to assert that sys dstrain and other strain parameters do exactly what they advertise
1
202,353
15,281,345,502
IssuesEvent
2021-02-23 07:59:09
YM-rakennettu-ymparisto/kaavatietomalli
https://api.github.com/repos/YM-rakennettu-ymparisto/kaavatietomalli
closed
Espoon ruutuyleiskaavan digitointi
Testaus
Digitoidaan Espoon ruutuyleiskaavaa tietomallimuotoon. Tätä ei kuitenkaan viedä tietokantaan tai julkaista rajapinnassa, koska kaavaa ei voi saattaa ladattavaan muotoon tässä vaiheessa kaavoitusprosessia. Matala prioriteetti, koska rajoituksen takia vain rajattu hyöty testaukselle
1.0
Espoon ruutuyleiskaavan digitointi - Digitoidaan Espoon ruutuyleiskaavaa tietomallimuotoon. Tätä ei kuitenkaan viedä tietokantaan tai julkaista rajapinnassa, koska kaavaa ei voi saattaa ladattavaan muotoon tässä vaiheessa kaavoitusprosessia. Matala prioriteetti, koska rajoituksen takia vain rajattu hyöty testaukselle
test
espoon ruutuyleiskaavan digitointi digitoidaan espoon ruutuyleiskaavaa tietomallimuotoon tätä ei kuitenkaan viedä tietokantaan tai julkaista rajapinnassa koska kaavaa ei voi saattaa ladattavaan muotoon tässä vaiheessa kaavoitusprosessia matala prioriteetti koska rajoituksen takia vain rajattu hyöty testaukselle
1
120,762
12,093,722,101
IssuesEvent
2020-04-19 20:47:58
insites-co/insites-ui
https://api.github.com/repos/insites-co/insites-ui
closed
Radio and Checkbox docs examples
bug documentation good first issue help wanted
Currently our Radio and Checkbox examples are broken. The fix would be pretty easy - to move them to a separate components in Documentation repo and provide `useState` to them. I'd really appreciate some help here. https://insites-ui-docs.netlify.app/form/checkbox/ https://insites-ui-docs.netlify.app/form/radio/ https://github.com/insites-co/insites-ui-docs/blob/master/content/form/checkbox.mdx https://github.com/insites-co/insites-ui-docs/blob/master/content/form/radio.mdx
1.0
Radio and Checkbox docs examples - Currently our Radio and Checkbox examples are broken. The fix would be pretty easy - to move them to a separate components in Documentation repo and provide `useState` to them. I'd really appreciate some help here. https://insites-ui-docs.netlify.app/form/checkbox/ https://insites-ui-docs.netlify.app/form/radio/ https://github.com/insites-co/insites-ui-docs/blob/master/content/form/checkbox.mdx https://github.com/insites-co/insites-ui-docs/blob/master/content/form/radio.mdx
non_test
radio and checkbox docs examples currently our radio and checkbox examples are broken the fix would be pretty easy to move them to a separate components in documentation repo and provide usestate to them i d really appreciate some help here
0
299,660
25,916,269,860
IssuesEvent
2022-12-15 17:37:27
worMatty/deathrun_toolkit
https://api.github.com/repos/worMatty/deathrun_toolkit
closed
[Bug] Activator health scaling doesn't work sometimes
bug needs testing
When used on command it does not work. When triggered by the map it only seems to work on next round.
1.0
[Bug] Activator health scaling doesn't work sometimes - When used on command it does not work. When triggered by the map it only seems to work on next round.
test
activator health scaling doesn t work sometimes when used on command it does not work when triggered by the map it only seems to work on next round
1
8,262
5,540,592,182
IssuesEvent
2017-03-22 10:31:19
ESPD/ESPD-Service
https://api.github.com/repos/ESPD/ESPD-Service
closed
popup for sessions longer than 30 minutes and change the workflow
enhancement usability
1. Provide a popup after 30 Minutes saying that: "This service has not been used for more than 45 Minutes. Do you want to save the ESPD?" The drawback of this is that of the user is having lunch he will come back to late and the data might have been gone already. 2. Change workflow: The most elegant way is probably to change the workflow. On the finish page we do not allow to save the "XML" immediately but to say "Next". They will then be presented an overview. At the bottom we would offer to buttons "Save as PDF" and "Save as XML".
True
popup for sessions longer than 30 minutes and change the workflow - 1. Provide a popup after 30 Minutes saying that: "This service has not been used for more than 45 Minutes. Do you want to save the ESPD?" The drawback of this is that of the user is having lunch he will come back to late and the data might have been gone already. 2. Change workflow: The most elegant way is probably to change the workflow. On the finish page we do not allow to save the "XML" immediately but to say "Next". They will then be presented an overview. At the bottom we would offer to buttons "Save as PDF" and "Save as XML".
non_test
popup for sessions longer than minutes and change the workflow provide a popup after minutes saying that this service has not been used for more than minutes do you want to save the espd the drawback of this is that of the user is having lunch he will come back to late and the data might have been gone already change workflow the most elegant way is probably to change the workflow on the finish page we do not allow to save the xml immediately but to say next they will then be presented an overview at the bottom we would offer to buttons save as pdf and save as xml
0
327,814
9,981,610,280
IssuesEvent
2019-07-10 07:56:14
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
[0.8.2.7] Can't change authorization settings on wooden elevator call post.
Fixed Medium Priority QA Verified
From version 0.8.2.0 it has become possible to give access to the Elevator to another player. But you cann't give acces to its additional buttons. It's necessary to add the possibility of authorization in Wooden Elevator Call Post. Or that the buttons receive authorization rights from the Elevator. If I give access to the Elevator, then automatically I give access to the buttons attached to it.
1.0
[0.8.2.7] Can't change authorization settings on wooden elevator call post. - From version 0.8.2.0 it has become possible to give access to the Elevator to another player. But you cann't give acces to its additional buttons. It's necessary to add the possibility of authorization in Wooden Elevator Call Post. Or that the buttons receive authorization rights from the Elevator. If I give access to the Elevator, then automatically I give access to the buttons attached to it.
non_test
can t change authorization settings on wooden elevator call post from version it has become possible to give access to the elevator to another player but you cann t give acces to its additional buttons it s necessary to add the possibility of authorization in wooden elevator call post or that the buttons receive authorization rights from the elevator if i give access to the elevator then automatically i give access to the buttons attached to it
0
215,437
16,671,639,567
IssuesEvent
2021-06-07 11:39:53
WPChill/download-monitor
https://api.github.com/repos/WPChill/download-monitor
closed
On archives and other places where get_post_permalink is used the permalink is not OK
needs testing
Should filter the return with the following: `add_filter( 'post_type_link', array( $this, 'dlm_archive_permalink_structure' ), 15, 2 );` function something( $post_link, $post ) { if ( 'dlm_download' == $post->post_type ) { // fetch download object try{ /** @var DLM_Download $download */ $download = download_monitor()->service( 'download_repository' )->retrieve_single( $post->ID ); return $download->get_the_download_link(); } catch ( Exception $e ){ } } return $post_link; }
1.0
On archives and other places where get_post_permalink is used the permalink is not OK - Should filter the return with the following: `add_filter( 'post_type_link', array( $this, 'dlm_archive_permalink_structure' ), 15, 2 );` function something( $post_link, $post ) { if ( 'dlm_download' == $post->post_type ) { // fetch download object try{ /** @var DLM_Download $download */ $download = download_monitor()->service( 'download_repository' )->retrieve_single( $post->ID ); return $download->get_the_download_link(); } catch ( Exception $e ){ } } return $post_link; }
test
on archives and other places where get post permalink is used the permalink is not ok should filter the return with the following add filter post type link array this dlm archive permalink structure function something post link post if dlm download post post type fetch download object try var dlm download download download download monitor service download repository retrieve single post id return download get the download link catch exception e return post link
1
98,956
20,827,599,660
IssuesEvent
2022-03-19 00:02:34
pulumi/pulumi
https://api.github.com/repos/pulumi/pulumi
closed
[codegen/go] Add support for specifying Go dependencies
kind/enhancement language/go area/codegen
## Hello! <!-- Please leave this section as-is, it's designed to help others in the community know how to interact with our GitHub issues. --> - Vote on this issue by adding a 👍 reaction - If you want to implement this feature, comment to let us know (we'll work with you on design, scheduling, etc.) ## Issue details There are situations where an MLC provider may make use of other resource providers inside its implementation (e.g. `pulumi-random`, `pulumi-docker`, etc.), but not exposed publicly. Such dependencies _aren't_ listed in the provider's associated Go SDK's `go.mod` since nothing in the SDK itself refers to them. Which can be problematic for plugin discovery. Since these dependencies are not listed in the Go SDK's `go.mod`, the associated plugins will not be installed automatically for users of the Go SDK. Note that the Go SDK codegen does not generate the `go.mod` file -- this must be created manually. And manually adding such dependencies to the SDK's `go.mod` is not sufficient to address this because `go mod tidy` will remove any unused dependencies. To avoid this, we should add a new capability to the Go codegen that emits internal dummy imports (referencing some well known exported type) for any Go dependencies listed in the schema. That way, the dependencies can be manually added to `go.mod` and won't be stripped when `go mod tidy` is run.
1.0
[codegen/go] Add support for specifying Go dependencies - ## Hello! <!-- Please leave this section as-is, it's designed to help others in the community know how to interact with our GitHub issues. --> - Vote on this issue by adding a 👍 reaction - If you want to implement this feature, comment to let us know (we'll work with you on design, scheduling, etc.) ## Issue details There are situations where an MLC provider may make use of other resource providers inside its implementation (e.g. `pulumi-random`, `pulumi-docker`, etc.), but not exposed publicly. Such dependencies _aren't_ listed in the provider's associated Go SDK's `go.mod` since nothing in the SDK itself refers to them. Which can be problematic for plugin discovery. Since these dependencies are not listed in the Go SDK's `go.mod`, the associated plugins will not be installed automatically for users of the Go SDK. Note that the Go SDK codegen does not generate the `go.mod` file -- this must be created manually. And manually adding such dependencies to the SDK's `go.mod` is not sufficient to address this because `go mod tidy` will remove any unused dependencies. To avoid this, we should add a new capability to the Go codegen that emits internal dummy imports (referencing some well known exported type) for any Go dependencies listed in the schema. That way, the dependencies can be manually added to `go.mod` and won't be stripped when `go mod tidy` is run.
non_test
add support for specifying go dependencies hello vote on this issue by adding a 👍 reaction if you want to implement this feature comment to let us know we ll work with you on design scheduling etc issue details there are situations where an mlc provider may make use of other resource providers inside its implementation e g pulumi random pulumi docker etc but not exposed publicly such dependencies aren t listed in the provider s associated go sdk s go mod since nothing in the sdk itself refers to them which can be problematic for plugin discovery since these dependencies are not listed in the go sdk s go mod the associated plugins will not be installed automatically for users of the go sdk note that the go sdk codegen does not generate the go mod file this must be created manually and manually adding such dependencies to the sdk s go mod is not sufficient to address this because go mod tidy will remove any unused dependencies to avoid this we should add a new capability to the go codegen that emits internal dummy imports referencing some well known exported type for any go dependencies listed in the schema that way the dependencies can be manually added to go mod and won t be stripped when go mod tidy is run
0
40,524
10,545,604,739
IssuesEvent
2019-10-02 19:31:31
zowe/zowe-install-packaging
https://api.github.com/repos/zowe/zowe-install-packaging
closed
Write a tool to clean up artifactory history
CI/CD build
Need to clean up branch/PR builds regularly to free up storage.
1.0
Write a tool to clean up artifactory history - Need to clean up branch/PR builds regularly to free up storage.
non_test
write a tool to clean up artifactory history need to clean up branch pr builds regularly to free up storage
0
65,376
6,961,596,123
IssuesEvent
2017-12-08 10:05:51
radare/radare2
https://api.github.com/repos/radare/radare2
closed
anal.noncode is not respected
anal bug has-test
Following the new broken tests for anal.calls/anal.noncode -> https://github.com/radare/radare2-regressions/commit/8a22c7323815c85fd7e07230924e461f21aa2ab1 . ![image](https://cloud.githubusercontent.com/assets/1408600/26532220/21f6641a-43fb-11e7-9786-5ef93923ef70.png)
1.0
anal.noncode is not respected - Following the new broken tests for anal.calls/anal.noncode -> https://github.com/radare/radare2-regressions/commit/8a22c7323815c85fd7e07230924e461f21aa2ab1 . ![image](https://cloud.githubusercontent.com/assets/1408600/26532220/21f6641a-43fb-11e7-9786-5ef93923ef70.png)
test
anal noncode is not respected following the new broken tests for anal calls anal noncode
1
122,041
12,139,823,331
IssuesEvent
2020-04-23 19:29:39
amanzi/amanzi
https://api.github.com/repos/amanzi/amanzi
closed
Input Spec/Multiple point definitions
documentation
**typedef_point** contains an attribute "coordinate" and is used for the region type point as a typedef_labeled_point (typedef_point with additional attribute "name"): `<point name="abc" coordinate="(x, y, z)"/>` Points are also defined as elements under polygons as direct coordinates: <xs:complexType name="typedef_polygonal_surface"> <xs:sequence> <xs:element name="point" type="typedef_coordinate" minOccurs="0" maxOccurs="unbounded" /> ... `<point>(x, y, z)</point>` Validating a file with a polygon currently gives schema validation error. Recommending we change the polygon point type to typedef_point for consistency: `<point coordinate="(x, y, z)"/>`
1.0
Input Spec/Multiple point definitions - **typedef_point** contains an attribute "coordinate" and is used for the region type point as a typedef_labeled_point (typedef_point with additional attribute "name"): `<point name="abc" coordinate="(x, y, z)"/>` Points are also defined as elements under polygons as direct coordinates: <xs:complexType name="typedef_polygonal_surface"> <xs:sequence> <xs:element name="point" type="typedef_coordinate" minOccurs="0" maxOccurs="unbounded" /> ... `<point>(x, y, z)</point>` Validating a file with a polygon currently gives schema validation error. Recommending we change the polygon point type to typedef_point for consistency: `<point coordinate="(x, y, z)"/>`
non_test
input spec multiple point definitions typedef point contains an attribute coordinate and is used for the region type point as a typedef labeled point typedef point with additional attribute name points are also defined as elements under polygons as direct coordinates x y z validating a file with a polygon currently gives schema validation error recommending we change the polygon point type to typedef point for consistency
0
757,687
26,524,509,622
IssuesEvent
2023-01-19 07:26:43
etjump/etjump
https://api.github.com/repos/etjump/etjump
opened
Expose g_save, g_ghostPlayers and g_nofatigue to vote system
Feature Priority: Normal
Players should be able to vote on these settings.
1.0
Expose g_save, g_ghostPlayers and g_nofatigue to vote system - Players should be able to vote on these settings.
non_test
expose g save g ghostplayers and g nofatigue to vote system players should be able to vote on these settings
0
541,969
15,836,760,477
IssuesEvent
2021-04-06 19:48:04
GoogleChrome/lighthouse
https://api.github.com/repos/GoogleChrome/lighthouse
closed
Warnings: Lighthouse was unable to automatically check the `font-display` value for the origin https://collegeadmission.co. Please help me for this issue fix it...
needs-more-info needs-priority
Warnings: Lighthouse was unable to automatically check the `font-display` value for the origin https://collegeadmission.co. Please help me with this issue to fix it... Already checking with...Below Given... link rel="preload" as="style" href="https://fonts.googleapis.com/css2?family=Open+Sans&display=block link rel="preload" as="font" href="https://fonts.gstatic.com/s/opensans/v18/mem8YaGs126MiZpBA-UFVZ0bf8pkAg.woff2" type="font/woff" crossorigin="anonymous" @font-face { font-family:'Open Sans'; src: url(https://fonts.gstatic.com/s/opensans/v18/mem8YaGs126MiZpBA-UFVZ0bf8pkAg.woff2) format("woff2"); font-display: auto; } ![image](https://user-images.githubusercontent.com/5756147/103191588-8cbb2c00-48fb-11eb-9c8f-a59711022fbb.png)
1.0
Warnings: Lighthouse was unable to automatically check the `font-display` value for the origin https://collegeadmission.co. Please help me for this issue fix it... - Warnings: Lighthouse was unable to automatically check the `font-display` value for the origin https://collegeadmission.co. Please help me with this issue to fix it... Already checking with...Below Given... link rel="preload" as="style" href="https://fonts.googleapis.com/css2?family=Open+Sans&display=block link rel="preload" as="font" href="https://fonts.gstatic.com/s/opensans/v18/mem8YaGs126MiZpBA-UFVZ0bf8pkAg.woff2" type="font/woff" crossorigin="anonymous" @font-face { font-family:'Open Sans'; src: url(https://fonts.gstatic.com/s/opensans/v18/mem8YaGs126MiZpBA-UFVZ0bf8pkAg.woff2) format("woff2"); font-display: auto; } ![image](https://user-images.githubusercontent.com/5756147/103191588-8cbb2c00-48fb-11eb-9c8f-a59711022fbb.png)
non_test
warnings lighthouse was unable to automatically check the font display value for the origin please help me for this issue fix it warnings lighthouse was unable to automatically check the font display value for the origin please help me with this issue to fix it already checking with below given link rel preload as style href link rel preload as font href type font woff crossorigin anonymous font face font family open sans src url format font display auto
0
82,943
7,857,118,365
IssuesEvent
2018-06-21 09:44:35
brave/browser-laptop
https://api.github.com/repos/brave/browser-laptop
closed
Prompt for upgrade to regular build if ads are off for extended period
initiative/bat-ads initiative/bat-ads/ads-test release/blocking
The ads test build isn't up to date with other security and privacy features in Brave or Muon. Nobody should use this build unless they're actively participating in the ads test. If someone uses this build for an hour without turning on the ads feature, they should be prompted to "upgrade" to the latest current release build of Brave, by way of a notification bar — just like normal software upgrades. cf., https://github.com/brave-intl/internal/issues/57
1.0
Prompt for upgrade to regular build if ads are off for extended period - The ads test build isn't up to date with other security and privacy features in Brave or Muon. Nobody should use this build unless they're actively participating in the ads test. If someone uses this build for an hour without turning on the ads feature, they should be prompted to "upgrade" to the latest current release build of Brave, by way of a notification bar — just like normal software upgrades. cf., https://github.com/brave-intl/internal/issues/57
test
prompt for upgrade to regular build if ads are off for extended period the ads test build isn t up to date with other security and privacy features in brave or muon nobody should use this build unless they re actively participating in the ads test if someone uses this build for an hour without turning on the ads feature they should be prompted to upgrade to the latest current release build of brave by way of a notification bar — just like normal software upgrades cf
1
19,706
2,622,161,395
IssuesEvent
2015-03-04 00:10:13
byzhang/terrastore
https://api.github.com/repos/byzhang/terrastore
opened
Improve underlying http stack
auto-migrated Priority-Medium Project-Terrastore Type-Enhancement
``` improve underlying http sever/client stack see comarison agains apache http://groups.google.com/group/terrastore-discussions/browse_thread/thread/d2cf6 58e9dd01036 ``` Original issue reported on code.google.com by `Andrei.Pozolotin` on 10 Mar 2011 at 4:18
1.0
Improve underlying http stack - ``` improve underlying http sever/client stack see comarison agains apache http://groups.google.com/group/terrastore-discussions/browse_thread/thread/d2cf6 58e9dd01036 ``` Original issue reported on code.google.com by `Andrei.Pozolotin` on 10 Mar 2011 at 4:18
non_test
improve underlying http stack improve underlying http sever client stack see comarison agains apache original issue reported on code google com by andrei pozolotin on mar at
0
208,279
16,108,666,311
IssuesEvent
2021-04-27 18:03:02
spring-io/spring-asciidoctor-backends
https://api.github.com/repos/spring-io/spring-asciidoctor-backends
closed
Wrong Spring Maven repo in README.adoc
type: documentation
README.adoc says: ``` maven { url "https://repo.spring.io/release" } ``` but must be: ``` maven { url "https://repo.spring.io/milestone" } ``` since there are only milestone releases for now.
1.0
Wrong Spring Maven repo in README.adoc - README.adoc says: ``` maven { url "https://repo.spring.io/release" } ``` but must be: ``` maven { url "https://repo.spring.io/milestone" } ``` since there are only milestone releases for now.
non_test
wrong spring maven repo in readme adoc readme adoc says maven url but must be maven url since there are only milestone releases for now
0
209,231
23,705,172,307
IssuesEvent
2022-08-29 23:49:44
itziklavon/springdemo
https://api.github.com/repos/itziklavon/springdemo
closed
swagger-jersey-jaxrs-1.5.21.jar: 1 vulnerabilities (highest severity is: 7.5)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>swagger-jersey-jaxrs-1.5.21.jar</b></p></summary> <p></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/itziklavon/springdemo/commit/98e4930dff8ac42633a041eaf411f0caf6a38b4c">98e4930dff8ac42633a041eaf411f0caf6a38b4c</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2017-18640](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-18640) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | snakeyaml-1.19.jar | Transitive | 1.6.0 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2017-18640</summary> ### Vulnerable Library - <b>snakeyaml-1.19.jar</b></p> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar</p> <p> Dependency Hierarchy: - swagger-jersey-jaxrs-1.5.21.jar (Root Library) - swagger-jaxrs-1.5.21.jar - swagger-core-1.5.21.jar - jackson-dataformat-yaml-2.9.6.jar - :x: **snakeyaml-1.19.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/itziklavon/springdemo/commit/98e4930dff8ac42633a041eaf411f0caf6a38b4c">98e4930dff8ac42633a041eaf411f0caf6a38b4c</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> The Alias feature in SnakeYAML before 1.26 allows entity expansion during a load operation, a related issue to CVE-2003-1564. <p>Publish Date: 2019-12-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-18640>CVE-2017-18640</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-18640">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-18640</a></p> <p>Release Date: 2019-12-12</p> <p>Fix Resolution (org.yaml:snakeyaml): 1.26</p> <p>Direct dependency fix Resolution (io.swagger:swagger-jersey-jaxrs): 1.6.0</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
True
swagger-jersey-jaxrs-1.5.21.jar: 1 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>swagger-jersey-jaxrs-1.5.21.jar</b></p></summary> <p></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/itziklavon/springdemo/commit/98e4930dff8ac42633a041eaf411f0caf6a38b4c">98e4930dff8ac42633a041eaf411f0caf6a38b4c</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2017-18640](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-18640) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | snakeyaml-1.19.jar | Transitive | 1.6.0 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2017-18640</summary> ### Vulnerable Library - <b>snakeyaml-1.19.jar</b></p> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar</p> <p> Dependency Hierarchy: - swagger-jersey-jaxrs-1.5.21.jar (Root Library) - swagger-jaxrs-1.5.21.jar - swagger-core-1.5.21.jar - jackson-dataformat-yaml-2.9.6.jar - :x: **snakeyaml-1.19.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/itziklavon/springdemo/commit/98e4930dff8ac42633a041eaf411f0caf6a38b4c">98e4930dff8ac42633a041eaf411f0caf6a38b4c</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> The Alias feature in SnakeYAML before 1.26 allows entity expansion during a load operation, a related issue to CVE-2003-1564. <p>Publish Date: 2019-12-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-18640>CVE-2017-18640</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-18640">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-18640</a></p> <p>Release Date: 2019-12-12</p> <p>Fix Resolution (org.yaml:snakeyaml): 1.26</p> <p>Direct dependency fix Resolution (io.swagger:swagger-jersey-jaxrs): 1.6.0</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
non_test
swagger jersey jaxrs jar vulnerabilities highest severity is vulnerable library swagger jersey jaxrs jar path to dependency file pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high snakeyaml jar transitive details cve vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar dependency hierarchy swagger jersey jaxrs jar root library swagger jaxrs jar swagger core jar jackson dataformat yaml jar x snakeyaml jar vulnerable library found in head commit a href found in base branch master vulnerability details the alias feature in snakeyaml before allows entity expansion during a load operation a related issue to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org yaml snakeyaml direct dependency fix resolution io swagger swagger jersey jaxrs step up your open source security game with mend
0
280,363
8,680,840,699
IssuesEvent
2018-12-01 14:42:04
firecracker-microvm/firecracker
https://api.github.com/repos/firecracker-microvm/firecracker
closed
Remove DeviceState from NetworkInterfaceConfig
Contribute: Good First Issue Priority: High Quality: Improvement
The device state is a leftover from our previous device model, when we wanted to be able to hot plug/unplug devices. This is not something that we are planing to implement in the near future and should be remove. The DeviceState structure is defined in `vmm/src/vmm_config/mod.rs` and it is used in the NetworkInterfaceConfig structure definition in `vmm/src/vmm_config/net.rs`. With this change, the swagger definition (api_server/swagger/firecracker.yaml) also needs to be updated.
1.0
Remove DeviceState from NetworkInterfaceConfig - The device state is a leftover from our previous device model, when we wanted to be able to hot plug/unplug devices. This is not something that we are planing to implement in the near future and should be remove. The DeviceState structure is defined in `vmm/src/vmm_config/mod.rs` and it is used in the NetworkInterfaceConfig structure definition in `vmm/src/vmm_config/net.rs`. With this change, the swagger definition (api_server/swagger/firecracker.yaml) also needs to be updated.
non_test
remove devicestate from networkinterfaceconfig the device state is a leftover from our previous device model when we wanted to be able to hot plug unplug devices this is not something that we are planing to implement in the near future and should be remove the devicestate structure is defined in vmm src vmm config mod rs and it is used in the networkinterfaceconfig structure definition in vmm src vmm config net rs with this change the swagger definition api server swagger firecracker yaml also needs to be updated
0
85,758
7,988,977,256
IssuesEvent
2018-07-19 12:56:02
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
github.com/cockroachdb/cockroach/pkg/storage: TestStoreRangeSystemSplits failed under stress
A-core-kv C-test-failure O-robot
SHA: https://github.com/cockroachdb/cockroach/commits/97c3622c2a6f4ec579ed93655b9d8a83732ac654 Parameters: ``` TAGS= GOFLAGS=-race ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=693452&tab=buildLog ``` I180602 07:21:25.666097 111393 gossip/gossip.go:333 [n1] NodeDescriptor set to node_id:1 address:<network_field:"" address_field:"" > attrs:<> locality:<> ServerVersion:<major_val:0 minor_val:0 patch:0 unstable:0 > I180602 07:21:25.755244 111393 storage/store.go:1454 [s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available I180602 07:21:25.822572 111432 storage/replica_command.go:863 [split,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2] W180602 07:21:25.865205 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer I180602 07:21:25.915992 111562 storage/replica_command.go:863 [split,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3] W180602 07:21:25.934571 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:25.938941 111461 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "unnamed" id=17deafae key=/Table/SystemConfigSpan/Start rw=true pri=0.03184783 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,53 orig=0.000000123,53 max=0.000000123,54 wto=false rop=false seq=1 I180602 07:21:26.066952 111572 storage/replica_command.go:863 [split,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4] W180602 07:21:26.094922 111464 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "unnamed" id=17deafae key=/Table/SystemConfigSpan/Start rw=true pri=0.03184783 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,53 orig=0.000000123,53 max=0.000000123,54 wto=false rop=false seq=1 W180602 07:21:26.143966 111457 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=46177098 key=/Local/Range/System/NodeLiveness/RangeDescriptor rw=true pri=0.06770799 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,182 orig=0.000000123,182 max=0.000000123,196 wto=false rop=false seq=1 W180602 07:21:26.171835 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer I180602 07:21:26.190567 111466 storage/replica_command.go:863 [split,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5] W180602 07:21:26.200246 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:26.223842 111587 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "unnamed" id=17deafae key=/Table/SystemConfigSpan/Start rw=true pri=0.03184783 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,53 orig=0.000000123,53 max=0.000000123,54 wto=false rop=false seq=1 W180602 07:21:26.259843 103909 gossip/gossip.go:1298 [n1] first range unavailable; trying remaining resolvers I180602 07:21:26.262515 111416 gossip/client.go:129 [n1] started gossip client to 127.0.0.1:44307 I180602 07:21:26.304909 111402 storage/replica_command.go:863 [split,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6] W180602 07:21:26.307892 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer W180602 07:21:26.340668 111552 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "unnamed" id=17deafae key=/Table/SystemConfigSpan/Start rw=true pri=0.03184783 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,53 orig=0.000000123,53 max=0.000000123,54 wto=false rop=false seq=1 I180602 07:21:26.413526 111603 storage/replica_command.go:863 [split,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7] W180602 07:21:26.434307 111573 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "unnamed" id=17deafae key=/Table/SystemConfigSpan/Start rw=true pri=0.03184783 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,53 orig=0.000000123,53 max=0.000000123,54 wto=false rop=false seq=1 I180602 07:21:26.489843 111577 storage/replica_command.go:863 [split,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8] I180602 07:21:26.601373 111652 storage/replica_command.go:863 [split,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9] I180602 07:21:26.700193 111406 storage/replica_command.go:863 [split,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10] W180602 07:21:26.748760 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer W180602 07:21:26.761617 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer E180602 07:21:26.803145 111637 storage/consistency_queue.go:125 [replica consistency checker,s1,r1/1:/{Min-System/}] computing own checksum: could not resolve node ID 1: dummy resolver E180602 07:21:26.803792 111637 storage/queue.go:778 [replica consistency checker,s1,r1/1:/{Min-System/}] computing own checksum: could not resolve node ID 1: dummy resolver W180602 07:21:26.812209 104013 gossip/gossip.go:1298 [n2] first range unavailable; trying remaining resolvers I180602 07:21:26.818283 111638 storage/replica_command.go:863 [split,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11] I180602 07:21:26.918273 111580 storage/replica_command.go:863 [split,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12] W180602 07:21:27.012870 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer I180602 07:21:27.016534 104022 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming I180602 07:21:27.038033 111616 storage/replica_command.go:863 [split,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13] I180602 07:21:27.156325 111649 storage/replica_command.go:863 [split,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14] W180602 07:21:27.223175 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer W180602 07:21:27.281224 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer I180602 07:21:27.300168 111690 storage/replica_command.go:863 [split,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15] W180602 07:21:27.311006 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer I180602 07:21:27.420307 111719 storage/replica_command.go:863 [split,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16] I180602 07:21:27.542361 111701 storage/replica_command.go:863 [split,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17] W180602 07:21:27.645111 111725 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=da9ad7de key=/Local/Range/Table/19/RangeDescriptor rw=true pri=0.05901391 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,1879 orig=0.000000123,1879 max=0.000000123,1893 wto=false rop=false seq=1 W180602 07:21:27.659094 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer I180602 07:21:27.684752 111680 storage/replica_command.go:863 [split,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18] W180602 07:21:27.763939 111780 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=125a67e3 key=/Local/Range/Table/20/RangeDescriptor rw=true pri=0.03635720 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2111 orig=0.000000123,2042 max=0.000000123,2084 wto=false rop=false seq=1 W180602 07:21:27.766883 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer E180602 07:21:27.806309 111766 storage/consistency_queue.go:125 [replica consistency checker,s1,r4/1:/System/{NodeLive…-tsd}] computing own checksum: could not resolve node ID 1: dummy resolver E180602 07:21:27.806975 111766 storage/queue.go:778 [replica consistency checker,s1,r4/1:/System/{NodeLive…-tsd}] computing own checksum: could not resolve node ID 1: dummy resolver I180602 07:21:27.818124 111796 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065 I180602 07:21:27.826497 111796 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming I180602 07:21:27.874237 111726 storage/replica_command.go:863 [split,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19] W180602 07:21:27.982386 111770 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=c476a094 key=/Local/Range/Table/21/RangeDescriptor rw=true pri=0.02374100 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2252 orig=0.000000123,2252 max=0.000000123,2275 wto=false rop=false seq=1 I180602 07:21:28.029652 111810 storage/replica_command.go:863 [split,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20] W180602 07:21:28.031230 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:28.118611 111774 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=fcaa46c5 key=/Local/Range/Table/22/RangeDescriptor rw=true pri=0.00146925 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2397 orig=0.000000123,2397 max=0.000000123,2414 wto=false rop=false seq=1 I180602 07:21:28.154689 111775 storage/replica_command.go:863 [split,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21] W180602 07:21:28.213520 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer I180602 07:21:28.274397 111813 storage/replica_command.go:863 [split,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22] W180602 07:21:28.338665 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer W180602 07:21:28.352719 111831 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=a8866910 key=/Local/Range/Table/50/RangeDescriptor rw=true pri=0.01775119 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2619 orig=0.000000123,2619 max=0.000000123,2648 wto=false rop=false seq=1 I180602 07:21:28.382776 111710 storage/replica_command.go:863 [split,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23] W180602 07:21:28.458973 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer W180602 07:21:28.477987 111834 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=3ad9f13b key=/Local/Range/Table/51/RangeDescriptor rw=true pri=0.02323743 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2754 orig=0.000000123,2754 max=0.000000123,2801 wto=false rop=false seq=1 I180602 07:21:28.508700 111631 storage/replica_command.go:863 [split,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r24] W180602 07:21:28.583750 111755 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=199fc42e key=/Local/Range/Table/52/RangeDescriptor rw=true pri=0.01267702 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2880 orig=0.000000123,2880 max=0.000000123,2925 wto=false rop=false seq=1 W180602 07:21:28.621310 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer I180602 07:21:28.623619 111837 storage/replica_command.go:863 [split,s1,r24/1:/{Table/53-Max}] initiating a split of this range at key /Table/54 [r25] W180602 07:21:28.689811 111758 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=4806ee51 key=/Local/Range/Table/53/RangeDescriptor rw=true pri=0.03011487 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,3011 orig=0.000000123,3011 max=0.000000123,3022 wto=false rop=false seq=1 E180602 07:21:28.791818 111633 storage/consistency_queue.go:125 [replica consistency checker,s1,r8/1:/Table/1{1-2}] computing own checksum: could not resolve node ID 1: dummy resolver E180602 07:21:28.792442 111633 storage/queue.go:778 [replica consistency checker,s1,r8/1:/Table/1{1-2}] computing own checksum: could not resolve node ID 1: dummy resolver I180602 07:21:28.806116 111842 storage/replica_command.go:863 [split,s1,r25/1:/{Table/54-Max}] initiating a split of this range at key /Table/55 [r26] I180602 07:21:28.844513 111819 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065 W180602 07:21:28.898378 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer I180602 07:21:28.909067 111819 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming I180602 07:21:28.950153 111879 storage/replica_command.go:863 [split,s1,r26/1:/{Table/55-Max}] initiating a split of this range at key /Table/56 [r27] I180602 07:21:29.067577 111852 storage/replica_command.go:863 [split,s1,r27/1:/{Table/56-Max}] initiating a split of this range at key /Table/57 [r28] W180602 07:21:29.149687 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:29.178637 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer I180602 07:21:29.223093 111393 util/stop/stopper.go:471 quiescing; tasks left: 1 storage.replicate: purgatory processing replica 1 storage.intentResolver: processing intents I180602 07:21:29.223977 111393 util/stop/stopper.go:471 quiescing; tasks left: 1 storage.intentResolver: processing intents W180602 07:21:29.406961 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer W180602 07:21:29.434693 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer W180602 07:21:29.729400 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer W180602 07:21:29.802959 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer I180602 07:21:29.845634 111896 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065 I180602 07:21:29.856527 111896 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming W180602 07:21:30.184916 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:30.217264 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:30.330989 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer W180602 07:21:30.492147 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer W180602 07:21:30.719824 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer W180602 07:21:30.843026 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer I180602 07:21:30.846414 111868 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065 I180602 07:21:30.853735 111868 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming W180602 07:21:31.088347 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:31.202071 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer W180602 07:21:31.306937 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:31.411590 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer W180602 07:21:31.763071 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer I180602 07:21:31.847405 111959 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065 I180602 07:21:31.855401 111959 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming W180602 07:21:31.957920 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer W180602 07:21:32.120898 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:32.275896 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer W180602 07:21:32.398697 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:32.482426 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer W180602 07:21:32.785552 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer W180602 07:21:32.843238 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer I180602 07:21:32.848515 111974 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065 I180602 07:21:32.872364 111974 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming W180602 07:21:33.014392 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:33.313420 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer W180602 07:21:33.441125 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer W180602 07:21:33.506666 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:33.761849 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer I180602 07:21:33.849460 111914 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065 W180602 07:21:33.851714 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer I180602 07:21:33.855960 111914 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming W180602 07:21:33.986173 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:34.319167 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer leaktest.go:110: Leaked goroutine: goroutine 111416 [select]: github.com/cockroachdb/cockroach/pkg/gossip.(*client).gossip(0xc422408dd0, 0x3548840, 0xc420da95f0, 0xc420e05980, 0x3562720, 0xc4212e0090, 0xc4207ea3f0, 0xc4232828f0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/client.go:323 +0x400 github.com/cockroachdb/cockroach/pkg/gossip.(*client).startLocked.func1(0x3548840, 0xc420da95f0) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/client.go:130 +0x5fd github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc4212e0030, 0xc4207ea3f0, 0xc421028280) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:193 +0x14b created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:186 +0xbb leaktest.go:110: Leaked goroutine: goroutine 111417 [select]: github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.newClientStream.func5(0xc4201ba300, 0xc420fc2780, 0x3548840, 0xc420da9680) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/stream.go:321 +0x120 created by github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.newClientStream /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/stream.go:320 +0x1200 leaktest.go:110: Leaked goroutine: goroutine 111425 [select]: github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*recvBufferReader).read(0xc4211e79a0, 0xc4203e0f10, 0x5, 0x5, 0xc421cea660, 0x843cb0, 0x2810ab0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:142 +0x19c github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*recvBufferReader).Read(0xc4211e79a0, 0xc4203e0f10, 0x5, 0x5, 0x842f6c, 0xc421cea660, 0xc4233bd8c8) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:131 +0x82 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*transportReader).Read(0xc420da97a0, 0xc4203e0f10, 0x5, 0x5, 0xf25293, 0xc421cea660, 0x1fffe000000636) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:394 +0x71 io.ReadAtLeast(0x3525540, 0xc420da97a0, 0xc4203e0f10, 0x5, 0x5, 0x5, 0xc4233bd9a0, 0xf4daa9, 0xc421dec900) /usr/local/go/src/io/io.go:309 +0x94 io.ReadFull(0x3525540, 0xc420da97a0, 0xc4203e0f10, 0x5, 0x5, 0x15, 0x20, 0x153446de41c1c962) /usr/local/go/src/io/io.go:327 +0x73 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*Stream).Read(0xc4218e4c00, 0xc4203e0f10, 0x5, 0x5, 0xc421cfeea8, 0xc421187338, 0xc420ffc0a0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:378 +0x15d github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*parser).recvMsg(0xc4203e0f00, 0x7fffffff, 0xb, 0x0, 0xc422322660, 0xc420255500, 0x0, 0xe) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/rpc_util.go:452 +0x99 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.recv(0xc4203e0f00, 0x2ac8eb7c2b10, 0x5068e68, 0xc4218e4c00, 0x0, 0x0, 0x2de9b60, 0xc422f60640, 0x7fffffff, 0xc422bcb800, ...) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/rpc_util.go:578 +0x61 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*csAttempt).recvMsg(0xc422408ea0, 0x2de9b60, 0xc422f60640, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/stream.go:539 +0x27c github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*clientStream).RecvMsg(0xc420fc2780, 0x2de9b60, 0xc422f60640, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/stream.go:405 +0x67 github.com/cockroachdb/cockroach/pkg/gossip.(*gossipGossipClient).Recv(0xc4212e0090, 0x3548840, 0xc423150150, 0xc420e05980) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/gossip.pb.go:191 +0x83 github.com/cockroachdb/cockroach/pkg/gossip.(*client).gossip.func2.1(0x3562720, 0xc4212e0090, 0xc422408dd0, 0x3548840, 0xc423150150, 0xc420e05980, 0xc420a70e00, 0x30) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/client.go:311 +0x43 github.com/cockroachdb/cockroach/pkg/gossip.(*client).gossip.func2(0x3548840, 0xc423150150) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/client.go:319 +0xe6 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc4212e0180, 0xc4207ea3f0, 0xc421028600) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:193 +0x14b created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:186 +0xbb leaktest.go:110: Leaked goroutine: goroutine 111565 [select]: github.com/cockroachdb/cockroach/pkg/gossip.(*server).Gossip(0xc4226a6780, 0x3562780, 0xc422e85ec0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/server.go:185 +0xa61 github.com/cockroachdb/cockroach/pkg/gossip._Gossip_Gossip_Handler(0x2ebd7c0, 0xc4226a6780, 0x3559420, 0xc4241a2a50, 0xc4200fc2c8, 0xc4239ab200) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/gossip.pb.go:208 +0xd3 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).processStreamingRPC(0xc4208fca80, 0x3565de0, 0xc420653800, 0xc4211a1d00, 0xc4211bad20, 0x4448820, 0x0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:1160 +0x132a github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).handleStream(0xc4208fca80, 0x3565de0, 0xc420653800, 0xc4211a1d00, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:1253 +0x12f9 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).serveStreams.func1.1(0xc421c10bd0, 0xc4208fca80, 0x3565de0, 0xc420653800, 0xc4211a1d00) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:680 +0xad created by github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).serveStreams.func1 /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:678 +0xb9 leaktest.go:110: Leaked goroutine: goroutine 111566 [select]: github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*recvBufferReader).read(0xc420871270, 0xc421a9f4b0, 0x5, 0x5, 0xc4219a0500, 0x843cb0, 0x2810ab0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:142 +0x19c github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*recvBufferReader).Read(0xc420871270, 0xc421a9f4b0, 0x5, 0x5, 0x842f6c, 0xc4219a0500, 0xc4209ef610) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:131 +0x82 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*transportReader).Read(0xc4239ab1a0, 0xc421a9f4b0, 0x5, 0x5, 0xf25293, 0xc4219a0500, 0x1fffe000001251) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:394 +0x71 io.ReadAtLeast(0x3525540, 0xc4239ab1a0, 0xc421a9f4b0, 0x5, 0x5, 0x5, 0xc4209ef6e8, 0xf4f539, 0xc420653800) /usr/local/go/src/io/io.go:309 +0x94 io.ReadFull(0x3525540, 0xc4239ab1a0, 0xc421a9f4b0, 0x5, 0x5, 0xc4229b6280, 0xc4209ef7d0, 0x110410f) /usr/local/go/src/io/io.go:327 +0x73 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*Stream).Read(0xc4211a1d00, 0xc421a9f4b0, 0x5, 0x5, 0xc42004e500, 0xc42004e500, 0xc4209ef840) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:378 +0x15d github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*parser).recvMsg(0xc421a9f4a0, 0x7fffffff, 0xb, 0x0, 0x2de9b60, 0xc420354320, 0x0, 0xe) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/rpc_util.go:452 +0x99 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.recv(0xc421a9f4a0, 0x2ac8eb7c2b10, 0x5068e68, 0xc4211a1d00, 0x0, 0x0, 0x2e0a100, 0xc422f600a0, 0x7fffffff, 0xc422bca3c0, ...) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/rpc_util.go:578 +0x61 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*serverStream).RecvMsg(0xc4241a2a50, 0x2e0a100, 0xc422f600a0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/stream.go:757 +0x1e9 github.com/cockroachdb/cockroach/pkg/gossip.(*gossipGossipServer).Recv(0xc422e85ec0, 0xc400000000, 0x1836efc, 0xc4209efb38) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/gossip.pb.go:227 +0x83 github.com/cockroachdb/cockroach/pkg/gossip.(Gossip_GossipServer).Recv-fm(0xc4226a67c8, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/server.go:148 +0x4b github.com/cockroachdb/cockroach/pkg/gossip.(*server).gossipReceiver(0xc4226a6780, 0x3548840, 0xc4231500f0, 0xc420b3aab8, 0xc420cff700, 0xc420cc5f50, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/server.go:326 +0x67c github.com/cockroachdb/cockroach/pkg/gossip.(*server).Gossip.func3.1(0x3548840, 0xc4231500f0) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/server.go:148 +0x12d github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc422e85ef0, 0xc4207ea3f0, 0xc420cff7c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:193 +0x14b created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:186 +0xbb ```
1.0
github.com/cockroachdb/cockroach/pkg/storage: TestStoreRangeSystemSplits failed under stress - SHA: https://github.com/cockroachdb/cockroach/commits/97c3622c2a6f4ec579ed93655b9d8a83732ac654 Parameters: ``` TAGS= GOFLAGS=-race ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=693452&tab=buildLog ``` I180602 07:21:25.666097 111393 gossip/gossip.go:333 [n1] NodeDescriptor set to node_id:1 address:<network_field:"" address_field:"" > attrs:<> locality:<> ServerVersion:<major_val:0 minor_val:0 patch:0 unstable:0 > I180602 07:21:25.755244 111393 storage/store.go:1454 [s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available I180602 07:21:25.822572 111432 storage/replica_command.go:863 [split,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2] W180602 07:21:25.865205 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer I180602 07:21:25.915992 111562 storage/replica_command.go:863 [split,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3] W180602 07:21:25.934571 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:25.938941 111461 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "unnamed" id=17deafae key=/Table/SystemConfigSpan/Start rw=true pri=0.03184783 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,53 orig=0.000000123,53 max=0.000000123,54 wto=false rop=false seq=1 I180602 07:21:26.066952 111572 storage/replica_command.go:863 [split,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4] W180602 07:21:26.094922 111464 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "unnamed" id=17deafae key=/Table/SystemConfigSpan/Start rw=true pri=0.03184783 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,53 orig=0.000000123,53 max=0.000000123,54 wto=false rop=false seq=1 W180602 07:21:26.143966 111457 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=46177098 key=/Local/Range/System/NodeLiveness/RangeDescriptor rw=true pri=0.06770799 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,182 orig=0.000000123,182 max=0.000000123,196 wto=false rop=false seq=1 W180602 07:21:26.171835 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer I180602 07:21:26.190567 111466 storage/replica_command.go:863 [split,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5] W180602 07:21:26.200246 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:26.223842 111587 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "unnamed" id=17deafae key=/Table/SystemConfigSpan/Start rw=true pri=0.03184783 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,53 orig=0.000000123,53 max=0.000000123,54 wto=false rop=false seq=1 W180602 07:21:26.259843 103909 gossip/gossip.go:1298 [n1] first range unavailable; trying remaining resolvers I180602 07:21:26.262515 111416 gossip/client.go:129 [n1] started gossip client to 127.0.0.1:44307 I180602 07:21:26.304909 111402 storage/replica_command.go:863 [split,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6] W180602 07:21:26.307892 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer W180602 07:21:26.340668 111552 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "unnamed" id=17deafae key=/Table/SystemConfigSpan/Start rw=true pri=0.03184783 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,53 orig=0.000000123,53 max=0.000000123,54 wto=false rop=false seq=1 I180602 07:21:26.413526 111603 storage/replica_command.go:863 [split,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7] W180602 07:21:26.434307 111573 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "unnamed" id=17deafae key=/Table/SystemConfigSpan/Start rw=true pri=0.03184783 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,53 orig=0.000000123,53 max=0.000000123,54 wto=false rop=false seq=1 I180602 07:21:26.489843 111577 storage/replica_command.go:863 [split,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8] I180602 07:21:26.601373 111652 storage/replica_command.go:863 [split,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9] I180602 07:21:26.700193 111406 storage/replica_command.go:863 [split,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10] W180602 07:21:26.748760 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer W180602 07:21:26.761617 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer E180602 07:21:26.803145 111637 storage/consistency_queue.go:125 [replica consistency checker,s1,r1/1:/{Min-System/}] computing own checksum: could not resolve node ID 1: dummy resolver E180602 07:21:26.803792 111637 storage/queue.go:778 [replica consistency checker,s1,r1/1:/{Min-System/}] computing own checksum: could not resolve node ID 1: dummy resolver W180602 07:21:26.812209 104013 gossip/gossip.go:1298 [n2] first range unavailable; trying remaining resolvers I180602 07:21:26.818283 111638 storage/replica_command.go:863 [split,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11] I180602 07:21:26.918273 111580 storage/replica_command.go:863 [split,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12] W180602 07:21:27.012870 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer I180602 07:21:27.016534 104022 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming I180602 07:21:27.038033 111616 storage/replica_command.go:863 [split,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13] I180602 07:21:27.156325 111649 storage/replica_command.go:863 [split,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14] W180602 07:21:27.223175 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer W180602 07:21:27.281224 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer I180602 07:21:27.300168 111690 storage/replica_command.go:863 [split,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15] W180602 07:21:27.311006 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer I180602 07:21:27.420307 111719 storage/replica_command.go:863 [split,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16] I180602 07:21:27.542361 111701 storage/replica_command.go:863 [split,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17] W180602 07:21:27.645111 111725 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=da9ad7de key=/Local/Range/Table/19/RangeDescriptor rw=true pri=0.05901391 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,1879 orig=0.000000123,1879 max=0.000000123,1893 wto=false rop=false seq=1 W180602 07:21:27.659094 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer I180602 07:21:27.684752 111680 storage/replica_command.go:863 [split,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18] W180602 07:21:27.763939 111780 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=125a67e3 key=/Local/Range/Table/20/RangeDescriptor rw=true pri=0.03635720 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2111 orig=0.000000123,2042 max=0.000000123,2084 wto=false rop=false seq=1 W180602 07:21:27.766883 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer E180602 07:21:27.806309 111766 storage/consistency_queue.go:125 [replica consistency checker,s1,r4/1:/System/{NodeLive…-tsd}] computing own checksum: could not resolve node ID 1: dummy resolver E180602 07:21:27.806975 111766 storage/queue.go:778 [replica consistency checker,s1,r4/1:/System/{NodeLive…-tsd}] computing own checksum: could not resolve node ID 1: dummy resolver I180602 07:21:27.818124 111796 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065 I180602 07:21:27.826497 111796 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming I180602 07:21:27.874237 111726 storage/replica_command.go:863 [split,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19] W180602 07:21:27.982386 111770 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=c476a094 key=/Local/Range/Table/21/RangeDescriptor rw=true pri=0.02374100 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2252 orig=0.000000123,2252 max=0.000000123,2275 wto=false rop=false seq=1 I180602 07:21:28.029652 111810 storage/replica_command.go:863 [split,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20] W180602 07:21:28.031230 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:28.118611 111774 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=fcaa46c5 key=/Local/Range/Table/22/RangeDescriptor rw=true pri=0.00146925 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2397 orig=0.000000123,2397 max=0.000000123,2414 wto=false rop=false seq=1 I180602 07:21:28.154689 111775 storage/replica_command.go:863 [split,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21] W180602 07:21:28.213520 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer I180602 07:21:28.274397 111813 storage/replica_command.go:863 [split,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22] W180602 07:21:28.338665 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer W180602 07:21:28.352719 111831 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=a8866910 key=/Local/Range/Table/50/RangeDescriptor rw=true pri=0.01775119 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2619 orig=0.000000123,2619 max=0.000000123,2648 wto=false rop=false seq=1 I180602 07:21:28.382776 111710 storage/replica_command.go:863 [split,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23] W180602 07:21:28.458973 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer W180602 07:21:28.477987 111834 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=3ad9f13b key=/Local/Range/Table/51/RangeDescriptor rw=true pri=0.02323743 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2754 orig=0.000000123,2754 max=0.000000123,2801 wto=false rop=false seq=1 I180602 07:21:28.508700 111631 storage/replica_command.go:863 [split,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r24] W180602 07:21:28.583750 111755 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=199fc42e key=/Local/Range/Table/52/RangeDescriptor rw=true pri=0.01267702 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2880 orig=0.000000123,2880 max=0.000000123,2925 wto=false rop=false seq=1 W180602 07:21:28.621310 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer I180602 07:21:28.623619 111837 storage/replica_command.go:863 [split,s1,r24/1:/{Table/53-Max}] initiating a split of this range at key /Table/54 [r25] W180602 07:21:28.689811 111758 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=4806ee51 key=/Local/Range/Table/53/RangeDescriptor rw=true pri=0.03011487 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,3011 orig=0.000000123,3011 max=0.000000123,3022 wto=false rop=false seq=1 E180602 07:21:28.791818 111633 storage/consistency_queue.go:125 [replica consistency checker,s1,r8/1:/Table/1{1-2}] computing own checksum: could not resolve node ID 1: dummy resolver E180602 07:21:28.792442 111633 storage/queue.go:778 [replica consistency checker,s1,r8/1:/Table/1{1-2}] computing own checksum: could not resolve node ID 1: dummy resolver I180602 07:21:28.806116 111842 storage/replica_command.go:863 [split,s1,r25/1:/{Table/54-Max}] initiating a split of this range at key /Table/55 [r26] I180602 07:21:28.844513 111819 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065 W180602 07:21:28.898378 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer I180602 07:21:28.909067 111819 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming I180602 07:21:28.950153 111879 storage/replica_command.go:863 [split,s1,r26/1:/{Table/55-Max}] initiating a split of this range at key /Table/56 [r27] I180602 07:21:29.067577 111852 storage/replica_command.go:863 [split,s1,r27/1:/{Table/56-Max}] initiating a split of this range at key /Table/57 [r28] W180602 07:21:29.149687 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:29.178637 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer I180602 07:21:29.223093 111393 util/stop/stopper.go:471 quiescing; tasks left: 1 storage.replicate: purgatory processing replica 1 storage.intentResolver: processing intents I180602 07:21:29.223977 111393 util/stop/stopper.go:471 quiescing; tasks left: 1 storage.intentResolver: processing intents W180602 07:21:29.406961 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer W180602 07:21:29.434693 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer W180602 07:21:29.729400 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer W180602 07:21:29.802959 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer I180602 07:21:29.845634 111896 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065 I180602 07:21:29.856527 111896 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming W180602 07:21:30.184916 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:30.217264 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:30.330989 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer W180602 07:21:30.492147 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer W180602 07:21:30.719824 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer W180602 07:21:30.843026 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer I180602 07:21:30.846414 111868 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065 I180602 07:21:30.853735 111868 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming W180602 07:21:31.088347 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:31.202071 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer W180602 07:21:31.306937 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:31.411590 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer W180602 07:21:31.763071 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer I180602 07:21:31.847405 111959 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065 I180602 07:21:31.855401 111959 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming W180602 07:21:31.957920 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer W180602 07:21:32.120898 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:32.275896 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer W180602 07:21:32.398697 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:32.482426 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer W180602 07:21:32.785552 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer W180602 07:21:32.843238 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer I180602 07:21:32.848515 111974 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065 I180602 07:21:32.872364 111974 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming W180602 07:21:33.014392 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:33.313420 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer W180602 07:21:33.441125 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer W180602 07:21:33.506666 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:33.761849 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer I180602 07:21:33.849460 111914 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065 W180602 07:21:33.851714 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer I180602 07:21:33.855960 111914 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming W180602 07:21:33.986173 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer W180602 07:21:34.319167 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer leaktest.go:110: Leaked goroutine: goroutine 111416 [select]: github.com/cockroachdb/cockroach/pkg/gossip.(*client).gossip(0xc422408dd0, 0x3548840, 0xc420da95f0, 0xc420e05980, 0x3562720, 0xc4212e0090, 0xc4207ea3f0, 0xc4232828f0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/client.go:323 +0x400 github.com/cockroachdb/cockroach/pkg/gossip.(*client).startLocked.func1(0x3548840, 0xc420da95f0) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/client.go:130 +0x5fd github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc4212e0030, 0xc4207ea3f0, 0xc421028280) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:193 +0x14b created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:186 +0xbb leaktest.go:110: Leaked goroutine: goroutine 111417 [select]: github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.newClientStream.func5(0xc4201ba300, 0xc420fc2780, 0x3548840, 0xc420da9680) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/stream.go:321 +0x120 created by github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.newClientStream /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/stream.go:320 +0x1200 leaktest.go:110: Leaked goroutine: goroutine 111425 [select]: github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*recvBufferReader).read(0xc4211e79a0, 0xc4203e0f10, 0x5, 0x5, 0xc421cea660, 0x843cb0, 0x2810ab0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:142 +0x19c github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*recvBufferReader).Read(0xc4211e79a0, 0xc4203e0f10, 0x5, 0x5, 0x842f6c, 0xc421cea660, 0xc4233bd8c8) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:131 +0x82 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*transportReader).Read(0xc420da97a0, 0xc4203e0f10, 0x5, 0x5, 0xf25293, 0xc421cea660, 0x1fffe000000636) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:394 +0x71 io.ReadAtLeast(0x3525540, 0xc420da97a0, 0xc4203e0f10, 0x5, 0x5, 0x5, 0xc4233bd9a0, 0xf4daa9, 0xc421dec900) /usr/local/go/src/io/io.go:309 +0x94 io.ReadFull(0x3525540, 0xc420da97a0, 0xc4203e0f10, 0x5, 0x5, 0x15, 0x20, 0x153446de41c1c962) /usr/local/go/src/io/io.go:327 +0x73 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*Stream).Read(0xc4218e4c00, 0xc4203e0f10, 0x5, 0x5, 0xc421cfeea8, 0xc421187338, 0xc420ffc0a0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:378 +0x15d github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*parser).recvMsg(0xc4203e0f00, 0x7fffffff, 0xb, 0x0, 0xc422322660, 0xc420255500, 0x0, 0xe) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/rpc_util.go:452 +0x99 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.recv(0xc4203e0f00, 0x2ac8eb7c2b10, 0x5068e68, 0xc4218e4c00, 0x0, 0x0, 0x2de9b60, 0xc422f60640, 0x7fffffff, 0xc422bcb800, ...) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/rpc_util.go:578 +0x61 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*csAttempt).recvMsg(0xc422408ea0, 0x2de9b60, 0xc422f60640, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/stream.go:539 +0x27c github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*clientStream).RecvMsg(0xc420fc2780, 0x2de9b60, 0xc422f60640, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/stream.go:405 +0x67 github.com/cockroachdb/cockroach/pkg/gossip.(*gossipGossipClient).Recv(0xc4212e0090, 0x3548840, 0xc423150150, 0xc420e05980) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/gossip.pb.go:191 +0x83 github.com/cockroachdb/cockroach/pkg/gossip.(*client).gossip.func2.1(0x3562720, 0xc4212e0090, 0xc422408dd0, 0x3548840, 0xc423150150, 0xc420e05980, 0xc420a70e00, 0x30) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/client.go:311 +0x43 github.com/cockroachdb/cockroach/pkg/gossip.(*client).gossip.func2(0x3548840, 0xc423150150) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/client.go:319 +0xe6 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc4212e0180, 0xc4207ea3f0, 0xc421028600) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:193 +0x14b created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:186 +0xbb leaktest.go:110: Leaked goroutine: goroutine 111565 [select]: github.com/cockroachdb/cockroach/pkg/gossip.(*server).Gossip(0xc4226a6780, 0x3562780, 0xc422e85ec0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/server.go:185 +0xa61 github.com/cockroachdb/cockroach/pkg/gossip._Gossip_Gossip_Handler(0x2ebd7c0, 0xc4226a6780, 0x3559420, 0xc4241a2a50, 0xc4200fc2c8, 0xc4239ab200) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/gossip.pb.go:208 +0xd3 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).processStreamingRPC(0xc4208fca80, 0x3565de0, 0xc420653800, 0xc4211a1d00, 0xc4211bad20, 0x4448820, 0x0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:1160 +0x132a github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).handleStream(0xc4208fca80, 0x3565de0, 0xc420653800, 0xc4211a1d00, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:1253 +0x12f9 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).serveStreams.func1.1(0xc421c10bd0, 0xc4208fca80, 0x3565de0, 0xc420653800, 0xc4211a1d00) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:680 +0xad created by github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).serveStreams.func1 /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:678 +0xb9 leaktest.go:110: Leaked goroutine: goroutine 111566 [select]: github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*recvBufferReader).read(0xc420871270, 0xc421a9f4b0, 0x5, 0x5, 0xc4219a0500, 0x843cb0, 0x2810ab0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:142 +0x19c github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*recvBufferReader).Read(0xc420871270, 0xc421a9f4b0, 0x5, 0x5, 0x842f6c, 0xc4219a0500, 0xc4209ef610) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:131 +0x82 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*transportReader).Read(0xc4239ab1a0, 0xc421a9f4b0, 0x5, 0x5, 0xf25293, 0xc4219a0500, 0x1fffe000001251) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:394 +0x71 io.ReadAtLeast(0x3525540, 0xc4239ab1a0, 0xc421a9f4b0, 0x5, 0x5, 0x5, 0xc4209ef6e8, 0xf4f539, 0xc420653800) /usr/local/go/src/io/io.go:309 +0x94 io.ReadFull(0x3525540, 0xc4239ab1a0, 0xc421a9f4b0, 0x5, 0x5, 0xc4229b6280, 0xc4209ef7d0, 0x110410f) /usr/local/go/src/io/io.go:327 +0x73 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*Stream).Read(0xc4211a1d00, 0xc421a9f4b0, 0x5, 0x5, 0xc42004e500, 0xc42004e500, 0xc4209ef840) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:378 +0x15d github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*parser).recvMsg(0xc421a9f4a0, 0x7fffffff, 0xb, 0x0, 0x2de9b60, 0xc420354320, 0x0, 0xe) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/rpc_util.go:452 +0x99 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.recv(0xc421a9f4a0, 0x2ac8eb7c2b10, 0x5068e68, 0xc4211a1d00, 0x0, 0x0, 0x2e0a100, 0xc422f600a0, 0x7fffffff, 0xc422bca3c0, ...) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/rpc_util.go:578 +0x61 github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*serverStream).RecvMsg(0xc4241a2a50, 0x2e0a100, 0xc422f600a0, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/stream.go:757 +0x1e9 github.com/cockroachdb/cockroach/pkg/gossip.(*gossipGossipServer).Recv(0xc422e85ec0, 0xc400000000, 0x1836efc, 0xc4209efb38) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/gossip.pb.go:227 +0x83 github.com/cockroachdb/cockroach/pkg/gossip.(Gossip_GossipServer).Recv-fm(0xc4226a67c8, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/server.go:148 +0x4b github.com/cockroachdb/cockroach/pkg/gossip.(*server).gossipReceiver(0xc4226a6780, 0x3548840, 0xc4231500f0, 0xc420b3aab8, 0xc420cff700, 0xc420cc5f50, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/server.go:326 +0x67c github.com/cockroachdb/cockroach/pkg/gossip.(*server).Gossip.func3.1(0x3548840, 0xc4231500f0) /go/src/github.com/cockroachdb/cockroach/pkg/gossip/server.go:148 +0x12d github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc422e85ef0, 0xc4207ea3f0, 0xc420cff7c0) /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:193 +0x14b created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:186 +0xbb ```
test
github com cockroachdb cockroach pkg storage teststorerangesystemsplits failed under stress sha parameters tags goflags race failed test gossip gossip go nodedescriptor set to node id address attrs locality serverversion storage store go failed initial metrics computation system config not yet available storage replica command go initiating a split of this range at key system storage store go could not gossip node liveness node unavailable try another peer storage replica command go initiating a split of this range at key system nodeliveness storage store go could not gossip first range descriptor node unavailable try another peer storage intent resolver go failed to push during intent resolution failed to push unnamed id key table systemconfigspan start rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key system nodelivenessmax storage intent resolver go failed to push during intent resolution failed to push unnamed id key table systemconfigspan start rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage intent resolver go failed to push during intent resolution failed to push split id key local range system nodeliveness rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage store go could not gossip node liveness node unavailable try another peer storage replica command go initiating a split of this range at key system tsd storage store go could not gossip first range descriptor node unavailable try another peer storage intent resolver go failed to push during intent resolution failed to push unnamed id key table systemconfigspan start rw true pri iso serializable stat pending epo ts orig max wto false rop false seq gossip gossip go first range unavailable trying remaining resolvers gossip client go started gossip client to storage replica command go initiating a split of this range at key system tse storage store go could not gossip system config node unavailable try another peer storage intent resolver go failed to push during intent resolution failed to push unnamed id key table systemconfigspan start rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key table systemconfigspan start storage intent resolver go failed to push during intent resolution failed to push unnamed id key table systemconfigspan start rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage store go could not gossip node liveness node unavailable try another peer storage store go could not gossip system config node unavailable try another peer storage consistency queue go computing own checksum could not resolve node id dummy resolver storage queue go computing own checksum could not resolve node id dummy resolver gossip gossip go first range unavailable trying remaining resolvers storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage store go could not gossip first range descriptor node unavailable try another peer gossip client go closing client to node stopping outgoing client to node already have incoming storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage store go could not gossip node liveness node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer storage replica command go initiating a split of this range at key table storage store go could not gossip system config node unavailable try another peer storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage intent resolver go failed to push during intent resolution failed to push split id key local range table rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage store go could not gossip node liveness node unavailable try another peer storage replica command go initiating a split of this range at key table storage intent resolver go failed to push during intent resolution failed to push split id key local range table rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage store go could not gossip system config node unavailable try another peer storage consistency queue go computing own checksum could not resolve node id dummy resolver storage queue go computing own checksum could not resolve node id dummy resolver gossip client go started gossip client to gossip client go closing client to node stopping outgoing client to node already have incoming storage replica command go initiating a split of this range at key table storage intent resolver go failed to push during intent resolution failed to push split id key local range table rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key table storage store go could not gossip first range descriptor node unavailable try another peer storage intent resolver go failed to push during intent resolution failed to push split id key local range table rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key table storage store go could not gossip first range descriptor node unavailable try another peer storage replica command go initiating a split of this range at key table storage store go could not gossip node liveness node unavailable try another peer storage intent resolver go failed to push during intent resolution failed to push split id key local range table rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key table storage store go could not gossip system config node unavailable try another peer storage intent resolver go failed to push during intent resolution failed to push split id key local range table rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key table storage intent resolver go failed to push during intent resolution failed to push split id key local range table rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage store go could not gossip node liveness node unavailable try another peer storage replica command go initiating a split of this range at key table storage intent resolver go failed to push during intent resolution failed to push split id key local range table rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage consistency queue go computing own checksum could not resolve node id dummy resolver storage queue go computing own checksum could not resolve node id dummy resolver storage replica command go initiating a split of this range at key table gossip client go started gossip client to storage store go could not gossip system config node unavailable try another peer gossip client go closing client to node stopping outgoing client to node already have incoming storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer util stop stopper go quiescing tasks left storage replicate purgatory processing replica storage intentresolver processing intents util stop stopper go quiescing tasks left storage intentresolver processing intents storage store go could not gossip system config node unavailable try another peer storage store go could not gossip node liveness node unavailable try another peer storage store go could not gossip node liveness node unavailable try another peer storage store go could not gossip system config node unavailable try another peer gossip client go started gossip client to gossip client go closing client to node stopping outgoing client to node already have incoming storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip system config node unavailable try another peer storage store go could not gossip node liveness node unavailable try another peer storage store go could not gossip system config node unavailable try another peer storage store go could not gossip node liveness node unavailable try another peer gossip client go started gossip client to gossip client go closing client to node stopping outgoing client to node already have incoming storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip system config node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip node liveness node unavailable try another peer storage store go could not gossip system config node unavailable try another peer gossip client go started gossip client to gossip client go closing client to node stopping outgoing client to node already have incoming storage store go could not gossip node liveness node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip system config node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip node liveness node unavailable try another peer storage store go could not gossip system config node unavailable try another peer storage store go could not gossip node liveness node unavailable try another peer gossip client go started gossip client to gossip client go closing client to node stopping outgoing client to node already have incoming storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip system config node unavailable try another peer storage store go could not gossip node liveness node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip node liveness node unavailable try another peer gossip client go started gossip client to storage store go could not gossip system config node unavailable try another peer gossip client go closing client to node stopping outgoing client to node already have incoming storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip system config node unavailable try another peer leaktest go leaked goroutine goroutine github com cockroachdb cockroach pkg gossip client gossip go src github com cockroachdb cockroach pkg gossip client go github com cockroachdb cockroach pkg gossip client startlocked go src github com cockroachdb cockroach pkg gossip client go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go leaktest go leaked goroutine goroutine github com cockroachdb cockroach vendor google golang org grpc newclientstream go src github com cockroachdb cockroach vendor google golang org grpc stream go created by github com cockroachdb cockroach vendor google golang org grpc newclientstream go src github com cockroachdb cockroach vendor google golang org grpc stream go leaktest go leaked goroutine goroutine github com cockroachdb cockroach vendor google golang org grpc transport recvbufferreader read go src github com cockroachdb cockroach vendor google golang org grpc transport transport go github com cockroachdb cockroach vendor google golang org grpc transport recvbufferreader read go src github com cockroachdb cockroach vendor google golang org grpc transport transport go github com cockroachdb cockroach vendor google golang org grpc transport transportreader read go src github com cockroachdb cockroach vendor google golang org grpc transport transport go io readatleast usr local go src io io go io readfull usr local go src io io go github com cockroachdb cockroach vendor google golang org grpc transport stream read go src github com cockroachdb cockroach vendor google golang org grpc transport transport go github com cockroachdb cockroach vendor google golang org grpc parser recvmsg go src github com cockroachdb cockroach vendor google golang org grpc rpc util go github com cockroachdb cockroach vendor google golang org grpc recv go src github com cockroachdb cockroach vendor google golang org grpc rpc util go github com cockroachdb cockroach vendor google golang org grpc csattempt recvmsg go src github com cockroachdb cockroach vendor google golang org grpc stream go github com cockroachdb cockroach vendor google golang org grpc clientstream recvmsg go src github com cockroachdb cockroach vendor google golang org grpc stream go github com cockroachdb cockroach pkg gossip gossipgossipclient recv go src github com cockroachdb cockroach pkg gossip gossip pb go github com cockroachdb cockroach pkg gossip client gossip go src github com cockroachdb cockroach pkg gossip client go github com cockroachdb cockroach pkg gossip client gossip go src github com cockroachdb cockroach pkg gossip client go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go leaktest go leaked goroutine goroutine github com cockroachdb cockroach pkg gossip server gossip go src github com cockroachdb cockroach pkg gossip server go github com cockroachdb cockroach pkg gossip gossip gossip handler go src github com cockroachdb cockroach pkg gossip gossip pb go github com cockroachdb cockroach vendor google golang org grpc server processstreamingrpc go src github com cockroachdb cockroach vendor google golang org grpc server go github com cockroachdb cockroach vendor google golang org grpc server handlestream go src github com cockroachdb cockroach vendor google golang org grpc server go github com cockroachdb cockroach vendor google golang org grpc server servestreams go src github com cockroachdb cockroach vendor google golang org grpc server go created by github com cockroachdb cockroach vendor google golang org grpc server servestreams go src github com cockroachdb cockroach vendor google golang org grpc server go leaktest go leaked goroutine goroutine github com cockroachdb cockroach vendor google golang org grpc transport recvbufferreader read go src github com cockroachdb cockroach vendor google golang org grpc transport transport go github com cockroachdb cockroach vendor google golang org grpc transport recvbufferreader read go src github com cockroachdb cockroach vendor google golang org grpc transport transport go github com cockroachdb cockroach vendor google golang org grpc transport transportreader read go src github com cockroachdb cockroach vendor google golang org grpc transport transport go io readatleast usr local go src io io go io readfull usr local go src io io go github com cockroachdb cockroach vendor google golang org grpc transport stream read go src github com cockroachdb cockroach vendor google golang org grpc transport transport go github com cockroachdb cockroach vendor google golang org grpc parser recvmsg go src github com cockroachdb cockroach vendor google golang org grpc rpc util go github com cockroachdb cockroach vendor google golang org grpc recv go src github com cockroachdb cockroach vendor google golang org grpc rpc util go github com cockroachdb cockroach vendor google golang org grpc serverstream recvmsg go src github com cockroachdb cockroach vendor google golang org grpc stream go github com cockroachdb cockroach pkg gossip gossipgossipserver recv go src github com cockroachdb cockroach pkg gossip gossip pb go github com cockroachdb cockroach pkg gossip gossip gossipserver recv fm go src github com cockroachdb cockroach pkg gossip server go github com cockroachdb cockroach pkg gossip server gossipreceiver go src github com cockroachdb cockroach pkg gossip server go github com cockroachdb cockroach pkg gossip server gossip go src github com cockroachdb cockroach pkg gossip server go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go
1
192,447
14,617,187,604
IssuesEvent
2020-12-22 14:25:09
github-vet/rangeloop-pointer-findings
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
closed
terraform-providers/terraform-provider-oci: oci/dataintegration_workspace_test.go; 16 LoC
fresh small test
Found a possible issue in [terraform-providers/terraform-provider-oci](https://www.github.com/terraform-providers/terraform-provider-oci) at [oci/dataintegration_workspace_test.go](https://github.com/terraform-providers/terraform-provider-oci/blob/507acd0ed6517dbca2fbcfb8100874929c8fd8e1/oci/dataintegration_workspace_test.go#L297-L312) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > reference to workspaceId is reassigned at line 301 [Click here to see the code in its original context.](https://github.com/terraform-providers/terraform-provider-oci/blob/507acd0ed6517dbca2fbcfb8100874929c8fd8e1/oci/dataintegration_workspace_test.go#L297-L312) <details> <summary>Click here to show the 16 line(s) of Go which triggered the analyzer.</summary> ```go for _, workspaceId := range workspaceIds { if ok := SweeperDefaultResourceId[workspaceId]; !ok { deleteWorkspaceRequest := oci_dataintegration.DeleteWorkspaceRequest{} deleteWorkspaceRequest.WorkspaceId = &workspaceId deleteWorkspaceRequest.RequestMetadata.RetryPolicy = getRetryPolicy(true, "dataintegration") _, error := dataIntegrationClient.DeleteWorkspace(context.Background(), deleteWorkspaceRequest) if error != nil { fmt.Printf("Error deleting Workspace %s %s, It is possible that the resource is already deleted. Please verify manually \n", workspaceId, error) continue } waitTillCondition(testAccProvider, &workspaceId, workspaceSweepWaitCondition, time.Duration(3*time.Minute), workspaceSweepResponseFetchOperation, "dataintegration", true) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 507acd0ed6517dbca2fbcfb8100874929c8fd8e1
1.0
terraform-providers/terraform-provider-oci: oci/dataintegration_workspace_test.go; 16 LoC - Found a possible issue in [terraform-providers/terraform-provider-oci](https://www.github.com/terraform-providers/terraform-provider-oci) at [oci/dataintegration_workspace_test.go](https://github.com/terraform-providers/terraform-provider-oci/blob/507acd0ed6517dbca2fbcfb8100874929c8fd8e1/oci/dataintegration_workspace_test.go#L297-L312) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > reference to workspaceId is reassigned at line 301 [Click here to see the code in its original context.](https://github.com/terraform-providers/terraform-provider-oci/blob/507acd0ed6517dbca2fbcfb8100874929c8fd8e1/oci/dataintegration_workspace_test.go#L297-L312) <details> <summary>Click here to show the 16 line(s) of Go which triggered the analyzer.</summary> ```go for _, workspaceId := range workspaceIds { if ok := SweeperDefaultResourceId[workspaceId]; !ok { deleteWorkspaceRequest := oci_dataintegration.DeleteWorkspaceRequest{} deleteWorkspaceRequest.WorkspaceId = &workspaceId deleteWorkspaceRequest.RequestMetadata.RetryPolicy = getRetryPolicy(true, "dataintegration") _, error := dataIntegrationClient.DeleteWorkspace(context.Background(), deleteWorkspaceRequest) if error != nil { fmt.Printf("Error deleting Workspace %s %s, It is possible that the resource is already deleted. Please verify manually \n", workspaceId, error) continue } waitTillCondition(testAccProvider, &workspaceId, workspaceSweepWaitCondition, time.Duration(3*time.Minute), workspaceSweepResponseFetchOperation, "dataintegration", true) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 507acd0ed6517dbca2fbcfb8100874929c8fd8e1
test
terraform providers terraform provider oci oci dataintegration workspace test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message reference to workspaceid is reassigned at line click here to show the line s of go which triggered the analyzer go for workspaceid range workspaceids if ok sweeperdefaultresourceid ok deleteworkspacerequest oci dataintegration deleteworkspacerequest deleteworkspacerequest workspaceid workspaceid deleteworkspacerequest requestmetadata retrypolicy getretrypolicy true dataintegration error dataintegrationclient deleteworkspace context background deleteworkspacerequest if error nil fmt printf error deleting workspace s s it is possible that the resource is already deleted please verify manually n workspaceid error continue waittillcondition testaccprovider workspaceid workspacesweepwaitcondition time duration time minute workspacesweepresponsefetchoperation dataintegration true leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
216,357
7,307,162,418
IssuesEvent
2018-02-28 01:18:45
Chingu-cohorts/voyage-wiki
https://api.github.com/repos/Chingu-cohorts/voyage-wiki
closed
bears-team-2 lost one of two PMs
priority:must have scope:story state:done type:team-repair
**_About You_** - Team Name: bears-team-2 - Slack Name: @heathercoraje - Timezone: Asia **_Issue Description & Expected Outcome:_** Heathercoraje has to leave due to time restrictions & local bootcamp. **_Symptoms:_** **_Steps to Recreate:_** **_Resolution:_** Will check in with bears-2 and see how they are doing. Will need to have the PM access switched.
1.0
bears-team-2 lost one of two PMs - **_About You_** - Team Name: bears-team-2 - Slack Name: @heathercoraje - Timezone: Asia **_Issue Description & Expected Outcome:_** Heathercoraje has to leave due to time restrictions & local bootcamp. **_Symptoms:_** **_Steps to Recreate:_** **_Resolution:_** Will check in with bears-2 and see how they are doing. Will need to have the PM access switched.
non_test
bears team lost one of two pms about you team name bears team slack name heathercoraje timezone asia issue description expected outcome heathercoraje has to leave due to time restrictions local bootcamp symptoms steps to recreate resolution will check in with bears and see how they are doing will need to have the pm access switched
0
110,987
9,486,809,422
IssuesEvent
2019-04-22 15:04:12
dojot/dojot
https://api.github.com/repos/dojot/dojot
closed
Device Management API - Devices
Status:ToTest Team:Backend Type:Feature
Devices (Endpoint, get device) Create new parameter to filter by type of attr, like Geo.
1.0
Device Management API - Devices - Devices (Endpoint, get device) Create new parameter to filter by type of attr, like Geo.
test
device management api devices devices endpoint get device create new parameter to filter by type of attr like geo
1
104,108
4,195,222,014
IssuesEvent
2016-06-25 15:56:29
Sistema-Integrado-Gestao-Academica/SiGA
https://api.github.com/repos/Sistema-Integrado-Gestao-Academica/SiGA
closed
Detalhamento de Custos - *Despesa*
[Medium Priority]
Como **secretário financeiro** desejo detalhar uma despesa numa rúbrica (natureza de despesa) para que eu tenha controle dos gastos na de um plano orçamento orçamentário. ------------ C.A.: * Atributos de um detalhamento de despesa: * Nota de Empenho (obg) * ex.: 2011NE005787 * Data de emissão - date * Nº Processo SEI * Valor
1.0
Detalhamento de Custos - *Despesa* - Como **secretário financeiro** desejo detalhar uma despesa numa rúbrica (natureza de despesa) para que eu tenha controle dos gastos na de um plano orçamento orçamentário. ------------ C.A.: * Atributos de um detalhamento de despesa: * Nota de Empenho (obg) * ex.: 2011NE005787 * Data de emissão - date * Nº Processo SEI * Valor
non_test
detalhamento de custos despesa como secretário financeiro desejo detalhar uma despesa numa rúbrica natureza de despesa para que eu tenha controle dos gastos na de um plano orçamento orçamentário c a atributos de um detalhamento de despesa nota de empenho obg ex data de emissão date nº processo sei valor
0
141,849
19,006,012,659
IssuesEvent
2021-11-23 00:02:50
svg-at-wss/WebGoat
https://api.github.com/repos/svg-at-wss/WebGoat
closed
CVE-2021-21341 (High) detected in xstream-1.4.5.jar - autoclosed
security vulnerability
## CVE-2021-21341 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.5.jar</b></p></summary> <p>XStream is a serialization library from Java objects to XML and back.</p> <p>Path to dependency file: WebGoat/webgoat-server/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar,m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar</p> <p> Dependency Hierarchy: - :x: **xstream-1.4.5.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/svg-at-wss/WebGoat/commit/04c4639ea5cac557f56e64148a0f6081e5be4ddb">04c4639ea5cac557f56e64148a0f6081e5be4ddb</a></p> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is vulnerability which may allow a remote attacker to allocate 100% CPU time on the target system depending on CPU type or parallel execution of such a payload resulting in a denial of service only by manipulating the processed input stream. No user is affected who followed the recommendation to setup XStream's security framework with a whitelist limited to the minimal required types. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.16. <p>Publish Date: 2021-03-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21341>CVE-2021-21341</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-2p3x-qw9c-25hh">https://github.com/x-stream/xstream/security/advisories/GHSA-2p3x-qw9c-25hh</a></p> <p>Release Date: 2021-03-23</p> <p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.16</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.thoughtworks.xstream","packageName":"xstream","packageVersion":"1.4.5","packageFilePaths":["/webgoat-server/pom.xml","/webgoat-lessons/vulnerable-components/pom.xml","/webgoat-integration-tests/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.thoughtworks.xstream:xstream:1.4.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.thoughtworks.xstream:xstream:1.4.16"}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2021-21341","vulnerabilityDetails":"XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is vulnerability which may allow a remote attacker to allocate 100% CPU time on the target system depending on CPU type or parallel execution of such a payload resulting in a denial of service only by manipulating the processed input stream. No user is affected who followed the recommendation to setup XStream\u0027s security framework with a whitelist limited to the minimal required types. If you rely on XStream\u0027s default blacklist of the Security Framework, you will have to use at least version 1.4.16.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21341","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-21341 (High) detected in xstream-1.4.5.jar - autoclosed - ## CVE-2021-21341 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.5.jar</b></p></summary> <p>XStream is a serialization library from Java objects to XML and back.</p> <p>Path to dependency file: WebGoat/webgoat-server/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar,m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar</p> <p> Dependency Hierarchy: - :x: **xstream-1.4.5.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/svg-at-wss/WebGoat/commit/04c4639ea5cac557f56e64148a0f6081e5be4ddb">04c4639ea5cac557f56e64148a0f6081e5be4ddb</a></p> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is vulnerability which may allow a remote attacker to allocate 100% CPU time on the target system depending on CPU type or parallel execution of such a payload resulting in a denial of service only by manipulating the processed input stream. No user is affected who followed the recommendation to setup XStream's security framework with a whitelist limited to the minimal required types. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.16. <p>Publish Date: 2021-03-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21341>CVE-2021-21341</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-2p3x-qw9c-25hh">https://github.com/x-stream/xstream/security/advisories/GHSA-2p3x-qw9c-25hh</a></p> <p>Release Date: 2021-03-23</p> <p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.16</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.thoughtworks.xstream","packageName":"xstream","packageVersion":"1.4.5","packageFilePaths":["/webgoat-server/pom.xml","/webgoat-lessons/vulnerable-components/pom.xml","/webgoat-integration-tests/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.thoughtworks.xstream:xstream:1.4.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.thoughtworks.xstream:xstream:1.4.16"}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2021-21341","vulnerabilityDetails":"XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is vulnerability which may allow a remote attacker to allocate 100% CPU time on the target system depending on CPU type or parallel execution of such a payload resulting in a denial of service only by manipulating the processed input stream. No user is affected who followed the recommendation to setup XStream\u0027s security framework with a whitelist limited to the minimal required types. If you rely on XStream\u0027s default blacklist of the Security Framework, you will have to use at least version 1.4.16.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21341","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_test
cve high detected in xstream jar autoclosed cve high severity vulnerability vulnerable library xstream jar xstream is a serialization library from java objects to xml and back path to dependency file webgoat webgoat server pom xml path to vulnerable library home wss scanner repository com thoughtworks xstream xstream xstream jar repository com thoughtworks xstream xstream xstream jar home wss scanner repository com thoughtworks xstream xstream xstream jar dependency hierarchy x xstream jar vulnerable library found in head commit a href found in base branch develop vulnerability details xstream is a java library to serialize objects to xml and back again in xstream before version there is vulnerability which may allow a remote attacker to allocate cpu time on the target system depending on cpu type or parallel execution of such a payload resulting in a denial of service only by manipulating the processed input stream no user is affected who followed the recommendation to setup xstream s security framework with a whitelist limited to the minimal required types if you rely on xstream s default blacklist of the security framework you will have to use at least version publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com thoughtworks xstream xstream rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com thoughtworks xstream xstream isminimumfixversionavailable true minimumfixversion com thoughtworks xstream xstream basebranches vulnerabilityidentifier cve vulnerabilitydetails xstream is a java library to serialize objects to xml and back again in xstream before version there is vulnerability which may allow a remote attacker to allocate cpu time on the target system depending on cpu type or parallel execution of such a payload resulting in a denial of service only by manipulating the processed input stream no user is affected who followed the recommendation to setup xstream security framework with a whitelist limited to the minimal required types if you rely on xstream default blacklist of the security framework you will have to use at least version vulnerabilityurl
0
190,116
14,533,451,479
IssuesEvent
2020-12-15 00:36:22
uoForms/App-CANBeWell
https://api.github.com/repos/uoForms/App-CANBeWell
closed
Cursor pointer in the Topics list in the topics tab
App CanBeWell 1.0 CanBeWell 2.0 To be tested bug
Cursor pointer in the Topics list in the topics tab
1.0
Cursor pointer in the Topics list in the topics tab - Cursor pointer in the Topics list in the topics tab
test
cursor pointer in the topics list in the topics tab cursor pointer in the topics list in the topics tab
1
49,852
6,042,919,987
IssuesEvent
2017-06-11 17:13:44
tuna/issues
https://api.github.com/repos/tuna/issues
closed
添加clojure语言库的repo
Accepted MirrorRequest Test Required
<!-- --> #### 项目名称与简介(Project Intro.) clojure语言库的repo #### 上游地址与镜像方法(How to Mirror) https://clojars.org/ https://github.com/clojars/clojars-web/wiki/Data #### 其他信息(Other) - 镜像大小(Mirror Size): 未知
1.0
添加clojure语言库的repo - <!-- --> #### 项目名称与简介(Project Intro.) clojure语言库的repo #### 上游地址与镜像方法(How to Mirror) https://clojars.org/ https://github.com/clojars/clojars-web/wiki/Data #### 其他信息(Other) - 镜像大小(Mirror Size): 未知
test
添加clojure语言库的repo 项目名称与简介 project intro clojure语言库的repo 上游地址与镜像方法 how to mirror 其他信息 other 镜像大小 mirror size 未知
1
11,602
3,211,393,940
IssuesEvent
2015-10-06 10:28:16
ceylon/ceylon-ide-eclipse
https://api.github.com/repos/ceylon/ceylon-ide-eclipse
closed
Running ceylon.net tests
bug launch/run/debug test plugin WAITING_FOR_ANSWER
If I try to run the ceylon.net tests (i.e. the module `test.ceylon.net` as a Ceylon Test) in the IDE I get the following error: ceylon test: Error while loading the test.ceylon.net/1.1.1 module: Declaration 'java.net.URI' could not be found in module 'test.ceylon.net' or its imported modules but was found in the non-imported module 'java.base' com.redhat.ceylon.model.loader.ModelResolutionException: Error while loading the test.ceylon.net/1.1.1 module: Declaration 'java.net.URI' could not be found in module 'test.ceylon.net' or its imported modules but was found in the non-imported module 'java.base' at com.redhat.ceylon.compiler.java.runtime.model.RuntimeModelLoader$1.reportError(RuntimeModelLoader.java:209) at com.redhat.ceylon.model.typechecker.model.UnknownType.reportErrors(UnknownType.java:94) at com.redhat.ceylon.compiler.java.runtime.metamodel.Metamodel.getMetamodel(Metamodel.java:504) at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeValue.<init>(FreeValue.java:43) at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeReference.<init>(FreeReference.java:17) at com.redhat.ceylon.compiler.java.runtime.metamodel.Metamodel.getOrCreateMetamodel(Metamodel.java:344) at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeClassOrInterface.init(FreeClassOrInterface.java:92) at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeClass.init(FreeClass.java:51) at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeClassOrInterface.checkInit(FreeClassOrInterface.java:144) at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeClassOrInterface.filteredMembers(FreeClassOrInterface.java:212) at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeClassOrInterface.annotatedMemberDeclarations(FreeClassOrInterface.java:186) at ceylon.test.core.findCandidatesInClass_.findCandidatesInClass(DefaultTestRunner.ceylon:206) at ceylon.test.core.findCandidatesInPackage_.findCandidatesInPackage(DefaultTestRunner.ceylon:200) at ceylon.test.core.findCandidatesInModule_.findCandidatesInModule(DefaultTestRunner.ceylon:191) at ceylon.test.core.findCandidatesInModuleLiteral_.findCandidatesInModuleLiteral(DefaultTestRunner.ceylon:259) at ceylon.test.core.findCandidatesInTypeLiteral_.findCandidatesInTypeLiteral(DefaultTestRunner.ceylon:233) at ceylon.test.core.findCandidates_.findCandidates(DefaultTestRunner.ceylon:183) at ceylon.test.core.createExecutors_.createExecutors(DefaultTestRunner.ceylon:85) at ceylon.test.core.DefaultTestRunner.initExecutors$priv$(DefaultTestRunner.ceylon:43) at ceylon.test.core.DefaultTestRunner.<init>(DefaultTestRunner.ceylon:46) at ceylon.test.createTestRunner_.createTestRunner(TestRunner.ceylon:49) at ceylon.test.createTestRunner_.createTestRunner(TestRunner.ceylon:48) at com.redhat.ceylon.testjvm.Runner.run(tool.ceylon:85) at com.redhat.ceylon.testjvm.run_.run(tool.ceylon:36) at com.redhat.ceylon.testjvm.run_.main(tool.ceylon) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at ceylon.modules.api.runtime.SecurityActions.invokeRunInternal(SecurityActions.java:57) at ceylon.modules.api.runtime.SecurityActions.invokeRun(SecurityActions.java:48) at ceylon.modules.api.runtime.AbstractRuntime.invokeRun(AbstractRuntime.java:110) at ceylon.modules.api.runtime.AbstractRuntime.execute(AbstractRuntime.java:176) at ceylon.modules.api.runtime.AbstractRuntime.execute(AbstractRuntime.java:160) at ceylon.modules.Main.execute(Main.java:69) at ceylon.modules.Main.main(Main.java:42) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.jboss.modules.Module.run(Module.java:312) at org.jboss.modules.Main.main(Main.java:460) at ceylon.modules.bootstrap.CeylonRunTool.run(CeylonRunTool.java:246) at ceylon.modules.bootstrap.CeylonTestTool.run(CeylonTestTool.java:173) at com.redhat.ceylon.common.tools.CeylonTool.run(CeylonTool.java:390) at com.redhat.ceylon.common.tools.CeylonTool.execute(CeylonTool.java:327) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at com.redhat.ceylon.launcher.Launcher.runInJava7Checked(Launcher.java:114) at com.redhat.ceylon.launcher.Launcher.run(Launcher.java:41) at com.redhat.ceylon.launcher.Launcher.run(Launcher.java:34) at com.redhat.ceylon.launcher.Launcher.main(Launcher.java:27) The thing is that the `test.ceylon.net` module.ceylon does `import java.base "7";` The tests run fine in the CLI.
1.0
Running ceylon.net tests - If I try to run the ceylon.net tests (i.e. the module `test.ceylon.net` as a Ceylon Test) in the IDE I get the following error: ceylon test: Error while loading the test.ceylon.net/1.1.1 module: Declaration 'java.net.URI' could not be found in module 'test.ceylon.net' or its imported modules but was found in the non-imported module 'java.base' com.redhat.ceylon.model.loader.ModelResolutionException: Error while loading the test.ceylon.net/1.1.1 module: Declaration 'java.net.URI' could not be found in module 'test.ceylon.net' or its imported modules but was found in the non-imported module 'java.base' at com.redhat.ceylon.compiler.java.runtime.model.RuntimeModelLoader$1.reportError(RuntimeModelLoader.java:209) at com.redhat.ceylon.model.typechecker.model.UnknownType.reportErrors(UnknownType.java:94) at com.redhat.ceylon.compiler.java.runtime.metamodel.Metamodel.getMetamodel(Metamodel.java:504) at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeValue.<init>(FreeValue.java:43) at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeReference.<init>(FreeReference.java:17) at com.redhat.ceylon.compiler.java.runtime.metamodel.Metamodel.getOrCreateMetamodel(Metamodel.java:344) at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeClassOrInterface.init(FreeClassOrInterface.java:92) at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeClass.init(FreeClass.java:51) at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeClassOrInterface.checkInit(FreeClassOrInterface.java:144) at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeClassOrInterface.filteredMembers(FreeClassOrInterface.java:212) at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeClassOrInterface.annotatedMemberDeclarations(FreeClassOrInterface.java:186) at ceylon.test.core.findCandidatesInClass_.findCandidatesInClass(DefaultTestRunner.ceylon:206) at ceylon.test.core.findCandidatesInPackage_.findCandidatesInPackage(DefaultTestRunner.ceylon:200) at ceylon.test.core.findCandidatesInModule_.findCandidatesInModule(DefaultTestRunner.ceylon:191) at ceylon.test.core.findCandidatesInModuleLiteral_.findCandidatesInModuleLiteral(DefaultTestRunner.ceylon:259) at ceylon.test.core.findCandidatesInTypeLiteral_.findCandidatesInTypeLiteral(DefaultTestRunner.ceylon:233) at ceylon.test.core.findCandidates_.findCandidates(DefaultTestRunner.ceylon:183) at ceylon.test.core.createExecutors_.createExecutors(DefaultTestRunner.ceylon:85) at ceylon.test.core.DefaultTestRunner.initExecutors$priv$(DefaultTestRunner.ceylon:43) at ceylon.test.core.DefaultTestRunner.<init>(DefaultTestRunner.ceylon:46) at ceylon.test.createTestRunner_.createTestRunner(TestRunner.ceylon:49) at ceylon.test.createTestRunner_.createTestRunner(TestRunner.ceylon:48) at com.redhat.ceylon.testjvm.Runner.run(tool.ceylon:85) at com.redhat.ceylon.testjvm.run_.run(tool.ceylon:36) at com.redhat.ceylon.testjvm.run_.main(tool.ceylon) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at ceylon.modules.api.runtime.SecurityActions.invokeRunInternal(SecurityActions.java:57) at ceylon.modules.api.runtime.SecurityActions.invokeRun(SecurityActions.java:48) at ceylon.modules.api.runtime.AbstractRuntime.invokeRun(AbstractRuntime.java:110) at ceylon.modules.api.runtime.AbstractRuntime.execute(AbstractRuntime.java:176) at ceylon.modules.api.runtime.AbstractRuntime.execute(AbstractRuntime.java:160) at ceylon.modules.Main.execute(Main.java:69) at ceylon.modules.Main.main(Main.java:42) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.jboss.modules.Module.run(Module.java:312) at org.jboss.modules.Main.main(Main.java:460) at ceylon.modules.bootstrap.CeylonRunTool.run(CeylonRunTool.java:246) at ceylon.modules.bootstrap.CeylonTestTool.run(CeylonTestTool.java:173) at com.redhat.ceylon.common.tools.CeylonTool.run(CeylonTool.java:390) at com.redhat.ceylon.common.tools.CeylonTool.execute(CeylonTool.java:327) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at com.redhat.ceylon.launcher.Launcher.runInJava7Checked(Launcher.java:114) at com.redhat.ceylon.launcher.Launcher.run(Launcher.java:41) at com.redhat.ceylon.launcher.Launcher.run(Launcher.java:34) at com.redhat.ceylon.launcher.Launcher.main(Launcher.java:27) The thing is that the `test.ceylon.net` module.ceylon does `import java.base "7";` The tests run fine in the CLI.
test
running ceylon net tests if i try to run the ceylon net tests i e the module test ceylon net as a ceylon test in the ide i get the following error ceylon test error while loading the test ceylon net module declaration java net uri could not be found in module test ceylon net or its imported modules but was found in the non imported module java base com redhat ceylon model loader modelresolutionexception error while loading the test ceylon net module declaration java net uri could not be found in module test ceylon net or its imported modules but was found in the non imported module java base at com redhat ceylon compiler java runtime model runtimemodelloader reporterror runtimemodelloader java at com redhat ceylon model typechecker model unknowntype reporterrors unknowntype java at com redhat ceylon compiler java runtime metamodel metamodel getmetamodel metamodel java at com redhat ceylon compiler java runtime metamodel freevalue freevalue java at com redhat ceylon compiler java runtime metamodel freereference freereference java at com redhat ceylon compiler java runtime metamodel metamodel getorcreatemetamodel metamodel java at com redhat ceylon compiler java runtime metamodel freeclassorinterface init freeclassorinterface java at com redhat ceylon compiler java runtime metamodel freeclass init freeclass java at com redhat ceylon compiler java runtime metamodel freeclassorinterface checkinit freeclassorinterface java at com redhat ceylon compiler java runtime metamodel freeclassorinterface filteredmembers freeclassorinterface java at com redhat ceylon compiler java runtime metamodel freeclassorinterface annotatedmemberdeclarations freeclassorinterface java at ceylon test core findcandidatesinclass findcandidatesinclass defaulttestrunner ceylon at ceylon test core findcandidatesinpackage findcandidatesinpackage defaulttestrunner ceylon at ceylon test core findcandidatesinmodule findcandidatesinmodule defaulttestrunner ceylon at ceylon test core findcandidatesinmoduleliteral findcandidatesinmoduleliteral defaulttestrunner ceylon at ceylon test core findcandidatesintypeliteral findcandidatesintypeliteral defaulttestrunner ceylon at ceylon test core findcandidates findcandidates defaulttestrunner ceylon at ceylon test core createexecutors createexecutors defaulttestrunner ceylon at ceylon test core defaulttestrunner initexecutors priv defaulttestrunner ceylon at ceylon test core defaulttestrunner defaulttestrunner ceylon at ceylon test createtestrunner createtestrunner testrunner ceylon at ceylon test createtestrunner createtestrunner testrunner ceylon at com redhat ceylon testjvm runner run tool ceylon at com redhat ceylon testjvm run run tool ceylon at com redhat ceylon testjvm run main tool ceylon at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at ceylon modules api runtime securityactions invokeruninternal securityactions java at ceylon modules api runtime securityactions invokerun securityactions java at ceylon modules api runtime abstractruntime invokerun abstractruntime java at ceylon modules api runtime abstractruntime execute abstractruntime java at ceylon modules api runtime abstractruntime execute abstractruntime java at ceylon modules main execute main java at ceylon modules main main main java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org jboss modules module run module java at org jboss modules main main main java at ceylon modules bootstrap ceylonruntool run ceylonruntool java at ceylon modules bootstrap ceylontesttool run ceylontesttool java at com redhat ceylon common tools ceylontool run ceylontool java at com redhat ceylon common tools ceylontool execute ceylontool java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com redhat ceylon launcher launcher launcher java at com redhat ceylon launcher launcher run launcher java at com redhat ceylon launcher launcher run launcher java at com redhat ceylon launcher launcher main launcher java the thing is that the test ceylon net module ceylon does import java base the tests run fine in the cli
1
119,137
10,026,416,381
IssuesEvent
2019-07-17 06:34:22
Sinapse-Energia/AP-Sinapse
https://api.github.com/repos/Sinapse-Energia/AP-Sinapse
closed
[AP][ALL] Integrate 6LP features into the main branch
Device: All Priority: high Size: 13 Status: done - to be tested Type: integration
Now, we have two parallel branches and we miss several important features in each one. It is necessary to integrate the 6LP RF as a communication southbound in the main branch in order to have all the developed features for CMC, EN and AP accessible
1.0
[AP][ALL] Integrate 6LP features into the main branch - Now, we have two parallel branches and we miss several important features in each one. It is necessary to integrate the 6LP RF as a communication southbound in the main branch in order to have all the developed features for CMC, EN and AP accessible
test
integrate features into the main branch now we have two parallel branches and we miss several important features in each one it is necessary to integrate the rf as a communication southbound in the main branch in order to have all the developed features for cmc en and ap accessible
1
229,327
25,319,035,562
IssuesEvent
2022-11-18 01:06:19
DavidSpek/kubeflow
https://api.github.com/repos/DavidSpek/kubeflow
opened
CVE-2022-45198 (High) detected in Pillow-8.3.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl
security vulnerability
## CVE-2022-45198 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-8.3.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl</b></p></summary> <p>Python Imaging Library (Fork)</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/8e/7a/b047f6f80fdb02c0cca1d3761d71e9800bcf6d4874b71c9e6548ec59e156/Pillow-8.3.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/8e/7a/b047f6f80fdb02c0cca1d3761d71e9800bcf6d4874b71c9e6548ec59e156/Pillow-8.3.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl</a></p> <p>Path to dependency file: /components/crud-web-apps/common/backend</p> <p>Path to vulnerable library: /components/crud-web-apps/common/backend</p> <p> Dependency Hierarchy: - :x: **Pillow-8.3.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Pillow before 9.2.0 performs Improper Handling of Highly Compressed GIF Data (Data Amplification). <p>Publish Date: 2022-11-14 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-45198>CVE-2022-45198</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-11-14</p> <p>Fix Resolution: Pillow - 9.2.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-45198 (High) detected in Pillow-8.3.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl - ## CVE-2022-45198 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-8.3.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl</b></p></summary> <p>Python Imaging Library (Fork)</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/8e/7a/b047f6f80fdb02c0cca1d3761d71e9800bcf6d4874b71c9e6548ec59e156/Pillow-8.3.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/8e/7a/b047f6f80fdb02c0cca1d3761d71e9800bcf6d4874b71c9e6548ec59e156/Pillow-8.3.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl</a></p> <p>Path to dependency file: /components/crud-web-apps/common/backend</p> <p>Path to vulnerable library: /components/crud-web-apps/common/backend</p> <p> Dependency Hierarchy: - :x: **Pillow-8.3.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Pillow before 9.2.0 performs Improper Handling of Highly Compressed GIF Data (Data Amplification). <p>Publish Date: 2022-11-14 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-45198>CVE-2022-45198</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-11-14</p> <p>Fix Resolution: Pillow - 9.2.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in pillow manylinux whl cve high severity vulnerability vulnerable library pillow manylinux whl python imaging library fork library home page a href path to dependency file components crud web apps common backend path to vulnerable library components crud web apps common backend dependency hierarchy x pillow manylinux whl vulnerable library found in base branch master vulnerability details pillow before performs improper handling of highly compressed gif data data amplification publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution pillow step up your open source security game with mend
0
31,661
13,610,571,679
IssuesEvent
2020-09-23 07:35:06
microsoft/botframework-solutions
https://api.github.com/repos/microsoft/botframework-solutions
closed
[MS Teams] Unable to upload attachments to bot
Bot Services Needs Triage Type: Bug customer-replied-to customer-reported
Hi, We have 1v1 bot available to our users through MS Teams. Since roughly this week, it is no longer possible to upload attachments to the bot - it fails with error (see below). The uploading itself is done simply by opening up chat window with bot, clicking on paperclip icon, clicking on _Upload from my computer_ and finally choosing the file to upload: ![att01](https://user-images.githubusercontent.com/64434438/93337320-4862bf00-f829-11ea-9fbc-2043a1d96ab4.png) The progress bar then appears: ![att03](https://user-images.githubusercontent.com/64434438/93337465-75af6d00-f829-11ea-913e-77375f2612a1.png) After uploading completes, the following error popup is displayed and attachment is not sent over: ![att02](https://user-images.githubusercontent.com/64434438/93337211-1d786b00-f829-11ea-9a2c-3786461cc77c.png) Additional notes: * Choosing _Onedrive_ option does not show the error popup, but no file is sent - nothing happens. * The file type and size does not matter. But all the tests were done with png and txt file with no more than few kilobytes. * The error happens in both the desktop and web version of Teams client. * Same error happens for all of our Teams enabled bots, including one that was set up just last week. * The error does not happen if inline image is pasted and sent over. * The Teams tenant is based in Europe. * This should be irrelevant, but the bot itself is powered by .NET SDK of MS Bot Framework v4 and hosted in Azure with MS Teams enabled as one of its channels. For the reference, here is Teams package manifest: ``` { "$schema": "https://developer.microsoft.com/en-us/json-schemas/teams/v1.7/MicrosoftTeams.schema.json", "manifestVersion": "1.7", "version": "0.0.1", "id": "<removed>", "packageName": "<removed>", "developer": { "name": "<removed>", "websiteUrl": "<removed>", "privacyUrl": "<removed>", "termsOfUseUrl": "<removed>" }, "icons": { "color": "color.png", "outline": "outline.png" }, "name": { "short": "<removed>", "full": "<removed>" }, "description": { "short": "<removed>", "full": "<removed>" }, "accentColor": "#761092", "bots": [ { "botId": "<removed>", "scopes": [ "personal" ], "supportsFiles": true, "isNotificationOnly": false } ], "permissions": [ "identity", "messageTeamMembers" ], "validDomains": [] } ``` Will appreciate any ideas, thanks.
1.0
[MS Teams] Unable to upload attachments to bot - Hi, We have 1v1 bot available to our users through MS Teams. Since roughly this week, it is no longer possible to upload attachments to the bot - it fails with error (see below). The uploading itself is done simply by opening up chat window with bot, clicking on paperclip icon, clicking on _Upload from my computer_ and finally choosing the file to upload: ![att01](https://user-images.githubusercontent.com/64434438/93337320-4862bf00-f829-11ea-9fbc-2043a1d96ab4.png) The progress bar then appears: ![att03](https://user-images.githubusercontent.com/64434438/93337465-75af6d00-f829-11ea-913e-77375f2612a1.png) After uploading completes, the following error popup is displayed and attachment is not sent over: ![att02](https://user-images.githubusercontent.com/64434438/93337211-1d786b00-f829-11ea-9a2c-3786461cc77c.png) Additional notes: * Choosing _Onedrive_ option does not show the error popup, but no file is sent - nothing happens. * The file type and size does not matter. But all the tests were done with png and txt file with no more than few kilobytes. * The error happens in both the desktop and web version of Teams client. * Same error happens for all of our Teams enabled bots, including one that was set up just last week. * The error does not happen if inline image is pasted and sent over. * The Teams tenant is based in Europe. * This should be irrelevant, but the bot itself is powered by .NET SDK of MS Bot Framework v4 and hosted in Azure with MS Teams enabled as one of its channels. For the reference, here is Teams package manifest: ``` { "$schema": "https://developer.microsoft.com/en-us/json-schemas/teams/v1.7/MicrosoftTeams.schema.json", "manifestVersion": "1.7", "version": "0.0.1", "id": "<removed>", "packageName": "<removed>", "developer": { "name": "<removed>", "websiteUrl": "<removed>", "privacyUrl": "<removed>", "termsOfUseUrl": "<removed>" }, "icons": { "color": "color.png", "outline": "outline.png" }, "name": { "short": "<removed>", "full": "<removed>" }, "description": { "short": "<removed>", "full": "<removed>" }, "accentColor": "#761092", "bots": [ { "botId": "<removed>", "scopes": [ "personal" ], "supportsFiles": true, "isNotificationOnly": false } ], "permissions": [ "identity", "messageTeamMembers" ], "validDomains": [] } ``` Will appreciate any ideas, thanks.
non_test
unable to upload attachments to bot hi we have bot available to our users through ms teams since roughly this week it is no longer possible to upload attachments to the bot it fails with error see below the uploading itself is done simply by opening up chat window with bot clicking on paperclip icon clicking on upload from my computer and finally choosing the file to upload the progress bar then appears after uploading completes the following error popup is displayed and attachment is not sent over additional notes choosing onedrive option does not show the error popup but no file is sent nothing happens the file type and size does not matter but all the tests were done with png and txt file with no more than few kilobytes the error happens in both the desktop and web version of teams client same error happens for all of our teams enabled bots including one that was set up just last week the error does not happen if inline image is pasted and sent over the teams tenant is based in europe this should be irrelevant but the bot itself is powered by net sdk of ms bot framework and hosted in azure with ms teams enabled as one of its channels for the reference here is teams package manifest schema manifestversion version id packagename developer name websiteurl privacyurl termsofuseurl icons color color png outline outline png name short full description short full accentcolor bots botid scopes personal supportsfiles true isnotificationonly false permissions identity messageteammembers validdomains will appreciate any ideas thanks
0
96,495
10,934,215,437
IssuesEvent
2019-11-24 09:50:31
hapijs/joi
https://api.github.com/repos/hapijs/joi
closed
Promise use in `.validateAsync`
documentation
<!-- ⚠️ ⚠️ ⚠️ ⚠️ ⚠️ ⚠️ You must complete this entire issue template to receive support. You MUST NOT remove, change, or replace the template with your own format. A missing or incomplete report will cause your issue to be closed without comment. Please respect the time and experience that went into this template. It is here for a reason. Thank you! ⚠️ ⚠️ ⚠️ ⚠️ ⚠️ ⚠️ --> #### Context * *node version*: 8.11.1 * *module version*: 16.1.7 #### What are you trying to achieve or the steps to reproduce ? <!-- Before opening a documentation issue, please consider opening a Pull Request instead for trivial changes such as typos, spelling, incorrect links, anchors, or other corrections that are easier to just fix than report using this template. Please do not spend valuable time proposing extensive changes to the documentation before first asking about it. We value your time and do not want to waste it. Just open an issue first using this template and ask if your proposed changes would be helpful. Make sure to wrap all code examples in backticks so that they display correctly. Before submitting an issue, make sure to click on the Preview tab above to verify everything is formatted correctly. --> I'm trying to use `.validateAsync()` and consume the results as a promise. I'm having trouble finding the documentation about when the promise resolves or rejects and what arguments will be supplied to `.then` and `.catch`. I assumed both `then` and `catch` would receive an object with `value`, `error`, and `warning` keys similar to the synchronous version. Testing and logging has shown this to not be the case. 1. What are the arguments to `.then`? Is it just the updated value or `undefined` if no transformations were applied? 2. What are the arguments to `.catch`? Is it just the error? 3. If `.catch` is called, are joi transformations applied to the original value?
1.0
Promise use in `.validateAsync` - <!-- ⚠️ ⚠️ ⚠️ ⚠️ ⚠️ ⚠️ You must complete this entire issue template to receive support. You MUST NOT remove, change, or replace the template with your own format. A missing or incomplete report will cause your issue to be closed without comment. Please respect the time and experience that went into this template. It is here for a reason. Thank you! ⚠️ ⚠️ ⚠️ ⚠️ ⚠️ ⚠️ --> #### Context * *node version*: 8.11.1 * *module version*: 16.1.7 #### What are you trying to achieve or the steps to reproduce ? <!-- Before opening a documentation issue, please consider opening a Pull Request instead for trivial changes such as typos, spelling, incorrect links, anchors, or other corrections that are easier to just fix than report using this template. Please do not spend valuable time proposing extensive changes to the documentation before first asking about it. We value your time and do not want to waste it. Just open an issue first using this template and ask if your proposed changes would be helpful. Make sure to wrap all code examples in backticks so that they display correctly. Before submitting an issue, make sure to click on the Preview tab above to verify everything is formatted correctly. --> I'm trying to use `.validateAsync()` and consume the results as a promise. I'm having trouble finding the documentation about when the promise resolves or rejects and what arguments will be supplied to `.then` and `.catch`. I assumed both `then` and `catch` would receive an object with `value`, `error`, and `warning` keys similar to the synchronous version. Testing and logging has shown this to not be the case. 1. What are the arguments to `.then`? Is it just the updated value or `undefined` if no transformations were applied? 2. What are the arguments to `.catch`? Is it just the error? 3. If `.catch` is called, are joi transformations applied to the original value?
non_test
promise use in validateasync ⚠️ ⚠️ ⚠️ ⚠️ ⚠️ ⚠️ you must complete this entire issue template to receive support you must not remove change or replace the template with your own format a missing or incomplete report will cause your issue to be closed without comment please respect the time and experience that went into this template it is here for a reason thank you ⚠️ ⚠️ ⚠️ ⚠️ ⚠️ ⚠️ context node version module version what are you trying to achieve or the steps to reproduce before opening a documentation issue please consider opening a pull request instead for trivial changes such as typos spelling incorrect links anchors or other corrections that are easier to just fix than report using this template please do not spend valuable time proposing extensive changes to the documentation before first asking about it we value your time and do not want to waste it just open an issue first using this template and ask if your proposed changes would be helpful make sure to wrap all code examples in backticks so that they display correctly before submitting an issue make sure to click on the preview tab above to verify everything is formatted correctly i m trying to use validateasync and consume the results as a promise i m having trouble finding the documentation about when the promise resolves or rejects and what arguments will be supplied to then and catch i assumed both then and catch would receive an object with value error and warning keys similar to the synchronous version testing and logging has shown this to not be the case what are the arguments to then is it just the updated value or undefined if no transformations were applied what are the arguments to catch is it just the error if catch is called are joi transformations applied to the original value
0
527,994
15,357,901,472
IssuesEvent
2021-03-01 14:13:44
airshipit/airshipctl
https://api.github.com/repos/airshipit/airshipctl
closed
Enhance Github Bot to add proper labels to move issues to InProgress
enhancement priority/medium size s
Is it possible for the bot to besides adding the WIP label , change the status to In Progress?
1.0
Enhance Github Bot to add proper labels to move issues to InProgress - Is it possible for the bot to besides adding the WIP label , change the status to In Progress?
non_test
enhance github bot to add proper labels to move issues to inprogress is it possible for the bot to besides adding the wip label change the status to in progress
0
133,912
10,866,832,406
IssuesEvent
2019-11-14 22:10:58
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
vSphere Legacy Template Normalizing
[zube]: To Test area/vsphere team/az
The old fields in vmwarevsphere templates were free form text fields and the finder methods in govmomi allowed for something like "My Network" to be a lookup when the real object path was "/CLUSTER_NAME/networks/My Network". We need a shim during the upgrade process which looks at a template and tries to upgrade the data to the full path we now use w/ the lookup api in 2.3.3+
1.0
vSphere Legacy Template Normalizing - The old fields in vmwarevsphere templates were free form text fields and the finder methods in govmomi allowed for something like "My Network" to be a lookup when the real object path was "/CLUSTER_NAME/networks/My Network". We need a shim during the upgrade process which looks at a template and tries to upgrade the data to the full path we now use w/ the lookup api in 2.3.3+
test
vsphere legacy template normalizing the old fields in vmwarevsphere templates were free form text fields and the finder methods in govmomi allowed for something like my network to be a lookup when the real object path was cluster name networks my network we need a shim during the upgrade process which looks at a template and tries to upgrade the data to the full path we now use w the lookup api in
1
343,868
24,788,506,136
IssuesEvent
2022-10-24 11:56:14
opted-eu/wp3inventory
https://api.github.com/repos/opted-eu/wp3inventory
opened
Contribution Guidelines for Tools
documentation OPTED Infrastructure
Collect here rules and guidelines for adding tools - "name" should always be as short as possible. Full name of abbreviations and acronyms are stored in "other names" and can be repeated in the "description" text - Affective Norms for English Words ➡️ ANEW - "The General Inquirer: A Computer System for Content Analysis and Retrieval Based on the Sentence as a Unit of Information" ➡️ General Inquirer Dictionaries - If there are duplicated entries, the older entry is retained, the newer entry is rejected. - exception: the new entry has significantly more information and is of higher quality - in any case, the "unique_name" field has to be checked afterwards in the canonical entry
1.0
Contribution Guidelines for Tools - Collect here rules and guidelines for adding tools - "name" should always be as short as possible. Full name of abbreviations and acronyms are stored in "other names" and can be repeated in the "description" text - Affective Norms for English Words ➡️ ANEW - "The General Inquirer: A Computer System for Content Analysis and Retrieval Based on the Sentence as a Unit of Information" ➡️ General Inquirer Dictionaries - If there are duplicated entries, the older entry is retained, the newer entry is rejected. - exception: the new entry has significantly more information and is of higher quality - in any case, the "unique_name" field has to be checked afterwards in the canonical entry
non_test
contribution guidelines for tools collect here rules and guidelines for adding tools name should always be as short as possible full name of abbreviations and acronyms are stored in other names and can be repeated in the description text affective norms for english words ➡️ anew the general inquirer a computer system for content analysis and retrieval based on the sentence as a unit of information ➡️ general inquirer dictionaries if there are duplicated entries the older entry is retained the newer entry is rejected exception the new entry has significantly more information and is of higher quality in any case the unique name field has to be checked afterwards in the canonical entry
0
287,935
24,876,046,079
IssuesEvent
2022-10-27 19:11:43
lowRISC/opentitan
https://api.github.com/repos/lowRISC/opentitan
closed
[rom-e2e] rom_e2e_asm_interrupt_handler
Type:Task SW:ROM Milestone:V2 Component:Rom/E2e/Test
**Testpoint name:** [rom_e2e_asm_interrupt_handler](https://cs.opensource.google/opentitan/opentitan/+/master:sw/device/silicon_creator/rom/data/rom_e2e_testplan.hjson?q=rom_e2e_asm_interrupt_handler) **Contact person:** @alphan **Description:** Verify that asm interrupt handler resets the chip. `CREATOR_SW_CFG_ROM_EXEC_EN` should be set to `0` and the chip should in a life cycle state where debugging is enabled, i.e. TEST, DEV, or RMA. - Connect a debugger. - Set a breakpoint on the asm handler. - Trigger an exception, e.g. by setting `PC` to an address (end of ROM) that will trigger an instruction access fault. - Verify that execution breaks at the asm handler.
1.0
[rom-e2e] rom_e2e_asm_interrupt_handler - **Testpoint name:** [rom_e2e_asm_interrupt_handler](https://cs.opensource.google/opentitan/opentitan/+/master:sw/device/silicon_creator/rom/data/rom_e2e_testplan.hjson?q=rom_e2e_asm_interrupt_handler) **Contact person:** @alphan **Description:** Verify that asm interrupt handler resets the chip. `CREATOR_SW_CFG_ROM_EXEC_EN` should be set to `0` and the chip should in a life cycle state where debugging is enabled, i.e. TEST, DEV, or RMA. - Connect a debugger. - Set a breakpoint on the asm handler. - Trigger an exception, e.g. by setting `PC` to an address (end of ROM) that will trigger an instruction access fault. - Verify that execution breaks at the asm handler.
test
rom asm interrupt handler testpoint name contact person alphan description verify that asm interrupt handler resets the chip creator sw cfg rom exec en should be set to and the chip should in a life cycle state where debugging is enabled i e test dev or rma connect a debugger set a breakpoint on the asm handler trigger an exception e g by setting pc to an address end of rom that will trigger an instruction access fault verify that execution breaks at the asm handler
1
271,269
23,593,577,763
IssuesEvent
2022-08-23 17:12:03
MPMG-DCC-UFMG/F01
https://api.github.com/repos/MPMG-DCC-UFMG/F01
closed
Teste de generalizacao para a tag Obras públicas - Dados para acompanhamento - Guaranésia
generalization test development template - Betha tag - Obras Públicas subtag - Dados para acompanhamento
DoD: Realizar o teste de Generalização do validador da tag Obras públicas - Dados para acompanhamento para o Município de Guaranésia.
1.0
Teste de generalizacao para a tag Obras públicas - Dados para acompanhamento - Guaranésia - DoD: Realizar o teste de Generalização do validador da tag Obras públicas - Dados para acompanhamento para o Município de Guaranésia.
test
teste de generalizacao para a tag obras públicas dados para acompanhamento guaranésia dod realizar o teste de generalização do validador da tag obras públicas dados para acompanhamento para o município de guaranésia
1
234,115
19,096,229,636
IssuesEvent
2021-11-29 16:57:33
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
kv/kvserver: TestRequestsOnLaggingReplica failed
C-test-failure O-robot branch-release-21.2
kv/kvserver.TestRequestsOnLaggingReplica [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3813272&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3813272&tab=artifacts#/) on release-21.2 @ [d60310f8bbc38b9d090847b2266a90ff11a300e6](https://github.com/cockroachdb/cockroach/commits/d60310f8bbc38b9d090847b2266a90ff11a300e6): ``` === RUN TestRequestsOnLaggingReplica test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestRequestsOnLaggingReplica546084043 test_log_scope.go:80: use -show-logs to present logs inline client_raft_test.go:1281: Error Trace: client_raft_test.go:1281 Error: Not equal: expected: 2 actual : 1 Test: TestRequestsOnLaggingReplica panic.go:613: -- test log scope end -- test logs left over in: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestRequestsOnLaggingReplica546084043 --- FAIL: TestRequestsOnLaggingReplica (5.76s) ``` <details><summary>Reproduce</summary> <p> To reproduce, try: ```bash make stressrace TESTS=TestRequestsOnLaggingReplica PKG=./pkg/kv/kvserver TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1 ``` Parameters in this failure: - GOFLAGS=-json </p> </details> <details><summary>Same failure on other branches</summary> <p> - #57932 kv/kvserver: TestRequestsOnLaggingReplica failed [C-test-failure O-robot branch-master] </p> </details> /cc @cockroachdb/kv andreimatei <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestRequestsOnLaggingReplica.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
1.0
kv/kvserver: TestRequestsOnLaggingReplica failed - kv/kvserver.TestRequestsOnLaggingReplica [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3813272&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3813272&tab=artifacts#/) on release-21.2 @ [d60310f8bbc38b9d090847b2266a90ff11a300e6](https://github.com/cockroachdb/cockroach/commits/d60310f8bbc38b9d090847b2266a90ff11a300e6): ``` === RUN TestRequestsOnLaggingReplica test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestRequestsOnLaggingReplica546084043 test_log_scope.go:80: use -show-logs to present logs inline client_raft_test.go:1281: Error Trace: client_raft_test.go:1281 Error: Not equal: expected: 2 actual : 1 Test: TestRequestsOnLaggingReplica panic.go:613: -- test log scope end -- test logs left over in: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestRequestsOnLaggingReplica546084043 --- FAIL: TestRequestsOnLaggingReplica (5.76s) ``` <details><summary>Reproduce</summary> <p> To reproduce, try: ```bash make stressrace TESTS=TestRequestsOnLaggingReplica PKG=./pkg/kv/kvserver TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1 ``` Parameters in this failure: - GOFLAGS=-json </p> </details> <details><summary>Same failure on other branches</summary> <p> - #57932 kv/kvserver: TestRequestsOnLaggingReplica failed [C-test-failure O-robot branch-master] </p> </details> /cc @cockroachdb/kv andreimatei <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestRequestsOnLaggingReplica.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
kv kvserver testrequestsonlaggingreplica failed kv kvserver testrequestsonlaggingreplica with on release run testrequestsonlaggingreplica test log scope go test logs captured to go src github com cockroachdb cockroach artifacts test log scope go use show logs to present logs inline client raft test go error trace client raft test go error not equal expected actual test testrequestsonlaggingreplica panic go test log scope end test logs left over in go src github com cockroachdb cockroach artifacts fail testrequestsonlaggingreplica reproduce to reproduce try bash make stressrace tests testrequestsonlaggingreplica pkg pkg kv kvserver testtimeout stressflags timeout parameters in this failure goflags json same failure on other branches kv kvserver testrequestsonlaggingreplica failed cc cockroachdb kv andreimatei
1
317,457
27,238,270,162
IssuesEvent
2023-02-21 18:00:52
DCS-LCSR/SignStream3
https://api.github.com/repos/DCS-LCSR/SignStream3
closed
June 1 Program seems to hang when Sign Bank search fails because DAI server is down
enhancement CRITICAL testing-June-1 severity MED usability concern
<img width="947" alt="Screen Shot 2021-06-09 at 4 24 05 AM" src="https://user-images.githubusercontent.com/13629122/121319927-aca74080-c8da-11eb-9245-bf8aa6d9c292.png"> It is unresponsive for a very long time.
1.0
June 1 Program seems to hang when Sign Bank search fails because DAI server is down - <img width="947" alt="Screen Shot 2021-06-09 at 4 24 05 AM" src="https://user-images.githubusercontent.com/13629122/121319927-aca74080-c8da-11eb-9245-bf8aa6d9c292.png"> It is unresponsive for a very long time.
test
june program seems to hang when sign bank search fails because dai server is down img width alt screen shot at am src it is unresponsive for a very long time
1
211,242
16,191,851,584
IssuesEvent
2021-05-04 09:35:45
ubtue/tuefind
https://api.github.com/repos/ubtue/tuefind
closed
Alertpakete: irreführende Anzeige "Druck"
System: RelBib ready for testing
In der Anzeige der Alertpakete ist immer "Druck" enthalten: https://www.relbib.de/Search/Results?type=AllFields&filter%5B%5D=%7Eformat%3A%22Subscription+Bundle%22 Das ist besonders irreführend bei dem Paket "Open Access Zeitschriften. Kann diese Anzeige entfernt bzw. korrigiert werden, falls der Aufwand nicht zu hoch? Z. B. so: 1. Anzeige "elektronisch", wenn nur elektronische Zss im Paket enthalten 2. Anzeige "elektronisch / Druck", wenn sowohl Druck- als auch elektronische Zss enthalten 3. Anzeige "Druck", wenn nur Druck-Zss enthalten Das müsste dann natürlich dynamisch funktionieren.
1.0
Alertpakete: irreführende Anzeige "Druck" - In der Anzeige der Alertpakete ist immer "Druck" enthalten: https://www.relbib.de/Search/Results?type=AllFields&filter%5B%5D=%7Eformat%3A%22Subscription+Bundle%22 Das ist besonders irreführend bei dem Paket "Open Access Zeitschriften. Kann diese Anzeige entfernt bzw. korrigiert werden, falls der Aufwand nicht zu hoch? Z. B. so: 1. Anzeige "elektronisch", wenn nur elektronische Zss im Paket enthalten 2. Anzeige "elektronisch / Druck", wenn sowohl Druck- als auch elektronische Zss enthalten 3. Anzeige "Druck", wenn nur Druck-Zss enthalten Das müsste dann natürlich dynamisch funktionieren.
test
alertpakete irreführende anzeige druck in der anzeige der alertpakete ist immer druck enthalten das ist besonders irreführend bei dem paket open access zeitschriften kann diese anzeige entfernt bzw korrigiert werden falls der aufwand nicht zu hoch z b so anzeige elektronisch wenn nur elektronische zss im paket enthalten anzeige elektronisch druck wenn sowohl druck als auch elektronische zss enthalten anzeige druck wenn nur druck zss enthalten das müsste dann natürlich dynamisch funktionieren
1
267,956
23,335,807,528
IssuesEvent
2022-08-09 09:49:44
mozilla-mobile/fenix
https://api.github.com/repos/mozilla-mobile/fenix
opened
Intermittent Unit Test failure - StartupTypeTelemetryTest.WHEN record is called THEN telemetry is recorded with the appropriate label
eng:intermittent-test
### Test Run: Test failed in https://github.com/mozilla-mobile/fenix/pull/26378 - https://firefoxci.taskcluster-artifacts.net/OIThuwo1TRSMqXOucAYXqg/0/public/reports/test/testDebugUnitTest/index.html ### Stacktrace: > java.lang.AssertionError: expected:<1> but was:<null> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:834) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.mozilla.fenix.perf.StartupTypeTelemetryTest.WHEN record is called THEN telemetry is recorded with the appropriate label(StartupTypeTelemetryTest.kt:96) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55) at org.junit.rules.RunRules.evaluate(RunRules.java:20) at org.robolectric.RobolectricTestRunner$HelperTestRunner$1.evaluate(RobolectricTestRunner.java:591) at org.robolectric.internal.SandboxTestRunner$2.lambda$evaluate$0(SandboxTestRunner.java:274) at org.robolectric.internal.bytecode.Sandbox.lambda$runOnMainThread$0(Sandbox.java:88) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) Suppressed: org.robolectric.android.internal.AndroidTestEnvironment$UnExecutedRunnablesException: Main looper has queued unexecuted runnables. This might be the cause of the test failure. You might need a shadowOf(Looper.getMainLooper()).idle() call. ### Build: 105
1.0
Intermittent Unit Test failure - StartupTypeTelemetryTest.WHEN record is called THEN telemetry is recorded with the appropriate label - ### Test Run: Test failed in https://github.com/mozilla-mobile/fenix/pull/26378 - https://firefoxci.taskcluster-artifacts.net/OIThuwo1TRSMqXOucAYXqg/0/public/reports/test/testDebugUnitTest/index.html ### Stacktrace: > java.lang.AssertionError: expected:<1> but was:<null> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:834) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.mozilla.fenix.perf.StartupTypeTelemetryTest.WHEN record is called THEN telemetry is recorded with the appropriate label(StartupTypeTelemetryTest.kt:96) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55) at org.junit.rules.RunRules.evaluate(RunRules.java:20) at org.robolectric.RobolectricTestRunner$HelperTestRunner$1.evaluate(RobolectricTestRunner.java:591) at org.robolectric.internal.SandboxTestRunner$2.lambda$evaluate$0(SandboxTestRunner.java:274) at org.robolectric.internal.bytecode.Sandbox.lambda$runOnMainThread$0(Sandbox.java:88) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) Suppressed: org.robolectric.android.internal.AndroidTestEnvironment$UnExecutedRunnablesException: Main looper has queued unexecuted runnables. This might be the cause of the test failure. You might need a shadowOf(Looper.getMainLooper()).idle() call. ### Build: 105
test
intermittent unit test failure startuptypetelemetrytest when record is called then telemetry is recorded with the appropriate label test run test failed in stacktrace java lang assertionerror expected but was at org junit assert fail assert java at org junit assert failnotequals assert java at org junit assert assertequals assert java at org junit assert assertequals assert java at org mozilla fenix perf startuptypetelemetrytest when record is called then telemetry is recorded with the appropriate label startuptypetelemetrytest kt at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements runbefores evaluate runbefores java at org junit rules testwatcher evaluate testwatcher java at org junit rules runrules evaluate runrules java at org robolectric robolectrictestrunner helpertestrunner evaluate robolectrictestrunner java at org robolectric internal sandboxtestrunner lambda evaluate sandboxtestrunner java at org robolectric internal bytecode sandbox lambda runonmainthread sandbox java at java base java util concurrent futuretask run futuretask java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java base java lang thread run thread java suppressed org robolectric android internal androidtestenvironment unexecutedrunnablesexception main looper has queued unexecuted runnables this might be the cause of the test failure you might need a shadowof looper getmainlooper idle call build
1
2,027
2,581,515,710
IssuesEvent
2015-02-14 03:54:28
easydigitaldownloads/Easy-Digital-Downloads
https://api.github.com/repos/easydigitaldownloads/Easy-Digital-Downloads
closed
Polylang support
Enhancement Needs Testing
I am looking to make [Polylang](https://wordpress.org/plugins/polylang/) work with EDD. One thing I need to make multilingual is the checkout page. I created on in English and one in German. So that [`edd_get_checkout_uri()`](https://github.com/easydigitaldownloads/Easy-Digital-Downloads/blob/master/includes/checkout/functions.php#L118) returns the correct link acording to the language I would need to run `$edd_options['purchase_page']` though `pll_get_post()`. The function `pll_get_post()` returns the post id depending on the language of the current page. So something like this ```php $edd_options = edd_get_settings(); // Translate post_id for pages in options $edd_options['purchase_page'] = pll_get_post( $edd_options['purchase_page'] ); ``` Is it possible to add a filter which I could hook into? ---------------- Two posts(downloads) are created so to be able to translate the content in two langues. At the moment depending on which language the site being viewed in that version of the download would be added to the basket. A solution would be to only allow one version of the downloads to be added to the cart. I see that you use [`edd_add_to_cart()`](https://github.com/easydigitaldownloads/Easy-Digital-Downloads/blob/master/includes/cart/functions.php#L121) to add to the cart. If I were able to run this line of code then I would be able to achieve this. ```php $download_id = pll_get_post( $download_id, pll_default_language() ); ``` Is it possible to add a filter in this place? I hope I am making sense. I can make a PR to add these two filters. I just wanted to discuss this before make it.
1.0
Polylang support - I am looking to make [Polylang](https://wordpress.org/plugins/polylang/) work with EDD. One thing I need to make multilingual is the checkout page. I created on in English and one in German. So that [`edd_get_checkout_uri()`](https://github.com/easydigitaldownloads/Easy-Digital-Downloads/blob/master/includes/checkout/functions.php#L118) returns the correct link acording to the language I would need to run `$edd_options['purchase_page']` though `pll_get_post()`. The function `pll_get_post()` returns the post id depending on the language of the current page. So something like this ```php $edd_options = edd_get_settings(); // Translate post_id for pages in options $edd_options['purchase_page'] = pll_get_post( $edd_options['purchase_page'] ); ``` Is it possible to add a filter which I could hook into? ---------------- Two posts(downloads) are created so to be able to translate the content in two langues. At the moment depending on which language the site being viewed in that version of the download would be added to the basket. A solution would be to only allow one version of the downloads to be added to the cart. I see that you use [`edd_add_to_cart()`](https://github.com/easydigitaldownloads/Easy-Digital-Downloads/blob/master/includes/cart/functions.php#L121) to add to the cart. If I were able to run this line of code then I would be able to achieve this. ```php $download_id = pll_get_post( $download_id, pll_default_language() ); ``` Is it possible to add a filter in this place? I hope I am making sense. I can make a PR to add these two filters. I just wanted to discuss this before make it.
test
polylang support i am looking to make work with edd one thing i need to make multilingual is the checkout page i created on in english and one in german so that returns the correct link acording to the language i would need to run edd options though pll get post the function pll get post returns the post id depending on the language of the current page so something like this php edd options edd get settings translate post id for pages in options edd options pll get post edd options is it possible to add a filter which i could hook into two posts downloads are created so to be able to translate the content in two langues at the moment depending on which language the site being viewed in that version of the download would be added to the basket a solution would be to only allow one version of the downloads to be added to the cart i see that you use to add to the cart if i were able to run this line of code then i would be able to achieve this php download id pll get post download id pll default language is it possible to add a filter in this place i hope i am making sense i can make a pr to add these two filters i just wanted to discuss this before make it
1
38,551
5,192,035,449
IssuesEvent
2017-01-22 03:15:38
mil-oss/fgsms
https://api.github.com/repos/mil-oss/fgsms
closed
clean up uddi smoke test
bug tests
currently still uses the old old config file setup. a singular config file is the preferred solution. also need to stand up the server, test publication against it
1.0
clean up uddi smoke test - currently still uses the old old config file setup. a singular config file is the preferred solution. also need to stand up the server, test publication against it
test
clean up uddi smoke test currently still uses the old old config file setup a singular config file is the preferred solution also need to stand up the server test publication against it
1
29,462
4,501,258,373
IssuesEvent
2016-09-01 08:47:03
mattbearman/lime
https://api.github.com/repos/mattbearman/lime
closed
BugMuncher Feedback Report
bug BugMuncher other test
## Details ## **Submitted:** January 05, 2016 15:38 **Category:** other **Sender Email:** info@bugmuncher.com **Website:** BugMuncher App **URL:** https://app.bugmuncher.com/profiles/1/reports/112905/edit **Operating System:** Mac OS X Yosemite **Browser:** Chrome 47.0.2526.106 **Browser Size:** 1278 x 1341 **User Agent:** Mozilla/5.0 (Macintosh; Intel Mac OS X 10_10_5) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/47.0.2526.106 Safari/537.36 **Description:** General ## Custom Data ## **user_id:** 1 **account_id:** 1 **plan:** BugMuncher Corporate ## Browser Plugins ## Shockwave Flash Widevine Content Decryption Module Chrome PDF Viewer Native Client ## Events ## **method:** GET **url:** https://app.bugmuncher.com/user/dashboard **timestamp:** Tue Jan 05 2016 15:32:17 GMT+0000 (GMT) **type:** page_load --- **method:** GET **url:** https://app.bugmuncher.com/profiles/1/reports/112905/edit **timestamp:** Tue Jan 05 2016 15:36:01 GMT+0000 (GMT) **type:** page_load --- **content:** Feedback Button Clicked **timestamp:** Tue Jan 05 2016 15:38:42 GMT+0000 (GMT) **type:** bugmuncher --- **type:** bugmuncher **content:** Feedback Report Submitted **timestamp:** Tue Jan 05 2016 15:38:49 GMT+0000 (GMT) ---
1.0
BugMuncher Feedback Report - ## Details ## **Submitted:** January 05, 2016 15:38 **Category:** other **Sender Email:** info@bugmuncher.com **Website:** BugMuncher App **URL:** https://app.bugmuncher.com/profiles/1/reports/112905/edit **Operating System:** Mac OS X Yosemite **Browser:** Chrome 47.0.2526.106 **Browser Size:** 1278 x 1341 **User Agent:** Mozilla/5.0 (Macintosh; Intel Mac OS X 10_10_5) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/47.0.2526.106 Safari/537.36 **Description:** General ## Custom Data ## **user_id:** 1 **account_id:** 1 **plan:** BugMuncher Corporate ## Browser Plugins ## Shockwave Flash Widevine Content Decryption Module Chrome PDF Viewer Native Client ## Events ## **method:** GET **url:** https://app.bugmuncher.com/user/dashboard **timestamp:** Tue Jan 05 2016 15:32:17 GMT+0000 (GMT) **type:** page_load --- **method:** GET **url:** https://app.bugmuncher.com/profiles/1/reports/112905/edit **timestamp:** Tue Jan 05 2016 15:36:01 GMT+0000 (GMT) **type:** page_load --- **content:** Feedback Button Clicked **timestamp:** Tue Jan 05 2016 15:38:42 GMT+0000 (GMT) **type:** bugmuncher --- **type:** bugmuncher **content:** Feedback Report Submitted **timestamp:** Tue Jan 05 2016 15:38:49 GMT+0000 (GMT) ---
test
bugmuncher feedback report details submitted january category other sender email info bugmuncher com website bugmuncher app url operating system mac os x yosemite browser chrome browser size x user agent mozilla macintosh intel mac os x applewebkit khtml like gecko chrome safari description general custom data user id account id plan bugmuncher corporate browser plugins shockwave flash widevine content decryption module chrome pdf viewer native client events method get url timestamp tue jan gmt gmt type page load method get url timestamp tue jan gmt gmt type page load content feedback button clicked timestamp tue jan gmt gmt type bugmuncher type bugmuncher content feedback report submitted timestamp tue jan gmt gmt
1
439,966
30,724,101,161
IssuesEvent
2023-07-27 18:11:01
typescript-eslint/typescript-eslint
https://api.github.com/repos/typescript-eslint/typescript-eslint
closed
Docs: Mention preference for small, self-contained rule test cases
documentation accepting prs
### Before You File a Documentation Request Please Confirm You Have Done The Following... - [X] I have looked for existing [open or closed documentation requests](https://github.com/typescript-eslint/typescript-eslint/issues?q=is%3Aissue+label%3Adocumentation) that match my proposal. - [X] I have [read the FAQ](https://typescript-eslint.io/linting/troubleshooting) and my problem is not listed. ### Suggested Changes In general, in our `packages/eslint-plugin/tests/rules/**/*.test.ts` tests, I'm under the impression we generally ask for test cases that are single-purpose. Good: ```ts { code: `willCauseComplaint(1);` errors: [ { /* ... */ } ] } ``` Not so good, should be split up: ```ts { code: ` willCauseComplaint(1); willCauseComplaint(2); ` errors: [ { /* ... */ }, { /* ... */ } ] } ``` I see the latter form come up sometimes in PRs. Let's document our preference for the former? ### Affected URL(s) https://typescript-eslint.io/maintenance/pull-requests Or perhaps we'd want to split out a page dedicated to good practices for ESLint rule PRs? Since they're such a large percentage of our PRs & have specific practices associated. Aside: if this is accepted by other maintainers we should file an issue to fix up old test files that violate this preference. E.g. `no-floating-promises.test.ts`.
1.0
Docs: Mention preference for small, self-contained rule test cases - ### Before You File a Documentation Request Please Confirm You Have Done The Following... - [X] I have looked for existing [open or closed documentation requests](https://github.com/typescript-eslint/typescript-eslint/issues?q=is%3Aissue+label%3Adocumentation) that match my proposal. - [X] I have [read the FAQ](https://typescript-eslint.io/linting/troubleshooting) and my problem is not listed. ### Suggested Changes In general, in our `packages/eslint-plugin/tests/rules/**/*.test.ts` tests, I'm under the impression we generally ask for test cases that are single-purpose. Good: ```ts { code: `willCauseComplaint(1);` errors: [ { /* ... */ } ] } ``` Not so good, should be split up: ```ts { code: ` willCauseComplaint(1); willCauseComplaint(2); ` errors: [ { /* ... */ }, { /* ... */ } ] } ``` I see the latter form come up sometimes in PRs. Let's document our preference for the former? ### Affected URL(s) https://typescript-eslint.io/maintenance/pull-requests Or perhaps we'd want to split out a page dedicated to good practices for ESLint rule PRs? Since they're such a large percentage of our PRs & have specific practices associated. Aside: if this is accepted by other maintainers we should file an issue to fix up old test files that violate this preference. E.g. `no-floating-promises.test.ts`.
non_test
docs mention preference for small self contained rule test cases before you file a documentation request please confirm you have done the following i have looked for existing that match my proposal i have and my problem is not listed suggested changes in general in our packages eslint plugin tests rules test ts tests i m under the impression we generally ask for test cases that are single purpose good ts code willcausecomplaint errors not so good should be split up ts code willcausecomplaint willcausecomplaint errors i see the latter form come up sometimes in prs let s document our preference for the former affected url s or perhaps we d want to split out a page dedicated to good practices for eslint rule prs since they re such a large percentage of our prs have specific practices associated aside if this is accepted by other maintainers we should file an issue to fix up old test files that violate this preference e g no floating promises test ts
0
10,159
26,419,783,803
IssuesEvent
2023-01-13 19:13:14
MicrosoftDocs/architecture-center
https://api.github.com/repos/MicrosoftDocs/architecture-center
closed
Not All Page Appears When Export To PDF
assigned-to-author doc-idea triaged architecture-center/svc Pri1 design-pattern/subsvc
When I export to PDF some page doesn't appears for example: - Circuit Breaker Pattern - Claim Check - Compensating Transaction - Competing Consumers - Compute Resource Consolidation - CQRS - Deployment Stamps - Event Sourcing - External Configuration Store - Federated Identity Best regards --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 0a4ecd26-cb7b-4cb2-ae3a-3748ada92771 * Version Independent ID: 97f96179-a22b-8401-a2f4-264a86b5d2f5 * Content: [Cloud design patterns - Azure Architecture Center](https://docs.microsoft.com/en-us/azure/architecture/patterns/) * Content Source: [docs/patterns/index.md](https://github.com/microsoftdocs/architecture-center/blob/main/docs/patterns/index.md) * Service: **architecture-center** * Sub-service: **design-pattern** * GitHub Login: @EdPrice-MSFT * Microsoft Alias: **masashin**
1.0
Not All Page Appears When Export To PDF - When I export to PDF some page doesn't appears for example: - Circuit Breaker Pattern - Claim Check - Compensating Transaction - Competing Consumers - Compute Resource Consolidation - CQRS - Deployment Stamps - Event Sourcing - External Configuration Store - Federated Identity Best regards --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 0a4ecd26-cb7b-4cb2-ae3a-3748ada92771 * Version Independent ID: 97f96179-a22b-8401-a2f4-264a86b5d2f5 * Content: [Cloud design patterns - Azure Architecture Center](https://docs.microsoft.com/en-us/azure/architecture/patterns/) * Content Source: [docs/patterns/index.md](https://github.com/microsoftdocs/architecture-center/blob/main/docs/patterns/index.md) * Service: **architecture-center** * Sub-service: **design-pattern** * GitHub Login: @EdPrice-MSFT * Microsoft Alias: **masashin**
non_test
not all page appears when export to pdf when i export to pdf some page doesn t appears for example circuit breaker pattern claim check compensating transaction competing consumers compute resource consolidation cqrs deployment stamps event sourcing external configuration store federated identity best regards document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service architecture center sub service design pattern github login edprice msft microsoft alias masashin
0
302,651
26,158,716,148
IssuesEvent
2022-12-31 06:24:36
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: cdc/bank failed
C-test-failure O-robot O-roachtest release-blocker branch-release-22.2
roachtest.cdc/bank [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=artifacts#/cdc/bank) on release-22.2 @ [07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1](https://github.com/cockroachdb/cockroach/commits/07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1): ``` test artifacts and logs in: /artifacts/cdc/bank/run_1 (test_impl.go:286).Fatal: pq: Use of CHANGEFEED requires an enterprise license. Your evaluation license expired on December 30, 2022. If you're interested in getting a new license, please contact subscriptions@cockroachlabs.com and we can help you out. (test_impl.go:286).Fatal: cluster.RunE: context canceled ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #94463 roachtest: cdc/bank failed [C-test-failure O-roachtest O-robot T-cdc branch-master release-blocker] </p> </details> /cc @cockroachdb/cdc <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*cdc/bank.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
2.0
roachtest: cdc/bank failed - roachtest.cdc/bank [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=artifacts#/cdc/bank) on release-22.2 @ [07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1](https://github.com/cockroachdb/cockroach/commits/07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1): ``` test artifacts and logs in: /artifacts/cdc/bank/run_1 (test_impl.go:286).Fatal: pq: Use of CHANGEFEED requires an enterprise license. Your evaluation license expired on December 30, 2022. If you're interested in getting a new license, please contact subscriptions@cockroachlabs.com and we can help you out. (test_impl.go:286).Fatal: cluster.RunE: context canceled ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #94463 roachtest: cdc/bank failed [C-test-failure O-roachtest O-robot T-cdc branch-master release-blocker] </p> </details> /cc @cockroachdb/cdc <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*cdc/bank.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
roachtest cdc bank failed roachtest cdc bank with on release test artifacts and logs in artifacts cdc bank run test impl go fatal pq use of changefeed requires an enterprise license your evaluation license expired on december if you re interested in getting a new license please contact subscriptions cockroachlabs com and we can help you out test impl go fatal cluster rune context canceled parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see same failure on other branches roachtest cdc bank failed cc cockroachdb cdc
1
156,307
12,304,814,925
IssuesEvent
2020-05-11 21:10:49
futurewei-cloud/arktos
https://api.github.com/repos/futurewei-cloud/arktos
closed
Endpoints failed to get IP and port with error FailedToUpdateEndpoint
area/controller-manager area/performance-test
<!-- Please use this template while reporting a bug and provide as much info as possible. Not doing so may result in your bug not being addressed in a timely manner. Thanks! If the matter is security related, please disclose it privately via https://kubernetes.io/security/ --> **What happened**: 1. run kube-up.sh to start arktos ``` export MASTER_DISK_SIZE=300GB KUBE_GCE_ZONE=us-west2-b MASTER_SIZE=n1-highmem-32 NODE_SIZE=n1-highmem-8 NUM_NODES=3 NODE_DISK_SIZE=200GB KUBE_GCE_NETWORK=default ./cluster/kube-up.sh ``` 2. after arktos start successfully, check all endpoints status, get the info: ``` NAMESPACE NAME ENDPOINTS AGE SERVICEGROUPID default kubernetes 35.236.127.13:443 64s 0 kube-system default-http-backend <none> 29s <none> kube-system heapster <none> 29s <none> kube-system kube-controller-manager <none> 64s <none> kube-system kube-scheduler <none> 54s <none> kube-system kubernetes-dashboard <none> 29s <none> kube-system metrics-server <none> 29s <none> ``` ``` Name: metrics-server Namespace: kube-system Labels: addonmanager.kubernetes.io/mode=Reconcile kubernetes.io/cluster-service=true kubernetes.io/name=Metrics-server Annotations: endpoints.kubernetes.io/last-change-trigger-time: 2020-04-30T02:32:24Z Subsets: Events: Type Reason Age From Message ---- ------ ---- ---- ------- Warning FailedToUpdateEndpoint 3m1s (x25 over 3m55s) endpoint-controller Failed to update endpoint default/kube-system/metrics-server: Endpoints "metrics-server" is invalid: subsets[0].addresses[0].ip: Invalid value: "169.254.123.2": may not be in the link-local range (169.254.0.0/16) ``` **What you expected to happen**: Endpoints should get ipaddress and ports as below: ``` NAMESPACE NAME ENDPOINTS AGE default kubernetes 10.40.0.2:443 2m26s kube-system default-http-backend 10.64.7.4:8080 2m11s kube-system kube-controller-manager <none> 2m25s kube-system kube-dns 10.64.1.2:53,10.64.7.3:53,10.64.1.2:9153 + 3 more... 2m11s kube-system kube-scheduler <none> 2m24s kube-system kubernetes-dashboard 10.64.4.4:8443 2m11s kube-system metrics-server 10.64.6.2:443 2m11s ``` **How to reproduce it (as minimally and precisely as possible)**: **Anything else we need to know?**: **Environment**: - Arktos version (use `kubectl version`): - Cloud provider or hardware configuration: - OS (e.g: `cat /etc/os-release`): - Kernel (e.g. `uname -a`): - Install tools: - Network plugin and version (if this is a network-related bug): - Others:
1.0
Endpoints failed to get IP and port with error FailedToUpdateEndpoint - <!-- Please use this template while reporting a bug and provide as much info as possible. Not doing so may result in your bug not being addressed in a timely manner. Thanks! If the matter is security related, please disclose it privately via https://kubernetes.io/security/ --> **What happened**: 1. run kube-up.sh to start arktos ``` export MASTER_DISK_SIZE=300GB KUBE_GCE_ZONE=us-west2-b MASTER_SIZE=n1-highmem-32 NODE_SIZE=n1-highmem-8 NUM_NODES=3 NODE_DISK_SIZE=200GB KUBE_GCE_NETWORK=default ./cluster/kube-up.sh ``` 2. after arktos start successfully, check all endpoints status, get the info: ``` NAMESPACE NAME ENDPOINTS AGE SERVICEGROUPID default kubernetes 35.236.127.13:443 64s 0 kube-system default-http-backend <none> 29s <none> kube-system heapster <none> 29s <none> kube-system kube-controller-manager <none> 64s <none> kube-system kube-scheduler <none> 54s <none> kube-system kubernetes-dashboard <none> 29s <none> kube-system metrics-server <none> 29s <none> ``` ``` Name: metrics-server Namespace: kube-system Labels: addonmanager.kubernetes.io/mode=Reconcile kubernetes.io/cluster-service=true kubernetes.io/name=Metrics-server Annotations: endpoints.kubernetes.io/last-change-trigger-time: 2020-04-30T02:32:24Z Subsets: Events: Type Reason Age From Message ---- ------ ---- ---- ------- Warning FailedToUpdateEndpoint 3m1s (x25 over 3m55s) endpoint-controller Failed to update endpoint default/kube-system/metrics-server: Endpoints "metrics-server" is invalid: subsets[0].addresses[0].ip: Invalid value: "169.254.123.2": may not be in the link-local range (169.254.0.0/16) ``` **What you expected to happen**: Endpoints should get ipaddress and ports as below: ``` NAMESPACE NAME ENDPOINTS AGE default kubernetes 10.40.0.2:443 2m26s kube-system default-http-backend 10.64.7.4:8080 2m11s kube-system kube-controller-manager <none> 2m25s kube-system kube-dns 10.64.1.2:53,10.64.7.3:53,10.64.1.2:9153 + 3 more... 2m11s kube-system kube-scheduler <none> 2m24s kube-system kubernetes-dashboard 10.64.4.4:8443 2m11s kube-system metrics-server 10.64.6.2:443 2m11s ``` **How to reproduce it (as minimally and precisely as possible)**: **Anything else we need to know?**: **Environment**: - Arktos version (use `kubectl version`): - Cloud provider or hardware configuration: - OS (e.g: `cat /etc/os-release`): - Kernel (e.g. `uname -a`): - Install tools: - Network plugin and version (if this is a network-related bug): - Others:
test
endpoints failed to get ip and port with error failedtoupdateendpoint please use this template while reporting a bug and provide as much info as possible not doing so may result in your bug not being addressed in a timely manner thanks if the matter is security related please disclose it privately via what happened run kube up sh to start arktos export master disk size kube gce zone us b master size highmem node size highmem num nodes node disk size kube gce network default cluster kube up sh after arktos start successfully check all endpoints status get the info namespace name endpoints age servicegroupid default kubernetes kube system default http backend kube system heapster kube system kube controller manager kube system kube scheduler kube system kubernetes dashboard kube system metrics server name metrics server namespace kube system labels addonmanager kubernetes io mode reconcile kubernetes io cluster service true kubernetes io name metrics server annotations endpoints kubernetes io last change trigger time subsets events type reason age from message warning failedtoupdateendpoint over endpoint controller failed to update endpoint default kube system metrics server endpoints metrics server is invalid subsets addresses ip invalid value may not be in the link local range what you expected to happen endpoints should get ipaddress and ports as below namespace name endpoints age default kubernetes kube system default http backend kube system kube controller manager kube system kube dns more kube system kube scheduler kube system kubernetes dashboard kube system metrics server how to reproduce it as minimally and precisely as possible anything else we need to know environment arktos version use kubectl version cloud provider or hardware configuration os e g cat etc os release kernel e g uname a install tools network plugin and version if this is a network related bug others
1
228,067
18,154,148,894
IssuesEvent
2021-09-26 19:37:28
pandas-dev/pandas
https://api.github.com/repos/pandas-dev/pandas
opened
TST: Add doctests on cython files
Testing Docs CI
Adding `--doctest-cython` to `ci/run_tests.sh` to cover doctests in cython files. Lots of them are failing at the moment. But I think few of them (like now, today), can be skipped using `# doctest: +SKIP`. Below is the current list of failures atm. ```python ===================================================================================== short test summary info ===================================================================================== FAILED pandas/_libs/hashtable.cpython-38-darwin.so::pandas._libs.hashtable.Int64Factorizer.factorize FAILED pandas/_libs/hashtable.cpython-38-darwin.so::pandas._libs.hashtable.ObjectFactorizer.factorize FAILED pandas/_libs/hashtable.cpython-38-darwin.so::pandas._libs.hashtable.__test__.Int64Factorizer.factorize (line 134) FAILED pandas/_libs/hashtable.cpython-38-darwin.so::pandas._libs.hashtable.__test__.ObjectFactorizer.factorize (line 88) FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.__test__.infer_dtype (line 1299) FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.__test__.is_iterator (line 242) FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.__test__.is_list_like (line 1051) FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.__test__.is_scalar (line 149) FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.infer_dtype FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.is_iterator FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.is_list_like FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.is_scalar FAILED pandas/_libs/tslibs/dtypes.cpython-38-darwin.so::pandas._libs.tslibs.dtypes.Resolution.from_attrname FAILED pandas/_libs/tslibs/dtypes.cpython-38-darwin.so::pandas._libs.tslibs.dtypes.Resolution.get_reso_from_freq FAILED pandas/_libs/tslibs/dtypes.cpython-38-darwin.so::pandas._libs.tslibs.dtypes.__test__.Resolution.from_attrname (line 223) FAILED pandas/_libs/tslibs/dtypes.cpython-38-darwin.so::pandas._libs.tslibs.dtypes.__test__.Resolution.get_reso_from_freq (line 238) FAILED pandas/_libs/tslibs/nattype.cpython-38-darwin.so::pandas._libs.tslibs.nattype.NaTType.now FAILED pandas/_libs/tslibs/nattype.cpython-38-darwin.so::pandas._libs.tslibs.nattype.NaTType.today FAILED pandas/_libs/tslibs/nattype.cpython-38-darwin.so::pandas._libs.tslibs.nattype.NaTType.utcfromtimestamp FAILED pandas/_libs/tslibs/nattype.cpython-38-darwin.so::pandas._libs.tslibs.nattype.NaTType.utcnow FAILED pandas/_libs/tslibs/period.cpython-38-darwin.so::pandas._libs.tslibs.period._Period.strftime FAILED pandas/_libs/tslibs/period.cpython-38-darwin.so::pandas._libs.tslibs.period.__test__._Period.strftime (line 2243) FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.Timestamp.now FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.Timestamp.today FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.Timestamp.utcfromtimestamp FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.Timestamp.utcnow FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps._Timestamp.to_period FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.__test__.Timestamp.now (line 1048) FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.__test__.Timestamp.today (line 1075) FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.__test__.Timestamp.utcfromtimestamp (line 1115) FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.__test__.Timestamp.utcnow (line 1101) FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.__test__._Timestamp.to_period (line 905) FAILED pandas/_libs/tslibs/timezones.cpython-38-darwin.so::pandas._libs.tslibs.timezones.__test__.tz_standardize (line 375) FAILED pandas/_libs/tslibs/timezones.cpython-38-darwin.so::pandas._libs.tslibs.timezones.tz_standardize ================================================================================= 34 failed, 106 passed in 1.10s ==================================================================================
1.0
TST: Add doctests on cython files - Adding `--doctest-cython` to `ci/run_tests.sh` to cover doctests in cython files. Lots of them are failing at the moment. But I think few of them (like now, today), can be skipped using `# doctest: +SKIP`. Below is the current list of failures atm. ```python ===================================================================================== short test summary info ===================================================================================== FAILED pandas/_libs/hashtable.cpython-38-darwin.so::pandas._libs.hashtable.Int64Factorizer.factorize FAILED pandas/_libs/hashtable.cpython-38-darwin.so::pandas._libs.hashtable.ObjectFactorizer.factorize FAILED pandas/_libs/hashtable.cpython-38-darwin.so::pandas._libs.hashtable.__test__.Int64Factorizer.factorize (line 134) FAILED pandas/_libs/hashtable.cpython-38-darwin.so::pandas._libs.hashtable.__test__.ObjectFactorizer.factorize (line 88) FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.__test__.infer_dtype (line 1299) FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.__test__.is_iterator (line 242) FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.__test__.is_list_like (line 1051) FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.__test__.is_scalar (line 149) FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.infer_dtype FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.is_iterator FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.is_list_like FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.is_scalar FAILED pandas/_libs/tslibs/dtypes.cpython-38-darwin.so::pandas._libs.tslibs.dtypes.Resolution.from_attrname FAILED pandas/_libs/tslibs/dtypes.cpython-38-darwin.so::pandas._libs.tslibs.dtypes.Resolution.get_reso_from_freq FAILED pandas/_libs/tslibs/dtypes.cpython-38-darwin.so::pandas._libs.tslibs.dtypes.__test__.Resolution.from_attrname (line 223) FAILED pandas/_libs/tslibs/dtypes.cpython-38-darwin.so::pandas._libs.tslibs.dtypes.__test__.Resolution.get_reso_from_freq (line 238) FAILED pandas/_libs/tslibs/nattype.cpython-38-darwin.so::pandas._libs.tslibs.nattype.NaTType.now FAILED pandas/_libs/tslibs/nattype.cpython-38-darwin.so::pandas._libs.tslibs.nattype.NaTType.today FAILED pandas/_libs/tslibs/nattype.cpython-38-darwin.so::pandas._libs.tslibs.nattype.NaTType.utcfromtimestamp FAILED pandas/_libs/tslibs/nattype.cpython-38-darwin.so::pandas._libs.tslibs.nattype.NaTType.utcnow FAILED pandas/_libs/tslibs/period.cpython-38-darwin.so::pandas._libs.tslibs.period._Period.strftime FAILED pandas/_libs/tslibs/period.cpython-38-darwin.so::pandas._libs.tslibs.period.__test__._Period.strftime (line 2243) FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.Timestamp.now FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.Timestamp.today FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.Timestamp.utcfromtimestamp FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.Timestamp.utcnow FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps._Timestamp.to_period FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.__test__.Timestamp.now (line 1048) FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.__test__.Timestamp.today (line 1075) FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.__test__.Timestamp.utcfromtimestamp (line 1115) FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.__test__.Timestamp.utcnow (line 1101) FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.__test__._Timestamp.to_period (line 905) FAILED pandas/_libs/tslibs/timezones.cpython-38-darwin.so::pandas._libs.tslibs.timezones.__test__.tz_standardize (line 375) FAILED pandas/_libs/tslibs/timezones.cpython-38-darwin.so::pandas._libs.tslibs.timezones.tz_standardize ================================================================================= 34 failed, 106 passed in 1.10s ==================================================================================
test
tst add doctests on cython files adding doctest cython to ci run tests sh to cover doctests in cython files lots of them are failing at the moment but i think few of them like now today can be skipped using doctest skip below is the current list of failures atm python short test summary info failed pandas libs hashtable cpython darwin so pandas libs hashtable factorize failed pandas libs hashtable cpython darwin so pandas libs hashtable objectfactorizer factorize failed pandas libs hashtable cpython darwin so pandas libs hashtable test factorize line failed pandas libs hashtable cpython darwin so pandas libs hashtable test objectfactorizer factorize line failed pandas libs lib cpython darwin so pandas libs lib test infer dtype line failed pandas libs lib cpython darwin so pandas libs lib test is iterator line failed pandas libs lib cpython darwin so pandas libs lib test is list like line failed pandas libs lib cpython darwin so pandas libs lib test is scalar line failed pandas libs lib cpython darwin so pandas libs lib infer dtype failed pandas libs lib cpython darwin so pandas libs lib is iterator failed pandas libs lib cpython darwin so pandas libs lib is list like failed pandas libs lib cpython darwin so pandas libs lib is scalar failed pandas libs tslibs dtypes cpython darwin so pandas libs tslibs dtypes resolution from attrname failed pandas libs tslibs dtypes cpython darwin so pandas libs tslibs dtypes resolution get reso from freq failed pandas libs tslibs dtypes cpython darwin so pandas libs tslibs dtypes test resolution from attrname line failed pandas libs tslibs dtypes cpython darwin so pandas libs tslibs dtypes test resolution get reso from freq line failed pandas libs tslibs nattype cpython darwin so pandas libs tslibs nattype nattype now failed pandas libs tslibs nattype cpython darwin so pandas libs tslibs nattype nattype today failed pandas libs tslibs nattype cpython darwin so pandas libs tslibs nattype nattype utcfromtimestamp failed pandas libs tslibs nattype cpython darwin so pandas libs tslibs nattype nattype utcnow failed pandas libs tslibs period cpython darwin so pandas libs tslibs period period strftime failed pandas libs tslibs period cpython darwin so pandas libs tslibs period test period strftime line failed pandas libs tslibs timestamps cpython darwin so pandas libs tslibs timestamps timestamp now failed pandas libs tslibs timestamps cpython darwin so pandas libs tslibs timestamps timestamp today failed pandas libs tslibs timestamps cpython darwin so pandas libs tslibs timestamps timestamp utcfromtimestamp failed pandas libs tslibs timestamps cpython darwin so pandas libs tslibs timestamps timestamp utcnow failed pandas libs tslibs timestamps cpython darwin so pandas libs tslibs timestamps timestamp to period failed pandas libs tslibs timestamps cpython darwin so pandas libs tslibs timestamps test timestamp now line failed pandas libs tslibs timestamps cpython darwin so pandas libs tslibs timestamps test timestamp today line failed pandas libs tslibs timestamps cpython darwin so pandas libs tslibs timestamps test timestamp utcfromtimestamp line failed pandas libs tslibs timestamps cpython darwin so pandas libs tslibs timestamps test timestamp utcnow line failed pandas libs tslibs timestamps cpython darwin so pandas libs tslibs timestamps test timestamp to period line failed pandas libs tslibs timezones cpython darwin so pandas libs tslibs timezones test tz standardize line failed pandas libs tslibs timezones cpython darwin so pandas libs tslibs timezones tz standardize failed passed in
1
114,732
9,753,401,510
IssuesEvent
2019-06-04 09:03:24
spacemeshos/go-spacemesh
https://api.github.com/repos/spacemeshos/go-spacemesh
closed
Create query for logs per node after test ends
automation tests
# Overview / Motivation we want developers to easily debug logs from a test run, usually, this involves watching logs from different parts of the test isolated but simultaneously. # The Task create a query that gets all the relevant logs for that test grouped by the client, node or pod that wrote them. # Implementation Notes TODO: Add links to relevant resources, specs, related issues, etc... # Contribution Guidelines Important: Issue assignment to developers will be by the order of their application and proficiency level according to the tasks complexity. We will not assign tasks to developers who have'nt introduced themselves on our Gitter [dev channel](https://gitter.im/spacemesh-os/Lobby) 1. Introduce yourself on go-spacemesh [dev chat channel](https://gitter.im/spacemesh-os/Lobby) - ask our team any question you may have about this task 2. Fork branch `develop` to your own repo and work in your repo 3. You must document all methods, enums and types with [godoc comments](https://blog.golang.org/godoc-documenting-go-code) 4. You must write go unit tests for all types and methods when submitting a component, and integration tests if you submit a feature 5. When ready for code review, submit a PR from your repo back to branch `develop` 6. Attach relevant issue to PR
1.0
Create query for logs per node after test ends - # Overview / Motivation we want developers to easily debug logs from a test run, usually, this involves watching logs from different parts of the test isolated but simultaneously. # The Task create a query that gets all the relevant logs for that test grouped by the client, node or pod that wrote them. # Implementation Notes TODO: Add links to relevant resources, specs, related issues, etc... # Contribution Guidelines Important: Issue assignment to developers will be by the order of their application and proficiency level according to the tasks complexity. We will not assign tasks to developers who have'nt introduced themselves on our Gitter [dev channel](https://gitter.im/spacemesh-os/Lobby) 1. Introduce yourself on go-spacemesh [dev chat channel](https://gitter.im/spacemesh-os/Lobby) - ask our team any question you may have about this task 2. Fork branch `develop` to your own repo and work in your repo 3. You must document all methods, enums and types with [godoc comments](https://blog.golang.org/godoc-documenting-go-code) 4. You must write go unit tests for all types and methods when submitting a component, and integration tests if you submit a feature 5. When ready for code review, submit a PR from your repo back to branch `develop` 6. Attach relevant issue to PR
test
create query for logs per node after test ends overview motivation we want developers to easily debug logs from a test run usually this involves watching logs from different parts of the test isolated but simultaneously the task create a query that gets all the relevant logs for that test grouped by the client node or pod that wrote them implementation notes todo add links to relevant resources specs related issues etc contribution guidelines important issue assignment to developers will be by the order of their application and proficiency level according to the tasks complexity we will not assign tasks to developers who have nt introduced themselves on our gitter introduce yourself on go spacemesh ask our team any question you may have about this task fork branch develop to your own repo and work in your repo you must document all methods enums and types with you must write go unit tests for all types and methods when submitting a component and integration tests if you submit a feature when ready for code review submit a pr from your repo back to branch develop attach relevant issue to pr
1
76,562
7,539,839,902
IssuesEvent
2018-04-17 02:51:22
sudar/bulk-delete
https://api.github.com/repos/sudar/bulk-delete
opened
User Post Meta - Migrate to new architecture
1 - Ready task tests
- [ ] Migrate User Post Meta - [ ] Manually Test the changes - [ ] Manually Test that the User User Meta Pro add-on works ## Tests - [ ] Add tests for deleting single user meta fields from 'admin' user role - [ ] Add tests for deleting single user meta fields from 'subscriber' user role - [ ] Add tests for deleting multiple user meta fields from 'admin' user role - [ ] Add tests for deleting multiple user meta fields from 'subscriber' user role - [ ] Add tests for deleting user meta from users in batches
1.0
User Post Meta - Migrate to new architecture - - [ ] Migrate User Post Meta - [ ] Manually Test the changes - [ ] Manually Test that the User User Meta Pro add-on works ## Tests - [ ] Add tests for deleting single user meta fields from 'admin' user role - [ ] Add tests for deleting single user meta fields from 'subscriber' user role - [ ] Add tests for deleting multiple user meta fields from 'admin' user role - [ ] Add tests for deleting multiple user meta fields from 'subscriber' user role - [ ] Add tests for deleting user meta from users in batches
test
user post meta migrate to new architecture migrate user post meta manually test the changes manually test that the user user meta pro add on works tests add tests for deleting single user meta fields from admin user role add tests for deleting single user meta fields from subscriber user role add tests for deleting multiple user meta fields from admin user role add tests for deleting multiple user meta fields from subscriber user role add tests for deleting user meta from users in batches
1
299,847
25,930,862,464
IssuesEvent
2022-12-16 09:53:37
parca-dev/parca-agent
https://api.github.com/repos/parca-dev/parca-agent
opened
Testing the Agent
stack-quality testing
## The state of testing affairs At the moment we have: - **unit-tests**: to ensure the correctness of small well-contained pieces of code, such as our [Kernel symbolizer](https://github.com/parca-dev/parca-agent/blob/main/pkg/ksym/ksym_test.go). This helps us ensure that the behaviour is correct, but not less importantly has everything in place for us to add regression tests if we encounter bugs that we then fix; - **e2e tests**: running on Minikube that have the potential of testing our Kubernetes / systems metadata collection; These two are great and are very necessary. However, we have some blindspots, that are only going to get worse once we start shipping more advanced features, such as the [DWARF-based unwinder](https://github.com/parca-dev/parca-agent/issues/768). In my opinion, we are missing three main things: 1. A way to ensure that the **kernels we want to support work** (it's easy to introduce regressions, such as a feature that's not supported in a kernel we care about, or some code pattern that the verifier in some kernels rejects. These two are things I've spotted before, not fabricated examples) 1. Ensuring that resource usage, e.g. **memory locking etc is within what we expect** (we had to recently [remediate this](https://github.com/parca-dev/parca-agent/pull/1074)) 1. **Correctness testing**: on any kernel we support, running test programs and our unwinder and verifying that we get correct stack traces ## Propose additions ### VM testing For both 1) and - partially - 2) we need to add some testing in VMs with different kernels. Instead of running all of our integration tests (which may be flaky) or unit tests (which don't add much value in different kernels), I would like to propose the following for the kernels we support: - Can create the maps, so we ensure that our logic to create a map of size set at runtime is correct (note: we'll make this logic smarter to load as much as we can for a given host with a ceiling of 400-500MB); - Load our BPF program, to ensure that the verifier is happy with our code; This will also test that the BTF-pulling system for hosts that don't have it locally works, too (https://github.com/parca-dev/parca-agent/pull/1131). Ideally, we'll publish a prototype of this by the end of the day with the following goals: - quick iteration times: each VM should take less than 1min from spinning up to reporting results - simplicity: simple to understand, and hopefully simple(r) to debug - can be run locally, but also in CI. We would need to decide if we always want to run it in CI or only in certain occasions as it might slow things down (e.g. once a day, or just if bpf/ gets modified) ### Stack correctness tests (This will be something to tackle early next year) Running on the current machine, asserting that stacks are what we expect. Something similar to [this](https://github.com/javierhonduco/rbperf/blob/main/src/rbperf.rs#L597-L624). They can run on the host's kernel as we'll be testing that the programs can still be loaded in all the kernels we support with the VM testing infra cc/ @v-thakkar @Sylfrena @kakkoyun @brancz
1.0
Testing the Agent - ## The state of testing affairs At the moment we have: - **unit-tests**: to ensure the correctness of small well-contained pieces of code, such as our [Kernel symbolizer](https://github.com/parca-dev/parca-agent/blob/main/pkg/ksym/ksym_test.go). This helps us ensure that the behaviour is correct, but not less importantly has everything in place for us to add regression tests if we encounter bugs that we then fix; - **e2e tests**: running on Minikube that have the potential of testing our Kubernetes / systems metadata collection; These two are great and are very necessary. However, we have some blindspots, that are only going to get worse once we start shipping more advanced features, such as the [DWARF-based unwinder](https://github.com/parca-dev/parca-agent/issues/768). In my opinion, we are missing three main things: 1. A way to ensure that the **kernels we want to support work** (it's easy to introduce regressions, such as a feature that's not supported in a kernel we care about, or some code pattern that the verifier in some kernels rejects. These two are things I've spotted before, not fabricated examples) 1. Ensuring that resource usage, e.g. **memory locking etc is within what we expect** (we had to recently [remediate this](https://github.com/parca-dev/parca-agent/pull/1074)) 1. **Correctness testing**: on any kernel we support, running test programs and our unwinder and verifying that we get correct stack traces ## Propose additions ### VM testing For both 1) and - partially - 2) we need to add some testing in VMs with different kernels. Instead of running all of our integration tests (which may be flaky) or unit tests (which don't add much value in different kernels), I would like to propose the following for the kernels we support: - Can create the maps, so we ensure that our logic to create a map of size set at runtime is correct (note: we'll make this logic smarter to load as much as we can for a given host with a ceiling of 400-500MB); - Load our BPF program, to ensure that the verifier is happy with our code; This will also test that the BTF-pulling system for hosts that don't have it locally works, too (https://github.com/parca-dev/parca-agent/pull/1131). Ideally, we'll publish a prototype of this by the end of the day with the following goals: - quick iteration times: each VM should take less than 1min from spinning up to reporting results - simplicity: simple to understand, and hopefully simple(r) to debug - can be run locally, but also in CI. We would need to decide if we always want to run it in CI or only in certain occasions as it might slow things down (e.g. once a day, or just if bpf/ gets modified) ### Stack correctness tests (This will be something to tackle early next year) Running on the current machine, asserting that stacks are what we expect. Something similar to [this](https://github.com/javierhonduco/rbperf/blob/main/src/rbperf.rs#L597-L624). They can run on the host's kernel as we'll be testing that the programs can still be loaded in all the kernels we support with the VM testing infra cc/ @v-thakkar @Sylfrena @kakkoyun @brancz
test
testing the agent the state of testing affairs at the moment we have unit tests to ensure the correctness of small well contained pieces of code such as our this helps us ensure that the behaviour is correct but not less importantly has everything in place for us to add regression tests if we encounter bugs that we then fix tests running on minikube that have the potential of testing our kubernetes systems metadata collection these two are great and are very necessary however we have some blindspots that are only going to get worse once we start shipping more advanced features such as the in my opinion we are missing three main things a way to ensure that the kernels we want to support work it s easy to introduce regressions such as a feature that s not supported in a kernel we care about or some code pattern that the verifier in some kernels rejects these two are things i ve spotted before not fabricated examples ensuring that resource usage e g memory locking etc is within what we expect we had to recently correctness testing on any kernel we support running test programs and our unwinder and verifying that we get correct stack traces propose additions vm testing for both and partially we need to add some testing in vms with different kernels instead of running all of our integration tests which may be flaky or unit tests which don t add much value in different kernels i would like to propose the following for the kernels we support can create the maps so we ensure that our logic to create a map of size set at runtime is correct note we ll make this logic smarter to load as much as we can for a given host with a ceiling of load our bpf program to ensure that the verifier is happy with our code this will also test that the btf pulling system for hosts that don t have it locally works too ideally we ll publish a prototype of this by the end of the day with the following goals quick iteration times each vm should take less than from spinning up to reporting results simplicity simple to understand and hopefully simple r to debug can be run locally but also in ci we would need to decide if we always want to run it in ci or only in certain occasions as it might slow things down e g once a day or just if bpf gets modified stack correctness tests this will be something to tackle early next year running on the current machine asserting that stacks are what we expect something similar to they can run on the host s kernel as we ll be testing that the programs can still be loaded in all the kernels we support with the vm testing infra cc v thakkar sylfrena kakkoyun brancz
1
2,216
4,613,932,496
IssuesEvent
2016-09-25 09:08:09
OpenSRP/opensrp-client
https://api.github.com/repos/OpenSRP/opensrp-client
closed
handling cursor close events in register fragments
BANGLADESH enhancement FATAL Functional Requirement High Priority
currently cursor closing is not handled properly for cursor adaptors and hence this is required when exiting a register in order to save memory
1.0
handling cursor close events in register fragments - currently cursor closing is not handled properly for cursor adaptors and hence this is required when exiting a register in order to save memory
non_test
handling cursor close events in register fragments currently cursor closing is not handled properly for cursor adaptors and hence this is required when exiting a register in order to save memory
0
311,349
23,383,260,307
IssuesEvent
2022-08-11 11:32:59
alphagov/govuk-prototype-kit
https://api.github.com/repos/alphagov/govuk-prototype-kit
opened
Document how to show parts of a page or component based on data
documentation awaiting triage
From support: Need to know how to show parts of a page or component based on data, for example: ``` {% if data['over-18'] %} [ Over 18 content here ] {% else %} [ Under 18 content here ] {% endif %} ``` and ``` {{ govukRadios({ idPrefix: "where-do-you-live", name: "where-do-you-live", fieldset: { legend: { text: "Where do you live?", isPageHeading: true, classes: "govuk-fieldset__legend--l" } }, items: [ { value: "england", text: "England" }, { value: "scotland", text: "Scotland" }, { value: "wales", text: "Wales" }, { value: "northern-ireland", text: "Northern Ireland" } if (data['example'] == true) ] }) }} ``` with the key part being `if (data['example'] == true)` ## Related documentation Not sure we have any ## Suggestion None yet apart from covering the topics as above ## Evidence (where applicable) From multiple support requests
1.0
Document how to show parts of a page or component based on data - From support: Need to know how to show parts of a page or component based on data, for example: ``` {% if data['over-18'] %} [ Over 18 content here ] {% else %} [ Under 18 content here ] {% endif %} ``` and ``` {{ govukRadios({ idPrefix: "where-do-you-live", name: "where-do-you-live", fieldset: { legend: { text: "Where do you live?", isPageHeading: true, classes: "govuk-fieldset__legend--l" } }, items: [ { value: "england", text: "England" }, { value: "scotland", text: "Scotland" }, { value: "wales", text: "Wales" }, { value: "northern-ireland", text: "Northern Ireland" } if (data['example'] == true) ] }) }} ``` with the key part being `if (data['example'] == true)` ## Related documentation Not sure we have any ## Suggestion None yet apart from covering the topics as above ## Evidence (where applicable) From multiple support requests
non_test
document how to show parts of a page or component based on data from support need to know how to show parts of a page or component based on data for example if data else endif and govukradios idprefix where do you live name where do you live fieldset legend text where do you live ispageheading true classes govuk fieldset legend l items value england text england value scotland text scotland value wales text wales value northern ireland text northern ireland if data true with the key part being if data true related documentation not sure we have any suggestion none yet apart from covering the topics as above evidence where applicable from multiple support requests
0
248,844
7,936,846,663
IssuesEvent
2018-07-09 10:47:38
status-im/status-react
https://api.github.com/repos/status-im/status-react
closed
Implement User Profile screen
desktop feature high-priority
[comment]: # (Please replace ... with your information. Remove < and >) ### User Story As a user, I want to see nice profile screen with photo, options, etc ### Description *Type*: Feature *Summary*: implement user profile screen according to design zpl.io/25qZqB3 #### Expected behavior User profile looks like here: zpl.io/25qZqB3 #### Actual behavior ![Uploading Prof.png…]() ### Reproduction [comment]: # (Describe how we can replicate the bug step by step.) - Open Status - Create account - Go to "Profile" *Summary*: - [ ] Add profile icon - [ ] Left panel width should be 340px (fixed size), now is 280px. - [ ] User name should be centered in left panel (now is aligned to the left) ### Additional Information * Status version: [desktop build 29.06/2018](https://jenkins.status.im/job/status-react/job/desktop/job/manual/2/)
1.0
Implement User Profile screen - [comment]: # (Please replace ... with your information. Remove < and >) ### User Story As a user, I want to see nice profile screen with photo, options, etc ### Description *Type*: Feature *Summary*: implement user profile screen according to design zpl.io/25qZqB3 #### Expected behavior User profile looks like here: zpl.io/25qZqB3 #### Actual behavior ![Uploading Prof.png…]() ### Reproduction [comment]: # (Describe how we can replicate the bug step by step.) - Open Status - Create account - Go to "Profile" *Summary*: - [ ] Add profile icon - [ ] Left panel width should be 340px (fixed size), now is 280px. - [ ] User name should be centered in left panel (now is aligned to the left) ### Additional Information * Status version: [desktop build 29.06/2018](https://jenkins.status.im/job/status-react/job/desktop/job/manual/2/)
non_test
implement user profile screen please replace with your information remove user story as a user i want to see nice profile screen with photo options etc description type feature summary implement user profile screen according to design zpl io expected behavior user profile looks like here zpl io actual behavior reproduction describe how we can replicate the bug step by step open status create account go to profile summary add profile icon left panel width should be fixed size now is user name should be centered in left panel now is aligned to the left additional information status version
0
747,098
26,073,171,664
IssuesEvent
2022-12-24 04:35:17
tinkerbell/boots
https://api.github.com/repos/tinkerbell/boots
closed
Make use of params command optional for phone-home
kind/feature priority/backlog
<!--- Provide a general summary of the issue in the Title above --> Working through getting my physical servers working with tinkerbell to start doing some experimentation.. they fail saying the “params” command is not found. Leased from datacenter so flashing or updating ipxe might not be possible or ideal. ## Expected Behaviour <!--- If you're describing a bug, tell us what should happen --> <!--- If you're suggesting a change/improvement, tell us how it should work --> Boots ## Current Behaviour <!--- If describing a bug, tell us what happens instead of the expected behavior --> <!--- If suggesting a change/improvement, explain the difference from current behavior --> Errors out at params not found ## Possible Solution <!--- Not obligatory, but suggest a fix/reason for the bug, --> <!--- or ideas how to implement the addition or change --> Maybe hardware option? Or use query string ## Steps to Reproduce (for bugs) <!--- Provide a link to a live example, or an unambiguous set of steps to --> <!--- reproduce this bug. Include code to reproduce, if relevant --> 1. Run ipxe with out params built in 2. Try to boot ## Context <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most useful in the real world --> My solution was just to remove the params here: https://github.com/tinkerbell/boots/blob/master/ipxe/script.go#L34 Then just let it phone home even with out params. Seems like using ?body=${body}&type=${body} might could work. I’m also not even sure if needed? Looking at the phone home code seems like it’s mostly ignored? ## Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * Operating System and version (e.g. Linux, Windows, MacOS): * How are you running Tinkerbell? Using Vagrant & VirtualBox, Vagrant & Libvirt, on Packet using Terraform, or give details: tinkerbell it’s self kvm on the network * Link to your project or a code example to reproduce issue:
1.0
Make use of params command optional for phone-home - <!--- Provide a general summary of the issue in the Title above --> Working through getting my physical servers working with tinkerbell to start doing some experimentation.. they fail saying the “params” command is not found. Leased from datacenter so flashing or updating ipxe might not be possible or ideal. ## Expected Behaviour <!--- If you're describing a bug, tell us what should happen --> <!--- If you're suggesting a change/improvement, tell us how it should work --> Boots ## Current Behaviour <!--- If describing a bug, tell us what happens instead of the expected behavior --> <!--- If suggesting a change/improvement, explain the difference from current behavior --> Errors out at params not found ## Possible Solution <!--- Not obligatory, but suggest a fix/reason for the bug, --> <!--- or ideas how to implement the addition or change --> Maybe hardware option? Or use query string ## Steps to Reproduce (for bugs) <!--- Provide a link to a live example, or an unambiguous set of steps to --> <!--- reproduce this bug. Include code to reproduce, if relevant --> 1. Run ipxe with out params built in 2. Try to boot ## Context <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most useful in the real world --> My solution was just to remove the params here: https://github.com/tinkerbell/boots/blob/master/ipxe/script.go#L34 Then just let it phone home even with out params. Seems like using ?body=${body}&type=${body} might could work. I’m also not even sure if needed? Looking at the phone home code seems like it’s mostly ignored? ## Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * Operating System and version (e.g. Linux, Windows, MacOS): * How are you running Tinkerbell? Using Vagrant & VirtualBox, Vagrant & Libvirt, on Packet using Terraform, or give details: tinkerbell it’s self kvm on the network * Link to your project or a code example to reproduce issue:
non_test
make use of params command optional for phone home working through getting my physical servers working with tinkerbell to start doing some experimentation they fail saying the “params” command is not found leased from datacenter so flashing or updating ipxe might not be possible or ideal expected behaviour boots current behaviour errors out at params not found possible solution maybe hardware option or use query string steps to reproduce for bugs run ipxe with out params built in try to boot context my solution was just to remove the params here then just let it phone home even with out params seems like using body body type body might could work i’m also not even sure if needed looking at the phone home code seems like it’s mostly ignored your environment operating system and version e g linux windows macos how are you running tinkerbell using vagrant virtualbox vagrant libvirt on packet using terraform or give details tinkerbell it’s self kvm on the network link to your project or a code example to reproduce issue
0
43,372
11,209,803,538
IssuesEvent
2020-01-06 11:27:02
go-gitea/gitea
https://api.github.com/repos/go-gitea/gitea
opened
Use nfpm for package release
kind/build
In https://github.com/go-gitea/gitea/pull/6671, it was discussed to use nfpm via https://github.com/techknowlogick/drone-nfpm I don't think this has move forward. I just open this issue to track the progress and keep a track of this choice since the last issue is closed. I re-think of it because I stumble upon a good tools to track packages in various distrib. [![Packaging status](https://repology.org/badge/vertical-allrepos/gitea.svg)](https://repology.org/project/gitea/versions)
1.0
Use nfpm for package release - In https://github.com/go-gitea/gitea/pull/6671, it was discussed to use nfpm via https://github.com/techknowlogick/drone-nfpm I don't think this has move forward. I just open this issue to track the progress and keep a track of this choice since the last issue is closed. I re-think of it because I stumble upon a good tools to track packages in various distrib. [![Packaging status](https://repology.org/badge/vertical-allrepos/gitea.svg)](https://repology.org/project/gitea/versions)
non_test
use nfpm for package release in it was discussed to use nfpm via i don t think this has move forward i just open this issue to track the progress and keep a track of this choice since the last issue is closed i re think of it because i stumble upon a good tools to track packages in various distrib
0
113,554
11,808,508,387
IssuesEvent
2020-03-19 13:30:35
nestauk/covres
https://api.github.com/repos/nestauk/covres
opened
Exploratory data analysis
documentation eda
Exploratory analysis of various CORD-19 datasets. Includes: * Data structure (sp for full-text) * Variables * Missing values * Duplicates * Sources * Time coverage * Status of identifiers (researcher names and institutions) * Salient terms Produces a tidy dataset for additional enrichment and analysis.
1.0
Exploratory data analysis - Exploratory analysis of various CORD-19 datasets. Includes: * Data structure (sp for full-text) * Variables * Missing values * Duplicates * Sources * Time coverage * Status of identifiers (researcher names and institutions) * Salient terms Produces a tidy dataset for additional enrichment and analysis.
non_test
exploratory data analysis exploratory analysis of various cord datasets includes data structure sp for full text variables missing values duplicates sources time coverage status of identifiers researcher names and institutions salient terms produces a tidy dataset for additional enrichment and analysis
0
41,080
5,331,941,748
IssuesEvent
2017-02-15 20:47:23
UCHIC/iUTAHData
https://api.github.com/repos/UCHIC/iUTAHData
closed
Update Site because of transition to HydroShare (and other updates)
ReadyForTesting
Need to update: Homepage: - [x] Change text: https://docs.google.com/document/d/1m4E31ygXoFX7WWt_FTZ6jJESuaJcDhHXiApuTR1OrlU/edit Software Development: - [x] Remove the sentence under "Software Development. - [x] Change text: https://docs.google.com/document/d/12LYwc4OkIcGI_Zb4H6h0Lasvgcomxuy4kjmShHB5Vhw/edit Data Policy and Publication: - [x] Change text: https://docs.google.com/document/d/1zPIV21-NX9wSMgP_pXBJ8D1XRLeNtxh84U4cKmml4RE/edit - [x] Under the "Data" menu, change the "Dataset Repository" item to "iUTAH Data on HydroShare" and link to https://www.hydroshare.org/search/?q=&selected_facets=owners_names_exact:iUTAH%20Data%20Manager - [x] Under the "Data" menu, add an item "Survey Data" and link to data.iutahepscor.org/surveys - [x] Personnel: Remove Tony Castronova with (in two places). - [x] Training Materials: Remove the last two videos (Search & Discovery and Submittal & Publication). - [x] Documentation: Update text in several places with some new links, too: https://docs.google.com/document/d/10bjsKWS81JzLoFlJORltH_qQAzQcnLSooYEB3VAoMKc/edit#
1.0
Update Site because of transition to HydroShare (and other updates) - Need to update: Homepage: - [x] Change text: https://docs.google.com/document/d/1m4E31ygXoFX7WWt_FTZ6jJESuaJcDhHXiApuTR1OrlU/edit Software Development: - [x] Remove the sentence under "Software Development. - [x] Change text: https://docs.google.com/document/d/12LYwc4OkIcGI_Zb4H6h0Lasvgcomxuy4kjmShHB5Vhw/edit Data Policy and Publication: - [x] Change text: https://docs.google.com/document/d/1zPIV21-NX9wSMgP_pXBJ8D1XRLeNtxh84U4cKmml4RE/edit - [x] Under the "Data" menu, change the "Dataset Repository" item to "iUTAH Data on HydroShare" and link to https://www.hydroshare.org/search/?q=&selected_facets=owners_names_exact:iUTAH%20Data%20Manager - [x] Under the "Data" menu, add an item "Survey Data" and link to data.iutahepscor.org/surveys - [x] Personnel: Remove Tony Castronova with (in two places). - [x] Training Materials: Remove the last two videos (Search & Discovery and Submittal & Publication). - [x] Documentation: Update text in several places with some new links, too: https://docs.google.com/document/d/10bjsKWS81JzLoFlJORltH_qQAzQcnLSooYEB3VAoMKc/edit#
test
update site because of transition to hydroshare and other updates need to update homepage change text software development remove the sentence under software development change text data policy and publication change text under the data menu change the dataset repository item to iutah data on hydroshare and link to under the data menu add an item survey data and link to data iutahepscor org surveys personnel remove tony castronova with in two places training materials remove the last two videos search discovery and submittal publication documentation update text in several places with some new links too
1
130,565
27,722,257,181
IssuesEvent
2023-03-14 21:43:39
Rothamsted/knetminer
https://api.github.com/repos/Rothamsted/knetminer
opened
I the /qtl API still used? If yes, it needs review
code review code quality
I can't see any call in the UI/client to `/qtl`. On the server, it does almost the same things that `/genome` does, worse its implementation is conflated with `/genome`, within `OndexLocalDataSource.handleMainSearch()`. Is this API used, within the UI or externally? If not, it's better to clean-up things by removing it (including `QtlResponse` and the conditional code which is run when the API itself is invoked. I guess @KeywanHP is the best to know it.
2.0
I the /qtl API still used? If yes, it needs review - I can't see any call in the UI/client to `/qtl`. On the server, it does almost the same things that `/genome` does, worse its implementation is conflated with `/genome`, within `OndexLocalDataSource.handleMainSearch()`. Is this API used, within the UI or externally? If not, it's better to clean-up things by removing it (including `QtlResponse` and the conditional code which is run when the API itself is invoked. I guess @KeywanHP is the best to know it.
non_test
i the qtl api still used if yes it needs review i can t see any call in the ui client to qtl on the server it does almost the same things that genome does worse its implementation is conflated with genome within ondexlocaldatasource handlemainsearch is this api used within the ui or externally if not it s better to clean up things by removing it including qtlresponse and the conditional code which is run when the api itself is invoked i guess keywanhp is the best to know it
0
271,423
29,498,277,336
IssuesEvent
2023-06-02 19:01:55
bcgov/rsbc-digital-forms
https://api.github.com/repos/bcgov/rsbc-digital-forms
closed
Critical vulnerabilities found in rsbc-roadside-forms-backend-api on PR 60
security
# Report ## Results - **[ERROR]** **[CVE-2022-29361]** [[HELP LINK](https://avd.aquasec.com/nvd/cve-2022-29361)] `Package: Werkzeug Installed Version: 2.0.2 Vulnerability CVE-2022-29361 Severity: CRITICAL Fixed Version: 2.1.1 Link: \[CVE-2022-29361\]\(https:\/\/avd.aquasec.com\/nvd\/cve-2022-29361\)` - https://github.com/bcgov/rsbc-digital-forms/blob/feature/dockerscanv2/opt/app-root/lib/python3.6/site-packages/Werkzeug-2.0.2.dist-info/METADATA#L1-L1 ## Suppressed Results Nothing here. ## Rules information <!-- Rule Info --> <details><summary>Rules details</summary> - CVE-2022-29361 [undefined] > CVE-2022-29361 </details> ## Tool information - Name: Trivy - Organization: undefined - Version: undefined
True
Critical vulnerabilities found in rsbc-roadside-forms-backend-api on PR 60 - # Report ## Results - **[ERROR]** **[CVE-2022-29361]** [[HELP LINK](https://avd.aquasec.com/nvd/cve-2022-29361)] `Package: Werkzeug Installed Version: 2.0.2 Vulnerability CVE-2022-29361 Severity: CRITICAL Fixed Version: 2.1.1 Link: \[CVE-2022-29361\]\(https:\/\/avd.aquasec.com\/nvd\/cve-2022-29361\)` - https://github.com/bcgov/rsbc-digital-forms/blob/feature/dockerscanv2/opt/app-root/lib/python3.6/site-packages/Werkzeug-2.0.2.dist-info/METADATA#L1-L1 ## Suppressed Results Nothing here. ## Rules information <!-- Rule Info --> <details><summary>Rules details</summary> - CVE-2022-29361 [undefined] > CVE-2022-29361 </details> ## Tool information - Name: Trivy - Organization: undefined - Version: undefined
non_test
critical vulnerabilities found in rsbc roadside forms backend api on pr report results package werkzeug installed version vulnerability cve severity critical fixed version link https avd aquasec com nvd cve suppressed results nothing here rules information rules details cve cve tool information name trivy organization undefined version undefined
0
92,887
26,794,706,592
IssuesEvent
2023-02-01 11:01:10
zeek/zeek
https://api.github.com/repos/zeek/zeek
closed
libspicy and libhilti should build as shared objects
Area: Build/Platform
They're being built as static libraries now which means that the size of the following binaries is needlessly large... ``` -rwxr-xr-x 1 root root 439337760 Jan 6 05:33 hilti-config -rwxr-xr-x 1 root root 442081168 Jan 6 05:33 hiltic -rwxr-xr-x 1 root root 725743872 Jan 6 05:36 spicy-config -rwxr-xr-x 1 root root 728677696 Jan 6 05:36 spicyc -rwxr-xr-x 1 root root 728891008 Jan 6 05:36 spicy-driver -rwxr-xr-x 1 root root 731454632 Jan 6 05:36 spicy-dump -rwxr-xr-x 1 root root 751046864 Jan 6 05:39 spicyz -rwxr-xr-x 1 root root 1048532000 Jan 6 05:40 zeek ``` Other than zeek, those binaries should all be very small but with a default build they're huge right now. Even if you strip them they're rather large due to the large amount of code that is in libhilti and libspicy.
1.0
libspicy and libhilti should build as shared objects - They're being built as static libraries now which means that the size of the following binaries is needlessly large... ``` -rwxr-xr-x 1 root root 439337760 Jan 6 05:33 hilti-config -rwxr-xr-x 1 root root 442081168 Jan 6 05:33 hiltic -rwxr-xr-x 1 root root 725743872 Jan 6 05:36 spicy-config -rwxr-xr-x 1 root root 728677696 Jan 6 05:36 spicyc -rwxr-xr-x 1 root root 728891008 Jan 6 05:36 spicy-driver -rwxr-xr-x 1 root root 731454632 Jan 6 05:36 spicy-dump -rwxr-xr-x 1 root root 751046864 Jan 6 05:39 spicyz -rwxr-xr-x 1 root root 1048532000 Jan 6 05:40 zeek ``` Other than zeek, those binaries should all be very small but with a default build they're huge right now. Even if you strip them they're rather large due to the large amount of code that is in libhilti and libspicy.
non_test
libspicy and libhilti should build as shared objects they re being built as static libraries now which means that the size of the following binaries is needlessly large rwxr xr x root root jan hilti config rwxr xr x root root jan hiltic rwxr xr x root root jan spicy config rwxr xr x root root jan spicyc rwxr xr x root root jan spicy driver rwxr xr x root root jan spicy dump rwxr xr x root root jan spicyz rwxr xr x root root jan zeek other than zeek those binaries should all be very small but with a default build they re huge right now even if you strip them they re rather large due to the large amount of code that is in libhilti and libspicy
0
164,272
12,795,271,185
IssuesEvent
2020-07-02 08:28:41
rancher/rke2
https://api.github.com/repos/rancher/rke2
closed
nginx-ingress-controller service is in pending state
[zube]: To Test
Version: rke2 v0.0.1-alpha.4 Issue: nginx-ingress-controller service is in pending state. Since we dont have servicelb it should not expected to be of type LoadBalancer ``` kube-system nginx-ingress-controller LoadBalancer 10.43.5.161 <pending> 80:30782/TCP,443:30488/TCP 4h32m ```
1.0
nginx-ingress-controller service is in pending state - Version: rke2 v0.0.1-alpha.4 Issue: nginx-ingress-controller service is in pending state. Since we dont have servicelb it should not expected to be of type LoadBalancer ``` kube-system nginx-ingress-controller LoadBalancer 10.43.5.161 <pending> 80:30782/TCP,443:30488/TCP 4h32m ```
test
nginx ingress controller service is in pending state version alpha issue nginx ingress controller service is in pending state since we dont have servicelb it should not expected to be of type loadbalancer kube system nginx ingress controller loadbalancer tcp tcp
1
342,312
30,613,395,184
IssuesEvent
2023-07-23 22:14:28
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
pkg/sql/schemachanger/schemachanger_test: TestValidateMixedVersionElements_drop_column_computed_index failed
C-test-failure O-robot branch-master T-sql-foundations
pkg/sql/schemachanger/schemachanger_test.TestValidateMixedVersionElements_drop_column_computed_index [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/11029922?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/11029922?buildTab=artifacts#/) on master @ [d293bb3acd9b7f67ac793ddb6a0836e637864629](https://github.com/cockroachdb/cockroach/commits/d293bb3acd9b7f67ac793ddb6a0836e637864629): ``` panic.go:522: -- test log scope end -- test logs left over in: /artifacts/tmp/_tmp/2d9af299258c11d4d3ade298f6c5b40e/logTestValidateMixedVersionElements_drop_column_computed_index1756802228 --- FAIL: TestValidateMixedVersionElements_drop_column_computed_index (249.43s) === RUN TestValidateMixedVersionElements_drop_column_computed_index/pause_upgrade_and_resume_at_each_stage --- FAIL: TestValidateMixedVersionElements_drop_column_computed_index/pause_upgrade_and_resume_at_each_stage (248.42s) === RUN TestValidateMixedVersionElements_drop_column_computed_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3 sql_runner.go:104: Error Trace: github.com/cockroachdb/cockroach/pkg/testutils/sqlutils/sql_runner.go:117 github.com/cockroachdb/cockroach/pkg/testutils/sqlutils/sql_runner.go:312 github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/end_to_end.go:494 github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1629 github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1730 Error: Received unexpected error: query 'SELECT status, job_type, description FROM [SHOW JOBS] WHERE job_type IN ('SCHEMA CHANGE', 'TYPEDESC SCHEMA CHANGE', 'NEW SCHEMA CHANGE') AND status NOT IN ('succeeded', 'failed')': expected: (1) attached stack trace -- stack trace: | github.com/cockroachdb/cockroach/pkg/testutils/sqlutils.(*SQLRunner).CheckQueryResultsRetry.func1 | github.com/cockroachdb/cockroach/pkg/testutils/sqlutils/sql_runner.go:315 | github.com/cockroachdb/cockroach/pkg/testutils.SucceedsWithinError.func1 | github.com/cockroachdb/cockroach/pkg/testutils/soon.go:75 | github.com/cockroachdb/cockroach/pkg/util/retry.ForDuration | github.com/cockroachdb/cockroach/pkg/util/retry/retry.go:213 | github.com/cockroachdb/cockroach/pkg/testutils.SucceedsWithinError | github.com/cockroachdb/cockroach/pkg/testutils/soon.go:81 | github.com/cockroachdb/cockroach/pkg/testutils/sqlutils.(*SQLRunner).succeedsWithin | github.com/cockroachdb/cockroach/pkg/testutils/sqlutils/sql_runner.go:117 | github.com/cockroachdb/cockroach/pkg/testutils/sqlutils.(*SQLRunner).CheckQueryResultsRetry | github.com/cockroachdb/cockroach/pkg/testutils/sqlutils/sql_runner.go:312 | github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest.waitForSchemaChangesToFinish | github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/end_to_end.go:494 | github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest.executeSchemaChangeTxn | github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1629 | github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest.ValidateMixedVersionElements.func5.1 | github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1730 | testing.tRunner | GOROOT/src/testing/testing.go:1446 | runtime.goexit | GOROOT/src/runtime/asm_arm64.s:1172 Wraps: (2) query 'SELECT status, job_type, description FROM [SHOW JOBS] WHERE job_type IN ('SCHEMA CHANGE', 'TYPEDESC SCHEMA CHANGE', 'NEW SCHEMA CHANGE') AND status NOT IN ('succeeded', 'failed')': expected: | | got: | paused, NEW SCHEMA CHANGE, ALTER TABLE defaultdb.public.t DROP COLUMN j CASCADE Error types: (1) *withstack.withStack (2) *errutil.leafError cumulative.go:1725: Error Trace: github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1725 GOROOT/src/runtime/asm_arm64.s:1172 Error: Received unexpected error: dial tcp 127.0.0.1:38973: connect: connection refused Test: TestValidateMixedVersionElements_drop_column_computed_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3 --- FAIL: TestValidateMixedVersionElements_drop_column_computed_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3 (51.57s) ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/sql-foundations <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestValidateMixedVersionElements_drop_column_computed_index.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
1.0
pkg/sql/schemachanger/schemachanger_test: TestValidateMixedVersionElements_drop_column_computed_index failed - pkg/sql/schemachanger/schemachanger_test.TestValidateMixedVersionElements_drop_column_computed_index [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/11029922?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/11029922?buildTab=artifacts#/) on master @ [d293bb3acd9b7f67ac793ddb6a0836e637864629](https://github.com/cockroachdb/cockroach/commits/d293bb3acd9b7f67ac793ddb6a0836e637864629): ``` panic.go:522: -- test log scope end -- test logs left over in: /artifacts/tmp/_tmp/2d9af299258c11d4d3ade298f6c5b40e/logTestValidateMixedVersionElements_drop_column_computed_index1756802228 --- FAIL: TestValidateMixedVersionElements_drop_column_computed_index (249.43s) === RUN TestValidateMixedVersionElements_drop_column_computed_index/pause_upgrade_and_resume_at_each_stage --- FAIL: TestValidateMixedVersionElements_drop_column_computed_index/pause_upgrade_and_resume_at_each_stage (248.42s) === RUN TestValidateMixedVersionElements_drop_column_computed_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3 sql_runner.go:104: Error Trace: github.com/cockroachdb/cockroach/pkg/testutils/sqlutils/sql_runner.go:117 github.com/cockroachdb/cockroach/pkg/testutils/sqlutils/sql_runner.go:312 github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/end_to_end.go:494 github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1629 github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1730 Error: Received unexpected error: query 'SELECT status, job_type, description FROM [SHOW JOBS] WHERE job_type IN ('SCHEMA CHANGE', 'TYPEDESC SCHEMA CHANGE', 'NEW SCHEMA CHANGE') AND status NOT IN ('succeeded', 'failed')': expected: (1) attached stack trace -- stack trace: | github.com/cockroachdb/cockroach/pkg/testutils/sqlutils.(*SQLRunner).CheckQueryResultsRetry.func1 | github.com/cockroachdb/cockroach/pkg/testutils/sqlutils/sql_runner.go:315 | github.com/cockroachdb/cockroach/pkg/testutils.SucceedsWithinError.func1 | github.com/cockroachdb/cockroach/pkg/testutils/soon.go:75 | github.com/cockroachdb/cockroach/pkg/util/retry.ForDuration | github.com/cockroachdb/cockroach/pkg/util/retry/retry.go:213 | github.com/cockroachdb/cockroach/pkg/testutils.SucceedsWithinError | github.com/cockroachdb/cockroach/pkg/testutils/soon.go:81 | github.com/cockroachdb/cockroach/pkg/testutils/sqlutils.(*SQLRunner).succeedsWithin | github.com/cockroachdb/cockroach/pkg/testutils/sqlutils/sql_runner.go:117 | github.com/cockroachdb/cockroach/pkg/testutils/sqlutils.(*SQLRunner).CheckQueryResultsRetry | github.com/cockroachdb/cockroach/pkg/testutils/sqlutils/sql_runner.go:312 | github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest.waitForSchemaChangesToFinish | github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/end_to_end.go:494 | github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest.executeSchemaChangeTxn | github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1629 | github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest.ValidateMixedVersionElements.func5.1 | github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1730 | testing.tRunner | GOROOT/src/testing/testing.go:1446 | runtime.goexit | GOROOT/src/runtime/asm_arm64.s:1172 Wraps: (2) query 'SELECT status, job_type, description FROM [SHOW JOBS] WHERE job_type IN ('SCHEMA CHANGE', 'TYPEDESC SCHEMA CHANGE', 'NEW SCHEMA CHANGE') AND status NOT IN ('succeeded', 'failed')': expected: | | got: | paused, NEW SCHEMA CHANGE, ALTER TABLE defaultdb.public.t DROP COLUMN j CASCADE Error types: (1) *withstack.withStack (2) *errutil.leafError cumulative.go:1725: Error Trace: github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1725 GOROOT/src/runtime/asm_arm64.s:1172 Error: Received unexpected error: dial tcp 127.0.0.1:38973: connect: connection refused Test: TestValidateMixedVersionElements_drop_column_computed_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3 --- FAIL: TestValidateMixedVersionElements_drop_column_computed_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3 (51.57s) ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/sql-foundations <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestValidateMixedVersionElements_drop_column_computed_index.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
pkg sql schemachanger schemachanger test testvalidatemixedversionelements drop column computed index failed pkg sql schemachanger schemachanger test testvalidatemixedversionelements drop column computed index with on master panic go test log scope end test logs left over in artifacts tmp tmp logtestvalidatemixedversionelements drop column computed fail testvalidatemixedversionelements drop column computed index run testvalidatemixedversionelements drop column computed index pause upgrade and resume at each stage fail testvalidatemixedversionelements drop column computed index pause upgrade and resume at each stage run testvalidatemixedversionelements drop column computed index pause upgrade and resume at each stage postcommitnonrevertiblephase of sql runner go error trace github com cockroachdb cockroach pkg testutils sqlutils sql runner go github com cockroachdb cockroach pkg testutils sqlutils sql runner go github com cockroachdb cockroach pkg sql schemachanger sctest end to end go github com cockroachdb cockroach pkg sql schemachanger sctest cumulative go github com cockroachdb cockroach pkg sql schemachanger sctest cumulative go error received unexpected error query select status job type description from where job type in schema change typedesc schema change new schema change and status not in succeeded failed expected attached stack trace stack trace github com cockroachdb cockroach pkg testutils sqlutils sqlrunner checkqueryresultsretry github com cockroachdb cockroach pkg testutils sqlutils sql runner go github com cockroachdb cockroach pkg testutils succeedswithinerror github com cockroachdb cockroach pkg testutils soon go github com cockroachdb cockroach pkg util retry forduration github com cockroachdb cockroach pkg util retry retry go github com cockroachdb cockroach pkg testutils succeedswithinerror github com cockroachdb cockroach pkg testutils soon go github com cockroachdb cockroach pkg testutils sqlutils sqlrunner succeedswithin github com cockroachdb cockroach pkg testutils sqlutils sql runner go github com cockroachdb cockroach pkg testutils sqlutils sqlrunner checkqueryresultsretry github com cockroachdb cockroach pkg testutils sqlutils sql runner go github com cockroachdb cockroach pkg sql schemachanger sctest waitforschemachangestofinish github com cockroachdb cockroach pkg sql schemachanger sctest end to end go github com cockroachdb cockroach pkg sql schemachanger sctest executeschemachangetxn github com cockroachdb cockroach pkg sql schemachanger sctest cumulative go github com cockroachdb cockroach pkg sql schemachanger sctest validatemixedversionelements github com cockroachdb cockroach pkg sql schemachanger sctest cumulative go testing trunner goroot src testing testing go runtime goexit goroot src runtime asm s wraps query select status job type description from where job type in schema change typedesc schema change new schema change and status not in succeeded failed expected got paused new schema change alter table defaultdb public t drop column j cascade error types withstack withstack errutil leaferror cumulative go error trace github com cockroachdb cockroach pkg sql schemachanger sctest cumulative go goroot src runtime asm s error received unexpected error dial tcp connect connection refused test testvalidatemixedversionelements drop column computed index pause upgrade and resume at each stage postcommitnonrevertiblephase of fail testvalidatemixedversionelements drop column computed index pause upgrade and resume at each stage postcommitnonrevertiblephase of help see also cc cockroachdb sql foundations
1