Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 4
112
| repo_url
stringlengths 33
141
| action
stringclasses 3
values | title
stringlengths 1
1.02k
| labels
stringlengths 4
1.54k
| body
stringlengths 1
262k
| index
stringclasses 17
values | text_combine
stringlengths 95
262k
| label
stringclasses 2
values | text
stringlengths 96
252k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
98,162
| 12,300,036,481
|
IssuesEvent
|
2020-05-11 13:23:35
|
Opentrons/opentrons
|
https://api.github.com/repos/Opentrons/opentrons
|
closed
|
[PD] Thermocycler Graph of Profile
|
WIP design protocol designer
|
# User Story
As a user, I should be able to reference a standard graph/visualization of the profile I’ve created which shows the profile’s temperature, time spans, and cycles.
# UI Design
_In progress_
# Acceptance Criteria
- [ ] The graph is made up of equal width/height blocks. These blocks contain:
- [ ] A horizontal line, the height of which represents temperature
- [ ] A temperature
- [ ] A duration
- [ ] Blocks that are together in a cycle have an arrow underneath with the cycle # in it. These blocks also have a different fill than individual blocks.
- [ ] The horizontal lines in each block are connected to their neighbors by a line that spans the gutter between blocks
|
2.0
|
[PD] Thermocycler Graph of Profile - # User Story
As a user, I should be able to reference a standard graph/visualization of the profile I’ve created which shows the profile’s temperature, time spans, and cycles.
# UI Design
_In progress_
# Acceptance Criteria
- [ ] The graph is made up of equal width/height blocks. These blocks contain:
- [ ] A horizontal line, the height of which represents temperature
- [ ] A temperature
- [ ] A duration
- [ ] Blocks that are together in a cycle have an arrow underneath with the cycle # in it. These blocks also have a different fill than individual blocks.
- [ ] The horizontal lines in each block are connected to their neighbors by a line that spans the gutter between blocks
|
non_test
|
thermocycler graph of profile user story as a user i should be able to reference a standard graph visualization of the profile i’ve created which shows the profile’s temperature time spans and cycles ui design in progress acceptance criteria the graph is made up of equal width height blocks these blocks contain a horizontal line the height of which represents temperature a temperature a duration blocks that are together in a cycle have an arrow underneath with the cycle in it these blocks also have a different fill than individual blocks the horizontal lines in each block are connected to their neighbors by a line that spans the gutter between blocks
| 0
|
346,012
| 10,383,016,236
|
IssuesEvent
|
2019-09-10 08:46:54
|
red-hat-storage/ocs-ci
|
https://api.github.com/repos/red-hat-storage/ocs-ci
|
opened
|
ocs_operator_storage_cluster_cr points to nonexisting url
|
High Priority bug
|
Installation of upstream version via operator is failing on following error:
```
E AssertionError: Couldn't load URL: https://raw.githubusercontent.com/openshift/ocs-operator/master/deploy/crds/ocs_v1alpha1_storagecluster_cr.yaml content! Status: 404.
```
The URL is set in default config [ocs_ci/framework/conf/default_config.yaml](https://github.com/red-hat-storage/ocs-ci/blob/master/ocs_ci/framework/conf/default_config.yaml#L44).
The original file does not exists any more, but it seems to be just renamed to `ocs_v1_storagecluster_cr.yaml`:
https://github.com/openshift/ocs-operator/tree/master/deploy/crds
|
1.0
|
ocs_operator_storage_cluster_cr points to nonexisting url - Installation of upstream version via operator is failing on following error:
```
E AssertionError: Couldn't load URL: https://raw.githubusercontent.com/openshift/ocs-operator/master/deploy/crds/ocs_v1alpha1_storagecluster_cr.yaml content! Status: 404.
```
The URL is set in default config [ocs_ci/framework/conf/default_config.yaml](https://github.com/red-hat-storage/ocs-ci/blob/master/ocs_ci/framework/conf/default_config.yaml#L44).
The original file does not exists any more, but it seems to be just renamed to `ocs_v1_storagecluster_cr.yaml`:
https://github.com/openshift/ocs-operator/tree/master/deploy/crds
|
non_test
|
ocs operator storage cluster cr points to nonexisting url installation of upstream version via operator is failing on following error e assertionerror couldn t load url content status the url is set in default config the original file does not exists any more but it seems to be just renamed to ocs storagecluster cr yaml
| 0
|
30,983
| 6,385,767,674
|
IssuesEvent
|
2017-08-03 09:23:27
|
bridgedotnet/Bridge
|
https://api.github.com/repos/bridgedotnet/Bridge
|
closed
|
Cannot cast null to nullable
|
defect
|
Cannot cast null to nullable.
### Steps To Reproduce
https://dev.deck.net/16c8630e00f0554844f810648e73f9e9/third
https://dotnetfiddle.net/FeYHFH
```c#
public class Program
{
public static void Main()
{
Console.WriteLine((((object)null) as Int64?).HasValue ? "Failed" : "Passed");
}
}
```
|
1.0
|
Cannot cast null to nullable - Cannot cast null to nullable.
### Steps To Reproduce
https://dev.deck.net/16c8630e00f0554844f810648e73f9e9/third
https://dotnetfiddle.net/FeYHFH
```c#
public class Program
{
public static void Main()
{
Console.WriteLine((((object)null) as Int64?).HasValue ? "Failed" : "Passed");
}
}
```
|
non_test
|
cannot cast null to nullable cannot cast null to nullable steps to reproduce c public class program public static void main console writeline object null as hasvalue failed passed
| 0
|
821,359
| 30,819,127,865
|
IssuesEvent
|
2023-08-01 15:11:40
|
opendatahub-io/data-science-pipelines-operator
|
https://api.github.com/repos/opendatahub-io/data-science-pipelines-operator
|
opened
|
Use openshift-goimports to sort go importants
|
triage/accepted priority/normal
|
### Feature description
We should follow openshift best practices when organizing our go imports, we should utilize [openshift go imports for this](https://github.com/openshift-eng/openshift-goimports).
Acceptance criteria:
* Figure out how best to include this as part of development workflow, (can this be part of pre-commit?)
* add a gh action to verify that PRs adhere to the imports via openshift go imports.
* Also add a documentation to dspo readme on how to develop/run go imports when developing.
### Describe alternatives you've considered
_No response_
### Anything else?
_No response_
|
1.0
|
Use openshift-goimports to sort go importants - ### Feature description
We should follow openshift best practices when organizing our go imports, we should utilize [openshift go imports for this](https://github.com/openshift-eng/openshift-goimports).
Acceptance criteria:
* Figure out how best to include this as part of development workflow, (can this be part of pre-commit?)
* add a gh action to verify that PRs adhere to the imports via openshift go imports.
* Also add a documentation to dspo readme on how to develop/run go imports when developing.
### Describe alternatives you've considered
_No response_
### Anything else?
_No response_
|
non_test
|
use openshift goimports to sort go importants feature description we should follow openshift best practices when organizing our go imports we should utilize acceptance criteria figure out how best to include this as part of development workflow can this be part of pre commit add a gh action to verify that prs adhere to the imports via openshift go imports also add a documentation to dspo readme on how to develop run go imports when developing describe alternatives you ve considered no response anything else no response
| 0
|
304,486
| 9,332,750,125
|
IssuesEvent
|
2019-03-28 12:59:57
|
mlibrary/heliotrope
|
https://api.github.com/repos/mlibrary/heliotrope
|
closed
|
Vanilla Fulcrum
|
EPIC low priority refactor systems
|
Separate Fulcrum from Heliotrope such that following the setup instructions in the README.md creates a generic 'Vanilla Fulcrum' application.
|
1.0
|
Vanilla Fulcrum - Separate Fulcrum from Heliotrope such that following the setup instructions in the README.md creates a generic 'Vanilla Fulcrum' application.
|
non_test
|
vanilla fulcrum separate fulcrum from heliotrope such that following the setup instructions in the readme md creates a generic vanilla fulcrum application
| 0
|
68,508
| 3,288,906,608
|
IssuesEvent
|
2015-10-29 16:51:42
|
INN/Largo
|
https://api.github.com/repos/INN/Largo
|
closed
|
Pulling .rst function documentation out of .php
|
priority: low type: question
|
#### What's currently working.
Using __[`doxphp`](https://github.com/avalanche123/doxphp)__ (a php phar), it's possible to generate a `.json` representation of documentation directly from our `.php` files.
From there, we can generate `*.rst` using the __`doxphp2sphinx`__ renderer supplied in the same phar.
$ doxphp < functions.php | doxphp2sphinx > functions.rst
As a proof of concept, this is currently in our sphinx Makefile for a few select files [here](https://github.com/INN/Largo/blob/develop/docs/Makefile#L179-L185). With `make php`, It will generate an `.rst` files for each specified `.php` file. With `make html`, those files will compile to `.html` [like this](http://largo.readthedocs.org/api/inc/helpers.html). It requires you to have **`doxphp`** installed.
**tl;dr:** `.php > .json > .rst > .html`
We could modify this pipeline, writing either our own `.php > .json` parser (seems unnecessary) or sphinx extension to render the included `.json > .rst` in a different way. In theory, this shouldn't be too hard.
#### Questions.
1. Is organizing function documentation like this the best practice? (i.e is the way we organize our code is the best way to organize our function documentation?)
2. Not all files have functions that need documentation. I'd assume most of what should be included resides in the `./inc/` folder, but is there anything else that should be included? Is there anything in `./inc/` that shouldn't be included?
#### Others.
What others are doing
* [WordPress Codex](http://codex.wordpress.org/): Organizes function reference one function per page. I tend to like this format and think our users would be familiar with it. It also would allow us to add examples and longer form documentation for those functions that need it.
* [WooThemes](http://docs.woothemes.com/): Seems to have their documentation all over the place.
- Their **WooCommerce** plugin uses [APIgen](http://www.apigen.org/) to generate documentation for their WooCommerce plugin. and keeps it separate from
- **WooCodex** seems to mirror the structure of WordPress for some of their more [commonly used](http://docs.woothemes.com/documentation/woocodex/) functions ([like this](http://docs.woothemes.com/document/woocommerce_breadcrumb/)).
* Others?
|
1.0
|
Pulling .rst function documentation out of .php - #### What's currently working.
Using __[`doxphp`](https://github.com/avalanche123/doxphp)__ (a php phar), it's possible to generate a `.json` representation of documentation directly from our `.php` files.
From there, we can generate `*.rst` using the __`doxphp2sphinx`__ renderer supplied in the same phar.
$ doxphp < functions.php | doxphp2sphinx > functions.rst
As a proof of concept, this is currently in our sphinx Makefile for a few select files [here](https://github.com/INN/Largo/blob/develop/docs/Makefile#L179-L185). With `make php`, It will generate an `.rst` files for each specified `.php` file. With `make html`, those files will compile to `.html` [like this](http://largo.readthedocs.org/api/inc/helpers.html). It requires you to have **`doxphp`** installed.
**tl;dr:** `.php > .json > .rst > .html`
We could modify this pipeline, writing either our own `.php > .json` parser (seems unnecessary) or sphinx extension to render the included `.json > .rst` in a different way. In theory, this shouldn't be too hard.
#### Questions.
1. Is organizing function documentation like this the best practice? (i.e is the way we organize our code is the best way to organize our function documentation?)
2. Not all files have functions that need documentation. I'd assume most of what should be included resides in the `./inc/` folder, but is there anything else that should be included? Is there anything in `./inc/` that shouldn't be included?
#### Others.
What others are doing
* [WordPress Codex](http://codex.wordpress.org/): Organizes function reference one function per page. I tend to like this format and think our users would be familiar with it. It also would allow us to add examples and longer form documentation for those functions that need it.
* [WooThemes](http://docs.woothemes.com/): Seems to have their documentation all over the place.
- Their **WooCommerce** plugin uses [APIgen](http://www.apigen.org/) to generate documentation for their WooCommerce plugin. and keeps it separate from
- **WooCodex** seems to mirror the structure of WordPress for some of their more [commonly used](http://docs.woothemes.com/documentation/woocodex/) functions ([like this](http://docs.woothemes.com/document/woocommerce_breadcrumb/)).
* Others?
|
non_test
|
pulling rst function documentation out of php what s currently working using a php phar it s possible to generate a json representation of documentation directly from our php files from there we can generate rst using the renderer supplied in the same phar doxphp functions rst as a proof of concept this is currently in our sphinx makefile for a few select files with make php it will generate an rst files for each specified php file with make html those files will compile to html it requires you to have doxphp installed tl dr php json rst html we could modify this pipeline writing either our own php json parser seems unnecessary or sphinx extension to render the included json rst in a different way in theory this shouldn t be too hard questions is organizing function documentation like this the best practice i e is the way we organize our code is the best way to organize our function documentation not all files have functions that need documentation i d assume most of what should be included resides in the inc folder but is there anything else that should be included is there anything in inc that shouldn t be included others what others are doing organizes function reference one function per page i tend to like this format and think our users would be familiar with it it also would allow us to add examples and longer form documentation for those functions that need it seems to have their documentation all over the place their woocommerce plugin uses to generate documentation for their woocommerce plugin and keeps it separate from woocodex seems to mirror the structure of wordpress for some of their more functions others
| 0
|
170,442
| 6,444,659,752
|
IssuesEvent
|
2017-08-12 15:16:17
|
drckf/paysage
|
https://api.github.com/repos/drckf/paysage
|
closed
|
Dropout RBMs
|
Priority: Medium
|
Dropout RBMs are discussed in the original dropout paper with results superior to normal RBMs.
|
1.0
|
Dropout RBMs - Dropout RBMs are discussed in the original dropout paper with results superior to normal RBMs.
|
non_test
|
dropout rbms dropout rbms are discussed in the original dropout paper with results superior to normal rbms
| 0
|
329,055
| 28,146,631,717
|
IssuesEvent
|
2023-04-02 15:03:16
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
closed
|
Fix elementwise.test_count_nonzero
|
Sub Task Ivy API Experimental Failing Test
|
| | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4588728136/jobs/8103151468" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4588728136/jobs/8103151468" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4588728136/jobs/8103151468" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4588728136/jobs/8103151468" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|
1.0
|
Fix elementwise.test_count_nonzero - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4588728136/jobs/8103151468" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4588728136/jobs/8103151468" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4588728136/jobs/8103151468" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4588728136/jobs/8103151468" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|
test
|
fix elementwise test count nonzero tensorflow img src torch img src numpy img src jax img src
| 1
|
613,759
| 19,097,793,857
|
IssuesEvent
|
2021-11-29 18:36:20
|
ballerina-platform/ballerina-lang
|
https://api.github.com/repos/ballerina-platform/ballerina-lang
|
closed
|
Able to push a package with a local repo dependancy to central
|
Type/Bug Priority/High Status/Blocked Team/DevTools SwanLakeDump Area/ProjectAPI
|
**Description:**
Even though a similar package name does or does not exist in central, should not be able to push a package with a local dependency to the central. Currently, it's allowed to push for both scenarios.
**Steps to reproduce:**
**Affected Versions:**
**OS, DB, other environment details and versions:**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
|
1.0
|
Able to push a package with a local repo dependancy to central - **Description:**
Even though a similar package name does or does not exist in central, should not be able to push a package with a local dependency to the central. Currently, it's allowed to push for both scenarios.
**Steps to reproduce:**
**Affected Versions:**
**OS, DB, other environment details and versions:**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
|
non_test
|
able to push a package with a local repo dependancy to central description even though a similar package name does or does not exist in central should not be able to push a package with a local dependency to the central currently it s allowed to push for both scenarios steps to reproduce affected versions os db other environment details and versions related issues optional suggested labels optional suggested assignees optional
| 0
|
516,420
| 14,981,892,391
|
IssuesEvent
|
2021-01-28 15:21:49
|
jetstack/cert-manager
|
https://api.github.com/repos/jetstack/cert-manager
|
closed
|
backup instructions in the docs are incorrect
|
kind/bug kind/feature priority/important-longterm
|
**Describe the bug**:
This page has instruction for backup up the certs and restoring them:
https://docs.cert-manager.io/en/latest/tasks/backup-restore-crds.html
The instruction don't work as is. The backup part work, but the restore won't work for many reasons.
```
status.conditions.lastTransitionTime in body must be of type string: "null"
```
a quick hack here is to replace `lastTransitionTime: null` with some valid date, but filtering the status part would be better. That fixed I get
```
Error from server (Conflict): Operation cannot be fulfilled on issuers.certmanager.k8s.io "letsencrypt-prod": the object has been modified; please apply your changes to the latest version and try again
Error from server (Conflict): Operation cannot be fulfilled on certificates.certmanager.k8s.io "cloud-master": the object has been modified; please apply your changes to the latest version and try again
```
**Expected behaviour**:
Restore should be tested and working
**Environment details:**:
I was upgrading from 0.5.2 to 0.6.7 (in order to upgrade further).
/kind bug
|
1.0
|
backup instructions in the docs are incorrect - **Describe the bug**:
This page has instruction for backup up the certs and restoring them:
https://docs.cert-manager.io/en/latest/tasks/backup-restore-crds.html
The instruction don't work as is. The backup part work, but the restore won't work for many reasons.
```
status.conditions.lastTransitionTime in body must be of type string: "null"
```
a quick hack here is to replace `lastTransitionTime: null` with some valid date, but filtering the status part would be better. That fixed I get
```
Error from server (Conflict): Operation cannot be fulfilled on issuers.certmanager.k8s.io "letsencrypt-prod": the object has been modified; please apply your changes to the latest version and try again
Error from server (Conflict): Operation cannot be fulfilled on certificates.certmanager.k8s.io "cloud-master": the object has been modified; please apply your changes to the latest version and try again
```
**Expected behaviour**:
Restore should be tested and working
**Environment details:**:
I was upgrading from 0.5.2 to 0.6.7 (in order to upgrade further).
/kind bug
|
non_test
|
backup instructions in the docs are incorrect describe the bug this page has instruction for backup up the certs and restoring them the instruction don t work as is the backup part work but the restore won t work for many reasons status conditions lasttransitiontime in body must be of type string null a quick hack here is to replace lasttransitiontime null with some valid date but filtering the status part would be better that fixed i get error from server conflict operation cannot be fulfilled on issuers certmanager io letsencrypt prod the object has been modified please apply your changes to the latest version and try again error from server conflict operation cannot be fulfilled on certificates certmanager io cloud master the object has been modified please apply your changes to the latest version and try again expected behaviour restore should be tested and working environment details i was upgrading from to in order to upgrade further kind bug
| 0
|
199,125
| 15,024,780,032
|
IssuesEvent
|
2021-02-01 20:07:48
|
CARTAvis/carta-backend
|
https://api.github.com/repos/CARTAvis/carta-backend
|
closed
|
extra channel info derived from header with ra-dec-stokes
|
awaiting testing bug
|
If the image has 3 axes as ra-dec-stokes, there is an extra channel info in the file info tab.


image from @zarda
|
1.0
|
extra channel info derived from header with ra-dec-stokes - If the image has 3 axes as ra-dec-stokes, there is an extra channel info in the file info tab.


image from @zarda
|
test
|
extra channel info derived from header with ra dec stokes if the image has axes as ra dec stokes there is an extra channel info in the file info tab image from zarda
| 1
|
147,582
| 13,210,679,832
|
IssuesEvent
|
2020-08-15 18:14:54
|
Gizra/og
|
https://api.github.com/repos/Gizra/og
|
closed
|
og_ungroup() documentation incorrectly informs about @return value.
|
Documentation Drupal 7
|
The current documentation of the og_ungroup() function states that an entity is returned but it's not currently the case:
```
/**
* Delete an association (e.g. unsubscribe) of an entity to a group.
*
* @param $group_type
* The entity type (e.g. "node").
* @param $gid
* The group entity object or ID, to ungroup.
* @param $entity_type
* (optional) The entity type (e.g. "node" or "user").
* @param $etid
* (optional) The entity object or ID, to ungroup.
*
* @return
* The entity with the fields updated.
*/
function og_ungroup($group_type, $gid, $entity_type = 'user', $etid = NULL) {
if (is_object($gid)) {
list($gid) = entity_extract_ids($group_type, $gid);
}
if ($entity_type == 'user' && empty($etid)) {
global $user;
$etid = $user->uid;
}
elseif (is_object($etid)) {
list($etid) = entity_extract_ids($entity_type, $etid);
}
if ($og_membership = og_get_membership($group_type, $gid, $entity_type, $etid)) {
$og_membership->delete();
}
}
```
Not providing a patch / PR because I don't know if what needs to be corrected is the documentation to follow the code, or to change the code to return something (perhaps more usefull...)
|
1.0
|
og_ungroup() documentation incorrectly informs about @return value. - The current documentation of the og_ungroup() function states that an entity is returned but it's not currently the case:
```
/**
* Delete an association (e.g. unsubscribe) of an entity to a group.
*
* @param $group_type
* The entity type (e.g. "node").
* @param $gid
* The group entity object or ID, to ungroup.
* @param $entity_type
* (optional) The entity type (e.g. "node" or "user").
* @param $etid
* (optional) The entity object or ID, to ungroup.
*
* @return
* The entity with the fields updated.
*/
function og_ungroup($group_type, $gid, $entity_type = 'user', $etid = NULL) {
if (is_object($gid)) {
list($gid) = entity_extract_ids($group_type, $gid);
}
if ($entity_type == 'user' && empty($etid)) {
global $user;
$etid = $user->uid;
}
elseif (is_object($etid)) {
list($etid) = entity_extract_ids($entity_type, $etid);
}
if ($og_membership = og_get_membership($group_type, $gid, $entity_type, $etid)) {
$og_membership->delete();
}
}
```
Not providing a patch / PR because I don't know if what needs to be corrected is the documentation to follow the code, or to change the code to return something (perhaps more usefull...)
|
non_test
|
og ungroup documentation incorrectly informs about return value the current documentation of the og ungroup function states that an entity is returned but it s not currently the case delete an association e g unsubscribe of an entity to a group param group type the entity type e g node param gid the group entity object or id to ungroup param entity type optional the entity type e g node or user param etid optional the entity object or id to ungroup return the entity with the fields updated function og ungroup group type gid entity type user etid null if is object gid list gid entity extract ids group type gid if entity type user empty etid global user etid user uid elseif is object etid list etid entity extract ids entity type etid if og membership og get membership group type gid entity type etid og membership delete not providing a patch pr because i don t know if what needs to be corrected is the documentation to follow the code or to change the code to return something perhaps more usefull
| 0
|
351,773
| 32,025,684,178
|
IssuesEvent
|
2023-09-22 08:40:25
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
closed
|
Fix solving_equations_and_inverting_matrices.test_numpy_tensorinv
|
NumPy Frontend Sub Task Failing Test
|
| | |
|---|---|
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6271730752"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6265043606"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/6265043606"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6265043606"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/6265043606"><img src=https://img.shields.io/badge/-success-success></a>
|
1.0
|
Fix solving_equations_and_inverting_matrices.test_numpy_tensorinv - | | |
|---|---|
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6271730752"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6265043606"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/6265043606"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6265043606"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/6265043606"><img src=https://img.shields.io/badge/-success-success></a>
|
test
|
fix solving equations and inverting matrices test numpy tensorinv paddle a href src numpy a href src jax a href src tensorflow a href src torch a href src
| 1
|
13,294
| 2,750,721,889
|
IssuesEvent
|
2015-04-24 01:41:59
|
micheldumontier/semanticscience
|
https://api.github.com/repos/micheldumontier/semanticscience
|
closed
|
"is source of" is both property and class
|
auto-migrated Priority-Medium Type-Defect
|
```
http://semanticscience.org/resource/SIO_000219 'is source of' in sio-bio.owl
seems to be both a property and (probably incorrect) a class. 'drug regulatory
authority' is a subclass of this entity.
Viewed in TopBraid Composer Free Edition without inferencing.
```
Original issue reported on code.google.com by `matthias...@gmail.com` on 29 Jun 2012 at 9:22
|
1.0
|
"is source of" is both property and class - ```
http://semanticscience.org/resource/SIO_000219 'is source of' in sio-bio.owl
seems to be both a property and (probably incorrect) a class. 'drug regulatory
authority' is a subclass of this entity.
Viewed in TopBraid Composer Free Edition without inferencing.
```
Original issue reported on code.google.com by `matthias...@gmail.com` on 29 Jun 2012 at 9:22
|
non_test
|
is source of is both property and class is source of in sio bio owl seems to be both a property and probably incorrect a class drug regulatory authority is a subclass of this entity viewed in topbraid composer free edition without inferencing original issue reported on code google com by matthias gmail com on jun at
| 0
|
268,009
| 23,339,202,079
|
IssuesEvent
|
2022-08-09 12:44:53
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
Failing ES Promotion: Chrome X-Pack UI Functional Tests - ML anomaly_detection.x-pack/test/functional/apps/ml/anomaly_detection/anomaly_explorer·ts - machine learning - anomaly detection anomaly explorer with farequote based multi metric job renders View By swim lane
|
blocker :ml skipped-test failed-es-promotion Team:ML v7.17.6
|
**Chrome X-Pack UI Functional Tests - ML anomaly_detection**
**x-pack/test/functional/apps/ml/anomaly_detection/anomaly_explorer.ts**
**machine learning - anomaly detection anomaly explorer with farequote based multi metric job renders View By swim lane**
This failure is preventing the promotion of the current Elasticsearch nightly snapshot.
For more information on the Elasticsearch snapshot promotion process including how to reproduce using the unverified nightly ES build: https://www.elastic.co/guide/en/kibana/master/development-es-snapshots.html
* [Failed promotion job](https://buildkite.com/elastic/kibana-elasticsearch-snapshot-verify/builds/1539#01827dcc-e0f1-4a59-a15c-acc245fdaa48)
* [Test Failure](https://buildkite.com/organizations/elastic/pipelines/kibana-elasticsearch-snapshot-verify/builds/1539/jobs/01827dcc-e0f1-4a59-a15c-acc245fdaa48/artifacts/01827df9-0925-4dfb-8d8a-f996e6024ebf)
```
Error: Expected swim lane y labels to be AAL,VRD,EGF,SWR,AMX,JZA,TRS,ACA,BAW,ASA, got AAL,EGF,VRD,SWR,JZA,AMX,TRS,ACA,BAW,ASA
at Assertion.assert (node_modules/@kbn/expect/expect.js:100:11)
at Assertion.eql (node_modules/@kbn/expect/expect.js:244:8)
at Object.assertAxisLabels (x-pack/test/functional/services/ml/swim_lane.ts:88:31)
at runMicrotasks (<anonymous>)
at processTicksAndRejections (node:internal/process/task_queues:96:5)
at Context.<anonymous> (x-pack/test/functional/apps/ml/anomaly_detection/anomaly_explorer.ts:167:11)
at Object.apply (node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) {
actual: '[\n' +
' "AAL"\n' +
' "EGF"\n' +
' "VRD"\n' +
' "SWR"\n' +
' "JZA"\n' +
' "AMX"\n' +
' "TRS"\n' +
' "ACA"\n' +
' "BAW"\n' +
' "ASA"\n' +
']',
expected: '[\n' +
' "AAL"\n' +
' "VRD"\n' +
' "EGF"\n' +
' "SWR"\n' +
' "AMX"\n' +
' "JZA"\n' +
' "TRS"\n' +
' "ACA"\n' +
' "BAW"\n' +
' "ASA"\n' +
']',
showDiff: true
}
```
|
1.0
|
Failing ES Promotion: Chrome X-Pack UI Functional Tests - ML anomaly_detection.x-pack/test/functional/apps/ml/anomaly_detection/anomaly_explorer·ts - machine learning - anomaly detection anomaly explorer with farequote based multi metric job renders View By swim lane - **Chrome X-Pack UI Functional Tests - ML anomaly_detection**
**x-pack/test/functional/apps/ml/anomaly_detection/anomaly_explorer.ts**
**machine learning - anomaly detection anomaly explorer with farequote based multi metric job renders View By swim lane**
This failure is preventing the promotion of the current Elasticsearch nightly snapshot.
For more information on the Elasticsearch snapshot promotion process including how to reproduce using the unverified nightly ES build: https://www.elastic.co/guide/en/kibana/master/development-es-snapshots.html
* [Failed promotion job](https://buildkite.com/elastic/kibana-elasticsearch-snapshot-verify/builds/1539#01827dcc-e0f1-4a59-a15c-acc245fdaa48)
* [Test Failure](https://buildkite.com/organizations/elastic/pipelines/kibana-elasticsearch-snapshot-verify/builds/1539/jobs/01827dcc-e0f1-4a59-a15c-acc245fdaa48/artifacts/01827df9-0925-4dfb-8d8a-f996e6024ebf)
```
Error: Expected swim lane y labels to be AAL,VRD,EGF,SWR,AMX,JZA,TRS,ACA,BAW,ASA, got AAL,EGF,VRD,SWR,JZA,AMX,TRS,ACA,BAW,ASA
at Assertion.assert (node_modules/@kbn/expect/expect.js:100:11)
at Assertion.eql (node_modules/@kbn/expect/expect.js:244:8)
at Object.assertAxisLabels (x-pack/test/functional/services/ml/swim_lane.ts:88:31)
at runMicrotasks (<anonymous>)
at processTicksAndRejections (node:internal/process/task_queues:96:5)
at Context.<anonymous> (x-pack/test/functional/apps/ml/anomaly_detection/anomaly_explorer.ts:167:11)
at Object.apply (node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) {
actual: '[\n' +
' "AAL"\n' +
' "EGF"\n' +
' "VRD"\n' +
' "SWR"\n' +
' "JZA"\n' +
' "AMX"\n' +
' "TRS"\n' +
' "ACA"\n' +
' "BAW"\n' +
' "ASA"\n' +
']',
expected: '[\n' +
' "AAL"\n' +
' "VRD"\n' +
' "EGF"\n' +
' "SWR"\n' +
' "AMX"\n' +
' "JZA"\n' +
' "TRS"\n' +
' "ACA"\n' +
' "BAW"\n' +
' "ASA"\n' +
']',
showDiff: true
}
```
|
test
|
failing es promotion chrome x pack ui functional tests ml anomaly detection x pack test functional apps ml anomaly detection anomaly explorer·ts machine learning anomaly detection anomaly explorer with farequote based multi metric job renders view by swim lane chrome x pack ui functional tests ml anomaly detection x pack test functional apps ml anomaly detection anomaly explorer ts machine learning anomaly detection anomaly explorer with farequote based multi metric job renders view by swim lane this failure is preventing the promotion of the current elasticsearch nightly snapshot for more information on the elasticsearch snapshot promotion process including how to reproduce using the unverified nightly es build error expected swim lane y labels to be aal vrd egf swr amx jza trs aca baw asa got aal egf vrd swr jza amx trs aca baw asa at assertion assert node modules kbn expect expect js at assertion eql node modules kbn expect expect js at object assertaxislabels x pack test functional services ml swim lane ts at runmicrotasks at processticksandrejections node internal process task queues at context x pack test functional apps ml anomaly detection anomaly explorer ts at object apply node modules kbn test target node functional test runner lib mocha wrap function js actual n aal n egf n vrd n swr n jza n amx n trs n aca n baw n asa n expected n aal n vrd n egf n swr n amx n jza n trs n aca n baw n asa n showdiff true
| 1
|
121,884
| 10,197,017,424
|
IssuesEvent
|
2019-08-12 22:32:15
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
teamcity: failed test: gossip/restart
|
C-test-failure O-robot
|
The following tests appear to have failed on master (roachtest): acceptance/gossip/restart
You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+gossip/restart).
[#1436098](https://teamcity.cockroachdb.com/viewLog.html?buildId=1436098):
```
acceptance/gossip/restart
--- FAIL: roachtest/acceptance/gossip/restart (26.519s)
test artifacts and logs in: artifacts/acceptance/gossip/restart/run_1
gossip.go:226,gossip.go:286,acceptance.go:69,test_runner.go:691: dial tcp 127.0.0.1:26261: connect: connection refused
test artifacts and logs in: artifacts/acceptance/gossip/restart/run_1
gossip.go:226,gossip.go:286,acceptance.go:69,test_runner.go:691: dial tcp 127.0.0.1:26261: connect: connection refused
```
Please assign, take a look and update the issue accordingly.
|
1.0
|
teamcity: failed test: gossip/restart - The following tests appear to have failed on master (roachtest): acceptance/gossip/restart
You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+gossip/restart).
[#1436098](https://teamcity.cockroachdb.com/viewLog.html?buildId=1436098):
```
acceptance/gossip/restart
--- FAIL: roachtest/acceptance/gossip/restart (26.519s)
test artifacts and logs in: artifacts/acceptance/gossip/restart/run_1
gossip.go:226,gossip.go:286,acceptance.go:69,test_runner.go:691: dial tcp 127.0.0.1:26261: connect: connection refused
test artifacts and logs in: artifacts/acceptance/gossip/restart/run_1
gossip.go:226,gossip.go:286,acceptance.go:69,test_runner.go:691: dial tcp 127.0.0.1:26261: connect: connection refused
```
Please assign, take a look and update the issue accordingly.
|
test
|
teamcity failed test gossip restart the following tests appear to have failed on master roachtest acceptance gossip restart you may want to check acceptance gossip restart fail roachtest acceptance gossip restart test artifacts and logs in artifacts acceptance gossip restart run gossip go gossip go acceptance go test runner go dial tcp connect connection refused test artifacts and logs in artifacts acceptance gossip restart run gossip go gossip go acceptance go test runner go dial tcp connect connection refused please assign take a look and update the issue accordingly
| 1
|
90,532
| 15,856,201,240
|
IssuesEvent
|
2021-04-08 01:46:26
|
AnhaaD/hacknightvol4
|
https://api.github.com/repos/AnhaaD/hacknightvol4
|
opened
|
CVE-2019-20149 (High) detected in kind-of-6.0.2.tgz
|
security vulnerability
|
## CVE-2019-20149 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>kind-of-6.0.2.tgz</b></p></summary>
<p>Get the native type of a value.</p>
<p>Library home page: <a href="https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz">https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz</a></p>
<p>Path to dependency file: /hacknightvol4/package.json</p>
<p>Path to vulnerable library: hacknightvol4/node_modules/snapdragon-node/node_modules/kind-of/package.json</p>
<p>
Dependency Hierarchy:
- nodemon-1.17.5.tgz (Root Library)
- chokidar-2.1.6.tgz
- anymatch-2.0.0.tgz
- micromatch-3.1.10.tgz
- :x: **kind-of-6.0.2.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
ctorName in index.js in kind-of v6.0.2 allows external user input to overwrite certain internal attributes via a conflicting name, as demonstrated by 'constructor': {'name':'Symbol'}. Hence, a crafted payload can overwrite this builtin attribute to manipulate the type detection result.
<p>Publish Date: 2019-12-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20149>CVE-2019-20149</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-20149">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-20149</a></p>
<p>Release Date: 2019-12-30</p>
<p>Fix Resolution: 6.0.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-20149 (High) detected in kind-of-6.0.2.tgz - ## CVE-2019-20149 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>kind-of-6.0.2.tgz</b></p></summary>
<p>Get the native type of a value.</p>
<p>Library home page: <a href="https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz">https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz</a></p>
<p>Path to dependency file: /hacknightvol4/package.json</p>
<p>Path to vulnerable library: hacknightvol4/node_modules/snapdragon-node/node_modules/kind-of/package.json</p>
<p>
Dependency Hierarchy:
- nodemon-1.17.5.tgz (Root Library)
- chokidar-2.1.6.tgz
- anymatch-2.0.0.tgz
- micromatch-3.1.10.tgz
- :x: **kind-of-6.0.2.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
ctorName in index.js in kind-of v6.0.2 allows external user input to overwrite certain internal attributes via a conflicting name, as demonstrated by 'constructor': {'name':'Symbol'}. Hence, a crafted payload can overwrite this builtin attribute to manipulate the type detection result.
<p>Publish Date: 2019-12-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20149>CVE-2019-20149</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-20149">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-20149</a></p>
<p>Release Date: 2019-12-30</p>
<p>Fix Resolution: 6.0.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve high detected in kind of tgz cve high severity vulnerability vulnerable library kind of tgz get the native type of a value library home page a href path to dependency file package json path to vulnerable library node modules snapdragon node node modules kind of package json dependency hierarchy nodemon tgz root library chokidar tgz anymatch tgz micromatch tgz x kind of tgz vulnerable library vulnerability details ctorname in index js in kind of allows external user input to overwrite certain internal attributes via a conflicting name as demonstrated by constructor name symbol hence a crafted payload can overwrite this builtin attribute to manipulate the type detection result publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
264,340
| 8,307,891,840
|
IssuesEvent
|
2018-09-23 14:51:36
|
publiclab/plots2
|
https://api.github.com/repos/publiclab/plots2
|
opened
|
Digest email bug
|
bug priority
|
The digest emails are currently not being sent.
`ArgumentError: wrong number of arguments (given 0, expected 1)` is shown on Sidekiq dashboard (https://publiclab.org/sidekiq/retries).
|
1.0
|
Digest email bug - The digest emails are currently not being sent.
`ArgumentError: wrong number of arguments (given 0, expected 1)` is shown on Sidekiq dashboard (https://publiclab.org/sidekiq/retries).
|
non_test
|
digest email bug the digest emails are currently not being sent argumenterror wrong number of arguments given expected is shown on sidekiq dashboard
| 0
|
412,971
| 12,058,928,646
|
IssuesEvent
|
2020-04-15 18:20:39
|
zulip/zulip
|
https://api.github.com/repos/zulip/zulip
|
closed
|
Optimize rate_limiter performance for get_events queries
|
area: production in progress priority: high
|
See https://chat.zulip.org/#narrow/stream/3-backend/topic/profiling.20get_events/near/816860 for profiling details, but basically, currently a get_events request spends 1.4ms/request talking to redis for our rate limiter, which is somewhere between 15% and 50% of the total request runtime (my measurement technique is susceptible to issues like the first request on a code path being extra expensive). Since get_events is our most scalability-critical endpoint, this is a big deal.
We should do some rethinking of the redis internals for our rate limiter. I have a few ideas:
* Writing an alternative rate-limiter implementation for `get_events `specifically that's entirely in-process and would be basically instant. Since the Tornado system has a relatively strong constraint that a given user always connect to the same server, this might be fairly cheap to implement and would bring that 1.4ms to probably 50us or less. (And gate it on `RUNNING_INSIDE_TORNADO`).
* Look at rewriting our redis transactions to be more efficient for the highest-traffic cases (E.g. user is not close to limit, or user is way over limit). E.g. maybe `incr_rateimit` should automatically return the `api_calls_left` result rather than requiring 2 transactions.
* Looking at https://github.com/popravich/python-redis-benchmark, there may be some alternative async IO redis clients we could consider migrating to, and possibly some that are just faster. Given how little code we have interacting with redis directly, this might be an easy port to do; I'm not sure whether or not it would help. (And unlike the in-process hack approach, this would have side benefits to non-Tornado endpoints).
|
1.0
|
Optimize rate_limiter performance for get_events queries - See https://chat.zulip.org/#narrow/stream/3-backend/topic/profiling.20get_events/near/816860 for profiling details, but basically, currently a get_events request spends 1.4ms/request talking to redis for our rate limiter, which is somewhere between 15% and 50% of the total request runtime (my measurement technique is susceptible to issues like the first request on a code path being extra expensive). Since get_events is our most scalability-critical endpoint, this is a big deal.
We should do some rethinking of the redis internals for our rate limiter. I have a few ideas:
* Writing an alternative rate-limiter implementation for `get_events `specifically that's entirely in-process and would be basically instant. Since the Tornado system has a relatively strong constraint that a given user always connect to the same server, this might be fairly cheap to implement and would bring that 1.4ms to probably 50us or less. (And gate it on `RUNNING_INSIDE_TORNADO`).
* Look at rewriting our redis transactions to be more efficient for the highest-traffic cases (E.g. user is not close to limit, or user is way over limit). E.g. maybe `incr_rateimit` should automatically return the `api_calls_left` result rather than requiring 2 transactions.
* Looking at https://github.com/popravich/python-redis-benchmark, there may be some alternative async IO redis clients we could consider migrating to, and possibly some that are just faster. Given how little code we have interacting with redis directly, this might be an easy port to do; I'm not sure whether or not it would help. (And unlike the in-process hack approach, this would have side benefits to non-Tornado endpoints).
|
non_test
|
optimize rate limiter performance for get events queries see for profiling details but basically currently a get events request spends request talking to redis for our rate limiter which is somewhere between and of the total request runtime my measurement technique is susceptible to issues like the first request on a code path being extra expensive since get events is our most scalability critical endpoint this is a big deal we should do some rethinking of the redis internals for our rate limiter i have a few ideas writing an alternative rate limiter implementation for get events specifically that s entirely in process and would be basically instant since the tornado system has a relatively strong constraint that a given user always connect to the same server this might be fairly cheap to implement and would bring that to probably or less and gate it on running inside tornado look at rewriting our redis transactions to be more efficient for the highest traffic cases e g user is not close to limit or user is way over limit e g maybe incr rateimit should automatically return the api calls left result rather than requiring transactions looking at there may be some alternative async io redis clients we could consider migrating to and possibly some that are just faster given how little code we have interacting with redis directly this might be an easy port to do i m not sure whether or not it would help and unlike the in process hack approach this would have side benefits to non tornado endpoints
| 0
|
4,195
| 4,968,932,741
|
IssuesEvent
|
2016-12-05 11:33:47
|
core-wg/oscoap
|
https://api.github.com/repos/core-wg/oscoap
|
closed
|
Replay window is an input parameter?
|
core-object-security-00
|
Section 3.2 - Why is the replay window an input - this seems to be odd as you would not pre-fill part of the window. I assume this is really just the Replay Window Size.
|
True
|
Replay window is an input parameter? - Section 3.2 - Why is the replay window an input - this seems to be odd as you would not pre-fill part of the window. I assume this is really just the Replay Window Size.
|
non_test
|
replay window is an input parameter section why is the replay window an input this seems to be odd as you would not pre fill part of the window i assume this is really just the replay window size
| 0
|
138,488
| 11,202,492,452
|
IssuesEvent
|
2020-01-04 12:59:02
|
searchkit/searchkit
|
https://api.github.com/repos/searchkit/searchkit
|
closed
|
rangeFormatter function not used in filter
|
2.3.0-9 Ready For Testing stale
|
<img width="286" alt="screenshot 2017-06-13 15 59 49" src="https://user-images.githubusercontent.com/7115982/27072441-f190638e-5052-11e7-9b15-1a5bb8c7018d.png">
<img width="433" alt="screenshot 2017-06-13 15 59 53" src="https://user-images.githubusercontent.com/7115982/27072443-f29c5710-5052-11e7-837b-43106a0e66c8.png">
rangeFormatter function used in rangeFilter but not used in top filter
|
1.0
|
rangeFormatter function not used in filter - <img width="286" alt="screenshot 2017-06-13 15 59 49" src="https://user-images.githubusercontent.com/7115982/27072441-f190638e-5052-11e7-9b15-1a5bb8c7018d.png">
<img width="433" alt="screenshot 2017-06-13 15 59 53" src="https://user-images.githubusercontent.com/7115982/27072443-f29c5710-5052-11e7-837b-43106a0e66c8.png">
rangeFormatter function used in rangeFilter but not used in top filter
|
test
|
rangeformatter function not used in filter img width alt screenshot src img width alt screenshot src rangeformatter function used in rangefilter but not used in top filter
| 1
|
263,587
| 28,047,514,574
|
IssuesEvent
|
2023-03-29 01:03:58
|
tabacws-sandbox/juice-shop-checkPR
|
https://api.github.com/repos/tabacws-sandbox/juice-shop-checkPR
|
closed
|
check-dependencies-1.1.0.tgz: 1 vulnerabilities (highest severity is: 9.8) - autoclosed
|
Mend: dependency security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>check-dependencies-1.1.0.tgz</b></p></summary>
<p></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/set-value/package.json</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/tabacws-sandbox/juice-shop-checkPR/commit/898e55dce59f24513206f629f1dd595ca468b56f">898e55dce59f24513206f629f1dd595ca468b56f</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (check-dependencies version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2021-23440](https://www.mend.io/vulnerability-database/CVE-2021-23440) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> Critical | 9.8 | set-value-2.0.1.tgz | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the "Details" section below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> CVE-2021-23440</summary>
### Vulnerable Library - <b>set-value-2.0.1.tgz</b></p>
<p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p>
<p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-2.0.1.tgz">https://registry.npmjs.org/set-value/-/set-value-2.0.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/set-value/package.json</p>
<p>
Dependency Hierarchy:
- check-dependencies-1.1.0.tgz (Root Library)
- findup-sync-2.0.0.tgz
- micromatch-3.1.10.tgz
- snapdragon-0.8.2.tgz
- base-0.11.2.tgz
- cache-base-1.0.1.tgz
- :x: **set-value-2.0.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/tabacws-sandbox/juice-shop-checkPR/commit/898e55dce59f24513206f629f1dd595ca468b56f">898e55dce59f24513206f629f1dd595ca468b56f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
This affects the package set-value before <2.0.1, >=3.0.0 <4.0.1. A type confusion vulnerability can lead to a bypass of CVE-2019-10747 when the user-provided keys used in the path parameter are arrays.
Mend Note: After conducting further research, Mend has determined that all versions of set-value up to version 4.0.0 are vulnerable to CVE-2021-23440.
<p>Publish Date: 2021-09-12
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23440>CVE-2021-23440</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-09-12</p>
<p>Fix Resolution: set-value - 4.0.1
</p>
</p>
<p></p>
</details>
|
True
|
check-dependencies-1.1.0.tgz: 1 vulnerabilities (highest severity is: 9.8) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>check-dependencies-1.1.0.tgz</b></p></summary>
<p></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/set-value/package.json</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/tabacws-sandbox/juice-shop-checkPR/commit/898e55dce59f24513206f629f1dd595ca468b56f">898e55dce59f24513206f629f1dd595ca468b56f</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (check-dependencies version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2021-23440](https://www.mend.io/vulnerability-database/CVE-2021-23440) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> Critical | 9.8 | set-value-2.0.1.tgz | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the "Details" section below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> CVE-2021-23440</summary>
### Vulnerable Library - <b>set-value-2.0.1.tgz</b></p>
<p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p>
<p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-2.0.1.tgz">https://registry.npmjs.org/set-value/-/set-value-2.0.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/set-value/package.json</p>
<p>
Dependency Hierarchy:
- check-dependencies-1.1.0.tgz (Root Library)
- findup-sync-2.0.0.tgz
- micromatch-3.1.10.tgz
- snapdragon-0.8.2.tgz
- base-0.11.2.tgz
- cache-base-1.0.1.tgz
- :x: **set-value-2.0.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/tabacws-sandbox/juice-shop-checkPR/commit/898e55dce59f24513206f629f1dd595ca468b56f">898e55dce59f24513206f629f1dd595ca468b56f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
This affects the package set-value before <2.0.1, >=3.0.0 <4.0.1. A type confusion vulnerability can lead to a bypass of CVE-2019-10747 when the user-provided keys used in the path parameter are arrays.
Mend Note: After conducting further research, Mend has determined that all versions of set-value up to version 4.0.0 are vulnerable to CVE-2021-23440.
<p>Publish Date: 2021-09-12
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23440>CVE-2021-23440</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-09-12</p>
<p>Fix Resolution: set-value - 4.0.1
</p>
</p>
<p></p>
</details>
|
non_test
|
check dependencies tgz vulnerabilities highest severity is autoclosed vulnerable library check dependencies tgz path to dependency file package json path to vulnerable library node modules set value package json found in head commit a href vulnerabilities cve severity cvss dependency type fixed in check dependencies version remediation available critical set value tgz transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the details section below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library set value tgz create nested values and any intermediaries using dot notation a b c paths library home page a href path to dependency file package json path to vulnerable library node modules set value package json dependency hierarchy check dependencies tgz root library findup sync tgz micromatch tgz snapdragon tgz base tgz cache base tgz x set value tgz vulnerable library found in head commit a href found in base branch master vulnerability details this affects the package set value before a type confusion vulnerability can lead to a bypass of cve when the user provided keys used in the path parameter are arrays mend note after conducting further research mend has determined that all versions of set value up to version are vulnerable to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution set value
| 0
|
38,842
| 8,554,684,441
|
IssuesEvent
|
2018-11-08 07:32:00
|
openshiftio/openshift.io
|
https://api.github.com/repos/openshiftio/openshift.io
|
closed
|
Support Quick fixes with codeAction in case of CVEs flagged from LSP
|
area/analytics env/vs-code status/in-progress team/analytics type/user-story
|
### Planner Link : https://openshift.io/openshiftio/Openshift_io/plan/detail/948
Currently in VSCode Extension, we flag CVEs for any core dependencies, proposed would be to enable quick fixes to those with enabling codeAction from Language Server.
### Tasks:
- [x] Update language server and client version
- [x] Bind codeAction supportin Language Server
- [x] Register command in client to listen and apply edits
|
1.0
|
Support Quick fixes with codeAction in case of CVEs flagged from LSP - ### Planner Link : https://openshift.io/openshiftio/Openshift_io/plan/detail/948
Currently in VSCode Extension, we flag CVEs for any core dependencies, proposed would be to enable quick fixes to those with enabling codeAction from Language Server.
### Tasks:
- [x] Update language server and client version
- [x] Bind codeAction supportin Language Server
- [x] Register command in client to listen and apply edits
|
non_test
|
support quick fixes with codeaction in case of cves flagged from lsp planner link currently in vscode extension we flag cves for any core dependencies proposed would be to enable quick fixes to those with enabling codeaction from language server tasks update language server and client version bind codeaction supportin language server register command in client to listen and apply edits
| 0
|
245,192
| 20,751,961,035
|
IssuesEvent
|
2022-03-15 08:34:34
|
openvinotoolkit/openvino
|
https://api.github.com/repos/openvinotoolkit/openvino
|
closed
|
[Bug] (minor) test namespace ?
|
bug category: IE Tests support_request PSE
|
##### System information (version)
<!-- Example
- OpenVINO => 2020.4
- Operating System / Platform => Windows 64 Bit
- Compiler => Visual Studio 2017
- Problem classification: Model Conversion
- Framework: TensorFlow (if applicable)
- Model name: ResNet50 (if applicable)
-->
- OpenVINO=> :grey_question:
- Operating System / Platform => :grey_question:
- Compiler => :grey_question:
- Problem classification => :grey_question:
##### Detailed description
<!-- your description -->
I suppose this namespace was meant to be "LayerTest**s**Definitions", instead of "LayerTestDefinitions", to be aligned with the rest of the filters. I was getting an unexpected compile error and I barely noticed the missing 's':
https://github.com/openvinotoolkit/openvino/blob/e8d5cf43d0e153f4f52c7be133f71f65c9eb4512/src/tests/functional/shared_test_classes/include/shared_test_classes/single_layer/prior_box.hpp#L29
##### Steps to reproduce
<!--
Describe your problem and steps you've done before you got to this point.
to add code example fence it with triple backticks and optional file extension
```.cpp
// C++ code example
```
or attach as .txt or .zip file
-->
##### Issue submission checklist
- [x] I report the issue, it's not a question
<!--
OpenVINO team works with support forum, Stack Overflow and other communities
to discuss problems. Tickets with question without real issue statement will be
closed.
-->
- [ ] I checked the problem with documentation, FAQ, open issues, Stack Overflow, etc and have not found solution
<!--
Places to check:
* OpenVINO documentation: https://docs.openvinotoolkit.org/
* OpenVINO forum: https://community.intel.com/t5/Intel-Distribution-of-OpenVINO/bd-p/distribution-openvino-toolkit
* OpenVINO issue tracker: https://github.com/openvinotoolkit/openvino/issues?q=is%3Aissue
* Stack Overflow branch: https://stackoverflow.com/questions/tagged/openvino
-->
- [ ] There is reproducer code and related data files: images, videos, models, etc.
<!--
The best reproducer -- test case for OpenVINO that we can add to the library.
-->
|
1.0
|
[Bug] (minor) test namespace ? - ##### System information (version)
<!-- Example
- OpenVINO => 2020.4
- Operating System / Platform => Windows 64 Bit
- Compiler => Visual Studio 2017
- Problem classification: Model Conversion
- Framework: TensorFlow (if applicable)
- Model name: ResNet50 (if applicable)
-->
- OpenVINO=> :grey_question:
- Operating System / Platform => :grey_question:
- Compiler => :grey_question:
- Problem classification => :grey_question:
##### Detailed description
<!-- your description -->
I suppose this namespace was meant to be "LayerTest**s**Definitions", instead of "LayerTestDefinitions", to be aligned with the rest of the filters. I was getting an unexpected compile error and I barely noticed the missing 's':
https://github.com/openvinotoolkit/openvino/blob/e8d5cf43d0e153f4f52c7be133f71f65c9eb4512/src/tests/functional/shared_test_classes/include/shared_test_classes/single_layer/prior_box.hpp#L29
##### Steps to reproduce
<!--
Describe your problem and steps you've done before you got to this point.
to add code example fence it with triple backticks and optional file extension
```.cpp
// C++ code example
```
or attach as .txt or .zip file
-->
##### Issue submission checklist
- [x] I report the issue, it's not a question
<!--
OpenVINO team works with support forum, Stack Overflow and other communities
to discuss problems. Tickets with question without real issue statement will be
closed.
-->
- [ ] I checked the problem with documentation, FAQ, open issues, Stack Overflow, etc and have not found solution
<!--
Places to check:
* OpenVINO documentation: https://docs.openvinotoolkit.org/
* OpenVINO forum: https://community.intel.com/t5/Intel-Distribution-of-OpenVINO/bd-p/distribution-openvino-toolkit
* OpenVINO issue tracker: https://github.com/openvinotoolkit/openvino/issues?q=is%3Aissue
* Stack Overflow branch: https://stackoverflow.com/questions/tagged/openvino
-->
- [ ] There is reproducer code and related data files: images, videos, models, etc.
<!--
The best reproducer -- test case for OpenVINO that we can add to the library.
-->
|
test
|
minor test namespace system information version example openvino operating system platform windows bit compiler visual studio problem classification model conversion framework tensorflow if applicable model name if applicable openvino grey question operating system platform grey question compiler grey question problem classification grey question detailed description i suppose this namespace was meant to be layertest s definitions instead of layertestdefinitions to be aligned with the rest of the filters i was getting an unexpected compile error and i barely noticed the missing s steps to reproduce describe your problem and steps you ve done before you got to this point to add code example fence it with triple backticks and optional file extension cpp c code example or attach as txt or zip file issue submission checklist i report the issue it s not a question openvino team works with support forum stack overflow and other communities to discuss problems tickets with question without real issue statement will be closed i checked the problem with documentation faq open issues stack overflow etc and have not found solution places to check openvino documentation openvino forum openvino issue tracker stack overflow branch there is reproducer code and related data files images videos models etc the best reproducer test case for openvino that we can add to the library
| 1
|
654,339
| 21,648,352,441
|
IssuesEvent
|
2022-05-06 06:27:33
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
video.gazzetta.it - video or audio doesn't play
|
priority-normal browser-focus-geckoview engine-gecko
|
<!-- @browser: Firefox Mobile 100.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:100.0) Gecko/100.0 Firefox/100.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/103955 -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://video.gazzetta.it/video-reral-madrid-finale-ancelotti-abbraccia-florentino-perez/5460ed74-cc92-11ec-9cc1-9243b00d5567?vclk=videobar
**Browser / Version**: Firefox Mobile 100.0
**Operating System**: Android 11
**Tested Another Browser**: No
**Problem type**: Video or audio doesn't play
**Description**: The video or audio does not play
**Steps to Reproduce**:
Video does not play after clicking the play button
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/5/b4dbf693-5e8e-4678-b9ea-89544e9bb28a.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220428192727</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2022/5/d917ecb7-c99a-4fa5-8776-91389cf22c19)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
video.gazzetta.it - video or audio doesn't play - <!-- @browser: Firefox Mobile 100.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:100.0) Gecko/100.0 Firefox/100.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/103955 -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://video.gazzetta.it/video-reral-madrid-finale-ancelotti-abbraccia-florentino-perez/5460ed74-cc92-11ec-9cc1-9243b00d5567?vclk=videobar
**Browser / Version**: Firefox Mobile 100.0
**Operating System**: Android 11
**Tested Another Browser**: No
**Problem type**: Video or audio doesn't play
**Description**: The video or audio does not play
**Steps to Reproduce**:
Video does not play after clicking the play button
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/5/b4dbf693-5e8e-4678-b9ea-89544e9bb28a.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220428192727</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2022/5/d917ecb7-c99a-4fa5-8776-91389cf22c19)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_test
|
video gazzetta it video or audio doesn t play url browser version firefox mobile operating system android tested another browser no problem type video or audio doesn t play description the video or audio does not play steps to reproduce video does not play after clicking the play button view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel release hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
12,450
| 14,696,699,262
|
IssuesEvent
|
2021-01-04 00:27:12
|
Electroblob77/Wizardry
|
https://api.github.com/repos/Electroblob77/Wizardry
|
closed
|
MOD CONFLICTS
|
bug compatibility crash workaround exists
|
Minecraft version: 1.12.2
Wizardry version: 4.3.1
Environment: Singleplayer/LAN game/Server
Issue details: I was creating a modpack for minecraft. Obviosly it had to have my favorite mod: electroblob's wizardry, but the modpack contained morph and when i tried to use morph something went wrong and said that electroblob's wizardry was responsible. Can you look and see what is the problem please? I really want to create a modpack with that mod before christmas
Other mods involved: Morph (maybe ichunutil)
[crash-2020-12-18_18.45.35-client.txt](https://github.com/Electroblob77/Wizardry/files/5717584/crash-2020-12-18_18.45.35-client.txt)
|
True
|
MOD CONFLICTS -
Minecraft version: 1.12.2
Wizardry version: 4.3.1
Environment: Singleplayer/LAN game/Server
Issue details: I was creating a modpack for minecraft. Obviosly it had to have my favorite mod: electroblob's wizardry, but the modpack contained morph and when i tried to use morph something went wrong and said that electroblob's wizardry was responsible. Can you look and see what is the problem please? I really want to create a modpack with that mod before christmas
Other mods involved: Morph (maybe ichunutil)
[crash-2020-12-18_18.45.35-client.txt](https://github.com/Electroblob77/Wizardry/files/5717584/crash-2020-12-18_18.45.35-client.txt)
|
non_test
|
mod conflicts minecraft version wizardry version environment singleplayer lan game server issue details i was creating a modpack for minecraft obviosly it had to have my favorite mod electroblob s wizardry but the modpack contained morph and when i tried to use morph something went wrong and said that electroblob s wizardry was responsible can you look and see what is the problem please i really want to create a modpack with that mod before christmas other mods involved morph maybe ichunutil
| 0
|
65,196
| 7,861,027,508
|
IssuesEvent
|
2018-06-21 22:09:08
|
Opentrons/opentrons
|
https://api.github.com/repos/Opentrons/opentrons
|
opened
|
TempDeck: Real Time Status During Run
|
app design feature
|
As a Run App user, I would like to be able to see my TempDeck's status during a run.
## Acceptance Criteria
- [ ] Display current temperature in Run sidebar
- [ ] Display target temperature in Run sidebar
- [ ] Display status (string) in Run sidebar
## Implementation Notes
- Dependent on 'Get Modules / 'Info' endpoint, #1653, including current and target temps
- Will just poll this endpoint
|
1.0
|
TempDeck: Real Time Status During Run - As a Run App user, I would like to be able to see my TempDeck's status during a run.
## Acceptance Criteria
- [ ] Display current temperature in Run sidebar
- [ ] Display target temperature in Run sidebar
- [ ] Display status (string) in Run sidebar
## Implementation Notes
- Dependent on 'Get Modules / 'Info' endpoint, #1653, including current and target temps
- Will just poll this endpoint
|
non_test
|
tempdeck real time status during run as a run app user i would like to be able to see my tempdeck s status during a run acceptance criteria display current temperature in run sidebar display target temperature in run sidebar display status string in run sidebar implementation notes dependent on get modules info endpoint including current and target temps will just poll this endpoint
| 0
|
214,612
| 16,601,168,924
|
IssuesEvent
|
2021-06-01 19:40:41
|
FEniCS/dolfinx
|
https://api.github.com/repos/FEniCS/dolfinx
|
closed
|
Bump gmsh version
|
testing
|
Currently, we are not using the latest release of gmsh due to the Eigen aligment issues, #1277 .
Once we have sorted this out, we should get on to the latest version of gmsh.
With this, we should remove the download of gmsh binaries, and rather use `pip3`, as only the following is required to make gmsh installable with pip3 (tested using ubuntu:20.04 as base):
```
apt-get update
apt-get install libglu1 libxcursor-dev libxinerama1libxft2 python3-pip
pip3 install gmsh --user
export PATH=$PATH:/root/.local/bin
python3 -c "import gmsh; print(gmsh.__version__)"
```
|
1.0
|
Bump gmsh version - Currently, we are not using the latest release of gmsh due to the Eigen aligment issues, #1277 .
Once we have sorted this out, we should get on to the latest version of gmsh.
With this, we should remove the download of gmsh binaries, and rather use `pip3`, as only the following is required to make gmsh installable with pip3 (tested using ubuntu:20.04 as base):
```
apt-get update
apt-get install libglu1 libxcursor-dev libxinerama1libxft2 python3-pip
pip3 install gmsh --user
export PATH=$PATH:/root/.local/bin
python3 -c "import gmsh; print(gmsh.__version__)"
```
|
test
|
bump gmsh version currently we are not using the latest release of gmsh due to the eigen aligment issues once we have sorted this out we should get on to the latest version of gmsh with this we should remove the download of gmsh binaries and rather use as only the following is required to make gmsh installable with tested using ubuntu as base apt get update apt get install libxcursor dev pip install gmsh user export path path root local bin c import gmsh print gmsh version
| 1
|
207,284
| 15,802,467,440
|
IssuesEvent
|
2021-04-03 09:52:51
|
Level/abstract-leveldown
|
https://api.github.com/repos/Level/abstract-leveldown
|
closed
|
Remove isTypedArray from tests
|
test
|
Continuing from https://github.com/Level/abstract-leveldown/issues/121#issuecomment-343722989, in our tests we have various instances of:
https://github.com/Level/abstract-leveldown/blob/f9143c8bf5e9a4f42b0028ed612d8ede0d1fe513/test/get-test.js#L67-L69
Investigate if we still need this, because all implementations are now able to use `Buffer`, including in browsers.
|
1.0
|
Remove isTypedArray from tests - Continuing from https://github.com/Level/abstract-leveldown/issues/121#issuecomment-343722989, in our tests we have various instances of:
https://github.com/Level/abstract-leveldown/blob/f9143c8bf5e9a4f42b0028ed612d8ede0d1fe513/test/get-test.js#L67-L69
Investigate if we still need this, because all implementations are now able to use `Buffer`, including in browsers.
|
test
|
remove istypedarray from tests continuing from in our tests we have various instances of investigate if we still need this because all implementations are now able to use buffer including in browsers
| 1
|
286,117
| 31,239,893,296
|
IssuesEvent
|
2023-08-20 18:33:29
|
MValle21/lamby_site
|
https://api.github.com/repos/MValle21/lamby_site
|
opened
|
CVE-2020-23064 (Medium) detected in jquery-3.3.1.js
|
Mend: dependency security vulnerability
|
## CVE-2020-23064 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-3.3.1.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.3.1/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.3.1/jquery.js</a></p>
<p>Path to vulnerable library: /vendor/assets/javascripts/jquery-3.3.1.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-3.3.1.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/MValle21/lamby_site/commit/58d2ba7cfe9616216cb126c69803c5ccd10d32b9">58d2ba7cfe9616216cb126c69803c5ccd10d32b9</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Cross Site Scripting vulnerability in jQuery 2.2.0 through 3.x before 3.5.0 allows a remote attacker to execute arbitrary code via the <options> element.
<p>Publish Date: 2023-06-26
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-23064>CVE-2020-23064</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p>
<p>Release Date: 2023-06-26</p>
<p>Fix Resolution: jquery - 3.5.0</p>
</p>
</details>
<p></p>
|
True
|
CVE-2020-23064 (Medium) detected in jquery-3.3.1.js - ## CVE-2020-23064 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-3.3.1.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.3.1/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.3.1/jquery.js</a></p>
<p>Path to vulnerable library: /vendor/assets/javascripts/jquery-3.3.1.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-3.3.1.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/MValle21/lamby_site/commit/58d2ba7cfe9616216cb126c69803c5ccd10d32b9">58d2ba7cfe9616216cb126c69803c5ccd10d32b9</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Cross Site Scripting vulnerability in jQuery 2.2.0 through 3.x before 3.5.0 allows a remote attacker to execute arbitrary code via the <options> element.
<p>Publish Date: 2023-06-26
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-23064>CVE-2020-23064</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p>
<p>Release Date: 2023-06-26</p>
<p>Fix Resolution: jquery - 3.5.0</p>
</p>
</details>
<p></p>
|
non_test
|
cve medium detected in jquery js cve medium severity vulnerability vulnerable library jquery js javascript library for dom operations library home page a href path to vulnerable library vendor assets javascripts jquery js dependency hierarchy x jquery js vulnerable library found in head commit a href found in base branch master vulnerability details cross site scripting vulnerability in jquery through x before allows a remote attacker to execute arbitrary code via the element publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery
| 0
|
53,087
| 6,299,649,498
|
IssuesEvent
|
2017-07-21 00:00:05
|
intel-analytics/BigDL
|
https://api.github.com/repos/intel-analytics/BigDL
|
opened
|
python textclassifier doc needs update for pip install
|
0.2 release test document python
|
https://github.com/intel-analytics/BigDL/tree/master/pyspark/bigdl/models/textclassifier
only provide instructions for the manually installation, pip install users will not know where to find the bigdl home and jar files.
|
1.0
|
python textclassifier doc needs update for pip install - https://github.com/intel-analytics/BigDL/tree/master/pyspark/bigdl/models/textclassifier
only provide instructions for the manually installation, pip install users will not know where to find the bigdl home and jar files.
|
test
|
python textclassifier doc needs update for pip install only provide instructions for the manually installation pip install users will not know where to find the bigdl home and jar files
| 1
|
141,074
| 11,392,230,592
|
IssuesEvent
|
2020-01-30 01:44:49
|
ray-project/ray
|
https://api.github.com/repos/ray-project/ray
|
opened
|
test_gather_benchmark in async_test.py is flaky
|
tests failing
|
For example, see https://travis-ci.com/ray-project/ray/jobs/281702565.
The error is the following.
```
―――――――――――――――――――――――――――― test_gather_benchmark ―――――――――――――――――――――――――――――
init = None
def test_gather_benchmark(init):
@ray.remote
def f(n):
time.sleep(0.001 * n)
return 42
async def test_async():
sum_time = 0.
for _ in range(50):
tasks = [f.remote(n) for n in range(20)]
start = time.time()
futures = [async_api.as_future(obj_id) for obj_id in tasks]
await asyncio.gather(*futures)
sum_time += time.time() - start
return sum_time
def baseline():
sum_time = 0.
for _ in range(50):
tasks = [f.remote(n) for n in range(20)]
start = time.time()
ray.get(tasks)
sum_time += time.time() - start
return sum_time
# warm up
baseline()
# async get
sum_time_1 = asyncio.get_event_loop().run_until_complete(test_async())
# get
sum_time_2 = baseline()
# Ensure the new implementation is not too slow.
> assert sum_time_2 * 1.2 > sum_time_1
E assert (11.759512901306152 * 1.2) > 14.426274538040161
```
Also, a test that takes 14 seconds seems too slow. Can we shorten this test and also not test anything that depends on timing to avoid flakiness?
cc @sven1977 @suquark @simon-mo
|
1.0
|
test_gather_benchmark in async_test.py is flaky - For example, see https://travis-ci.com/ray-project/ray/jobs/281702565.
The error is the following.
```
―――――――――――――――――――――――――――― test_gather_benchmark ―――――――――――――――――――――――――――――
init = None
def test_gather_benchmark(init):
@ray.remote
def f(n):
time.sleep(0.001 * n)
return 42
async def test_async():
sum_time = 0.
for _ in range(50):
tasks = [f.remote(n) for n in range(20)]
start = time.time()
futures = [async_api.as_future(obj_id) for obj_id in tasks]
await asyncio.gather(*futures)
sum_time += time.time() - start
return sum_time
def baseline():
sum_time = 0.
for _ in range(50):
tasks = [f.remote(n) for n in range(20)]
start = time.time()
ray.get(tasks)
sum_time += time.time() - start
return sum_time
# warm up
baseline()
# async get
sum_time_1 = asyncio.get_event_loop().run_until_complete(test_async())
# get
sum_time_2 = baseline()
# Ensure the new implementation is not too slow.
> assert sum_time_2 * 1.2 > sum_time_1
E assert (11.759512901306152 * 1.2) > 14.426274538040161
```
Also, a test that takes 14 seconds seems too slow. Can we shorten this test and also not test anything that depends on timing to avoid flakiness?
cc @sven1977 @suquark @simon-mo
|
test
|
test gather benchmark in async test py is flaky for example see the error is the following ―――――――――――――――――――――――――――― test gather benchmark ――――――――――――――――――――――――――――― init none def test gather benchmark init ray remote def f n time sleep n return async def test async sum time for in range tasks start time time futures await asyncio gather futures sum time time time start return sum time def baseline sum time for in range tasks start time time ray get tasks sum time time time start return sum time warm up baseline async get sum time asyncio get event loop run until complete test async get sum time baseline ensure the new implementation is not too slow assert sum time sum time e assert also a test that takes seconds seems too slow can we shorten this test and also not test anything that depends on timing to avoid flakiness cc suquark simon mo
| 1
|
30,433
| 7,197,030,240
|
IssuesEvent
|
2018-02-05 07:13:21
|
GSA/code-gov-api
|
https://api.github.com/repos/GSA/code-gov-api
|
closed
|
Add sort functionality
|
[code.gov] [help wanted] [issue-type] enhancement
|
## Context
Sorting has not been implemented. The frontend would benefit from receiving sorted responses from the API.
There seems to be an initial attempt to implement this. Take a look at `services/searcher/index.js`
|
1.0
|
Add sort functionality - ## Context
Sorting has not been implemented. The frontend would benefit from receiving sorted responses from the API.
There seems to be an initial attempt to implement this. Take a look at `services/searcher/index.js`
|
non_test
|
add sort functionality context sorting has not been implemented the frontend would benefit from receiving sorted responses from the api there seems to be an initial attempt to implement this take a look at services searcher index js
| 0
|
221,159
| 17,295,127,443
|
IssuesEvent
|
2021-07-25 15:13:05
|
Chatterino/chatterino2
|
https://api.github.com/repos/Chatterino/chatterino2
|
closed
|
PubSub creates only 1 connection, strictly limitting amount of topics Chatterino can listen to
|
bug help wanted needs testing pubsub
|
As defined here, these are the limits of how many connections and listen requests can PubSub make:
https://github.com/Chatterino/chatterino2/blob/5ead70a314ebee8b93541e6cd6855ad48c0f42b9/src/providers/twitch/PubsubClient.hpp#L47-L48
While those do comply with [twitch docs](https://dev.twitch.tv/docs/pubsub#api-limits), we don't scale connections to 10 at max (`MAX_PUBSUB_CONNECTIONS` isn't used anywhere), making it sometimes impossible to listen to certain topics.
As a proof to my previous statement, I've added a some code ([git diff](https://haste.zneix.eu/raw/hypakuguhu)) to see how many times we actually try to listen to a PubSub topic:

Note: 100 listen attempts instead of expected 50 might be caused due to following found by @Bun

|
1.0
|
PubSub creates only 1 connection, strictly limitting amount of topics Chatterino can listen to - As defined here, these are the limits of how many connections and listen requests can PubSub make:
https://github.com/Chatterino/chatterino2/blob/5ead70a314ebee8b93541e6cd6855ad48c0f42b9/src/providers/twitch/PubsubClient.hpp#L47-L48
While those do comply with [twitch docs](https://dev.twitch.tv/docs/pubsub#api-limits), we don't scale connections to 10 at max (`MAX_PUBSUB_CONNECTIONS` isn't used anywhere), making it sometimes impossible to listen to certain topics.
As a proof to my previous statement, I've added a some code ([git diff](https://haste.zneix.eu/raw/hypakuguhu)) to see how many times we actually try to listen to a PubSub topic:

Note: 100 listen attempts instead of expected 50 might be caused due to following found by @Bun

|
test
|
pubsub creates only connection strictly limitting amount of topics chatterino can listen to as defined here these are the limits of how many connections and listen requests can pubsub make while those do comply with we don t scale connections to at max max pubsub connections isn t used anywhere making it sometimes impossible to listen to certain topics as a proof to my previous statement i ve added a some code to see how many times we actually try to listen to a pubsub topic note listen attempts instead of expected might be caused due to following found by bun
| 1
|
143,234
| 11,525,120,669
|
IssuesEvent
|
2020-02-15 05:48:50
|
mgba-emu/mgba
|
https://api.github.com/repos/mgba-emu/mgba
|
closed
|
Stuck using Qt Multimedia
|
category:needs retest severity:bug
|
Change the Audio Driver to Qt Multimedia and try to load any rom, it will stuck with a white screen.
In the second time you try to load a rom, you'll see this message:
```
using null output device, none available
```
Using qtmultimedia5-dev 5.2.1, xubuntu 14.04 amd64.
|
1.0
|
Stuck using Qt Multimedia - Change the Audio Driver to Qt Multimedia and try to load any rom, it will stuck with a white screen.
In the second time you try to load a rom, you'll see this message:
```
using null output device, none available
```
Using qtmultimedia5-dev 5.2.1, xubuntu 14.04 amd64.
|
test
|
stuck using qt multimedia change the audio driver to qt multimedia and try to load any rom it will stuck with a white screen in the second time you try to load a rom you ll see this message using null output device none available using dev xubuntu
| 1
|
148,869
| 11,869,362,500
|
IssuesEvent
|
2020-03-26 10:49:21
|
ansible/ansible
|
https://api.github.com/repos/ansible/ansible
|
closed
|
hashi_vault should report secret when an error occurs
|
affects_2.10 feature needs_triage support:community support:core test
|
<!--- Verify first that your feature was not already discussed on GitHub -->
<!--- Complete *all* sections as described, this form is processed automatically -->
##### SUMMARY
<!--- Describe the new feature/improvement briefly below -->
Currently, `hashi_vault` only reports a very basic exception when it can't access a secret. On large playbooks, where many secrets are accessed, this gives no indication which secret was failing.
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
<!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure -->
hashi_vault
##### ADDITIONAL INFORMATION
<!--- Describe how the feature would be used, why it is needed and what it would solve -->
The error reported by `hashi_vault` currently looks like this:
<!--- Paste example playbooks or commands between quotes below -->
```
failed: [localhost] (item=example) => {
"changed": false,
"item": "example",
"msg": "AnsibleError: An unhandled exception occurred while running the lookup plugin 'hashi_vault'. Error was a <class 'hvac.exceptions.Forbidden'>, original message: 1 error occurred:\n\t* permission denied\n\n"
}
```
This should be more like:
```
failed: [localhost] (item=example) => {
"changed": false,
"item": "example",
"msg": "AnsibleError: An unhandled exception occurred while running the lookup plugin 'hashi_vault'. Error was a <class 'hvac.exceptions.Forbidden'>, original message: 1 error occurred:\n\t* permission denied on secret secret=secret/example:secret\n\n"
}
```
Or, the hvac.exceptions.Forbidden exception should be caught by the module and handled with a proper error message.
<!--- HINT: You can also paste gist.github.com links for larger files -->
|
1.0
|
hashi_vault should report secret when an error occurs - <!--- Verify first that your feature was not already discussed on GitHub -->
<!--- Complete *all* sections as described, this form is processed automatically -->
##### SUMMARY
<!--- Describe the new feature/improvement briefly below -->
Currently, `hashi_vault` only reports a very basic exception when it can't access a secret. On large playbooks, where many secrets are accessed, this gives no indication which secret was failing.
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
<!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure -->
hashi_vault
##### ADDITIONAL INFORMATION
<!--- Describe how the feature would be used, why it is needed and what it would solve -->
The error reported by `hashi_vault` currently looks like this:
<!--- Paste example playbooks or commands between quotes below -->
```
failed: [localhost] (item=example) => {
"changed": false,
"item": "example",
"msg": "AnsibleError: An unhandled exception occurred while running the lookup plugin 'hashi_vault'. Error was a <class 'hvac.exceptions.Forbidden'>, original message: 1 error occurred:\n\t* permission denied\n\n"
}
```
This should be more like:
```
failed: [localhost] (item=example) => {
"changed": false,
"item": "example",
"msg": "AnsibleError: An unhandled exception occurred while running the lookup plugin 'hashi_vault'. Error was a <class 'hvac.exceptions.Forbidden'>, original message: 1 error occurred:\n\t* permission denied on secret secret=secret/example:secret\n\n"
}
```
Or, the hvac.exceptions.Forbidden exception should be caught by the module and handled with a proper error message.
<!--- HINT: You can also paste gist.github.com links for larger files -->
|
test
|
hashi vault should report secret when an error occurs summary currently hashi vault only reports a very basic exception when it can t access a secret on large playbooks where many secrets are accessed this gives no indication which secret was failing issue type feature idea component name hashi vault additional information the error reported by hashi vault currently looks like this failed item example changed false item example msg ansibleerror an unhandled exception occurred while running the lookup plugin hashi vault error was a original message error occurred n t permission denied n n this should be more like failed item example changed false item example msg ansibleerror an unhandled exception occurred while running the lookup plugin hashi vault error was a original message error occurred n t permission denied on secret secret secret example secret n n or the hvac exceptions forbidden exception should be caught by the module and handled with a proper error message
| 1
|
262,381
| 22,837,578,219
|
IssuesEvent
|
2022-07-12 18:13:49
|
metaplex-foundation/metaplex
|
https://api.github.com/repos/metaplex-foundation/metaplex
|
closed
|
[Bug]: CandyMachineUI InvalidRentPayingAccount
|
needs tests bug
|
### Which package is this bug report for?
candy machine ui
### Issue description
Deployed a collection of 1 NFT to mainnet and it works as it should.
I deployed the full collection of 1200 NFTs and when I try to mint I get an error
`InvalidRentPayingAccount` when awaitConfirmationSignature is called. Seems to fail on the Set and Verify Collection instruction.
Anyone run into this before ?
### Command
_No response_
### Relevant log output
```shell
From the console of candymachineui (running latest)
REST null result for jnt4qwRgHRRntsJKiwJD8dueztKgowGjoT5wfiMTZgccPXoux3EMEUWC4W9TpLLqV9avNezEWdVuNxpNFkvkqLN null
connection.tsx:496 Rejected via websocket InvalidRentPayingAccount
index.js:1 Timeout Error caught {err: 'InvalidRentPayingAccount', slot: 139694285, confirmations: 0}
Caught failure: Error: Transaction failed: Instruction: Set and Verify Collection
```
```
### Operating system
Linux
### Priority this issue should have
Medium (should be fixed soon)
### Check the Docs First
- [X] I have checked the docs and it didn't solve my issue
|
1.0
|
[Bug]: CandyMachineUI InvalidRentPayingAccount - ### Which package is this bug report for?
candy machine ui
### Issue description
Deployed a collection of 1 NFT to mainnet and it works as it should.
I deployed the full collection of 1200 NFTs and when I try to mint I get an error
`InvalidRentPayingAccount` when awaitConfirmationSignature is called. Seems to fail on the Set and Verify Collection instruction.
Anyone run into this before ?
### Command
_No response_
### Relevant log output
```shell
From the console of candymachineui (running latest)
REST null result for jnt4qwRgHRRntsJKiwJD8dueztKgowGjoT5wfiMTZgccPXoux3EMEUWC4W9TpLLqV9avNezEWdVuNxpNFkvkqLN null
connection.tsx:496 Rejected via websocket InvalidRentPayingAccount
index.js:1 Timeout Error caught {err: 'InvalidRentPayingAccount', slot: 139694285, confirmations: 0}
Caught failure: Error: Transaction failed: Instruction: Set and Verify Collection
```
```
### Operating system
Linux
### Priority this issue should have
Medium (should be fixed soon)
### Check the Docs First
- [X] I have checked the docs and it didn't solve my issue
|
test
|
candymachineui invalidrentpayingaccount which package is this bug report for candy machine ui issue description deployed a collection of nft to mainnet and it works as it should i deployed the full collection of nfts and when i try to mint i get an error invalidrentpayingaccount when awaitconfirmationsignature is called seems to fail on the set and verify collection instruction anyone run into this before command no response relevant log output shell from the console of candymachineui running latest rest null result for null connection tsx rejected via websocket invalidrentpayingaccount index js timeout error caught err invalidrentpayingaccount slot confirmations caught failure error transaction failed instruction set and verify collection operating system linux priority this issue should have medium should be fixed soon check the docs first i have checked the docs and it didn t solve my issue
| 1
|
240,684
| 20,070,485,120
|
IssuesEvent
|
2022-02-04 05:48:40
|
StollLab/EasySpin
|
https://api.github.com/repos/StollLab/EasySpin
|
closed
|
Add tests that compare DStrain etc to explicit averaging
|
test
|
The goal is to assert that `Sys.DStrain` (and other strain parameters) do exactly what they advertise.
|
1.0
|
Add tests that compare DStrain etc to explicit averaging - The goal is to assert that `Sys.DStrain` (and other strain parameters) do exactly what they advertise.
|
test
|
add tests that compare dstrain etc to explicit averaging the goal is to assert that sys dstrain and other strain parameters do exactly what they advertise
| 1
|
202,353
| 15,281,345,502
|
IssuesEvent
|
2021-02-23 07:59:09
|
YM-rakennettu-ymparisto/kaavatietomalli
|
https://api.github.com/repos/YM-rakennettu-ymparisto/kaavatietomalli
|
closed
|
Espoon ruutuyleiskaavan digitointi
|
Testaus
|
Digitoidaan Espoon ruutuyleiskaavaa tietomallimuotoon. Tätä ei kuitenkaan viedä tietokantaan tai julkaista rajapinnassa, koska kaavaa ei voi saattaa ladattavaan muotoon tässä vaiheessa kaavoitusprosessia. Matala prioriteetti, koska rajoituksen takia vain rajattu hyöty testaukselle
|
1.0
|
Espoon ruutuyleiskaavan digitointi - Digitoidaan Espoon ruutuyleiskaavaa tietomallimuotoon. Tätä ei kuitenkaan viedä tietokantaan tai julkaista rajapinnassa, koska kaavaa ei voi saattaa ladattavaan muotoon tässä vaiheessa kaavoitusprosessia. Matala prioriteetti, koska rajoituksen takia vain rajattu hyöty testaukselle
|
test
|
espoon ruutuyleiskaavan digitointi digitoidaan espoon ruutuyleiskaavaa tietomallimuotoon tätä ei kuitenkaan viedä tietokantaan tai julkaista rajapinnassa koska kaavaa ei voi saattaa ladattavaan muotoon tässä vaiheessa kaavoitusprosessia matala prioriteetti koska rajoituksen takia vain rajattu hyöty testaukselle
| 1
|
120,762
| 12,093,722,101
|
IssuesEvent
|
2020-04-19 20:47:58
|
insites-co/insites-ui
|
https://api.github.com/repos/insites-co/insites-ui
|
closed
|
Radio and Checkbox docs examples
|
bug documentation good first issue help wanted
|
Currently our Radio and Checkbox examples are broken. The fix would be pretty easy - to move them to a separate components in Documentation repo and provide `useState` to them. I'd really appreciate some help here.
https://insites-ui-docs.netlify.app/form/checkbox/
https://insites-ui-docs.netlify.app/form/radio/
https://github.com/insites-co/insites-ui-docs/blob/master/content/form/checkbox.mdx
https://github.com/insites-co/insites-ui-docs/blob/master/content/form/radio.mdx
|
1.0
|
Radio and Checkbox docs examples - Currently our Radio and Checkbox examples are broken. The fix would be pretty easy - to move them to a separate components in Documentation repo and provide `useState` to them. I'd really appreciate some help here.
https://insites-ui-docs.netlify.app/form/checkbox/
https://insites-ui-docs.netlify.app/form/radio/
https://github.com/insites-co/insites-ui-docs/blob/master/content/form/checkbox.mdx
https://github.com/insites-co/insites-ui-docs/blob/master/content/form/radio.mdx
|
non_test
|
radio and checkbox docs examples currently our radio and checkbox examples are broken the fix would be pretty easy to move them to a separate components in documentation repo and provide usestate to them i d really appreciate some help here
| 0
|
299,660
| 25,916,269,860
|
IssuesEvent
|
2022-12-15 17:37:27
|
worMatty/deathrun_toolkit
|
https://api.github.com/repos/worMatty/deathrun_toolkit
|
closed
|
[Bug] Activator health scaling doesn't work sometimes
|
bug needs testing
|
When used on command it does not work.
When triggered by the map it only seems to work on next round.
|
1.0
|
[Bug] Activator health scaling doesn't work sometimes - When used on command it does not work.
When triggered by the map it only seems to work on next round.
|
test
|
activator health scaling doesn t work sometimes when used on command it does not work when triggered by the map it only seems to work on next round
| 1
|
8,262
| 5,540,592,182
|
IssuesEvent
|
2017-03-22 10:31:19
|
ESPD/ESPD-Service
|
https://api.github.com/repos/ESPD/ESPD-Service
|
closed
|
popup for sessions longer than 30 minutes and change the workflow
|
enhancement usability
|
1. Provide a popup after 30 Minutes saying that: "This service has not been used for more than 45 Minutes. Do you want to save the ESPD?" The drawback of this is that of the user is having lunch he will come back to late and the data might have been gone already.
2. Change workflow: The most elegant way is probably to change the workflow. On the finish page we do not allow to save the "XML" immediately but to say "Next". They will then be presented an overview. At the bottom we would offer to buttons "Save as PDF" and "Save as XML".
|
True
|
popup for sessions longer than 30 minutes and change the workflow - 1. Provide a popup after 30 Minutes saying that: "This service has not been used for more than 45 Minutes. Do you want to save the ESPD?" The drawback of this is that of the user is having lunch he will come back to late and the data might have been gone already.
2. Change workflow: The most elegant way is probably to change the workflow. On the finish page we do not allow to save the "XML" immediately but to say "Next". They will then be presented an overview. At the bottom we would offer to buttons "Save as PDF" and "Save as XML".
|
non_test
|
popup for sessions longer than minutes and change the workflow provide a popup after minutes saying that this service has not been used for more than minutes do you want to save the espd the drawback of this is that of the user is having lunch he will come back to late and the data might have been gone already change workflow the most elegant way is probably to change the workflow on the finish page we do not allow to save the xml immediately but to say next they will then be presented an overview at the bottom we would offer to buttons save as pdf and save as xml
| 0
|
327,814
| 9,981,610,280
|
IssuesEvent
|
2019-07-10 07:56:14
|
StrangeLoopGames/EcoIssues
|
https://api.github.com/repos/StrangeLoopGames/EcoIssues
|
closed
|
[0.8.2.7] Can't change authorization settings on wooden elevator call post.
|
Fixed Medium Priority QA Verified
|
From version 0.8.2.0 it has become possible to give access to the Elevator to another player. But you cann't give acces to its additional buttons.
It's necessary to add the possibility of authorization in Wooden Elevator Call Post. Or that the buttons receive authorization rights from the Elevator. If I give access to the Elevator, then automatically I give access to the buttons attached to it.
|
1.0
|
[0.8.2.7] Can't change authorization settings on wooden elevator call post. - From version 0.8.2.0 it has become possible to give access to the Elevator to another player. But you cann't give acces to its additional buttons.
It's necessary to add the possibility of authorization in Wooden Elevator Call Post. Or that the buttons receive authorization rights from the Elevator. If I give access to the Elevator, then automatically I give access to the buttons attached to it.
|
non_test
|
can t change authorization settings on wooden elevator call post from version it has become possible to give access to the elevator to another player but you cann t give acces to its additional buttons it s necessary to add the possibility of authorization in wooden elevator call post or that the buttons receive authorization rights from the elevator if i give access to the elevator then automatically i give access to the buttons attached to it
| 0
|
215,437
| 16,671,639,567
|
IssuesEvent
|
2021-06-07 11:39:53
|
WPChill/download-monitor
|
https://api.github.com/repos/WPChill/download-monitor
|
closed
|
On archives and other places where get_post_permalink is used the permalink is not OK
|
needs testing
|
Should filter the return with the following:
`add_filter( 'post_type_link', array( $this, 'dlm_archive_permalink_structure' ), 15, 2 );`
function something( $post_link, $post ) {
if ( 'dlm_download' == $post->post_type ) {
// fetch download object
try{
/** @var DLM_Download $download */
$download = download_monitor()->service( 'download_repository' )->retrieve_single( $post->ID );
return $download->get_the_download_link();
}
catch ( Exception $e ){
}
}
return $post_link;
}
|
1.0
|
On archives and other places where get_post_permalink is used the permalink is not OK - Should filter the return with the following:
`add_filter( 'post_type_link', array( $this, 'dlm_archive_permalink_structure' ), 15, 2 );`
function something( $post_link, $post ) {
if ( 'dlm_download' == $post->post_type ) {
// fetch download object
try{
/** @var DLM_Download $download */
$download = download_monitor()->service( 'download_repository' )->retrieve_single( $post->ID );
return $download->get_the_download_link();
}
catch ( Exception $e ){
}
}
return $post_link;
}
|
test
|
on archives and other places where get post permalink is used the permalink is not ok should filter the return with the following add filter post type link array this dlm archive permalink structure function something post link post if dlm download post post type fetch download object try var dlm download download download download monitor service download repository retrieve single post id return download get the download link catch exception e return post link
| 1
|
98,956
| 20,827,599,660
|
IssuesEvent
|
2022-03-19 00:02:34
|
pulumi/pulumi
|
https://api.github.com/repos/pulumi/pulumi
|
closed
|
[codegen/go] Add support for specifying Go dependencies
|
kind/enhancement language/go area/codegen
|
## Hello!
<!-- Please leave this section as-is, it's designed to help others in the community know how to interact with our GitHub issues. -->
- Vote on this issue by adding a 👍 reaction
- If you want to implement this feature, comment to let us know (we'll work with you on design, scheduling, etc.)
## Issue details
There are situations where an MLC provider may make use of other resource providers inside its implementation (e.g. `pulumi-random`, `pulumi-docker`, etc.), but not exposed publicly. Such dependencies _aren't_ listed in the provider's associated Go SDK's `go.mod` since nothing in the SDK itself refers to them. Which can be problematic for plugin discovery. Since these dependencies are not listed in the Go SDK's `go.mod`, the associated plugins will not be installed automatically for users of the Go SDK.
Note that the Go SDK codegen does not generate the `go.mod` file -- this must be created manually. And manually adding such dependencies to the SDK's `go.mod` is not sufficient to address this because `go mod tidy` will remove any unused dependencies. To avoid this, we should add a new capability to the Go codegen that emits internal dummy imports (referencing some well known exported type) for any Go dependencies listed in the schema. That way, the dependencies can be manually added to `go.mod` and won't be stripped when `go mod tidy` is run.
|
1.0
|
[codegen/go] Add support for specifying Go dependencies - ## Hello!
<!-- Please leave this section as-is, it's designed to help others in the community know how to interact with our GitHub issues. -->
- Vote on this issue by adding a 👍 reaction
- If you want to implement this feature, comment to let us know (we'll work with you on design, scheduling, etc.)
## Issue details
There are situations where an MLC provider may make use of other resource providers inside its implementation (e.g. `pulumi-random`, `pulumi-docker`, etc.), but not exposed publicly. Such dependencies _aren't_ listed in the provider's associated Go SDK's `go.mod` since nothing in the SDK itself refers to them. Which can be problematic for plugin discovery. Since these dependencies are not listed in the Go SDK's `go.mod`, the associated plugins will not be installed automatically for users of the Go SDK.
Note that the Go SDK codegen does not generate the `go.mod` file -- this must be created manually. And manually adding such dependencies to the SDK's `go.mod` is not sufficient to address this because `go mod tidy` will remove any unused dependencies. To avoid this, we should add a new capability to the Go codegen that emits internal dummy imports (referencing some well known exported type) for any Go dependencies listed in the schema. That way, the dependencies can be manually added to `go.mod` and won't be stripped when `go mod tidy` is run.
|
non_test
|
add support for specifying go dependencies hello vote on this issue by adding a 👍 reaction if you want to implement this feature comment to let us know we ll work with you on design scheduling etc issue details there are situations where an mlc provider may make use of other resource providers inside its implementation e g pulumi random pulumi docker etc but not exposed publicly such dependencies aren t listed in the provider s associated go sdk s go mod since nothing in the sdk itself refers to them which can be problematic for plugin discovery since these dependencies are not listed in the go sdk s go mod the associated plugins will not be installed automatically for users of the go sdk note that the go sdk codegen does not generate the go mod file this must be created manually and manually adding such dependencies to the sdk s go mod is not sufficient to address this because go mod tidy will remove any unused dependencies to avoid this we should add a new capability to the go codegen that emits internal dummy imports referencing some well known exported type for any go dependencies listed in the schema that way the dependencies can be manually added to go mod and won t be stripped when go mod tidy is run
| 0
|
40,524
| 10,545,604,739
|
IssuesEvent
|
2019-10-02 19:31:31
|
zowe/zowe-install-packaging
|
https://api.github.com/repos/zowe/zowe-install-packaging
|
closed
|
Write a tool to clean up artifactory history
|
CI/CD build
|
Need to clean up branch/PR builds regularly to free up storage.
|
1.0
|
Write a tool to clean up artifactory history - Need to clean up branch/PR builds regularly to free up storage.
|
non_test
|
write a tool to clean up artifactory history need to clean up branch pr builds regularly to free up storage
| 0
|
65,376
| 6,961,596,123
|
IssuesEvent
|
2017-12-08 10:05:51
|
radare/radare2
|
https://api.github.com/repos/radare/radare2
|
closed
|
anal.noncode is not respected
|
anal bug has-test
|
Following the new broken tests for anal.calls/anal.noncode -> https://github.com/radare/radare2-regressions/commit/8a22c7323815c85fd7e07230924e461f21aa2ab1 .

|
1.0
|
anal.noncode is not respected - Following the new broken tests for anal.calls/anal.noncode -> https://github.com/radare/radare2-regressions/commit/8a22c7323815c85fd7e07230924e461f21aa2ab1 .

|
test
|
anal noncode is not respected following the new broken tests for anal calls anal noncode
| 1
|
122,041
| 12,139,823,331
|
IssuesEvent
|
2020-04-23 19:29:39
|
amanzi/amanzi
|
https://api.github.com/repos/amanzi/amanzi
|
closed
|
Input Spec/Multiple point definitions
|
documentation
|
**typedef_point** contains an attribute "coordinate" and is used for the region type point as a typedef_labeled_point (typedef_point with additional attribute "name"):
`<point name="abc" coordinate="(x, y, z)"/>`
Points are also defined as elements under polygons as direct coordinates:
<xs:complexType name="typedef_polygonal_surface">
<xs:sequence>
<xs:element name="point" type="typedef_coordinate" minOccurs="0" maxOccurs="unbounded" /> ...
`<point>(x, y, z)</point>`
Validating a file with a polygon currently gives schema validation error. Recommending we change the polygon point type to typedef_point for consistency:
`<point coordinate="(x, y, z)"/>`
|
1.0
|
Input Spec/Multiple point definitions - **typedef_point** contains an attribute "coordinate" and is used for the region type point as a typedef_labeled_point (typedef_point with additional attribute "name"):
`<point name="abc" coordinate="(x, y, z)"/>`
Points are also defined as elements under polygons as direct coordinates:
<xs:complexType name="typedef_polygonal_surface">
<xs:sequence>
<xs:element name="point" type="typedef_coordinate" minOccurs="0" maxOccurs="unbounded" /> ...
`<point>(x, y, z)</point>`
Validating a file with a polygon currently gives schema validation error. Recommending we change the polygon point type to typedef_point for consistency:
`<point coordinate="(x, y, z)"/>`
|
non_test
|
input spec multiple point definitions typedef point contains an attribute coordinate and is used for the region type point as a typedef labeled point typedef point with additional attribute name points are also defined as elements under polygons as direct coordinates x y z validating a file with a polygon currently gives schema validation error recommending we change the polygon point type to typedef point for consistency
| 0
|
757,687
| 26,524,509,622
|
IssuesEvent
|
2023-01-19 07:26:43
|
etjump/etjump
|
https://api.github.com/repos/etjump/etjump
|
opened
|
Expose g_save, g_ghostPlayers and g_nofatigue to vote system
|
Feature Priority: Normal
|
Players should be able to vote on these settings.
|
1.0
|
Expose g_save, g_ghostPlayers and g_nofatigue to vote system - Players should be able to vote on these settings.
|
non_test
|
expose g save g ghostplayers and g nofatigue to vote system players should be able to vote on these settings
| 0
|
541,969
| 15,836,760,477
|
IssuesEvent
|
2021-04-06 19:48:04
|
GoogleChrome/lighthouse
|
https://api.github.com/repos/GoogleChrome/lighthouse
|
closed
|
Warnings: Lighthouse was unable to automatically check the `font-display` value for the origin https://collegeadmission.co. Please help me for this issue fix it...
|
needs-more-info needs-priority
|
Warnings: Lighthouse was unable to automatically check the `font-display` value for the origin https://collegeadmission.co. Please help me with this issue to fix it...
Already checking with...Below Given...
link rel="preload" as="style" href="https://fonts.googleapis.com/css2?family=Open+Sans&display=block
link rel="preload" as="font" href="https://fonts.gstatic.com/s/opensans/v18/mem8YaGs126MiZpBA-UFVZ0bf8pkAg.woff2" type="font/woff" crossorigin="anonymous"
@font-face {
font-family:'Open Sans';
src: url(https://fonts.gstatic.com/s/opensans/v18/mem8YaGs126MiZpBA-UFVZ0bf8pkAg.woff2) format("woff2");
font-display: auto;
}

|
1.0
|
Warnings: Lighthouse was unable to automatically check the `font-display` value for the origin https://collegeadmission.co. Please help me for this issue fix it... - Warnings: Lighthouse was unable to automatically check the `font-display` value for the origin https://collegeadmission.co. Please help me with this issue to fix it...
Already checking with...Below Given...
link rel="preload" as="style" href="https://fonts.googleapis.com/css2?family=Open+Sans&display=block
link rel="preload" as="font" href="https://fonts.gstatic.com/s/opensans/v18/mem8YaGs126MiZpBA-UFVZ0bf8pkAg.woff2" type="font/woff" crossorigin="anonymous"
@font-face {
font-family:'Open Sans';
src: url(https://fonts.gstatic.com/s/opensans/v18/mem8YaGs126MiZpBA-UFVZ0bf8pkAg.woff2) format("woff2");
font-display: auto;
}

|
non_test
|
warnings lighthouse was unable to automatically check the font display value for the origin please help me for this issue fix it warnings lighthouse was unable to automatically check the font display value for the origin please help me with this issue to fix it already checking with below given link rel preload as style href link rel preload as font href type font woff crossorigin anonymous font face font family open sans src url format font display auto
| 0
|
82,943
| 7,857,118,365
|
IssuesEvent
|
2018-06-21 09:44:35
|
brave/browser-laptop
|
https://api.github.com/repos/brave/browser-laptop
|
closed
|
Prompt for upgrade to regular build if ads are off for extended period
|
initiative/bat-ads initiative/bat-ads/ads-test release/blocking
|
The ads test build isn't up to date with other security and privacy features in Brave or Muon. Nobody should use this build unless they're actively participating in the ads test. If someone uses this build for an hour without turning on the ads feature, they should be prompted to "upgrade" to the latest current release build of Brave, by way of a notification bar — just like normal software upgrades. cf., https://github.com/brave-intl/internal/issues/57
|
1.0
|
Prompt for upgrade to regular build if ads are off for extended period - The ads test build isn't up to date with other security and privacy features in Brave or Muon. Nobody should use this build unless they're actively participating in the ads test. If someone uses this build for an hour without turning on the ads feature, they should be prompted to "upgrade" to the latest current release build of Brave, by way of a notification bar — just like normal software upgrades. cf., https://github.com/brave-intl/internal/issues/57
|
test
|
prompt for upgrade to regular build if ads are off for extended period the ads test build isn t up to date with other security and privacy features in brave or muon nobody should use this build unless they re actively participating in the ads test if someone uses this build for an hour without turning on the ads feature they should be prompted to upgrade to the latest current release build of brave by way of a notification bar — just like normal software upgrades cf
| 1
|
19,706
| 2,622,161,395
|
IssuesEvent
|
2015-03-04 00:10:13
|
byzhang/terrastore
|
https://api.github.com/repos/byzhang/terrastore
|
opened
|
Improve underlying http stack
|
auto-migrated Priority-Medium Project-Terrastore Type-Enhancement
|
```
improve underlying http sever/client stack
see comarison agains apache
http://groups.google.com/group/terrastore-discussions/browse_thread/thread/d2cf6
58e9dd01036
```
Original issue reported on code.google.com by `Andrei.Pozolotin` on 10 Mar 2011 at 4:18
|
1.0
|
Improve underlying http stack - ```
improve underlying http sever/client stack
see comarison agains apache
http://groups.google.com/group/terrastore-discussions/browse_thread/thread/d2cf6
58e9dd01036
```
Original issue reported on code.google.com by `Andrei.Pozolotin` on 10 Mar 2011 at 4:18
|
non_test
|
improve underlying http stack improve underlying http sever client stack see comarison agains apache original issue reported on code google com by andrei pozolotin on mar at
| 0
|
208,279
| 16,108,666,311
|
IssuesEvent
|
2021-04-27 18:03:02
|
spring-io/spring-asciidoctor-backends
|
https://api.github.com/repos/spring-io/spring-asciidoctor-backends
|
closed
|
Wrong Spring Maven repo in README.adoc
|
type: documentation
|
README.adoc says:
```
maven {
url "https://repo.spring.io/release"
}
```
but must be:
```
maven {
url "https://repo.spring.io/milestone"
}
```
since there are only milestone releases for now.
|
1.0
|
Wrong Spring Maven repo in README.adoc - README.adoc says:
```
maven {
url "https://repo.spring.io/release"
}
```
but must be:
```
maven {
url "https://repo.spring.io/milestone"
}
```
since there are only milestone releases for now.
|
non_test
|
wrong spring maven repo in readme adoc readme adoc says maven url but must be maven url since there are only milestone releases for now
| 0
|
209,231
| 23,705,172,307
|
IssuesEvent
|
2022-08-29 23:49:44
|
itziklavon/springdemo
|
https://api.github.com/repos/itziklavon/springdemo
|
closed
|
swagger-jersey-jaxrs-1.5.21.jar: 1 vulnerabilities (highest severity is: 7.5)
|
security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>swagger-jersey-jaxrs-1.5.21.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/itziklavon/springdemo/commit/98e4930dff8ac42633a041eaf411f0caf6a38b4c">98e4930dff8ac42633a041eaf411f0caf6a38b4c</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2017-18640](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-18640) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | snakeyaml-1.19.jar | Transitive | 1.6.0 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2017-18640</summary>
### Vulnerable Library - <b>snakeyaml-1.19.jar</b></p>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar</p>
<p>
Dependency Hierarchy:
- swagger-jersey-jaxrs-1.5.21.jar (Root Library)
- swagger-jaxrs-1.5.21.jar
- swagger-core-1.5.21.jar
- jackson-dataformat-yaml-2.9.6.jar
- :x: **snakeyaml-1.19.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/itziklavon/springdemo/commit/98e4930dff8ac42633a041eaf411f0caf6a38b4c">98e4930dff8ac42633a041eaf411f0caf6a38b4c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The Alias feature in SnakeYAML before 1.26 allows entity expansion during a load operation, a related issue to CVE-2003-1564.
<p>Publish Date: 2019-12-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-18640>CVE-2017-18640</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-18640">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-18640</a></p>
<p>Release Date: 2019-12-12</p>
<p>Fix Resolution (org.yaml:snakeyaml): 1.26</p>
<p>Direct dependency fix Resolution (io.swagger:swagger-jersey-jaxrs): 1.6.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
True
|
swagger-jersey-jaxrs-1.5.21.jar: 1 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>swagger-jersey-jaxrs-1.5.21.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/itziklavon/springdemo/commit/98e4930dff8ac42633a041eaf411f0caf6a38b4c">98e4930dff8ac42633a041eaf411f0caf6a38b4c</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2017-18640](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-18640) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | snakeyaml-1.19.jar | Transitive | 1.6.0 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2017-18640</summary>
### Vulnerable Library - <b>snakeyaml-1.19.jar</b></p>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar</p>
<p>
Dependency Hierarchy:
- swagger-jersey-jaxrs-1.5.21.jar (Root Library)
- swagger-jaxrs-1.5.21.jar
- swagger-core-1.5.21.jar
- jackson-dataformat-yaml-2.9.6.jar
- :x: **snakeyaml-1.19.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/itziklavon/springdemo/commit/98e4930dff8ac42633a041eaf411f0caf6a38b4c">98e4930dff8ac42633a041eaf411f0caf6a38b4c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The Alias feature in SnakeYAML before 1.26 allows entity expansion during a load operation, a related issue to CVE-2003-1564.
<p>Publish Date: 2019-12-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-18640>CVE-2017-18640</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-18640">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-18640</a></p>
<p>Release Date: 2019-12-12</p>
<p>Fix Resolution (org.yaml:snakeyaml): 1.26</p>
<p>Direct dependency fix Resolution (io.swagger:swagger-jersey-jaxrs): 1.6.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
non_test
|
swagger jersey jaxrs jar vulnerabilities highest severity is vulnerable library swagger jersey jaxrs jar path to dependency file pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high snakeyaml jar transitive details cve vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar dependency hierarchy swagger jersey jaxrs jar root library swagger jaxrs jar swagger core jar jackson dataformat yaml jar x snakeyaml jar vulnerable library found in head commit a href found in base branch master vulnerability details the alias feature in snakeyaml before allows entity expansion during a load operation a related issue to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org yaml snakeyaml direct dependency fix resolution io swagger swagger jersey jaxrs step up your open source security game with mend
| 0
|
280,363
| 8,680,840,699
|
IssuesEvent
|
2018-12-01 14:42:04
|
firecracker-microvm/firecracker
|
https://api.github.com/repos/firecracker-microvm/firecracker
|
closed
|
Remove DeviceState from NetworkInterfaceConfig
|
Contribute: Good First Issue Priority: High Quality: Improvement
|
The device state is a leftover from our previous device model, when we wanted to be able to hot plug/unplug devices. This is not something that we are planing to implement in the near future and should be remove.
The DeviceState structure is defined in `vmm/src/vmm_config/mod.rs` and it is used in the NetworkInterfaceConfig structure definition in `vmm/src/vmm_config/net.rs`.
With this change, the swagger definition (api_server/swagger/firecracker.yaml) also needs to be updated.
|
1.0
|
Remove DeviceState from NetworkInterfaceConfig - The device state is a leftover from our previous device model, when we wanted to be able to hot plug/unplug devices. This is not something that we are planing to implement in the near future and should be remove.
The DeviceState structure is defined in `vmm/src/vmm_config/mod.rs` and it is used in the NetworkInterfaceConfig structure definition in `vmm/src/vmm_config/net.rs`.
With this change, the swagger definition (api_server/swagger/firecracker.yaml) also needs to be updated.
|
non_test
|
remove devicestate from networkinterfaceconfig the device state is a leftover from our previous device model when we wanted to be able to hot plug unplug devices this is not something that we are planing to implement in the near future and should be remove the devicestate structure is defined in vmm src vmm config mod rs and it is used in the networkinterfaceconfig structure definition in vmm src vmm config net rs with this change the swagger definition api server swagger firecracker yaml also needs to be updated
| 0
|
85,758
| 7,988,977,256
|
IssuesEvent
|
2018-07-19 12:56:02
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
github.com/cockroachdb/cockroach/pkg/storage: TestStoreRangeSystemSplits failed under stress
|
A-core-kv C-test-failure O-robot
|
SHA: https://github.com/cockroachdb/cockroach/commits/97c3622c2a6f4ec579ed93655b9d8a83732ac654
Parameters:
```
TAGS=
GOFLAGS=-race
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=693452&tab=buildLog
```
I180602 07:21:25.666097 111393 gossip/gossip.go:333 [n1] NodeDescriptor set to node_id:1 address:<network_field:"" address_field:"" > attrs:<> locality:<> ServerVersion:<major_val:0 minor_val:0 patch:0 unstable:0 >
I180602 07:21:25.755244 111393 storage/store.go:1454 [s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available
I180602 07:21:25.822572 111432 storage/replica_command.go:863 [split,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2]
W180602 07:21:25.865205 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
I180602 07:21:25.915992 111562 storage/replica_command.go:863 [split,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3]
W180602 07:21:25.934571 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:25.938941 111461 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "unnamed" id=17deafae key=/Table/SystemConfigSpan/Start rw=true pri=0.03184783 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,53 orig=0.000000123,53 max=0.000000123,54 wto=false rop=false seq=1
I180602 07:21:26.066952 111572 storage/replica_command.go:863 [split,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4]
W180602 07:21:26.094922 111464 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "unnamed" id=17deafae key=/Table/SystemConfigSpan/Start rw=true pri=0.03184783 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,53 orig=0.000000123,53 max=0.000000123,54 wto=false rop=false seq=1
W180602 07:21:26.143966 111457 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=46177098 key=/Local/Range/System/NodeLiveness/RangeDescriptor rw=true pri=0.06770799 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,182 orig=0.000000123,182 max=0.000000123,196 wto=false rop=false seq=1
W180602 07:21:26.171835 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
I180602 07:21:26.190567 111466 storage/replica_command.go:863 [split,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5]
W180602 07:21:26.200246 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:26.223842 111587 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "unnamed" id=17deafae key=/Table/SystemConfigSpan/Start rw=true pri=0.03184783 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,53 orig=0.000000123,53 max=0.000000123,54 wto=false rop=false seq=1
W180602 07:21:26.259843 103909 gossip/gossip.go:1298 [n1] first range unavailable; trying remaining resolvers
I180602 07:21:26.262515 111416 gossip/client.go:129 [n1] started gossip client to 127.0.0.1:44307
I180602 07:21:26.304909 111402 storage/replica_command.go:863 [split,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6]
W180602 07:21:26.307892 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer
W180602 07:21:26.340668 111552 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "unnamed" id=17deafae key=/Table/SystemConfigSpan/Start rw=true pri=0.03184783 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,53 orig=0.000000123,53 max=0.000000123,54 wto=false rop=false seq=1
I180602 07:21:26.413526 111603 storage/replica_command.go:863 [split,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7]
W180602 07:21:26.434307 111573 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "unnamed" id=17deafae key=/Table/SystemConfigSpan/Start rw=true pri=0.03184783 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,53 orig=0.000000123,53 max=0.000000123,54 wto=false rop=false seq=1
I180602 07:21:26.489843 111577 storage/replica_command.go:863 [split,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8]
I180602 07:21:26.601373 111652 storage/replica_command.go:863 [split,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9]
I180602 07:21:26.700193 111406 storage/replica_command.go:863 [split,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10]
W180602 07:21:26.748760 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
W180602 07:21:26.761617 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer
E180602 07:21:26.803145 111637 storage/consistency_queue.go:125 [replica consistency checker,s1,r1/1:/{Min-System/}] computing own checksum: could not resolve node ID 1: dummy resolver
E180602 07:21:26.803792 111637 storage/queue.go:778 [replica consistency checker,s1,r1/1:/{Min-System/}] computing own checksum: could not resolve node ID 1: dummy resolver
W180602 07:21:26.812209 104013 gossip/gossip.go:1298 [n2] first range unavailable; trying remaining resolvers
I180602 07:21:26.818283 111638 storage/replica_command.go:863 [split,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11]
I180602 07:21:26.918273 111580 storage/replica_command.go:863 [split,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12]
W180602 07:21:27.012870 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
I180602 07:21:27.016534 104022 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming
I180602 07:21:27.038033 111616 storage/replica_command.go:863 [split,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13]
I180602 07:21:27.156325 111649 storage/replica_command.go:863 [split,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14]
W180602 07:21:27.223175 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
W180602 07:21:27.281224 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
I180602 07:21:27.300168 111690 storage/replica_command.go:863 [split,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15]
W180602 07:21:27.311006 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer
I180602 07:21:27.420307 111719 storage/replica_command.go:863 [split,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16]
I180602 07:21:27.542361 111701 storage/replica_command.go:863 [split,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17]
W180602 07:21:27.645111 111725 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=da9ad7de key=/Local/Range/Table/19/RangeDescriptor rw=true pri=0.05901391 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,1879 orig=0.000000123,1879 max=0.000000123,1893 wto=false rop=false seq=1
W180602 07:21:27.659094 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
I180602 07:21:27.684752 111680 storage/replica_command.go:863 [split,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18]
W180602 07:21:27.763939 111780 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=125a67e3 key=/Local/Range/Table/20/RangeDescriptor rw=true pri=0.03635720 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2111 orig=0.000000123,2042 max=0.000000123,2084 wto=false rop=false seq=1
W180602 07:21:27.766883 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer
E180602 07:21:27.806309 111766 storage/consistency_queue.go:125 [replica consistency checker,s1,r4/1:/System/{NodeLive…-tsd}] computing own checksum: could not resolve node ID 1: dummy resolver
E180602 07:21:27.806975 111766 storage/queue.go:778 [replica consistency checker,s1,r4/1:/System/{NodeLive…-tsd}] computing own checksum: could not resolve node ID 1: dummy resolver
I180602 07:21:27.818124 111796 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065
I180602 07:21:27.826497 111796 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming
I180602 07:21:27.874237 111726 storage/replica_command.go:863 [split,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19]
W180602 07:21:27.982386 111770 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=c476a094 key=/Local/Range/Table/21/RangeDescriptor rw=true pri=0.02374100 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2252 orig=0.000000123,2252 max=0.000000123,2275 wto=false rop=false seq=1
I180602 07:21:28.029652 111810 storage/replica_command.go:863 [split,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20]
W180602 07:21:28.031230 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:28.118611 111774 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=fcaa46c5 key=/Local/Range/Table/22/RangeDescriptor rw=true pri=0.00146925 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2397 orig=0.000000123,2397 max=0.000000123,2414 wto=false rop=false seq=1
I180602 07:21:28.154689 111775 storage/replica_command.go:863 [split,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21]
W180602 07:21:28.213520 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
I180602 07:21:28.274397 111813 storage/replica_command.go:863 [split,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22]
W180602 07:21:28.338665 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
W180602 07:21:28.352719 111831 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=a8866910 key=/Local/Range/Table/50/RangeDescriptor rw=true pri=0.01775119 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2619 orig=0.000000123,2619 max=0.000000123,2648 wto=false rop=false seq=1
I180602 07:21:28.382776 111710 storage/replica_command.go:863 [split,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23]
W180602 07:21:28.458973 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer
W180602 07:21:28.477987 111834 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=3ad9f13b key=/Local/Range/Table/51/RangeDescriptor rw=true pri=0.02323743 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2754 orig=0.000000123,2754 max=0.000000123,2801 wto=false rop=false seq=1
I180602 07:21:28.508700 111631 storage/replica_command.go:863 [split,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r24]
W180602 07:21:28.583750 111755 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=199fc42e key=/Local/Range/Table/52/RangeDescriptor rw=true pri=0.01267702 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2880 orig=0.000000123,2880 max=0.000000123,2925 wto=false rop=false seq=1
W180602 07:21:28.621310 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
I180602 07:21:28.623619 111837 storage/replica_command.go:863 [split,s1,r24/1:/{Table/53-Max}] initiating a split of this range at key /Table/54 [r25]
W180602 07:21:28.689811 111758 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=4806ee51 key=/Local/Range/Table/53/RangeDescriptor rw=true pri=0.03011487 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,3011 orig=0.000000123,3011 max=0.000000123,3022 wto=false rop=false seq=1
E180602 07:21:28.791818 111633 storage/consistency_queue.go:125 [replica consistency checker,s1,r8/1:/Table/1{1-2}] computing own checksum: could not resolve node ID 1: dummy resolver
E180602 07:21:28.792442 111633 storage/queue.go:778 [replica consistency checker,s1,r8/1:/Table/1{1-2}] computing own checksum: could not resolve node ID 1: dummy resolver
I180602 07:21:28.806116 111842 storage/replica_command.go:863 [split,s1,r25/1:/{Table/54-Max}] initiating a split of this range at key /Table/55 [r26]
I180602 07:21:28.844513 111819 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065
W180602 07:21:28.898378 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer
I180602 07:21:28.909067 111819 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming
I180602 07:21:28.950153 111879 storage/replica_command.go:863 [split,s1,r26/1:/{Table/55-Max}] initiating a split of this range at key /Table/56 [r27]
I180602 07:21:29.067577 111852 storage/replica_command.go:863 [split,s1,r27/1:/{Table/56-Max}] initiating a split of this range at key /Table/57 [r28]
W180602 07:21:29.149687 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:29.178637 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
I180602 07:21:29.223093 111393 util/stop/stopper.go:471 quiescing; tasks left:
1 storage.replicate: purgatory processing replica
1 storage.intentResolver: processing intents
I180602 07:21:29.223977 111393 util/stop/stopper.go:471 quiescing; tasks left:
1 storage.intentResolver: processing intents
W180602 07:21:29.406961 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer
W180602 07:21:29.434693 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
W180602 07:21:29.729400 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
W180602 07:21:29.802959 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer
I180602 07:21:29.845634 111896 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065
I180602 07:21:29.856527 111896 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming
W180602 07:21:30.184916 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:30.217264 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:30.330989 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer
W180602 07:21:30.492147 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
W180602 07:21:30.719824 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer
W180602 07:21:30.843026 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
I180602 07:21:30.846414 111868 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065
I180602 07:21:30.853735 111868 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming
W180602 07:21:31.088347 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:31.202071 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer
W180602 07:21:31.306937 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:31.411590 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
W180602 07:21:31.763071 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer
I180602 07:21:31.847405 111959 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065
I180602 07:21:31.855401 111959 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming
W180602 07:21:31.957920 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
W180602 07:21:32.120898 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:32.275896 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer
W180602 07:21:32.398697 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:32.482426 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
W180602 07:21:32.785552 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer
W180602 07:21:32.843238 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
I180602 07:21:32.848515 111974 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065
I180602 07:21:32.872364 111974 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming
W180602 07:21:33.014392 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:33.313420 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer
W180602 07:21:33.441125 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
W180602 07:21:33.506666 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:33.761849 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
I180602 07:21:33.849460 111914 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065
W180602 07:21:33.851714 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer
I180602 07:21:33.855960 111914 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming
W180602 07:21:33.986173 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:34.319167 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer
leaktest.go:110: Leaked goroutine: goroutine 111416 [select]:
github.com/cockroachdb/cockroach/pkg/gossip.(*client).gossip(0xc422408dd0, 0x3548840, 0xc420da95f0, 0xc420e05980, 0x3562720, 0xc4212e0090, 0xc4207ea3f0, 0xc4232828f0, 0x0, 0x0)
/go/src/github.com/cockroachdb/cockroach/pkg/gossip/client.go:323 +0x400
github.com/cockroachdb/cockroach/pkg/gossip.(*client).startLocked.func1(0x3548840, 0xc420da95f0)
/go/src/github.com/cockroachdb/cockroach/pkg/gossip/client.go:130 +0x5fd
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc4212e0030, 0xc4207ea3f0, 0xc421028280)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:193 +0x14b
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:186 +0xbb
leaktest.go:110: Leaked goroutine: goroutine 111417 [select]:
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.newClientStream.func5(0xc4201ba300, 0xc420fc2780, 0x3548840, 0xc420da9680)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/stream.go:321 +0x120
created by github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.newClientStream
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/stream.go:320 +0x1200
leaktest.go:110: Leaked goroutine: goroutine 111425 [select]:
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*recvBufferReader).read(0xc4211e79a0, 0xc4203e0f10, 0x5, 0x5, 0xc421cea660, 0x843cb0, 0x2810ab0)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:142 +0x19c
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*recvBufferReader).Read(0xc4211e79a0, 0xc4203e0f10, 0x5, 0x5, 0x842f6c, 0xc421cea660, 0xc4233bd8c8)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:131 +0x82
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*transportReader).Read(0xc420da97a0, 0xc4203e0f10, 0x5, 0x5, 0xf25293, 0xc421cea660, 0x1fffe000000636)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:394 +0x71
io.ReadAtLeast(0x3525540, 0xc420da97a0, 0xc4203e0f10, 0x5, 0x5, 0x5, 0xc4233bd9a0, 0xf4daa9, 0xc421dec900)
/usr/local/go/src/io/io.go:309 +0x94
io.ReadFull(0x3525540, 0xc420da97a0, 0xc4203e0f10, 0x5, 0x5, 0x15, 0x20, 0x153446de41c1c962)
/usr/local/go/src/io/io.go:327 +0x73
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*Stream).Read(0xc4218e4c00, 0xc4203e0f10, 0x5, 0x5, 0xc421cfeea8, 0xc421187338, 0xc420ffc0a0)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:378 +0x15d
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*parser).recvMsg(0xc4203e0f00, 0x7fffffff, 0xb, 0x0, 0xc422322660, 0xc420255500, 0x0, 0xe)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/rpc_util.go:452 +0x99
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.recv(0xc4203e0f00, 0x2ac8eb7c2b10, 0x5068e68, 0xc4218e4c00, 0x0, 0x0, 0x2de9b60, 0xc422f60640, 0x7fffffff, 0xc422bcb800, ...)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/rpc_util.go:578 +0x61
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*csAttempt).recvMsg(0xc422408ea0, 0x2de9b60, 0xc422f60640, 0x0, 0x0)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/stream.go:539 +0x27c
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*clientStream).RecvMsg(0xc420fc2780, 0x2de9b60, 0xc422f60640, 0x0, 0x0)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/stream.go:405 +0x67
github.com/cockroachdb/cockroach/pkg/gossip.(*gossipGossipClient).Recv(0xc4212e0090, 0x3548840, 0xc423150150, 0xc420e05980)
/go/src/github.com/cockroachdb/cockroach/pkg/gossip/gossip.pb.go:191 +0x83
github.com/cockroachdb/cockroach/pkg/gossip.(*client).gossip.func2.1(0x3562720, 0xc4212e0090, 0xc422408dd0, 0x3548840, 0xc423150150, 0xc420e05980, 0xc420a70e00, 0x30)
/go/src/github.com/cockroachdb/cockroach/pkg/gossip/client.go:311 +0x43
github.com/cockroachdb/cockroach/pkg/gossip.(*client).gossip.func2(0x3548840, 0xc423150150)
/go/src/github.com/cockroachdb/cockroach/pkg/gossip/client.go:319 +0xe6
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc4212e0180, 0xc4207ea3f0, 0xc421028600)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:193 +0x14b
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:186 +0xbb
leaktest.go:110: Leaked goroutine: goroutine 111565 [select]:
github.com/cockroachdb/cockroach/pkg/gossip.(*server).Gossip(0xc4226a6780, 0x3562780, 0xc422e85ec0, 0x0, 0x0)
/go/src/github.com/cockroachdb/cockroach/pkg/gossip/server.go:185 +0xa61
github.com/cockroachdb/cockroach/pkg/gossip._Gossip_Gossip_Handler(0x2ebd7c0, 0xc4226a6780, 0x3559420, 0xc4241a2a50, 0xc4200fc2c8, 0xc4239ab200)
/go/src/github.com/cockroachdb/cockroach/pkg/gossip/gossip.pb.go:208 +0xd3
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).processStreamingRPC(0xc4208fca80, 0x3565de0, 0xc420653800, 0xc4211a1d00, 0xc4211bad20, 0x4448820, 0x0, 0x0, 0x0)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:1160 +0x132a
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).handleStream(0xc4208fca80, 0x3565de0, 0xc420653800, 0xc4211a1d00, 0x0)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:1253 +0x12f9
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).serveStreams.func1.1(0xc421c10bd0, 0xc4208fca80, 0x3565de0, 0xc420653800, 0xc4211a1d00)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:680 +0xad
created by github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).serveStreams.func1
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:678 +0xb9
leaktest.go:110: Leaked goroutine: goroutine 111566 [select]:
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*recvBufferReader).read(0xc420871270, 0xc421a9f4b0, 0x5, 0x5, 0xc4219a0500, 0x843cb0, 0x2810ab0)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:142 +0x19c
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*recvBufferReader).Read(0xc420871270, 0xc421a9f4b0, 0x5, 0x5, 0x842f6c, 0xc4219a0500, 0xc4209ef610)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:131 +0x82
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*transportReader).Read(0xc4239ab1a0, 0xc421a9f4b0, 0x5, 0x5, 0xf25293, 0xc4219a0500, 0x1fffe000001251)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:394 +0x71
io.ReadAtLeast(0x3525540, 0xc4239ab1a0, 0xc421a9f4b0, 0x5, 0x5, 0x5, 0xc4209ef6e8, 0xf4f539, 0xc420653800)
/usr/local/go/src/io/io.go:309 +0x94
io.ReadFull(0x3525540, 0xc4239ab1a0, 0xc421a9f4b0, 0x5, 0x5, 0xc4229b6280, 0xc4209ef7d0, 0x110410f)
/usr/local/go/src/io/io.go:327 +0x73
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*Stream).Read(0xc4211a1d00, 0xc421a9f4b0, 0x5, 0x5, 0xc42004e500, 0xc42004e500, 0xc4209ef840)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:378 +0x15d
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*parser).recvMsg(0xc421a9f4a0, 0x7fffffff, 0xb, 0x0, 0x2de9b60, 0xc420354320, 0x0, 0xe)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/rpc_util.go:452 +0x99
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.recv(0xc421a9f4a0, 0x2ac8eb7c2b10, 0x5068e68, 0xc4211a1d00, 0x0, 0x0, 0x2e0a100, 0xc422f600a0, 0x7fffffff, 0xc422bca3c0, ...)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/rpc_util.go:578 +0x61
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*serverStream).RecvMsg(0xc4241a2a50, 0x2e0a100, 0xc422f600a0, 0x0, 0x0)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/stream.go:757 +0x1e9
github.com/cockroachdb/cockroach/pkg/gossip.(*gossipGossipServer).Recv(0xc422e85ec0, 0xc400000000, 0x1836efc, 0xc4209efb38)
/go/src/github.com/cockroachdb/cockroach/pkg/gossip/gossip.pb.go:227 +0x83
github.com/cockroachdb/cockroach/pkg/gossip.(Gossip_GossipServer).Recv-fm(0xc4226a67c8, 0x0, 0x0)
/go/src/github.com/cockroachdb/cockroach/pkg/gossip/server.go:148 +0x4b
github.com/cockroachdb/cockroach/pkg/gossip.(*server).gossipReceiver(0xc4226a6780, 0x3548840, 0xc4231500f0, 0xc420b3aab8, 0xc420cff700, 0xc420cc5f50, 0x0, 0x0)
/go/src/github.com/cockroachdb/cockroach/pkg/gossip/server.go:326 +0x67c
github.com/cockroachdb/cockroach/pkg/gossip.(*server).Gossip.func3.1(0x3548840, 0xc4231500f0)
/go/src/github.com/cockroachdb/cockroach/pkg/gossip/server.go:148 +0x12d
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc422e85ef0, 0xc4207ea3f0, 0xc420cff7c0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:193 +0x14b
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:186 +0xbb
```
|
1.0
|
github.com/cockroachdb/cockroach/pkg/storage: TestStoreRangeSystemSplits failed under stress - SHA: https://github.com/cockroachdb/cockroach/commits/97c3622c2a6f4ec579ed93655b9d8a83732ac654
Parameters:
```
TAGS=
GOFLAGS=-race
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=693452&tab=buildLog
```
I180602 07:21:25.666097 111393 gossip/gossip.go:333 [n1] NodeDescriptor set to node_id:1 address:<network_field:"" address_field:"" > attrs:<> locality:<> ServerVersion:<major_val:0 minor_val:0 patch:0 unstable:0 >
I180602 07:21:25.755244 111393 storage/store.go:1454 [s1] [n1,s1]: failed initial metrics computation: [n1,s1]: system config not yet available
I180602 07:21:25.822572 111432 storage/replica_command.go:863 [split,s1,r1/1:/M{in-ax}] initiating a split of this range at key /System/"" [r2]
W180602 07:21:25.865205 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
I180602 07:21:25.915992 111562 storage/replica_command.go:863 [split,s1,r2/1:/{System/-Max}] initiating a split of this range at key /System/NodeLiveness [r3]
W180602 07:21:25.934571 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:25.938941 111461 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "unnamed" id=17deafae key=/Table/SystemConfigSpan/Start rw=true pri=0.03184783 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,53 orig=0.000000123,53 max=0.000000123,54 wto=false rop=false seq=1
I180602 07:21:26.066952 111572 storage/replica_command.go:863 [split,s1,r3/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/NodeLivenessMax [r4]
W180602 07:21:26.094922 111464 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "unnamed" id=17deafae key=/Table/SystemConfigSpan/Start rw=true pri=0.03184783 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,53 orig=0.000000123,53 max=0.000000123,54 wto=false rop=false seq=1
W180602 07:21:26.143966 111457 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=46177098 key=/Local/Range/System/NodeLiveness/RangeDescriptor rw=true pri=0.06770799 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,182 orig=0.000000123,182 max=0.000000123,196 wto=false rop=false seq=1
W180602 07:21:26.171835 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
I180602 07:21:26.190567 111466 storage/replica_command.go:863 [split,s1,r4/1:/{System/NodeL…-Max}] initiating a split of this range at key /System/tsd [r5]
W180602 07:21:26.200246 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:26.223842 111587 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "unnamed" id=17deafae key=/Table/SystemConfigSpan/Start rw=true pri=0.03184783 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,53 orig=0.000000123,53 max=0.000000123,54 wto=false rop=false seq=1
W180602 07:21:26.259843 103909 gossip/gossip.go:1298 [n1] first range unavailable; trying remaining resolvers
I180602 07:21:26.262515 111416 gossip/client.go:129 [n1] started gossip client to 127.0.0.1:44307
I180602 07:21:26.304909 111402 storage/replica_command.go:863 [split,s1,r5/1:/{System/tsd-Max}] initiating a split of this range at key /System/"tse" [r6]
W180602 07:21:26.307892 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer
W180602 07:21:26.340668 111552 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "unnamed" id=17deafae key=/Table/SystemConfigSpan/Start rw=true pri=0.03184783 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,53 orig=0.000000123,53 max=0.000000123,54 wto=false rop=false seq=1
I180602 07:21:26.413526 111603 storage/replica_command.go:863 [split,s1,r6/1:/{System/tse-Max}] initiating a split of this range at key /Table/SystemConfigSpan/Start [r7]
W180602 07:21:26.434307 111573 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "unnamed" id=17deafae key=/Table/SystemConfigSpan/Start rw=true pri=0.03184783 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,53 orig=0.000000123,53 max=0.000000123,54 wto=false rop=false seq=1
I180602 07:21:26.489843 111577 storage/replica_command.go:863 [split,s1,r7/1:/{Table/System…-Max}] initiating a split of this range at key /Table/11 [r8]
I180602 07:21:26.601373 111652 storage/replica_command.go:863 [split,s1,r8/1:/{Table/11-Max}] initiating a split of this range at key /Table/12 [r9]
I180602 07:21:26.700193 111406 storage/replica_command.go:863 [split,s1,r9/1:/{Table/12-Max}] initiating a split of this range at key /Table/13 [r10]
W180602 07:21:26.748760 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
W180602 07:21:26.761617 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer
E180602 07:21:26.803145 111637 storage/consistency_queue.go:125 [replica consistency checker,s1,r1/1:/{Min-System/}] computing own checksum: could not resolve node ID 1: dummy resolver
E180602 07:21:26.803792 111637 storage/queue.go:778 [replica consistency checker,s1,r1/1:/{Min-System/}] computing own checksum: could not resolve node ID 1: dummy resolver
W180602 07:21:26.812209 104013 gossip/gossip.go:1298 [n2] first range unavailable; trying remaining resolvers
I180602 07:21:26.818283 111638 storage/replica_command.go:863 [split,s1,r10/1:/{Table/13-Max}] initiating a split of this range at key /Table/14 [r11]
I180602 07:21:26.918273 111580 storage/replica_command.go:863 [split,s1,r11/1:/{Table/14-Max}] initiating a split of this range at key /Table/15 [r12]
W180602 07:21:27.012870 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
I180602 07:21:27.016534 104022 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming
I180602 07:21:27.038033 111616 storage/replica_command.go:863 [split,s1,r12/1:/{Table/15-Max}] initiating a split of this range at key /Table/16 [r13]
I180602 07:21:27.156325 111649 storage/replica_command.go:863 [split,s1,r13/1:/{Table/16-Max}] initiating a split of this range at key /Table/17 [r14]
W180602 07:21:27.223175 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
W180602 07:21:27.281224 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
I180602 07:21:27.300168 111690 storage/replica_command.go:863 [split,s1,r14/1:/{Table/17-Max}] initiating a split of this range at key /Table/18 [r15]
W180602 07:21:27.311006 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer
I180602 07:21:27.420307 111719 storage/replica_command.go:863 [split,s1,r15/1:/{Table/18-Max}] initiating a split of this range at key /Table/19 [r16]
I180602 07:21:27.542361 111701 storage/replica_command.go:863 [split,s1,r16/1:/{Table/19-Max}] initiating a split of this range at key /Table/20 [r17]
W180602 07:21:27.645111 111725 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=da9ad7de key=/Local/Range/Table/19/RangeDescriptor rw=true pri=0.05901391 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,1879 orig=0.000000123,1879 max=0.000000123,1893 wto=false rop=false seq=1
W180602 07:21:27.659094 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
I180602 07:21:27.684752 111680 storage/replica_command.go:863 [split,s1,r17/1:/{Table/20-Max}] initiating a split of this range at key /Table/21 [r18]
W180602 07:21:27.763939 111780 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=125a67e3 key=/Local/Range/Table/20/RangeDescriptor rw=true pri=0.03635720 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2111 orig=0.000000123,2042 max=0.000000123,2084 wto=false rop=false seq=1
W180602 07:21:27.766883 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer
E180602 07:21:27.806309 111766 storage/consistency_queue.go:125 [replica consistency checker,s1,r4/1:/System/{NodeLive…-tsd}] computing own checksum: could not resolve node ID 1: dummy resolver
E180602 07:21:27.806975 111766 storage/queue.go:778 [replica consistency checker,s1,r4/1:/System/{NodeLive…-tsd}] computing own checksum: could not resolve node ID 1: dummy resolver
I180602 07:21:27.818124 111796 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065
I180602 07:21:27.826497 111796 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming
I180602 07:21:27.874237 111726 storage/replica_command.go:863 [split,s1,r18/1:/{Table/21-Max}] initiating a split of this range at key /Table/22 [r19]
W180602 07:21:27.982386 111770 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=c476a094 key=/Local/Range/Table/21/RangeDescriptor rw=true pri=0.02374100 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2252 orig=0.000000123,2252 max=0.000000123,2275 wto=false rop=false seq=1
I180602 07:21:28.029652 111810 storage/replica_command.go:863 [split,s1,r19/1:/{Table/22-Max}] initiating a split of this range at key /Table/23 [r20]
W180602 07:21:28.031230 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:28.118611 111774 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=fcaa46c5 key=/Local/Range/Table/22/RangeDescriptor rw=true pri=0.00146925 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2397 orig=0.000000123,2397 max=0.000000123,2414 wto=false rop=false seq=1
I180602 07:21:28.154689 111775 storage/replica_command.go:863 [split,s1,r20/1:/{Table/23-Max}] initiating a split of this range at key /Table/50 [r21]
W180602 07:21:28.213520 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
I180602 07:21:28.274397 111813 storage/replica_command.go:863 [split,s1,r21/1:/{Table/50-Max}] initiating a split of this range at key /Table/51 [r22]
W180602 07:21:28.338665 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
W180602 07:21:28.352719 111831 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=a8866910 key=/Local/Range/Table/50/RangeDescriptor rw=true pri=0.01775119 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2619 orig=0.000000123,2619 max=0.000000123,2648 wto=false rop=false seq=1
I180602 07:21:28.382776 111710 storage/replica_command.go:863 [split,s1,r22/1:/{Table/51-Max}] initiating a split of this range at key /Table/52 [r23]
W180602 07:21:28.458973 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer
W180602 07:21:28.477987 111834 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=3ad9f13b key=/Local/Range/Table/51/RangeDescriptor rw=true pri=0.02323743 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2754 orig=0.000000123,2754 max=0.000000123,2801 wto=false rop=false seq=1
I180602 07:21:28.508700 111631 storage/replica_command.go:863 [split,s1,r23/1:/{Table/52-Max}] initiating a split of this range at key /Table/53 [r24]
W180602 07:21:28.583750 111755 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=199fc42e key=/Local/Range/Table/52/RangeDescriptor rw=true pri=0.01267702 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,2880 orig=0.000000123,2880 max=0.000000123,2925 wto=false rop=false seq=1
W180602 07:21:28.621310 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
I180602 07:21:28.623619 111837 storage/replica_command.go:863 [split,s1,r24/1:/{Table/53-Max}] initiating a split of this range at key /Table/54 [r25]
W180602 07:21:28.689811 111758 storage/intent_resolver.go:638 [s1] failed to push during intent resolution: failed to push "split" id=4806ee51 key=/Local/Range/Table/53/RangeDescriptor rw=true pri=0.03011487 iso=SERIALIZABLE stat=PENDING epo=0 ts=0.000000123,3011 orig=0.000000123,3011 max=0.000000123,3022 wto=false rop=false seq=1
E180602 07:21:28.791818 111633 storage/consistency_queue.go:125 [replica consistency checker,s1,r8/1:/Table/1{1-2}] computing own checksum: could not resolve node ID 1: dummy resolver
E180602 07:21:28.792442 111633 storage/queue.go:778 [replica consistency checker,s1,r8/1:/Table/1{1-2}] computing own checksum: could not resolve node ID 1: dummy resolver
I180602 07:21:28.806116 111842 storage/replica_command.go:863 [split,s1,r25/1:/{Table/54-Max}] initiating a split of this range at key /Table/55 [r26]
I180602 07:21:28.844513 111819 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065
W180602 07:21:28.898378 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer
I180602 07:21:28.909067 111819 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming
I180602 07:21:28.950153 111879 storage/replica_command.go:863 [split,s1,r26/1:/{Table/55-Max}] initiating a split of this range at key /Table/56 [r27]
I180602 07:21:29.067577 111852 storage/replica_command.go:863 [split,s1,r27/1:/{Table/56-Max}] initiating a split of this range at key /Table/57 [r28]
W180602 07:21:29.149687 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:29.178637 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
I180602 07:21:29.223093 111393 util/stop/stopper.go:471 quiescing; tasks left:
1 storage.replicate: purgatory processing replica
1 storage.intentResolver: processing intents
I180602 07:21:29.223977 111393 util/stop/stopper.go:471 quiescing; tasks left:
1 storage.intentResolver: processing intents
W180602 07:21:29.406961 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer
W180602 07:21:29.434693 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
W180602 07:21:29.729400 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
W180602 07:21:29.802959 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer
I180602 07:21:29.845634 111896 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065
I180602 07:21:29.856527 111896 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming
W180602 07:21:30.184916 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:30.217264 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:30.330989 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer
W180602 07:21:30.492147 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
W180602 07:21:30.719824 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer
W180602 07:21:30.843026 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
I180602 07:21:30.846414 111868 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065
I180602 07:21:30.853735 111868 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming
W180602 07:21:31.088347 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:31.202071 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer
W180602 07:21:31.306937 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:31.411590 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
W180602 07:21:31.763071 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer
I180602 07:21:31.847405 111959 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065
I180602 07:21:31.855401 111959 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming
W180602 07:21:31.957920 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
W180602 07:21:32.120898 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:32.275896 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer
W180602 07:21:32.398697 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:32.482426 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
W180602 07:21:32.785552 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer
W180602 07:21:32.843238 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
I180602 07:21:32.848515 111974 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065
I180602 07:21:32.872364 111974 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming
W180602 07:21:33.014392 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:33.313420 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer
W180602 07:21:33.441125 103981 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
W180602 07:21:33.506666 103979 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:33.761849 104101 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip node liveness: node unavailable; try another peer
I180602 07:21:33.849460 111914 gossip/client.go:129 [n2] started gossip client to 127.0.0.1:45065
W180602 07:21:33.851714 104100 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip system config: node unavailable; try another peer
I180602 07:21:33.855960 111914 gossip/client.go:134 [n2] closing client to node 1 (127.0.0.1:45065): stopping outgoing client to node 1 (127.0.0.1:45065); already have incoming
W180602 07:21:33.986173 104099 storage/store.go:1550 [s2,r1/2:/M{in-ax}] could not gossip first range descriptor: node unavailable; try another peer
W180602 07:21:34.319167 103980 storage/store.go:1550 [s1,r1/1:/M{in-ax}] could not gossip system config: node unavailable; try another peer
leaktest.go:110: Leaked goroutine: goroutine 111416 [select]:
github.com/cockroachdb/cockroach/pkg/gossip.(*client).gossip(0xc422408dd0, 0x3548840, 0xc420da95f0, 0xc420e05980, 0x3562720, 0xc4212e0090, 0xc4207ea3f0, 0xc4232828f0, 0x0, 0x0)
/go/src/github.com/cockroachdb/cockroach/pkg/gossip/client.go:323 +0x400
github.com/cockroachdb/cockroach/pkg/gossip.(*client).startLocked.func1(0x3548840, 0xc420da95f0)
/go/src/github.com/cockroachdb/cockroach/pkg/gossip/client.go:130 +0x5fd
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc4212e0030, 0xc4207ea3f0, 0xc421028280)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:193 +0x14b
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:186 +0xbb
leaktest.go:110: Leaked goroutine: goroutine 111417 [select]:
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.newClientStream.func5(0xc4201ba300, 0xc420fc2780, 0x3548840, 0xc420da9680)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/stream.go:321 +0x120
created by github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.newClientStream
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/stream.go:320 +0x1200
leaktest.go:110: Leaked goroutine: goroutine 111425 [select]:
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*recvBufferReader).read(0xc4211e79a0, 0xc4203e0f10, 0x5, 0x5, 0xc421cea660, 0x843cb0, 0x2810ab0)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:142 +0x19c
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*recvBufferReader).Read(0xc4211e79a0, 0xc4203e0f10, 0x5, 0x5, 0x842f6c, 0xc421cea660, 0xc4233bd8c8)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:131 +0x82
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*transportReader).Read(0xc420da97a0, 0xc4203e0f10, 0x5, 0x5, 0xf25293, 0xc421cea660, 0x1fffe000000636)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:394 +0x71
io.ReadAtLeast(0x3525540, 0xc420da97a0, 0xc4203e0f10, 0x5, 0x5, 0x5, 0xc4233bd9a0, 0xf4daa9, 0xc421dec900)
/usr/local/go/src/io/io.go:309 +0x94
io.ReadFull(0x3525540, 0xc420da97a0, 0xc4203e0f10, 0x5, 0x5, 0x15, 0x20, 0x153446de41c1c962)
/usr/local/go/src/io/io.go:327 +0x73
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*Stream).Read(0xc4218e4c00, 0xc4203e0f10, 0x5, 0x5, 0xc421cfeea8, 0xc421187338, 0xc420ffc0a0)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:378 +0x15d
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*parser).recvMsg(0xc4203e0f00, 0x7fffffff, 0xb, 0x0, 0xc422322660, 0xc420255500, 0x0, 0xe)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/rpc_util.go:452 +0x99
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.recv(0xc4203e0f00, 0x2ac8eb7c2b10, 0x5068e68, 0xc4218e4c00, 0x0, 0x0, 0x2de9b60, 0xc422f60640, 0x7fffffff, 0xc422bcb800, ...)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/rpc_util.go:578 +0x61
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*csAttempt).recvMsg(0xc422408ea0, 0x2de9b60, 0xc422f60640, 0x0, 0x0)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/stream.go:539 +0x27c
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*clientStream).RecvMsg(0xc420fc2780, 0x2de9b60, 0xc422f60640, 0x0, 0x0)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/stream.go:405 +0x67
github.com/cockroachdb/cockroach/pkg/gossip.(*gossipGossipClient).Recv(0xc4212e0090, 0x3548840, 0xc423150150, 0xc420e05980)
/go/src/github.com/cockroachdb/cockroach/pkg/gossip/gossip.pb.go:191 +0x83
github.com/cockroachdb/cockroach/pkg/gossip.(*client).gossip.func2.1(0x3562720, 0xc4212e0090, 0xc422408dd0, 0x3548840, 0xc423150150, 0xc420e05980, 0xc420a70e00, 0x30)
/go/src/github.com/cockroachdb/cockroach/pkg/gossip/client.go:311 +0x43
github.com/cockroachdb/cockroach/pkg/gossip.(*client).gossip.func2(0x3548840, 0xc423150150)
/go/src/github.com/cockroachdb/cockroach/pkg/gossip/client.go:319 +0xe6
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc4212e0180, 0xc4207ea3f0, 0xc421028600)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:193 +0x14b
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:186 +0xbb
leaktest.go:110: Leaked goroutine: goroutine 111565 [select]:
github.com/cockroachdb/cockroach/pkg/gossip.(*server).Gossip(0xc4226a6780, 0x3562780, 0xc422e85ec0, 0x0, 0x0)
/go/src/github.com/cockroachdb/cockroach/pkg/gossip/server.go:185 +0xa61
github.com/cockroachdb/cockroach/pkg/gossip._Gossip_Gossip_Handler(0x2ebd7c0, 0xc4226a6780, 0x3559420, 0xc4241a2a50, 0xc4200fc2c8, 0xc4239ab200)
/go/src/github.com/cockroachdb/cockroach/pkg/gossip/gossip.pb.go:208 +0xd3
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).processStreamingRPC(0xc4208fca80, 0x3565de0, 0xc420653800, 0xc4211a1d00, 0xc4211bad20, 0x4448820, 0x0, 0x0, 0x0)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:1160 +0x132a
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).handleStream(0xc4208fca80, 0x3565de0, 0xc420653800, 0xc4211a1d00, 0x0)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:1253 +0x12f9
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).serveStreams.func1.1(0xc421c10bd0, 0xc4208fca80, 0x3565de0, 0xc420653800, 0xc4211a1d00)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:680 +0xad
created by github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*Server).serveStreams.func1
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/server.go:678 +0xb9
leaktest.go:110: Leaked goroutine: goroutine 111566 [select]:
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*recvBufferReader).read(0xc420871270, 0xc421a9f4b0, 0x5, 0x5, 0xc4219a0500, 0x843cb0, 0x2810ab0)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:142 +0x19c
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*recvBufferReader).Read(0xc420871270, 0xc421a9f4b0, 0x5, 0x5, 0x842f6c, 0xc4219a0500, 0xc4209ef610)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:131 +0x82
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*transportReader).Read(0xc4239ab1a0, 0xc421a9f4b0, 0x5, 0x5, 0xf25293, 0xc4219a0500, 0x1fffe000001251)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:394 +0x71
io.ReadAtLeast(0x3525540, 0xc4239ab1a0, 0xc421a9f4b0, 0x5, 0x5, 0x5, 0xc4209ef6e8, 0xf4f539, 0xc420653800)
/usr/local/go/src/io/io.go:309 +0x94
io.ReadFull(0x3525540, 0xc4239ab1a0, 0xc421a9f4b0, 0x5, 0x5, 0xc4229b6280, 0xc4209ef7d0, 0x110410f)
/usr/local/go/src/io/io.go:327 +0x73
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport.(*Stream).Read(0xc4211a1d00, 0xc421a9f4b0, 0x5, 0x5, 0xc42004e500, 0xc42004e500, 0xc4209ef840)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/transport/transport.go:378 +0x15d
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*parser).recvMsg(0xc421a9f4a0, 0x7fffffff, 0xb, 0x0, 0x2de9b60, 0xc420354320, 0x0, 0xe)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/rpc_util.go:452 +0x99
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.recv(0xc421a9f4a0, 0x2ac8eb7c2b10, 0x5068e68, 0xc4211a1d00, 0x0, 0x0, 0x2e0a100, 0xc422f600a0, 0x7fffffff, 0xc422bca3c0, ...)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/rpc_util.go:578 +0x61
github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc.(*serverStream).RecvMsg(0xc4241a2a50, 0x2e0a100, 0xc422f600a0, 0x0, 0x0)
/go/src/github.com/cockroachdb/cockroach/vendor/google.golang.org/grpc/stream.go:757 +0x1e9
github.com/cockroachdb/cockroach/pkg/gossip.(*gossipGossipServer).Recv(0xc422e85ec0, 0xc400000000, 0x1836efc, 0xc4209efb38)
/go/src/github.com/cockroachdb/cockroach/pkg/gossip/gossip.pb.go:227 +0x83
github.com/cockroachdb/cockroach/pkg/gossip.(Gossip_GossipServer).Recv-fm(0xc4226a67c8, 0x0, 0x0)
/go/src/github.com/cockroachdb/cockroach/pkg/gossip/server.go:148 +0x4b
github.com/cockroachdb/cockroach/pkg/gossip.(*server).gossipReceiver(0xc4226a6780, 0x3548840, 0xc4231500f0, 0xc420b3aab8, 0xc420cff700, 0xc420cc5f50, 0x0, 0x0)
/go/src/github.com/cockroachdb/cockroach/pkg/gossip/server.go:326 +0x67c
github.com/cockroachdb/cockroach/pkg/gossip.(*server).Gossip.func3.1(0x3548840, 0xc4231500f0)
/go/src/github.com/cockroachdb/cockroach/pkg/gossip/server.go:148 +0x12d
github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker.func1(0xc422e85ef0, 0xc4207ea3f0, 0xc420cff7c0)
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:193 +0x14b
created by github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunWorker
/go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:186 +0xbb
```
|
test
|
github com cockroachdb cockroach pkg storage teststorerangesystemsplits failed under stress sha parameters tags goflags race failed test gossip gossip go nodedescriptor set to node id address attrs locality serverversion storage store go failed initial metrics computation system config not yet available storage replica command go initiating a split of this range at key system storage store go could not gossip node liveness node unavailable try another peer storage replica command go initiating a split of this range at key system nodeliveness storage store go could not gossip first range descriptor node unavailable try another peer storage intent resolver go failed to push during intent resolution failed to push unnamed id key table systemconfigspan start rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key system nodelivenessmax storage intent resolver go failed to push during intent resolution failed to push unnamed id key table systemconfigspan start rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage intent resolver go failed to push during intent resolution failed to push split id key local range system nodeliveness rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage store go could not gossip node liveness node unavailable try another peer storage replica command go initiating a split of this range at key system tsd storage store go could not gossip first range descriptor node unavailable try another peer storage intent resolver go failed to push during intent resolution failed to push unnamed id key table systemconfigspan start rw true pri iso serializable stat pending epo ts orig max wto false rop false seq gossip gossip go first range unavailable trying remaining resolvers gossip client go started gossip client to storage replica command go initiating a split of this range at key system tse storage store go could not gossip system config node unavailable try another peer storage intent resolver go failed to push during intent resolution failed to push unnamed id key table systemconfigspan start rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key table systemconfigspan start storage intent resolver go failed to push during intent resolution failed to push unnamed id key table systemconfigspan start rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage store go could not gossip node liveness node unavailable try another peer storage store go could not gossip system config node unavailable try another peer storage consistency queue go computing own checksum could not resolve node id dummy resolver storage queue go computing own checksum could not resolve node id dummy resolver gossip gossip go first range unavailable trying remaining resolvers storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage store go could not gossip first range descriptor node unavailable try another peer gossip client go closing client to node stopping outgoing client to node already have incoming storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage store go could not gossip node liveness node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer storage replica command go initiating a split of this range at key table storage store go could not gossip system config node unavailable try another peer storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage intent resolver go failed to push during intent resolution failed to push split id key local range table rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage store go could not gossip node liveness node unavailable try another peer storage replica command go initiating a split of this range at key table storage intent resolver go failed to push during intent resolution failed to push split id key local range table rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage store go could not gossip system config node unavailable try another peer storage consistency queue go computing own checksum could not resolve node id dummy resolver storage queue go computing own checksum could not resolve node id dummy resolver gossip client go started gossip client to gossip client go closing client to node stopping outgoing client to node already have incoming storage replica command go initiating a split of this range at key table storage intent resolver go failed to push during intent resolution failed to push split id key local range table rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key table storage store go could not gossip first range descriptor node unavailable try another peer storage intent resolver go failed to push during intent resolution failed to push split id key local range table rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key table storage store go could not gossip first range descriptor node unavailable try another peer storage replica command go initiating a split of this range at key table storage store go could not gossip node liveness node unavailable try another peer storage intent resolver go failed to push during intent resolution failed to push split id key local range table rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key table storage store go could not gossip system config node unavailable try another peer storage intent resolver go failed to push during intent resolution failed to push split id key local range table rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage replica command go initiating a split of this range at key table storage intent resolver go failed to push during intent resolution failed to push split id key local range table rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage store go could not gossip node liveness node unavailable try another peer storage replica command go initiating a split of this range at key table storage intent resolver go failed to push during intent resolution failed to push split id key local range table rangedescriptor rw true pri iso serializable stat pending epo ts orig max wto false rop false seq storage consistency queue go computing own checksum could not resolve node id dummy resolver storage queue go computing own checksum could not resolve node id dummy resolver storage replica command go initiating a split of this range at key table gossip client go started gossip client to storage store go could not gossip system config node unavailable try another peer gossip client go closing client to node stopping outgoing client to node already have incoming storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer util stop stopper go quiescing tasks left storage replicate purgatory processing replica storage intentresolver processing intents util stop stopper go quiescing tasks left storage intentresolver processing intents storage store go could not gossip system config node unavailable try another peer storage store go could not gossip node liveness node unavailable try another peer storage store go could not gossip node liveness node unavailable try another peer storage store go could not gossip system config node unavailable try another peer gossip client go started gossip client to gossip client go closing client to node stopping outgoing client to node already have incoming storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip system config node unavailable try another peer storage store go could not gossip node liveness node unavailable try another peer storage store go could not gossip system config node unavailable try another peer storage store go could not gossip node liveness node unavailable try another peer gossip client go started gossip client to gossip client go closing client to node stopping outgoing client to node already have incoming storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip system config node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip node liveness node unavailable try another peer storage store go could not gossip system config node unavailable try another peer gossip client go started gossip client to gossip client go closing client to node stopping outgoing client to node already have incoming storage store go could not gossip node liveness node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip system config node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip node liveness node unavailable try another peer storage store go could not gossip system config node unavailable try another peer storage store go could not gossip node liveness node unavailable try another peer gossip client go started gossip client to gossip client go closing client to node stopping outgoing client to node already have incoming storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip system config node unavailable try another peer storage store go could not gossip node liveness node unavailable try another peer storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip node liveness node unavailable try another peer gossip client go started gossip client to storage store go could not gossip system config node unavailable try another peer gossip client go closing client to node stopping outgoing client to node already have incoming storage store go could not gossip first range descriptor node unavailable try another peer storage store go could not gossip system config node unavailable try another peer leaktest go leaked goroutine goroutine github com cockroachdb cockroach pkg gossip client gossip go src github com cockroachdb cockroach pkg gossip client go github com cockroachdb cockroach pkg gossip client startlocked go src github com cockroachdb cockroach pkg gossip client go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go leaktest go leaked goroutine goroutine github com cockroachdb cockroach vendor google golang org grpc newclientstream go src github com cockroachdb cockroach vendor google golang org grpc stream go created by github com cockroachdb cockroach vendor google golang org grpc newclientstream go src github com cockroachdb cockroach vendor google golang org grpc stream go leaktest go leaked goroutine goroutine github com cockroachdb cockroach vendor google golang org grpc transport recvbufferreader read go src github com cockroachdb cockroach vendor google golang org grpc transport transport go github com cockroachdb cockroach vendor google golang org grpc transport recvbufferreader read go src github com cockroachdb cockroach vendor google golang org grpc transport transport go github com cockroachdb cockroach vendor google golang org grpc transport transportreader read go src github com cockroachdb cockroach vendor google golang org grpc transport transport go io readatleast usr local go src io io go io readfull usr local go src io io go github com cockroachdb cockroach vendor google golang org grpc transport stream read go src github com cockroachdb cockroach vendor google golang org grpc transport transport go github com cockroachdb cockroach vendor google golang org grpc parser recvmsg go src github com cockroachdb cockroach vendor google golang org grpc rpc util go github com cockroachdb cockroach vendor google golang org grpc recv go src github com cockroachdb cockroach vendor google golang org grpc rpc util go github com cockroachdb cockroach vendor google golang org grpc csattempt recvmsg go src github com cockroachdb cockroach vendor google golang org grpc stream go github com cockroachdb cockroach vendor google golang org grpc clientstream recvmsg go src github com cockroachdb cockroach vendor google golang org grpc stream go github com cockroachdb cockroach pkg gossip gossipgossipclient recv go src github com cockroachdb cockroach pkg gossip gossip pb go github com cockroachdb cockroach pkg gossip client gossip go src github com cockroachdb cockroach pkg gossip client go github com cockroachdb cockroach pkg gossip client gossip go src github com cockroachdb cockroach pkg gossip client go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go leaktest go leaked goroutine goroutine github com cockroachdb cockroach pkg gossip server gossip go src github com cockroachdb cockroach pkg gossip server go github com cockroachdb cockroach pkg gossip gossip gossip handler go src github com cockroachdb cockroach pkg gossip gossip pb go github com cockroachdb cockroach vendor google golang org grpc server processstreamingrpc go src github com cockroachdb cockroach vendor google golang org grpc server go github com cockroachdb cockroach vendor google golang org grpc server handlestream go src github com cockroachdb cockroach vendor google golang org grpc server go github com cockroachdb cockroach vendor google golang org grpc server servestreams go src github com cockroachdb cockroach vendor google golang org grpc server go created by github com cockroachdb cockroach vendor google golang org grpc server servestreams go src github com cockroachdb cockroach vendor google golang org grpc server go leaktest go leaked goroutine goroutine github com cockroachdb cockroach vendor google golang org grpc transport recvbufferreader read go src github com cockroachdb cockroach vendor google golang org grpc transport transport go github com cockroachdb cockroach vendor google golang org grpc transport recvbufferreader read go src github com cockroachdb cockroach vendor google golang org grpc transport transport go github com cockroachdb cockroach vendor google golang org grpc transport transportreader read go src github com cockroachdb cockroach vendor google golang org grpc transport transport go io readatleast usr local go src io io go io readfull usr local go src io io go github com cockroachdb cockroach vendor google golang org grpc transport stream read go src github com cockroachdb cockroach vendor google golang org grpc transport transport go github com cockroachdb cockroach vendor google golang org grpc parser recvmsg go src github com cockroachdb cockroach vendor google golang org grpc rpc util go github com cockroachdb cockroach vendor google golang org grpc recv go src github com cockroachdb cockroach vendor google golang org grpc rpc util go github com cockroachdb cockroach vendor google golang org grpc serverstream recvmsg go src github com cockroachdb cockroach vendor google golang org grpc stream go github com cockroachdb cockroach pkg gossip gossipgossipserver recv go src github com cockroachdb cockroach pkg gossip gossip pb go github com cockroachdb cockroach pkg gossip gossip gossipserver recv fm go src github com cockroachdb cockroach pkg gossip server go github com cockroachdb cockroach pkg gossip server gossipreceiver go src github com cockroachdb cockroach pkg gossip server go github com cockroachdb cockroach pkg gossip server gossip go src github com cockroachdb cockroach pkg gossip server go github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go created by github com cockroachdb cockroach pkg util stop stopper runworker go src github com cockroachdb cockroach pkg util stop stopper go
| 1
|
192,447
| 14,617,187,604
|
IssuesEvent
|
2020-12-22 14:25:09
|
github-vet/rangeloop-pointer-findings
|
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
|
closed
|
terraform-providers/terraform-provider-oci: oci/dataintegration_workspace_test.go; 16 LoC
|
fresh small test
|
Found a possible issue in [terraform-providers/terraform-provider-oci](https://www.github.com/terraform-providers/terraform-provider-oci) at [oci/dataintegration_workspace_test.go](https://github.com/terraform-providers/terraform-provider-oci/blob/507acd0ed6517dbca2fbcfb8100874929c8fd8e1/oci/dataintegration_workspace_test.go#L297-L312)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> reference to workspaceId is reassigned at line 301
[Click here to see the code in its original context.](https://github.com/terraform-providers/terraform-provider-oci/blob/507acd0ed6517dbca2fbcfb8100874929c8fd8e1/oci/dataintegration_workspace_test.go#L297-L312)
<details>
<summary>Click here to show the 16 line(s) of Go which triggered the analyzer.</summary>
```go
for _, workspaceId := range workspaceIds {
if ok := SweeperDefaultResourceId[workspaceId]; !ok {
deleteWorkspaceRequest := oci_dataintegration.DeleteWorkspaceRequest{}
deleteWorkspaceRequest.WorkspaceId = &workspaceId
deleteWorkspaceRequest.RequestMetadata.RetryPolicy = getRetryPolicy(true, "dataintegration")
_, error := dataIntegrationClient.DeleteWorkspace(context.Background(), deleteWorkspaceRequest)
if error != nil {
fmt.Printf("Error deleting Workspace %s %s, It is possible that the resource is already deleted. Please verify manually \n", workspaceId, error)
continue
}
waitTillCondition(testAccProvider, &workspaceId, workspaceSweepWaitCondition, time.Duration(3*time.Minute),
workspaceSweepResponseFetchOperation, "dataintegration", true)
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 507acd0ed6517dbca2fbcfb8100874929c8fd8e1
|
1.0
|
terraform-providers/terraform-provider-oci: oci/dataintegration_workspace_test.go; 16 LoC -
Found a possible issue in [terraform-providers/terraform-provider-oci](https://www.github.com/terraform-providers/terraform-provider-oci) at [oci/dataintegration_workspace_test.go](https://github.com/terraform-providers/terraform-provider-oci/blob/507acd0ed6517dbca2fbcfb8100874929c8fd8e1/oci/dataintegration_workspace_test.go#L297-L312)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> reference to workspaceId is reassigned at line 301
[Click here to see the code in its original context.](https://github.com/terraform-providers/terraform-provider-oci/blob/507acd0ed6517dbca2fbcfb8100874929c8fd8e1/oci/dataintegration_workspace_test.go#L297-L312)
<details>
<summary>Click here to show the 16 line(s) of Go which triggered the analyzer.</summary>
```go
for _, workspaceId := range workspaceIds {
if ok := SweeperDefaultResourceId[workspaceId]; !ok {
deleteWorkspaceRequest := oci_dataintegration.DeleteWorkspaceRequest{}
deleteWorkspaceRequest.WorkspaceId = &workspaceId
deleteWorkspaceRequest.RequestMetadata.RetryPolicy = getRetryPolicy(true, "dataintegration")
_, error := dataIntegrationClient.DeleteWorkspace(context.Background(), deleteWorkspaceRequest)
if error != nil {
fmt.Printf("Error deleting Workspace %s %s, It is possible that the resource is already deleted. Please verify manually \n", workspaceId, error)
continue
}
waitTillCondition(testAccProvider, &workspaceId, workspaceSweepWaitCondition, time.Duration(3*time.Minute),
workspaceSweepResponseFetchOperation, "dataintegration", true)
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 507acd0ed6517dbca2fbcfb8100874929c8fd8e1
|
test
|
terraform providers terraform provider oci oci dataintegration workspace test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message reference to workspaceid is reassigned at line click here to show the line s of go which triggered the analyzer go for workspaceid range workspaceids if ok sweeperdefaultresourceid ok deleteworkspacerequest oci dataintegration deleteworkspacerequest deleteworkspacerequest workspaceid workspaceid deleteworkspacerequest requestmetadata retrypolicy getretrypolicy true dataintegration error dataintegrationclient deleteworkspace context background deleteworkspacerequest if error nil fmt printf error deleting workspace s s it is possible that the resource is already deleted please verify manually n workspaceid error continue waittillcondition testaccprovider workspaceid workspacesweepwaitcondition time duration time minute workspacesweepresponsefetchoperation dataintegration true leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
| 1
|
216,357
| 7,307,162,418
|
IssuesEvent
|
2018-02-28 01:18:45
|
Chingu-cohorts/voyage-wiki
|
https://api.github.com/repos/Chingu-cohorts/voyage-wiki
|
closed
|
bears-team-2 lost one of two PMs
|
priority:must have scope:story state:done type:team-repair
|
**_About You_**
- Team Name: bears-team-2
- Slack Name: @heathercoraje
- Timezone: Asia
**_Issue Description & Expected Outcome:_**
Heathercoraje has to leave due to time restrictions & local bootcamp.
**_Symptoms:_**
**_Steps to Recreate:_**
**_Resolution:_**
Will check in with bears-2 and see how they are doing. Will need to have the PM access switched.
|
1.0
|
bears-team-2 lost one of two PMs - **_About You_**
- Team Name: bears-team-2
- Slack Name: @heathercoraje
- Timezone: Asia
**_Issue Description & Expected Outcome:_**
Heathercoraje has to leave due to time restrictions & local bootcamp.
**_Symptoms:_**
**_Steps to Recreate:_**
**_Resolution:_**
Will check in with bears-2 and see how they are doing. Will need to have the PM access switched.
|
non_test
|
bears team lost one of two pms about you team name bears team slack name heathercoraje timezone asia issue description expected outcome heathercoraje has to leave due to time restrictions local bootcamp symptoms steps to recreate resolution will check in with bears and see how they are doing will need to have the pm access switched
| 0
|
110,987
| 9,486,809,422
|
IssuesEvent
|
2019-04-22 15:04:12
|
dojot/dojot
|
https://api.github.com/repos/dojot/dojot
|
closed
|
Device Management API - Devices
|
Status:ToTest Team:Backend Type:Feature
|
Devices (Endpoint, get device)
Create new parameter to filter by type of attr, like Geo.
|
1.0
|
Device Management API - Devices - Devices (Endpoint, get device)
Create new parameter to filter by type of attr, like Geo.
|
test
|
device management api devices devices endpoint get device create new parameter to filter by type of attr like geo
| 1
|
104,108
| 4,195,222,014
|
IssuesEvent
|
2016-06-25 15:56:29
|
Sistema-Integrado-Gestao-Academica/SiGA
|
https://api.github.com/repos/Sistema-Integrado-Gestao-Academica/SiGA
|
closed
|
Detalhamento de Custos - *Despesa*
|
[Medium Priority]
|
Como **secretário financeiro** desejo detalhar uma despesa numa rúbrica (natureza de despesa) para que eu tenha controle dos gastos na de um plano orçamento orçamentário.
------------
C.A.:
* Atributos de um detalhamento de despesa:
* Nota de Empenho (obg)
* ex.: 2011NE005787
* Data de emissão - date
* Nº Processo SEI
* Valor
|
1.0
|
Detalhamento de Custos - *Despesa* - Como **secretário financeiro** desejo detalhar uma despesa numa rúbrica (natureza de despesa) para que eu tenha controle dos gastos na de um plano orçamento orçamentário.
------------
C.A.:
* Atributos de um detalhamento de despesa:
* Nota de Empenho (obg)
* ex.: 2011NE005787
* Data de emissão - date
* Nº Processo SEI
* Valor
|
non_test
|
detalhamento de custos despesa como secretário financeiro desejo detalhar uma despesa numa rúbrica natureza de despesa para que eu tenha controle dos gastos na de um plano orçamento orçamentário c a atributos de um detalhamento de despesa nota de empenho obg ex data de emissão date nº processo sei valor
| 0
|
141,849
| 19,006,012,659
|
IssuesEvent
|
2021-11-23 00:02:50
|
svg-at-wss/WebGoat
|
https://api.github.com/repos/svg-at-wss/WebGoat
|
closed
|
CVE-2021-21341 (High) detected in xstream-1.4.5.jar - autoclosed
|
security vulnerability
|
## CVE-2021-21341 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.5.jar</b></p></summary>
<p>XStream is a serialization library from Java objects to XML and back.</p>
<p>Path to dependency file: WebGoat/webgoat-server/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar,m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar</p>
<p>
Dependency Hierarchy:
- :x: **xstream-1.4.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/svg-at-wss/WebGoat/commit/04c4639ea5cac557f56e64148a0f6081e5be4ddb">04c4639ea5cac557f56e64148a0f6081e5be4ddb</a></p>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is vulnerability which may allow a remote attacker to allocate 100% CPU time on the target system depending on CPU type or parallel execution of such a payload resulting in a denial of service only by manipulating the processed input stream. No user is affected who followed the recommendation to setup XStream's security framework with a whitelist limited to the minimal required types. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.16.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21341>CVE-2021-21341</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-2p3x-qw9c-25hh">https://github.com/x-stream/xstream/security/advisories/GHSA-2p3x-qw9c-25hh</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.16</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.thoughtworks.xstream","packageName":"xstream","packageVersion":"1.4.5","packageFilePaths":["/webgoat-server/pom.xml","/webgoat-lessons/vulnerable-components/pom.xml","/webgoat-integration-tests/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.thoughtworks.xstream:xstream:1.4.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.thoughtworks.xstream:xstream:1.4.16"}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2021-21341","vulnerabilityDetails":"XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is vulnerability which may allow a remote attacker to allocate 100% CPU time on the target system depending on CPU type or parallel execution of such a payload resulting in a denial of service only by manipulating the processed input stream. No user is affected who followed the recommendation to setup XStream\u0027s security framework with a whitelist limited to the minimal required types. If you rely on XStream\u0027s default blacklist of the Security Framework, you will have to use at least version 1.4.16.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21341","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2021-21341 (High) detected in xstream-1.4.5.jar - autoclosed - ## CVE-2021-21341 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.5.jar</b></p></summary>
<p>XStream is a serialization library from Java objects to XML and back.</p>
<p>Path to dependency file: WebGoat/webgoat-server/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar,m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar</p>
<p>
Dependency Hierarchy:
- :x: **xstream-1.4.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/svg-at-wss/WebGoat/commit/04c4639ea5cac557f56e64148a0f6081e5be4ddb">04c4639ea5cac557f56e64148a0f6081e5be4ddb</a></p>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is vulnerability which may allow a remote attacker to allocate 100% CPU time on the target system depending on CPU type or parallel execution of such a payload resulting in a denial of service only by manipulating the processed input stream. No user is affected who followed the recommendation to setup XStream's security framework with a whitelist limited to the minimal required types. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.16.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21341>CVE-2021-21341</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-2p3x-qw9c-25hh">https://github.com/x-stream/xstream/security/advisories/GHSA-2p3x-qw9c-25hh</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.16</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.thoughtworks.xstream","packageName":"xstream","packageVersion":"1.4.5","packageFilePaths":["/webgoat-server/pom.xml","/webgoat-lessons/vulnerable-components/pom.xml","/webgoat-integration-tests/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.thoughtworks.xstream:xstream:1.4.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.thoughtworks.xstream:xstream:1.4.16"}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2021-21341","vulnerabilityDetails":"XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is vulnerability which may allow a remote attacker to allocate 100% CPU time on the target system depending on CPU type or parallel execution of such a payload resulting in a denial of service only by manipulating the processed input stream. No user is affected who followed the recommendation to setup XStream\u0027s security framework with a whitelist limited to the minimal required types. If you rely on XStream\u0027s default blacklist of the Security Framework, you will have to use at least version 1.4.16.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21341","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_test
|
cve high detected in xstream jar autoclosed cve high severity vulnerability vulnerable library xstream jar xstream is a serialization library from java objects to xml and back path to dependency file webgoat webgoat server pom xml path to vulnerable library home wss scanner repository com thoughtworks xstream xstream xstream jar repository com thoughtworks xstream xstream xstream jar home wss scanner repository com thoughtworks xstream xstream xstream jar dependency hierarchy x xstream jar vulnerable library found in head commit a href found in base branch develop vulnerability details xstream is a java library to serialize objects to xml and back again in xstream before version there is vulnerability which may allow a remote attacker to allocate cpu time on the target system depending on cpu type or parallel execution of such a payload resulting in a denial of service only by manipulating the processed input stream no user is affected who followed the recommendation to setup xstream s security framework with a whitelist limited to the minimal required types if you rely on xstream s default blacklist of the security framework you will have to use at least version publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com thoughtworks xstream xstream rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com thoughtworks xstream xstream isminimumfixversionavailable true minimumfixversion com thoughtworks xstream xstream basebranches vulnerabilityidentifier cve vulnerabilitydetails xstream is a java library to serialize objects to xml and back again in xstream before version there is vulnerability which may allow a remote attacker to allocate cpu time on the target system depending on cpu type or parallel execution of such a payload resulting in a denial of service only by manipulating the processed input stream no user is affected who followed the recommendation to setup xstream security framework with a whitelist limited to the minimal required types if you rely on xstream default blacklist of the security framework you will have to use at least version vulnerabilityurl
| 0
|
190,116
| 14,533,451,479
|
IssuesEvent
|
2020-12-15 00:36:22
|
uoForms/App-CANBeWell
|
https://api.github.com/repos/uoForms/App-CANBeWell
|
closed
|
Cursor pointer in the Topics list in the topics tab
|
App CanBeWell 1.0 CanBeWell 2.0 To be tested bug
|
Cursor pointer in the Topics list in the topics tab
|
1.0
|
Cursor pointer in the Topics list in the topics tab - Cursor pointer in the Topics list in the topics tab
|
test
|
cursor pointer in the topics list in the topics tab cursor pointer in the topics list in the topics tab
| 1
|
49,852
| 6,042,919,987
|
IssuesEvent
|
2017-06-11 17:13:44
|
tuna/issues
|
https://api.github.com/repos/tuna/issues
|
closed
|
添加clojure语言库的repo
|
Accepted MirrorRequest Test Required
|
<!--
-->
#### 项目名称与简介(Project Intro.)
clojure语言库的repo
#### 上游地址与镜像方法(How to Mirror)
https://clojars.org/
https://github.com/clojars/clojars-web/wiki/Data
#### 其他信息(Other)
- 镜像大小(Mirror Size):
未知
|
1.0
|
添加clojure语言库的repo - <!--
-->
#### 项目名称与简介(Project Intro.)
clojure语言库的repo
#### 上游地址与镜像方法(How to Mirror)
https://clojars.org/
https://github.com/clojars/clojars-web/wiki/Data
#### 其他信息(Other)
- 镜像大小(Mirror Size):
未知
|
test
|
添加clojure语言库的repo 项目名称与简介 project intro clojure语言库的repo 上游地址与镜像方法 how to mirror 其他信息 other 镜像大小 mirror size 未知
| 1
|
11,602
| 3,211,393,940
|
IssuesEvent
|
2015-10-06 10:28:16
|
ceylon/ceylon-ide-eclipse
|
https://api.github.com/repos/ceylon/ceylon-ide-eclipse
|
closed
|
Running ceylon.net tests
|
bug launch/run/debug test plugin WAITING_FOR_ANSWER
|
If I try to run the ceylon.net tests (i.e. the module `test.ceylon.net` as a Ceylon Test) in the IDE I get the following error:
ceylon test: Error while loading the test.ceylon.net/1.1.1 module:
Declaration 'java.net.URI' could not be found in module 'test.ceylon.net' or
its imported modules but was found in the non-imported module 'java.base'
com.redhat.ceylon.model.loader.ModelResolutionException: Error while loading the test.ceylon.net/1.1.1 module:
Declaration 'java.net.URI' could not be found in module 'test.ceylon.net' or its imported modules but was found in the non-imported module 'java.base'
at com.redhat.ceylon.compiler.java.runtime.model.RuntimeModelLoader$1.reportError(RuntimeModelLoader.java:209)
at com.redhat.ceylon.model.typechecker.model.UnknownType.reportErrors(UnknownType.java:94)
at com.redhat.ceylon.compiler.java.runtime.metamodel.Metamodel.getMetamodel(Metamodel.java:504)
at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeValue.<init>(FreeValue.java:43)
at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeReference.<init>(FreeReference.java:17)
at com.redhat.ceylon.compiler.java.runtime.metamodel.Metamodel.getOrCreateMetamodel(Metamodel.java:344)
at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeClassOrInterface.init(FreeClassOrInterface.java:92)
at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeClass.init(FreeClass.java:51)
at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeClassOrInterface.checkInit(FreeClassOrInterface.java:144)
at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeClassOrInterface.filteredMembers(FreeClassOrInterface.java:212)
at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeClassOrInterface.annotatedMemberDeclarations(FreeClassOrInterface.java:186)
at ceylon.test.core.findCandidatesInClass_.findCandidatesInClass(DefaultTestRunner.ceylon:206)
at ceylon.test.core.findCandidatesInPackage_.findCandidatesInPackage(DefaultTestRunner.ceylon:200)
at ceylon.test.core.findCandidatesInModule_.findCandidatesInModule(DefaultTestRunner.ceylon:191)
at ceylon.test.core.findCandidatesInModuleLiteral_.findCandidatesInModuleLiteral(DefaultTestRunner.ceylon:259)
at ceylon.test.core.findCandidatesInTypeLiteral_.findCandidatesInTypeLiteral(DefaultTestRunner.ceylon:233)
at ceylon.test.core.findCandidates_.findCandidates(DefaultTestRunner.ceylon:183)
at ceylon.test.core.createExecutors_.createExecutors(DefaultTestRunner.ceylon:85)
at ceylon.test.core.DefaultTestRunner.initExecutors$priv$(DefaultTestRunner.ceylon:43)
at ceylon.test.core.DefaultTestRunner.<init>(DefaultTestRunner.ceylon:46)
at ceylon.test.createTestRunner_.createTestRunner(TestRunner.ceylon:49)
at ceylon.test.createTestRunner_.createTestRunner(TestRunner.ceylon:48)
at com.redhat.ceylon.testjvm.Runner.run(tool.ceylon:85)
at com.redhat.ceylon.testjvm.run_.run(tool.ceylon:36)
at com.redhat.ceylon.testjvm.run_.main(tool.ceylon)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at ceylon.modules.api.runtime.SecurityActions.invokeRunInternal(SecurityActions.java:57)
at ceylon.modules.api.runtime.SecurityActions.invokeRun(SecurityActions.java:48)
at ceylon.modules.api.runtime.AbstractRuntime.invokeRun(AbstractRuntime.java:110)
at ceylon.modules.api.runtime.AbstractRuntime.execute(AbstractRuntime.java:176)
at ceylon.modules.api.runtime.AbstractRuntime.execute(AbstractRuntime.java:160)
at ceylon.modules.Main.execute(Main.java:69)
at ceylon.modules.Main.main(Main.java:42)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at org.jboss.modules.Module.run(Module.java:312)
at org.jboss.modules.Main.main(Main.java:460)
at ceylon.modules.bootstrap.CeylonRunTool.run(CeylonRunTool.java:246)
at ceylon.modules.bootstrap.CeylonTestTool.run(CeylonTestTool.java:173)
at com.redhat.ceylon.common.tools.CeylonTool.run(CeylonTool.java:390)
at com.redhat.ceylon.common.tools.CeylonTool.execute(CeylonTool.java:327)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at com.redhat.ceylon.launcher.Launcher.runInJava7Checked(Launcher.java:114)
at com.redhat.ceylon.launcher.Launcher.run(Launcher.java:41)
at com.redhat.ceylon.launcher.Launcher.run(Launcher.java:34)
at com.redhat.ceylon.launcher.Launcher.main(Launcher.java:27)
The thing is that the `test.ceylon.net` module.ceylon does `import java.base "7";` The tests run fine in the CLI.
|
1.0
|
Running ceylon.net tests - If I try to run the ceylon.net tests (i.e. the module `test.ceylon.net` as a Ceylon Test) in the IDE I get the following error:
ceylon test: Error while loading the test.ceylon.net/1.1.1 module:
Declaration 'java.net.URI' could not be found in module 'test.ceylon.net' or
its imported modules but was found in the non-imported module 'java.base'
com.redhat.ceylon.model.loader.ModelResolutionException: Error while loading the test.ceylon.net/1.1.1 module:
Declaration 'java.net.URI' could not be found in module 'test.ceylon.net' or its imported modules but was found in the non-imported module 'java.base'
at com.redhat.ceylon.compiler.java.runtime.model.RuntimeModelLoader$1.reportError(RuntimeModelLoader.java:209)
at com.redhat.ceylon.model.typechecker.model.UnknownType.reportErrors(UnknownType.java:94)
at com.redhat.ceylon.compiler.java.runtime.metamodel.Metamodel.getMetamodel(Metamodel.java:504)
at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeValue.<init>(FreeValue.java:43)
at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeReference.<init>(FreeReference.java:17)
at com.redhat.ceylon.compiler.java.runtime.metamodel.Metamodel.getOrCreateMetamodel(Metamodel.java:344)
at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeClassOrInterface.init(FreeClassOrInterface.java:92)
at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeClass.init(FreeClass.java:51)
at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeClassOrInterface.checkInit(FreeClassOrInterface.java:144)
at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeClassOrInterface.filteredMembers(FreeClassOrInterface.java:212)
at com.redhat.ceylon.compiler.java.runtime.metamodel.FreeClassOrInterface.annotatedMemberDeclarations(FreeClassOrInterface.java:186)
at ceylon.test.core.findCandidatesInClass_.findCandidatesInClass(DefaultTestRunner.ceylon:206)
at ceylon.test.core.findCandidatesInPackage_.findCandidatesInPackage(DefaultTestRunner.ceylon:200)
at ceylon.test.core.findCandidatesInModule_.findCandidatesInModule(DefaultTestRunner.ceylon:191)
at ceylon.test.core.findCandidatesInModuleLiteral_.findCandidatesInModuleLiteral(DefaultTestRunner.ceylon:259)
at ceylon.test.core.findCandidatesInTypeLiteral_.findCandidatesInTypeLiteral(DefaultTestRunner.ceylon:233)
at ceylon.test.core.findCandidates_.findCandidates(DefaultTestRunner.ceylon:183)
at ceylon.test.core.createExecutors_.createExecutors(DefaultTestRunner.ceylon:85)
at ceylon.test.core.DefaultTestRunner.initExecutors$priv$(DefaultTestRunner.ceylon:43)
at ceylon.test.core.DefaultTestRunner.<init>(DefaultTestRunner.ceylon:46)
at ceylon.test.createTestRunner_.createTestRunner(TestRunner.ceylon:49)
at ceylon.test.createTestRunner_.createTestRunner(TestRunner.ceylon:48)
at com.redhat.ceylon.testjvm.Runner.run(tool.ceylon:85)
at com.redhat.ceylon.testjvm.run_.run(tool.ceylon:36)
at com.redhat.ceylon.testjvm.run_.main(tool.ceylon)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at ceylon.modules.api.runtime.SecurityActions.invokeRunInternal(SecurityActions.java:57)
at ceylon.modules.api.runtime.SecurityActions.invokeRun(SecurityActions.java:48)
at ceylon.modules.api.runtime.AbstractRuntime.invokeRun(AbstractRuntime.java:110)
at ceylon.modules.api.runtime.AbstractRuntime.execute(AbstractRuntime.java:176)
at ceylon.modules.api.runtime.AbstractRuntime.execute(AbstractRuntime.java:160)
at ceylon.modules.Main.execute(Main.java:69)
at ceylon.modules.Main.main(Main.java:42)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at org.jboss.modules.Module.run(Module.java:312)
at org.jboss.modules.Main.main(Main.java:460)
at ceylon.modules.bootstrap.CeylonRunTool.run(CeylonRunTool.java:246)
at ceylon.modules.bootstrap.CeylonTestTool.run(CeylonTestTool.java:173)
at com.redhat.ceylon.common.tools.CeylonTool.run(CeylonTool.java:390)
at com.redhat.ceylon.common.tools.CeylonTool.execute(CeylonTool.java:327)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at com.redhat.ceylon.launcher.Launcher.runInJava7Checked(Launcher.java:114)
at com.redhat.ceylon.launcher.Launcher.run(Launcher.java:41)
at com.redhat.ceylon.launcher.Launcher.run(Launcher.java:34)
at com.redhat.ceylon.launcher.Launcher.main(Launcher.java:27)
The thing is that the `test.ceylon.net` module.ceylon does `import java.base "7";` The tests run fine in the CLI.
|
test
|
running ceylon net tests if i try to run the ceylon net tests i e the module test ceylon net as a ceylon test in the ide i get the following error ceylon test error while loading the test ceylon net module declaration java net uri could not be found in module test ceylon net or its imported modules but was found in the non imported module java base com redhat ceylon model loader modelresolutionexception error while loading the test ceylon net module declaration java net uri could not be found in module test ceylon net or its imported modules but was found in the non imported module java base at com redhat ceylon compiler java runtime model runtimemodelloader reporterror runtimemodelloader java at com redhat ceylon model typechecker model unknowntype reporterrors unknowntype java at com redhat ceylon compiler java runtime metamodel metamodel getmetamodel metamodel java at com redhat ceylon compiler java runtime metamodel freevalue freevalue java at com redhat ceylon compiler java runtime metamodel freereference freereference java at com redhat ceylon compiler java runtime metamodel metamodel getorcreatemetamodel metamodel java at com redhat ceylon compiler java runtime metamodel freeclassorinterface init freeclassorinterface java at com redhat ceylon compiler java runtime metamodel freeclass init freeclass java at com redhat ceylon compiler java runtime metamodel freeclassorinterface checkinit freeclassorinterface java at com redhat ceylon compiler java runtime metamodel freeclassorinterface filteredmembers freeclassorinterface java at com redhat ceylon compiler java runtime metamodel freeclassorinterface annotatedmemberdeclarations freeclassorinterface java at ceylon test core findcandidatesinclass findcandidatesinclass defaulttestrunner ceylon at ceylon test core findcandidatesinpackage findcandidatesinpackage defaulttestrunner ceylon at ceylon test core findcandidatesinmodule findcandidatesinmodule defaulttestrunner ceylon at ceylon test core findcandidatesinmoduleliteral findcandidatesinmoduleliteral defaulttestrunner ceylon at ceylon test core findcandidatesintypeliteral findcandidatesintypeliteral defaulttestrunner ceylon at ceylon test core findcandidates findcandidates defaulttestrunner ceylon at ceylon test core createexecutors createexecutors defaulttestrunner ceylon at ceylon test core defaulttestrunner initexecutors priv defaulttestrunner ceylon at ceylon test core defaulttestrunner defaulttestrunner ceylon at ceylon test createtestrunner createtestrunner testrunner ceylon at ceylon test createtestrunner createtestrunner testrunner ceylon at com redhat ceylon testjvm runner run tool ceylon at com redhat ceylon testjvm run run tool ceylon at com redhat ceylon testjvm run main tool ceylon at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at ceylon modules api runtime securityactions invokeruninternal securityactions java at ceylon modules api runtime securityactions invokerun securityactions java at ceylon modules api runtime abstractruntime invokerun abstractruntime java at ceylon modules api runtime abstractruntime execute abstractruntime java at ceylon modules api runtime abstractruntime execute abstractruntime java at ceylon modules main execute main java at ceylon modules main main main java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org jboss modules module run module java at org jboss modules main main main java at ceylon modules bootstrap ceylonruntool run ceylonruntool java at ceylon modules bootstrap ceylontesttool run ceylontesttool java at com redhat ceylon common tools ceylontool run ceylontool java at com redhat ceylon common tools ceylontool execute ceylontool java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com redhat ceylon launcher launcher launcher java at com redhat ceylon launcher launcher run launcher java at com redhat ceylon launcher launcher run launcher java at com redhat ceylon launcher launcher main launcher java the thing is that the test ceylon net module ceylon does import java base the tests run fine in the cli
| 1
|
119,137
| 10,026,416,381
|
IssuesEvent
|
2019-07-17 06:34:22
|
Sinapse-Energia/AP-Sinapse
|
https://api.github.com/repos/Sinapse-Energia/AP-Sinapse
|
closed
|
[AP][ALL] Integrate 6LP features into the main branch
|
Device: All Priority: high Size: 13 Status: done - to be tested Type: integration
|
Now, we have two parallel branches and we miss several important features in each one.
It is necessary to integrate the 6LP RF as a communication southbound in the main branch in order to have all the developed features for CMC, EN and AP accessible
|
1.0
|
[AP][ALL] Integrate 6LP features into the main branch - Now, we have two parallel branches and we miss several important features in each one.
It is necessary to integrate the 6LP RF as a communication southbound in the main branch in order to have all the developed features for CMC, EN and AP accessible
|
test
|
integrate features into the main branch now we have two parallel branches and we miss several important features in each one it is necessary to integrate the rf as a communication southbound in the main branch in order to have all the developed features for cmc en and ap accessible
| 1
|
229,327
| 25,319,035,562
|
IssuesEvent
|
2022-11-18 01:06:19
|
DavidSpek/kubeflow
|
https://api.github.com/repos/DavidSpek/kubeflow
|
opened
|
CVE-2022-45198 (High) detected in Pillow-8.3.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl
|
security vulnerability
|
## CVE-2022-45198 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-8.3.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl</b></p></summary>
<p>Python Imaging Library (Fork)</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/8e/7a/b047f6f80fdb02c0cca1d3761d71e9800bcf6d4874b71c9e6548ec59e156/Pillow-8.3.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/8e/7a/b047f6f80fdb02c0cca1d3761d71e9800bcf6d4874b71c9e6548ec59e156/Pillow-8.3.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: /components/crud-web-apps/common/backend</p>
<p>Path to vulnerable library: /components/crud-web-apps/common/backend</p>
<p>
Dependency Hierarchy:
- :x: **Pillow-8.3.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Pillow before 9.2.0 performs Improper Handling of Highly Compressed GIF Data (Data Amplification).
<p>Publish Date: 2022-11-14
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-45198>CVE-2022-45198</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-11-14</p>
<p>Fix Resolution: Pillow - 9.2.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-45198 (High) detected in Pillow-8.3.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl - ## CVE-2022-45198 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-8.3.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl</b></p></summary>
<p>Python Imaging Library (Fork)</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/8e/7a/b047f6f80fdb02c0cca1d3761d71e9800bcf6d4874b71c9e6548ec59e156/Pillow-8.3.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/8e/7a/b047f6f80fdb02c0cca1d3761d71e9800bcf6d4874b71c9e6548ec59e156/Pillow-8.3.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: /components/crud-web-apps/common/backend</p>
<p>Path to vulnerable library: /components/crud-web-apps/common/backend</p>
<p>
Dependency Hierarchy:
- :x: **Pillow-8.3.1-cp37-cp37m-manylinux_2_5_x86_64.manylinux1_x86_64.whl** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Pillow before 9.2.0 performs Improper Handling of Highly Compressed GIF Data (Data Amplification).
<p>Publish Date: 2022-11-14
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-45198>CVE-2022-45198</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-11-14</p>
<p>Fix Resolution: Pillow - 9.2.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve high detected in pillow manylinux whl cve high severity vulnerability vulnerable library pillow manylinux whl python imaging library fork library home page a href path to dependency file components crud web apps common backend path to vulnerable library components crud web apps common backend dependency hierarchy x pillow manylinux whl vulnerable library found in base branch master vulnerability details pillow before performs improper handling of highly compressed gif data data amplification publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution pillow step up your open source security game with mend
| 0
|
31,661
| 13,610,571,679
|
IssuesEvent
|
2020-09-23 07:35:06
|
microsoft/botframework-solutions
|
https://api.github.com/repos/microsoft/botframework-solutions
|
closed
|
[MS Teams] Unable to upload attachments to bot
|
Bot Services Needs Triage Type: Bug customer-replied-to customer-reported
|
Hi,
We have 1v1 bot available to our users through MS Teams. Since roughly this week, it is no longer possible to upload attachments to the bot - it fails with error (see below).
The uploading itself is done simply by opening up chat window with bot, clicking on paperclip icon, clicking on _Upload from my computer_ and finally choosing the file to upload:

The progress bar then appears:

After uploading completes, the following error popup is displayed and attachment is not sent over:

Additional notes:
* Choosing _Onedrive_ option does not show the error popup, but no file is sent - nothing happens.
* The file type and size does not matter. But all the tests were done with png and txt file with no more than few kilobytes.
* The error happens in both the desktop and web version of Teams client.
* Same error happens for all of our Teams enabled bots, including one that was set up just last week.
* The error does not happen if inline image is pasted and sent over.
* The Teams tenant is based in Europe.
* This should be irrelevant, but the bot itself is powered by .NET SDK of MS Bot Framework v4 and hosted in Azure with MS Teams enabled as one of its channels.
For the reference, here is Teams package manifest:
```
{
"$schema": "https://developer.microsoft.com/en-us/json-schemas/teams/v1.7/MicrosoftTeams.schema.json",
"manifestVersion": "1.7",
"version": "0.0.1",
"id": "<removed>",
"packageName": "<removed>",
"developer": {
"name": "<removed>",
"websiteUrl": "<removed>",
"privacyUrl": "<removed>",
"termsOfUseUrl": "<removed>"
},
"icons": {
"color": "color.png",
"outline": "outline.png"
},
"name": {
"short": "<removed>",
"full": "<removed>"
},
"description": {
"short": "<removed>",
"full": "<removed>"
},
"accentColor": "#761092",
"bots": [
{
"botId": "<removed>",
"scopes": [
"personal"
],
"supportsFiles": true,
"isNotificationOnly": false
}
],
"permissions": [
"identity",
"messageTeamMembers"
],
"validDomains": []
}
```
Will appreciate any ideas, thanks.
|
1.0
|
[MS Teams] Unable to upload attachments to bot - Hi,
We have 1v1 bot available to our users through MS Teams. Since roughly this week, it is no longer possible to upload attachments to the bot - it fails with error (see below).
The uploading itself is done simply by opening up chat window with bot, clicking on paperclip icon, clicking on _Upload from my computer_ and finally choosing the file to upload:

The progress bar then appears:

After uploading completes, the following error popup is displayed and attachment is not sent over:

Additional notes:
* Choosing _Onedrive_ option does not show the error popup, but no file is sent - nothing happens.
* The file type and size does not matter. But all the tests were done with png and txt file with no more than few kilobytes.
* The error happens in both the desktop and web version of Teams client.
* Same error happens for all of our Teams enabled bots, including one that was set up just last week.
* The error does not happen if inline image is pasted and sent over.
* The Teams tenant is based in Europe.
* This should be irrelevant, but the bot itself is powered by .NET SDK of MS Bot Framework v4 and hosted in Azure with MS Teams enabled as one of its channels.
For the reference, here is Teams package manifest:
```
{
"$schema": "https://developer.microsoft.com/en-us/json-schemas/teams/v1.7/MicrosoftTeams.schema.json",
"manifestVersion": "1.7",
"version": "0.0.1",
"id": "<removed>",
"packageName": "<removed>",
"developer": {
"name": "<removed>",
"websiteUrl": "<removed>",
"privacyUrl": "<removed>",
"termsOfUseUrl": "<removed>"
},
"icons": {
"color": "color.png",
"outline": "outline.png"
},
"name": {
"short": "<removed>",
"full": "<removed>"
},
"description": {
"short": "<removed>",
"full": "<removed>"
},
"accentColor": "#761092",
"bots": [
{
"botId": "<removed>",
"scopes": [
"personal"
],
"supportsFiles": true,
"isNotificationOnly": false
}
],
"permissions": [
"identity",
"messageTeamMembers"
],
"validDomains": []
}
```
Will appreciate any ideas, thanks.
|
non_test
|
unable to upload attachments to bot hi we have bot available to our users through ms teams since roughly this week it is no longer possible to upload attachments to the bot it fails with error see below the uploading itself is done simply by opening up chat window with bot clicking on paperclip icon clicking on upload from my computer and finally choosing the file to upload the progress bar then appears after uploading completes the following error popup is displayed and attachment is not sent over additional notes choosing onedrive option does not show the error popup but no file is sent nothing happens the file type and size does not matter but all the tests were done with png and txt file with no more than few kilobytes the error happens in both the desktop and web version of teams client same error happens for all of our teams enabled bots including one that was set up just last week the error does not happen if inline image is pasted and sent over the teams tenant is based in europe this should be irrelevant but the bot itself is powered by net sdk of ms bot framework and hosted in azure with ms teams enabled as one of its channels for the reference here is teams package manifest schema manifestversion version id packagename developer name websiteurl privacyurl termsofuseurl icons color color png outline outline png name short full description short full accentcolor bots botid scopes personal supportsfiles true isnotificationonly false permissions identity messageteammembers validdomains will appreciate any ideas thanks
| 0
|
96,495
| 10,934,215,437
|
IssuesEvent
|
2019-11-24 09:50:31
|
hapijs/joi
|
https://api.github.com/repos/hapijs/joi
|
closed
|
Promise use in `.validateAsync`
|
documentation
|
<!--
⚠️ ⚠️ ⚠️ ⚠️ ⚠️ ⚠️
You must complete this entire issue template to receive support. You MUST NOT remove, change, or replace the template with your own format. A missing or incomplete report will cause your issue to be closed without comment. Please respect the time and experience that went into this template. It is here for a reason. Thank you!
⚠️ ⚠️ ⚠️ ⚠️ ⚠️ ⚠️
-->
#### Context
* *node version*: 8.11.1
* *module version*: 16.1.7
#### What are you trying to achieve or the steps to reproduce ?
<!--
Before opening a documentation issue, please consider opening a Pull Request instead for trivial changes such as typos, spelling, incorrect links, anchors, or other corrections that are easier to just fix than report using this template.
Please do not spend valuable time proposing extensive changes to the documentation before first asking about it. We value your time and do not want to waste it. Just open an issue first using this template and ask if your proposed changes would be helpful.
Make sure to wrap all code examples in backticks so that they display correctly. Before submitting an issue, make sure to click on the Preview tab above to verify everything is formatted correctly.
-->
I'm trying to use `.validateAsync()` and consume the results as a promise. I'm having trouble finding the documentation about when the promise resolves or rejects and what arguments will be supplied to `.then` and `.catch`.
I assumed both `then` and `catch` would receive an object with `value`, `error`, and `warning` keys similar to the synchronous version. Testing and logging has shown this to not be the case.
1. What are the arguments to `.then`? Is it just the updated value or `undefined` if no transformations were applied?
2. What are the arguments to `.catch`? Is it just the error?
3. If `.catch` is called, are joi transformations applied to the original value?
|
1.0
|
Promise use in `.validateAsync` - <!--
⚠️ ⚠️ ⚠️ ⚠️ ⚠️ ⚠️
You must complete this entire issue template to receive support. You MUST NOT remove, change, or replace the template with your own format. A missing or incomplete report will cause your issue to be closed without comment. Please respect the time and experience that went into this template. It is here for a reason. Thank you!
⚠️ ⚠️ ⚠️ ⚠️ ⚠️ ⚠️
-->
#### Context
* *node version*: 8.11.1
* *module version*: 16.1.7
#### What are you trying to achieve or the steps to reproduce ?
<!--
Before opening a documentation issue, please consider opening a Pull Request instead for trivial changes such as typos, spelling, incorrect links, anchors, or other corrections that are easier to just fix than report using this template.
Please do not spend valuable time proposing extensive changes to the documentation before first asking about it. We value your time and do not want to waste it. Just open an issue first using this template and ask if your proposed changes would be helpful.
Make sure to wrap all code examples in backticks so that they display correctly. Before submitting an issue, make sure to click on the Preview tab above to verify everything is formatted correctly.
-->
I'm trying to use `.validateAsync()` and consume the results as a promise. I'm having trouble finding the documentation about when the promise resolves or rejects and what arguments will be supplied to `.then` and `.catch`.
I assumed both `then` and `catch` would receive an object with `value`, `error`, and `warning` keys similar to the synchronous version. Testing and logging has shown this to not be the case.
1. What are the arguments to `.then`? Is it just the updated value or `undefined` if no transformations were applied?
2. What are the arguments to `.catch`? Is it just the error?
3. If `.catch` is called, are joi transformations applied to the original value?
|
non_test
|
promise use in validateasync ⚠️ ⚠️ ⚠️ ⚠️ ⚠️ ⚠️ you must complete this entire issue template to receive support you must not remove change or replace the template with your own format a missing or incomplete report will cause your issue to be closed without comment please respect the time and experience that went into this template it is here for a reason thank you ⚠️ ⚠️ ⚠️ ⚠️ ⚠️ ⚠️ context node version module version what are you trying to achieve or the steps to reproduce before opening a documentation issue please consider opening a pull request instead for trivial changes such as typos spelling incorrect links anchors or other corrections that are easier to just fix than report using this template please do not spend valuable time proposing extensive changes to the documentation before first asking about it we value your time and do not want to waste it just open an issue first using this template and ask if your proposed changes would be helpful make sure to wrap all code examples in backticks so that they display correctly before submitting an issue make sure to click on the preview tab above to verify everything is formatted correctly i m trying to use validateasync and consume the results as a promise i m having trouble finding the documentation about when the promise resolves or rejects and what arguments will be supplied to then and catch i assumed both then and catch would receive an object with value error and warning keys similar to the synchronous version testing and logging has shown this to not be the case what are the arguments to then is it just the updated value or undefined if no transformations were applied what are the arguments to catch is it just the error if catch is called are joi transformations applied to the original value
| 0
|
527,994
| 15,357,901,472
|
IssuesEvent
|
2021-03-01 14:13:44
|
airshipit/airshipctl
|
https://api.github.com/repos/airshipit/airshipctl
|
closed
|
Enhance Github Bot to add proper labels to move issues to InProgress
|
enhancement priority/medium size s
|
Is it possible for the bot to besides adding the WIP label , change the status to In Progress?
|
1.0
|
Enhance Github Bot to add proper labels to move issues to InProgress -
Is it possible for the bot to besides adding the WIP label , change the status to In Progress?
|
non_test
|
enhance github bot to add proper labels to move issues to inprogress is it possible for the bot to besides adding the wip label change the status to in progress
| 0
|
133,912
| 10,866,832,406
|
IssuesEvent
|
2019-11-14 22:10:58
|
rancher/rancher
|
https://api.github.com/repos/rancher/rancher
|
closed
|
vSphere Legacy Template Normalizing
|
[zube]: To Test area/vsphere team/az
|
The old fields in vmwarevsphere templates were free form text fields and the finder methods in govmomi allowed for something like "My Network" to be a lookup when the real object path was "/CLUSTER_NAME/networks/My Network". We need a shim during the upgrade process which looks at a template and tries to upgrade the data to the full path we now use w/ the lookup api in 2.3.3+
|
1.0
|
vSphere Legacy Template Normalizing - The old fields in vmwarevsphere templates were free form text fields and the finder methods in govmomi allowed for something like "My Network" to be a lookup when the real object path was "/CLUSTER_NAME/networks/My Network". We need a shim during the upgrade process which looks at a template and tries to upgrade the data to the full path we now use w/ the lookup api in 2.3.3+
|
test
|
vsphere legacy template normalizing the old fields in vmwarevsphere templates were free form text fields and the finder methods in govmomi allowed for something like my network to be a lookup when the real object path was cluster name networks my network we need a shim during the upgrade process which looks at a template and tries to upgrade the data to the full path we now use w the lookup api in
| 1
|
343,868
| 24,788,506,136
|
IssuesEvent
|
2022-10-24 11:56:14
|
opted-eu/wp3inventory
|
https://api.github.com/repos/opted-eu/wp3inventory
|
opened
|
Contribution Guidelines for Tools
|
documentation OPTED Infrastructure
|
Collect here rules and guidelines for adding tools
- "name" should always be as short as possible. Full name of abbreviations and acronyms are stored in "other names" and can be repeated in the "description" text
- Affective Norms for English Words ➡️ ANEW
- "The General Inquirer: A Computer System for Content Analysis and Retrieval Based on the Sentence as a Unit of Information" ➡️ General Inquirer Dictionaries
- If there are duplicated entries, the older entry is retained, the newer entry is rejected.
- exception: the new entry has significantly more information and is of higher quality
- in any case, the "unique_name" field has to be checked afterwards in the canonical entry
|
1.0
|
Contribution Guidelines for Tools - Collect here rules and guidelines for adding tools
- "name" should always be as short as possible. Full name of abbreviations and acronyms are stored in "other names" and can be repeated in the "description" text
- Affective Norms for English Words ➡️ ANEW
- "The General Inquirer: A Computer System for Content Analysis and Retrieval Based on the Sentence as a Unit of Information" ➡️ General Inquirer Dictionaries
- If there are duplicated entries, the older entry is retained, the newer entry is rejected.
- exception: the new entry has significantly more information and is of higher quality
- in any case, the "unique_name" field has to be checked afterwards in the canonical entry
|
non_test
|
contribution guidelines for tools collect here rules and guidelines for adding tools name should always be as short as possible full name of abbreviations and acronyms are stored in other names and can be repeated in the description text affective norms for english words ➡️ anew the general inquirer a computer system for content analysis and retrieval based on the sentence as a unit of information ➡️ general inquirer dictionaries if there are duplicated entries the older entry is retained the newer entry is rejected exception the new entry has significantly more information and is of higher quality in any case the unique name field has to be checked afterwards in the canonical entry
| 0
|
287,935
| 24,876,046,079
|
IssuesEvent
|
2022-10-27 19:11:43
|
lowRISC/opentitan
|
https://api.github.com/repos/lowRISC/opentitan
|
closed
|
[rom-e2e] rom_e2e_asm_interrupt_handler
|
Type:Task SW:ROM Milestone:V2 Component:Rom/E2e/Test
|
**Testpoint name:** [rom_e2e_asm_interrupt_handler](https://cs.opensource.google/opentitan/opentitan/+/master:sw/device/silicon_creator/rom/data/rom_e2e_testplan.hjson?q=rom_e2e_asm_interrupt_handler)
**Contact person:** @alphan
**Description:** Verify that asm interrupt handler resets the chip.
`CREATOR_SW_CFG_ROM_EXEC_EN` should be set to `0` and the chip should in a life cycle
state where debugging is enabled, i.e. TEST, DEV, or RMA.
- Connect a debugger.
- Set a breakpoint on the asm handler.
- Trigger an exception, e.g. by setting `PC` to an address (end of ROM) that will trigger an
instruction access fault.
- Verify that execution breaks at the asm handler.
|
1.0
|
[rom-e2e] rom_e2e_asm_interrupt_handler - **Testpoint name:** [rom_e2e_asm_interrupt_handler](https://cs.opensource.google/opentitan/opentitan/+/master:sw/device/silicon_creator/rom/data/rom_e2e_testplan.hjson?q=rom_e2e_asm_interrupt_handler)
**Contact person:** @alphan
**Description:** Verify that asm interrupt handler resets the chip.
`CREATOR_SW_CFG_ROM_EXEC_EN` should be set to `0` and the chip should in a life cycle
state where debugging is enabled, i.e. TEST, DEV, or RMA.
- Connect a debugger.
- Set a breakpoint on the asm handler.
- Trigger an exception, e.g. by setting `PC` to an address (end of ROM) that will trigger an
instruction access fault.
- Verify that execution breaks at the asm handler.
|
test
|
rom asm interrupt handler testpoint name contact person alphan description verify that asm interrupt handler resets the chip creator sw cfg rom exec en should be set to and the chip should in a life cycle state where debugging is enabled i e test dev or rma connect a debugger set a breakpoint on the asm handler trigger an exception e g by setting pc to an address end of rom that will trigger an instruction access fault verify that execution breaks at the asm handler
| 1
|
271,269
| 23,593,577,763
|
IssuesEvent
|
2022-08-23 17:12:03
|
MPMG-DCC-UFMG/F01
|
https://api.github.com/repos/MPMG-DCC-UFMG/F01
|
closed
|
Teste de generalizacao para a tag Obras públicas - Dados para acompanhamento - Guaranésia
|
generalization test development template - Betha tag - Obras Públicas subtag - Dados para acompanhamento
|
DoD: Realizar o teste de Generalização do validador da tag Obras públicas - Dados para acompanhamento para o Município de Guaranésia.
|
1.0
|
Teste de generalizacao para a tag Obras públicas - Dados para acompanhamento - Guaranésia - DoD: Realizar o teste de Generalização do validador da tag Obras públicas - Dados para acompanhamento para o Município de Guaranésia.
|
test
|
teste de generalizacao para a tag obras públicas dados para acompanhamento guaranésia dod realizar o teste de generalização do validador da tag obras públicas dados para acompanhamento para o município de guaranésia
| 1
|
234,115
| 19,096,229,636
|
IssuesEvent
|
2021-11-29 16:57:33
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
kv/kvserver: TestRequestsOnLaggingReplica failed
|
C-test-failure O-robot branch-release-21.2
|
kv/kvserver.TestRequestsOnLaggingReplica [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3813272&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3813272&tab=artifacts#/) on release-21.2 @ [d60310f8bbc38b9d090847b2266a90ff11a300e6](https://github.com/cockroachdb/cockroach/commits/d60310f8bbc38b9d090847b2266a90ff11a300e6):
```
=== RUN TestRequestsOnLaggingReplica
test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestRequestsOnLaggingReplica546084043
test_log_scope.go:80: use -show-logs to present logs inline
client_raft_test.go:1281:
Error Trace: client_raft_test.go:1281
Error: Not equal:
expected: 2
actual : 1
Test: TestRequestsOnLaggingReplica
panic.go:613: -- test log scope end --
test logs left over in: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestRequestsOnLaggingReplica546084043
--- FAIL: TestRequestsOnLaggingReplica (5.76s)
```
<details><summary>Reproduce</summary>
<p>
To reproduce, try:
```bash
make stressrace TESTS=TestRequestsOnLaggingReplica PKG=./pkg/kv/kvserver TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
Parameters in this failure:
- GOFLAGS=-json
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #57932 kv/kvserver: TestRequestsOnLaggingReplica failed [C-test-failure O-robot branch-master]
</p>
</details>
/cc @cockroachdb/kv andreimatei
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestRequestsOnLaggingReplica.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
1.0
|
kv/kvserver: TestRequestsOnLaggingReplica failed - kv/kvserver.TestRequestsOnLaggingReplica [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3813272&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3813272&tab=artifacts#/) on release-21.2 @ [d60310f8bbc38b9d090847b2266a90ff11a300e6](https://github.com/cockroachdb/cockroach/commits/d60310f8bbc38b9d090847b2266a90ff11a300e6):
```
=== RUN TestRequestsOnLaggingReplica
test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestRequestsOnLaggingReplica546084043
test_log_scope.go:80: use -show-logs to present logs inline
client_raft_test.go:1281:
Error Trace: client_raft_test.go:1281
Error: Not equal:
expected: 2
actual : 1
Test: TestRequestsOnLaggingReplica
panic.go:613: -- test log scope end --
test logs left over in: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestRequestsOnLaggingReplica546084043
--- FAIL: TestRequestsOnLaggingReplica (5.76s)
```
<details><summary>Reproduce</summary>
<p>
To reproduce, try:
```bash
make stressrace TESTS=TestRequestsOnLaggingReplica PKG=./pkg/kv/kvserver TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
Parameters in this failure:
- GOFLAGS=-json
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #57932 kv/kvserver: TestRequestsOnLaggingReplica failed [C-test-failure O-robot branch-master]
</p>
</details>
/cc @cockroachdb/kv andreimatei
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestRequestsOnLaggingReplica.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
test
|
kv kvserver testrequestsonlaggingreplica failed kv kvserver testrequestsonlaggingreplica with on release run testrequestsonlaggingreplica test log scope go test logs captured to go src github com cockroachdb cockroach artifacts test log scope go use show logs to present logs inline client raft test go error trace client raft test go error not equal expected actual test testrequestsonlaggingreplica panic go test log scope end test logs left over in go src github com cockroachdb cockroach artifacts fail testrequestsonlaggingreplica reproduce to reproduce try bash make stressrace tests testrequestsonlaggingreplica pkg pkg kv kvserver testtimeout stressflags timeout parameters in this failure goflags json same failure on other branches kv kvserver testrequestsonlaggingreplica failed cc cockroachdb kv andreimatei
| 1
|
317,457
| 27,238,270,162
|
IssuesEvent
|
2023-02-21 18:00:52
|
DCS-LCSR/SignStream3
|
https://api.github.com/repos/DCS-LCSR/SignStream3
|
closed
|
June 1 Program seems to hang when Sign Bank search fails because DAI server is down
|
enhancement CRITICAL testing-June-1 severity MED usability concern
|
<img width="947" alt="Screen Shot 2021-06-09 at 4 24 05 AM" src="https://user-images.githubusercontent.com/13629122/121319927-aca74080-c8da-11eb-9245-bf8aa6d9c292.png">
It is unresponsive for a very long time.
|
1.0
|
June 1 Program seems to hang when Sign Bank search fails because DAI server is down - <img width="947" alt="Screen Shot 2021-06-09 at 4 24 05 AM" src="https://user-images.githubusercontent.com/13629122/121319927-aca74080-c8da-11eb-9245-bf8aa6d9c292.png">
It is unresponsive for a very long time.
|
test
|
june program seems to hang when sign bank search fails because dai server is down img width alt screen shot at am src it is unresponsive for a very long time
| 1
|
211,242
| 16,191,851,584
|
IssuesEvent
|
2021-05-04 09:35:45
|
ubtue/tuefind
|
https://api.github.com/repos/ubtue/tuefind
|
closed
|
Alertpakete: irreführende Anzeige "Druck"
|
System: RelBib ready for testing
|
In der Anzeige der Alertpakete ist immer "Druck" enthalten:
https://www.relbib.de/Search/Results?type=AllFields&filter%5B%5D=%7Eformat%3A%22Subscription+Bundle%22
Das ist besonders irreführend bei dem Paket "Open Access Zeitschriften.
Kann diese Anzeige entfernt bzw. korrigiert werden, falls der Aufwand nicht zu hoch? Z. B. so:
1. Anzeige "elektronisch", wenn nur elektronische Zss im Paket enthalten
2. Anzeige "elektronisch / Druck", wenn sowohl Druck- als auch elektronische Zss enthalten
3. Anzeige "Druck", wenn nur Druck-Zss enthalten
Das müsste dann natürlich dynamisch funktionieren.
|
1.0
|
Alertpakete: irreführende Anzeige "Druck" - In der Anzeige der Alertpakete ist immer "Druck" enthalten:
https://www.relbib.de/Search/Results?type=AllFields&filter%5B%5D=%7Eformat%3A%22Subscription+Bundle%22
Das ist besonders irreführend bei dem Paket "Open Access Zeitschriften.
Kann diese Anzeige entfernt bzw. korrigiert werden, falls der Aufwand nicht zu hoch? Z. B. so:
1. Anzeige "elektronisch", wenn nur elektronische Zss im Paket enthalten
2. Anzeige "elektronisch / Druck", wenn sowohl Druck- als auch elektronische Zss enthalten
3. Anzeige "Druck", wenn nur Druck-Zss enthalten
Das müsste dann natürlich dynamisch funktionieren.
|
test
|
alertpakete irreführende anzeige druck in der anzeige der alertpakete ist immer druck enthalten das ist besonders irreführend bei dem paket open access zeitschriften kann diese anzeige entfernt bzw korrigiert werden falls der aufwand nicht zu hoch z b so anzeige elektronisch wenn nur elektronische zss im paket enthalten anzeige elektronisch druck wenn sowohl druck als auch elektronische zss enthalten anzeige druck wenn nur druck zss enthalten das müsste dann natürlich dynamisch funktionieren
| 1
|
267,956
| 23,335,807,528
|
IssuesEvent
|
2022-08-09 09:49:44
|
mozilla-mobile/fenix
|
https://api.github.com/repos/mozilla-mobile/fenix
|
opened
|
Intermittent Unit Test failure - StartupTypeTelemetryTest.WHEN record is called THEN telemetry is recorded with the appropriate label
|
eng:intermittent-test
|
### Test Run:
Test failed in https://github.com/mozilla-mobile/fenix/pull/26378 - https://firefoxci.taskcluster-artifacts.net/OIThuwo1TRSMqXOucAYXqg/0/public/reports/test/testDebugUnitTest/index.html
### Stacktrace:
> java.lang.AssertionError: expected:<1> but was:<null>
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:834)
at org.junit.Assert.assertEquals(Assert.java:118)
at org.junit.Assert.assertEquals(Assert.java:144)
at org.mozilla.fenix.perf.StartupTypeTelemetryTest.WHEN record is called THEN telemetry is recorded with the appropriate label(StartupTypeTelemetryTest.kt:96)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
at org.junit.rules.RunRules.evaluate(RunRules.java:20)
at org.robolectric.RobolectricTestRunner$HelperTestRunner$1.evaluate(RobolectricTestRunner.java:591)
at org.robolectric.internal.SandboxTestRunner$2.lambda$evaluate$0(SandboxTestRunner.java:274)
at org.robolectric.internal.bytecode.Sandbox.lambda$runOnMainThread$0(Sandbox.java:88)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:829)
Suppressed: org.robolectric.android.internal.AndroidTestEnvironment$UnExecutedRunnablesException: Main looper has queued unexecuted runnables. This might be the cause of the test failure. You might need a shadowOf(Looper.getMainLooper()).idle() call.
### Build: 105
|
1.0
|
Intermittent Unit Test failure - StartupTypeTelemetryTest.WHEN record is called THEN telemetry is recorded with the appropriate label - ### Test Run:
Test failed in https://github.com/mozilla-mobile/fenix/pull/26378 - https://firefoxci.taskcluster-artifacts.net/OIThuwo1TRSMqXOucAYXqg/0/public/reports/test/testDebugUnitTest/index.html
### Stacktrace:
> java.lang.AssertionError: expected:<1> but was:<null>
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:834)
at org.junit.Assert.assertEquals(Assert.java:118)
at org.junit.Assert.assertEquals(Assert.java:144)
at org.mozilla.fenix.perf.StartupTypeTelemetryTest.WHEN record is called THEN telemetry is recorded with the appropriate label(StartupTypeTelemetryTest.kt:96)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
at org.junit.rules.RunRules.evaluate(RunRules.java:20)
at org.robolectric.RobolectricTestRunner$HelperTestRunner$1.evaluate(RobolectricTestRunner.java:591)
at org.robolectric.internal.SandboxTestRunner$2.lambda$evaluate$0(SandboxTestRunner.java:274)
at org.robolectric.internal.bytecode.Sandbox.lambda$runOnMainThread$0(Sandbox.java:88)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:829)
Suppressed: org.robolectric.android.internal.AndroidTestEnvironment$UnExecutedRunnablesException: Main looper has queued unexecuted runnables. This might be the cause of the test failure. You might need a shadowOf(Looper.getMainLooper()).idle() call.
### Build: 105
|
test
|
intermittent unit test failure startuptypetelemetrytest when record is called then telemetry is recorded with the appropriate label test run test failed in stacktrace java lang assertionerror expected but was at org junit assert fail assert java at org junit assert failnotequals assert java at org junit assert assertequals assert java at org junit assert assertequals assert java at org mozilla fenix perf startuptypetelemetrytest when record is called then telemetry is recorded with the appropriate label startuptypetelemetrytest kt at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements runbefores evaluate runbefores java at org junit rules testwatcher evaluate testwatcher java at org junit rules runrules evaluate runrules java at org robolectric robolectrictestrunner helpertestrunner evaluate robolectrictestrunner java at org robolectric internal sandboxtestrunner lambda evaluate sandboxtestrunner java at org robolectric internal bytecode sandbox lambda runonmainthread sandbox java at java base java util concurrent futuretask run futuretask java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java base java lang thread run thread java suppressed org robolectric android internal androidtestenvironment unexecutedrunnablesexception main looper has queued unexecuted runnables this might be the cause of the test failure you might need a shadowof looper getmainlooper idle call build
| 1
|
2,027
| 2,581,515,710
|
IssuesEvent
|
2015-02-14 03:54:28
|
easydigitaldownloads/Easy-Digital-Downloads
|
https://api.github.com/repos/easydigitaldownloads/Easy-Digital-Downloads
|
closed
|
Polylang support
|
Enhancement Needs Testing
|
I am looking to make [Polylang](https://wordpress.org/plugins/polylang/) work with EDD.
One thing I need to make multilingual is the checkout page. I created on in English and one in German.
So that [`edd_get_checkout_uri()`](https://github.com/easydigitaldownloads/Easy-Digital-Downloads/blob/master/includes/checkout/functions.php#L118) returns the correct link acording to the language I would need to run `$edd_options['purchase_page']` though `pll_get_post()`. The function `pll_get_post()` returns the post id depending on the language of the current page.
So something like this
```php
$edd_options = edd_get_settings();
// Translate post_id for pages in options
$edd_options['purchase_page'] = pll_get_post( $edd_options['purchase_page'] );
```
Is it possible to add a filter which I could hook into?
----------------
Two posts(downloads) are created so to be able to translate the content in two langues. At the moment depending on which language the site being viewed in that version of the download would be added to the basket.
A solution would be to only allow one version of the downloads to be added to the cart. I see that you use [`edd_add_to_cart()`](https://github.com/easydigitaldownloads/Easy-Digital-Downloads/blob/master/includes/cart/functions.php#L121) to add to the cart. If I were able to run this line of code then I would be able to achieve this.
```php
$download_id = pll_get_post( $download_id, pll_default_language() );
```
Is it possible to add a filter in this place?
I hope I am making sense. I can make a PR to add these two filters. I just wanted to discuss this before make it.
|
1.0
|
Polylang support - I am looking to make [Polylang](https://wordpress.org/plugins/polylang/) work with EDD.
One thing I need to make multilingual is the checkout page. I created on in English and one in German.
So that [`edd_get_checkout_uri()`](https://github.com/easydigitaldownloads/Easy-Digital-Downloads/blob/master/includes/checkout/functions.php#L118) returns the correct link acording to the language I would need to run `$edd_options['purchase_page']` though `pll_get_post()`. The function `pll_get_post()` returns the post id depending on the language of the current page.
So something like this
```php
$edd_options = edd_get_settings();
// Translate post_id for pages in options
$edd_options['purchase_page'] = pll_get_post( $edd_options['purchase_page'] );
```
Is it possible to add a filter which I could hook into?
----------------
Two posts(downloads) are created so to be able to translate the content in two langues. At the moment depending on which language the site being viewed in that version of the download would be added to the basket.
A solution would be to only allow one version of the downloads to be added to the cart. I see that you use [`edd_add_to_cart()`](https://github.com/easydigitaldownloads/Easy-Digital-Downloads/blob/master/includes/cart/functions.php#L121) to add to the cart. If I were able to run this line of code then I would be able to achieve this.
```php
$download_id = pll_get_post( $download_id, pll_default_language() );
```
Is it possible to add a filter in this place?
I hope I am making sense. I can make a PR to add these two filters. I just wanted to discuss this before make it.
|
test
|
polylang support i am looking to make work with edd one thing i need to make multilingual is the checkout page i created on in english and one in german so that returns the correct link acording to the language i would need to run edd options though pll get post the function pll get post returns the post id depending on the language of the current page so something like this php edd options edd get settings translate post id for pages in options edd options pll get post edd options is it possible to add a filter which i could hook into two posts downloads are created so to be able to translate the content in two langues at the moment depending on which language the site being viewed in that version of the download would be added to the basket a solution would be to only allow one version of the downloads to be added to the cart i see that you use to add to the cart if i were able to run this line of code then i would be able to achieve this php download id pll get post download id pll default language is it possible to add a filter in this place i hope i am making sense i can make a pr to add these two filters i just wanted to discuss this before make it
| 1
|
38,551
| 5,192,035,449
|
IssuesEvent
|
2017-01-22 03:15:38
|
mil-oss/fgsms
|
https://api.github.com/repos/mil-oss/fgsms
|
closed
|
clean up uddi smoke test
|
bug tests
|
currently still uses the old old config file setup. a singular config file is the preferred solution. also need to stand up the server, test publication against it
|
1.0
|
clean up uddi smoke test - currently still uses the old old config file setup. a singular config file is the preferred solution. also need to stand up the server, test publication against it
|
test
|
clean up uddi smoke test currently still uses the old old config file setup a singular config file is the preferred solution also need to stand up the server test publication against it
| 1
|
29,462
| 4,501,258,373
|
IssuesEvent
|
2016-09-01 08:47:03
|
mattbearman/lime
|
https://api.github.com/repos/mattbearman/lime
|
closed
|
BugMuncher Feedback Report
|
bug BugMuncher other test
|
## Details ##
**Submitted:** January 05, 2016 15:38
**Category:** other
**Sender Email:** info@bugmuncher.com
**Website:** BugMuncher App
**URL:** https://app.bugmuncher.com/profiles/1/reports/112905/edit
**Operating System:** Mac OS X Yosemite
**Browser:** Chrome 47.0.2526.106
**Browser Size:** 1278 x 1341
**User Agent:**
Mozilla/5.0 (Macintosh; Intel Mac OS X 10_10_5) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/47.0.2526.106 Safari/537.36
**Description:**
General
## Custom Data ##
**user_id:** 1
**account_id:** 1
**plan:** BugMuncher Corporate
## Browser Plugins ##
Shockwave Flash
Widevine Content Decryption Module
Chrome PDF Viewer
Native Client
## Events ##
**method:** GET
**url:** https://app.bugmuncher.com/user/dashboard
**timestamp:** Tue Jan 05 2016 15:32:17 GMT+0000 (GMT)
**type:** page_load
---
**method:** GET
**url:** https://app.bugmuncher.com/profiles/1/reports/112905/edit
**timestamp:** Tue Jan 05 2016 15:36:01 GMT+0000 (GMT)
**type:** page_load
---
**content:** Feedback Button Clicked
**timestamp:** Tue Jan 05 2016 15:38:42 GMT+0000 (GMT)
**type:** bugmuncher
---
**type:** bugmuncher
**content:** Feedback Report Submitted
**timestamp:** Tue Jan 05 2016 15:38:49 GMT+0000 (GMT)
---
|
1.0
|
BugMuncher Feedback Report - ## Details ##
**Submitted:** January 05, 2016 15:38
**Category:** other
**Sender Email:** info@bugmuncher.com
**Website:** BugMuncher App
**URL:** https://app.bugmuncher.com/profiles/1/reports/112905/edit
**Operating System:** Mac OS X Yosemite
**Browser:** Chrome 47.0.2526.106
**Browser Size:** 1278 x 1341
**User Agent:**
Mozilla/5.0 (Macintosh; Intel Mac OS X 10_10_5) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/47.0.2526.106 Safari/537.36
**Description:**
General
## Custom Data ##
**user_id:** 1
**account_id:** 1
**plan:** BugMuncher Corporate
## Browser Plugins ##
Shockwave Flash
Widevine Content Decryption Module
Chrome PDF Viewer
Native Client
## Events ##
**method:** GET
**url:** https://app.bugmuncher.com/user/dashboard
**timestamp:** Tue Jan 05 2016 15:32:17 GMT+0000 (GMT)
**type:** page_load
---
**method:** GET
**url:** https://app.bugmuncher.com/profiles/1/reports/112905/edit
**timestamp:** Tue Jan 05 2016 15:36:01 GMT+0000 (GMT)
**type:** page_load
---
**content:** Feedback Button Clicked
**timestamp:** Tue Jan 05 2016 15:38:42 GMT+0000 (GMT)
**type:** bugmuncher
---
**type:** bugmuncher
**content:** Feedback Report Submitted
**timestamp:** Tue Jan 05 2016 15:38:49 GMT+0000 (GMT)
---
|
test
|
bugmuncher feedback report details submitted january category other sender email info bugmuncher com website bugmuncher app url operating system mac os x yosemite browser chrome browser size x user agent mozilla macintosh intel mac os x applewebkit khtml like gecko chrome safari description general custom data user id account id plan bugmuncher corporate browser plugins shockwave flash widevine content decryption module chrome pdf viewer native client events method get url timestamp tue jan gmt gmt type page load method get url timestamp tue jan gmt gmt type page load content feedback button clicked timestamp tue jan gmt gmt type bugmuncher type bugmuncher content feedback report submitted timestamp tue jan gmt gmt
| 1
|
439,966
| 30,724,101,161
|
IssuesEvent
|
2023-07-27 18:11:01
|
typescript-eslint/typescript-eslint
|
https://api.github.com/repos/typescript-eslint/typescript-eslint
|
closed
|
Docs: Mention preference for small, self-contained rule test cases
|
documentation accepting prs
|
### Before You File a Documentation Request Please Confirm You Have Done The Following...
- [X] I have looked for existing [open or closed documentation requests](https://github.com/typescript-eslint/typescript-eslint/issues?q=is%3Aissue+label%3Adocumentation) that match my proposal.
- [X] I have [read the FAQ](https://typescript-eslint.io/linting/troubleshooting) and my problem is not listed.
### Suggested Changes
In general, in our `packages/eslint-plugin/tests/rules/**/*.test.ts` tests, I'm under the impression we generally ask for test cases that are single-purpose.
Good:
```ts
{
code: `willCauseComplaint(1);`
errors: [
{ /* ... */ }
]
}
```
Not so good, should be split up:
```ts
{
code: `
willCauseComplaint(1);
willCauseComplaint(2);
`
errors: [
{ /* ... */ },
{ /* ... */ }
]
}
```
I see the latter form come up sometimes in PRs. Let's document our preference for the former?
### Affected URL(s)
https://typescript-eslint.io/maintenance/pull-requests
Or perhaps we'd want to split out a page dedicated to good practices for ESLint rule PRs? Since they're such a large percentage of our PRs & have specific practices associated.
Aside: if this is accepted by other maintainers we should file an issue to fix up old test files that violate this preference. E.g. `no-floating-promises.test.ts`.
|
1.0
|
Docs: Mention preference for small, self-contained rule test cases - ### Before You File a Documentation Request Please Confirm You Have Done The Following...
- [X] I have looked for existing [open or closed documentation requests](https://github.com/typescript-eslint/typescript-eslint/issues?q=is%3Aissue+label%3Adocumentation) that match my proposal.
- [X] I have [read the FAQ](https://typescript-eslint.io/linting/troubleshooting) and my problem is not listed.
### Suggested Changes
In general, in our `packages/eslint-plugin/tests/rules/**/*.test.ts` tests, I'm under the impression we generally ask for test cases that are single-purpose.
Good:
```ts
{
code: `willCauseComplaint(1);`
errors: [
{ /* ... */ }
]
}
```
Not so good, should be split up:
```ts
{
code: `
willCauseComplaint(1);
willCauseComplaint(2);
`
errors: [
{ /* ... */ },
{ /* ... */ }
]
}
```
I see the latter form come up sometimes in PRs. Let's document our preference for the former?
### Affected URL(s)
https://typescript-eslint.io/maintenance/pull-requests
Or perhaps we'd want to split out a page dedicated to good practices for ESLint rule PRs? Since they're such a large percentage of our PRs & have specific practices associated.
Aside: if this is accepted by other maintainers we should file an issue to fix up old test files that violate this preference. E.g. `no-floating-promises.test.ts`.
|
non_test
|
docs mention preference for small self contained rule test cases before you file a documentation request please confirm you have done the following i have looked for existing that match my proposal i have and my problem is not listed suggested changes in general in our packages eslint plugin tests rules test ts tests i m under the impression we generally ask for test cases that are single purpose good ts code willcausecomplaint errors not so good should be split up ts code willcausecomplaint willcausecomplaint errors i see the latter form come up sometimes in prs let s document our preference for the former affected url s or perhaps we d want to split out a page dedicated to good practices for eslint rule prs since they re such a large percentage of our prs have specific practices associated aside if this is accepted by other maintainers we should file an issue to fix up old test files that violate this preference e g no floating promises test ts
| 0
|
10,159
| 26,419,783,803
|
IssuesEvent
|
2023-01-13 19:13:14
|
MicrosoftDocs/architecture-center
|
https://api.github.com/repos/MicrosoftDocs/architecture-center
|
closed
|
Not All Page Appears When Export To PDF
|
assigned-to-author doc-idea triaged architecture-center/svc Pri1 design-pattern/subsvc
|
When I export to PDF some page doesn't appears for example:
- Circuit Breaker Pattern
- Claim Check
- Compensating Transaction
- Competing Consumers
- Compute Resource Consolidation
- CQRS
- Deployment Stamps
- Event Sourcing
- External Configuration Store
- Federated Identity
Best regards
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 0a4ecd26-cb7b-4cb2-ae3a-3748ada92771
* Version Independent ID: 97f96179-a22b-8401-a2f4-264a86b5d2f5
* Content: [Cloud design patterns - Azure Architecture Center](https://docs.microsoft.com/en-us/azure/architecture/patterns/)
* Content Source: [docs/patterns/index.md](https://github.com/microsoftdocs/architecture-center/blob/main/docs/patterns/index.md)
* Service: **architecture-center**
* Sub-service: **design-pattern**
* GitHub Login: @EdPrice-MSFT
* Microsoft Alias: **masashin**
|
1.0
|
Not All Page Appears When Export To PDF - When I export to PDF some page doesn't appears for example:
- Circuit Breaker Pattern
- Claim Check
- Compensating Transaction
- Competing Consumers
- Compute Resource Consolidation
- CQRS
- Deployment Stamps
- Event Sourcing
- External Configuration Store
- Federated Identity
Best regards
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 0a4ecd26-cb7b-4cb2-ae3a-3748ada92771
* Version Independent ID: 97f96179-a22b-8401-a2f4-264a86b5d2f5
* Content: [Cloud design patterns - Azure Architecture Center](https://docs.microsoft.com/en-us/azure/architecture/patterns/)
* Content Source: [docs/patterns/index.md](https://github.com/microsoftdocs/architecture-center/blob/main/docs/patterns/index.md)
* Service: **architecture-center**
* Sub-service: **design-pattern**
* GitHub Login: @EdPrice-MSFT
* Microsoft Alias: **masashin**
|
non_test
|
not all page appears when export to pdf when i export to pdf some page doesn t appears for example circuit breaker pattern claim check compensating transaction competing consumers compute resource consolidation cqrs deployment stamps event sourcing external configuration store federated identity best regards document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service architecture center sub service design pattern github login edprice msft microsoft alias masashin
| 0
|
302,651
| 26,158,716,148
|
IssuesEvent
|
2022-12-31 06:24:36
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
roachtest: cdc/bank failed
|
C-test-failure O-robot O-roachtest release-blocker branch-release-22.2
|
roachtest.cdc/bank [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=artifacts#/cdc/bank) on release-22.2 @ [07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1](https://github.com/cockroachdb/cockroach/commits/07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1):
```
test artifacts and logs in: /artifacts/cdc/bank/run_1
(test_impl.go:286).Fatal: pq: Use of CHANGEFEED requires an enterprise license. Your evaluation license expired on December 30, 2022. If you're interested in getting a new license, please contact subscriptions@cockroachlabs.com and we can help you out.
(test_impl.go:286).Fatal: cluster.RunE: context canceled
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #94463 roachtest: cdc/bank failed [C-test-failure O-roachtest O-robot T-cdc branch-master release-blocker]
</p>
</details>
/cc @cockroachdb/cdc
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*cdc/bank.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
2.0
|
roachtest: cdc/bank failed - roachtest.cdc/bank [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=artifacts#/cdc/bank) on release-22.2 @ [07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1](https://github.com/cockroachdb/cockroach/commits/07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1):
```
test artifacts and logs in: /artifacts/cdc/bank/run_1
(test_impl.go:286).Fatal: pq: Use of CHANGEFEED requires an enterprise license. Your evaluation license expired on December 30, 2022. If you're interested in getting a new license, please contact subscriptions@cockroachlabs.com and we can help you out.
(test_impl.go:286).Fatal: cluster.RunE: context canceled
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #94463 roachtest: cdc/bank failed [C-test-failure O-roachtest O-robot T-cdc branch-master release-blocker]
</p>
</details>
/cc @cockroachdb/cdc
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*cdc/bank.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
test
|
roachtest cdc bank failed roachtest cdc bank with on release test artifacts and logs in artifacts cdc bank run test impl go fatal pq use of changefeed requires an enterprise license your evaluation license expired on december if you re interested in getting a new license please contact subscriptions cockroachlabs com and we can help you out test impl go fatal cluster rune context canceled parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see same failure on other branches roachtest cdc bank failed cc cockroachdb cdc
| 1
|
156,307
| 12,304,814,925
|
IssuesEvent
|
2020-05-11 21:10:49
|
futurewei-cloud/arktos
|
https://api.github.com/repos/futurewei-cloud/arktos
|
closed
|
Endpoints failed to get IP and port with error FailedToUpdateEndpoint
|
area/controller-manager area/performance-test
|
<!-- Please use this template while reporting a bug and provide as much info as possible. Not doing so may result in your bug not being addressed in a timely manner. Thanks!
If the matter is security related, please disclose it privately via https://kubernetes.io/security/
-->
**What happened**:
1. run kube-up.sh to start arktos
```
export MASTER_DISK_SIZE=300GB KUBE_GCE_ZONE=us-west2-b MASTER_SIZE=n1-highmem-32 NODE_SIZE=n1-highmem-8 NUM_NODES=3 NODE_DISK_SIZE=200GB KUBE_GCE_NETWORK=default
./cluster/kube-up.sh
```
2. after arktos start successfully, check all endpoints status, get the info:
```
NAMESPACE NAME ENDPOINTS AGE SERVICEGROUPID
default kubernetes 35.236.127.13:443 64s 0
kube-system default-http-backend <none> 29s <none>
kube-system heapster <none> 29s <none>
kube-system kube-controller-manager <none> 64s <none>
kube-system kube-scheduler <none> 54s <none>
kube-system kubernetes-dashboard <none> 29s <none>
kube-system metrics-server <none> 29s <none>
```
```
Name: metrics-server
Namespace: kube-system
Labels: addonmanager.kubernetes.io/mode=Reconcile
kubernetes.io/cluster-service=true
kubernetes.io/name=Metrics-server
Annotations: endpoints.kubernetes.io/last-change-trigger-time: 2020-04-30T02:32:24Z
Subsets:
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Warning FailedToUpdateEndpoint 3m1s (x25 over 3m55s) endpoint-controller Failed to update endpoint default/kube-system/metrics-server: Endpoints "metrics-server" is invalid: subsets[0].addresses[0].ip: Invalid value: "169.254.123.2": may not be in the link-local range (169.254.0.0/16)
```
**What you expected to happen**:
Endpoints should get ipaddress and ports as below:
```
NAMESPACE NAME ENDPOINTS AGE
default kubernetes 10.40.0.2:443 2m26s
kube-system default-http-backend 10.64.7.4:8080 2m11s
kube-system kube-controller-manager <none> 2m25s
kube-system kube-dns 10.64.1.2:53,10.64.7.3:53,10.64.1.2:9153 + 3 more... 2m11s
kube-system kube-scheduler <none> 2m24s
kube-system kubernetes-dashboard 10.64.4.4:8443 2m11s
kube-system metrics-server 10.64.6.2:443 2m11s
```
**How to reproduce it (as minimally and precisely as possible)**:
**Anything else we need to know?**:
**Environment**:
- Arktos version (use `kubectl version`):
- Cloud provider or hardware configuration:
- OS (e.g: `cat /etc/os-release`):
- Kernel (e.g. `uname -a`):
- Install tools:
- Network plugin and version (if this is a network-related bug):
- Others:
|
1.0
|
Endpoints failed to get IP and port with error FailedToUpdateEndpoint - <!-- Please use this template while reporting a bug and provide as much info as possible. Not doing so may result in your bug not being addressed in a timely manner. Thanks!
If the matter is security related, please disclose it privately via https://kubernetes.io/security/
-->
**What happened**:
1. run kube-up.sh to start arktos
```
export MASTER_DISK_SIZE=300GB KUBE_GCE_ZONE=us-west2-b MASTER_SIZE=n1-highmem-32 NODE_SIZE=n1-highmem-8 NUM_NODES=3 NODE_DISK_SIZE=200GB KUBE_GCE_NETWORK=default
./cluster/kube-up.sh
```
2. after arktos start successfully, check all endpoints status, get the info:
```
NAMESPACE NAME ENDPOINTS AGE SERVICEGROUPID
default kubernetes 35.236.127.13:443 64s 0
kube-system default-http-backend <none> 29s <none>
kube-system heapster <none> 29s <none>
kube-system kube-controller-manager <none> 64s <none>
kube-system kube-scheduler <none> 54s <none>
kube-system kubernetes-dashboard <none> 29s <none>
kube-system metrics-server <none> 29s <none>
```
```
Name: metrics-server
Namespace: kube-system
Labels: addonmanager.kubernetes.io/mode=Reconcile
kubernetes.io/cluster-service=true
kubernetes.io/name=Metrics-server
Annotations: endpoints.kubernetes.io/last-change-trigger-time: 2020-04-30T02:32:24Z
Subsets:
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Warning FailedToUpdateEndpoint 3m1s (x25 over 3m55s) endpoint-controller Failed to update endpoint default/kube-system/metrics-server: Endpoints "metrics-server" is invalid: subsets[0].addresses[0].ip: Invalid value: "169.254.123.2": may not be in the link-local range (169.254.0.0/16)
```
**What you expected to happen**:
Endpoints should get ipaddress and ports as below:
```
NAMESPACE NAME ENDPOINTS AGE
default kubernetes 10.40.0.2:443 2m26s
kube-system default-http-backend 10.64.7.4:8080 2m11s
kube-system kube-controller-manager <none> 2m25s
kube-system kube-dns 10.64.1.2:53,10.64.7.3:53,10.64.1.2:9153 + 3 more... 2m11s
kube-system kube-scheduler <none> 2m24s
kube-system kubernetes-dashboard 10.64.4.4:8443 2m11s
kube-system metrics-server 10.64.6.2:443 2m11s
```
**How to reproduce it (as minimally and precisely as possible)**:
**Anything else we need to know?**:
**Environment**:
- Arktos version (use `kubectl version`):
- Cloud provider or hardware configuration:
- OS (e.g: `cat /etc/os-release`):
- Kernel (e.g. `uname -a`):
- Install tools:
- Network plugin and version (if this is a network-related bug):
- Others:
|
test
|
endpoints failed to get ip and port with error failedtoupdateendpoint please use this template while reporting a bug and provide as much info as possible not doing so may result in your bug not being addressed in a timely manner thanks if the matter is security related please disclose it privately via what happened run kube up sh to start arktos export master disk size kube gce zone us b master size highmem node size highmem num nodes node disk size kube gce network default cluster kube up sh after arktos start successfully check all endpoints status get the info namespace name endpoints age servicegroupid default kubernetes kube system default http backend kube system heapster kube system kube controller manager kube system kube scheduler kube system kubernetes dashboard kube system metrics server name metrics server namespace kube system labels addonmanager kubernetes io mode reconcile kubernetes io cluster service true kubernetes io name metrics server annotations endpoints kubernetes io last change trigger time subsets events type reason age from message warning failedtoupdateendpoint over endpoint controller failed to update endpoint default kube system metrics server endpoints metrics server is invalid subsets addresses ip invalid value may not be in the link local range what you expected to happen endpoints should get ipaddress and ports as below namespace name endpoints age default kubernetes kube system default http backend kube system kube controller manager kube system kube dns more kube system kube scheduler kube system kubernetes dashboard kube system metrics server how to reproduce it as minimally and precisely as possible anything else we need to know environment arktos version use kubectl version cloud provider or hardware configuration os e g cat etc os release kernel e g uname a install tools network plugin and version if this is a network related bug others
| 1
|
228,067
| 18,154,148,894
|
IssuesEvent
|
2021-09-26 19:37:28
|
pandas-dev/pandas
|
https://api.github.com/repos/pandas-dev/pandas
|
opened
|
TST: Add doctests on cython files
|
Testing Docs CI
|
Adding `--doctest-cython` to `ci/run_tests.sh` to cover doctests in cython files.
Lots of them are failing at the moment. But I think few of them (like now, today), can be skipped using `# doctest: +SKIP`.
Below is the current list of failures atm.
```python
===================================================================================== short test summary info =====================================================================================
FAILED pandas/_libs/hashtable.cpython-38-darwin.so::pandas._libs.hashtable.Int64Factorizer.factorize
FAILED pandas/_libs/hashtable.cpython-38-darwin.so::pandas._libs.hashtable.ObjectFactorizer.factorize
FAILED pandas/_libs/hashtable.cpython-38-darwin.so::pandas._libs.hashtable.__test__.Int64Factorizer.factorize (line 134)
FAILED pandas/_libs/hashtable.cpython-38-darwin.so::pandas._libs.hashtable.__test__.ObjectFactorizer.factorize (line 88)
FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.__test__.infer_dtype (line 1299)
FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.__test__.is_iterator (line 242)
FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.__test__.is_list_like (line 1051)
FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.__test__.is_scalar (line 149)
FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.infer_dtype
FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.is_iterator
FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.is_list_like
FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.is_scalar
FAILED pandas/_libs/tslibs/dtypes.cpython-38-darwin.so::pandas._libs.tslibs.dtypes.Resolution.from_attrname
FAILED pandas/_libs/tslibs/dtypes.cpython-38-darwin.so::pandas._libs.tslibs.dtypes.Resolution.get_reso_from_freq
FAILED pandas/_libs/tslibs/dtypes.cpython-38-darwin.so::pandas._libs.tslibs.dtypes.__test__.Resolution.from_attrname (line 223)
FAILED pandas/_libs/tslibs/dtypes.cpython-38-darwin.so::pandas._libs.tslibs.dtypes.__test__.Resolution.get_reso_from_freq (line 238)
FAILED pandas/_libs/tslibs/nattype.cpython-38-darwin.so::pandas._libs.tslibs.nattype.NaTType.now
FAILED pandas/_libs/tslibs/nattype.cpython-38-darwin.so::pandas._libs.tslibs.nattype.NaTType.today
FAILED pandas/_libs/tslibs/nattype.cpython-38-darwin.so::pandas._libs.tslibs.nattype.NaTType.utcfromtimestamp
FAILED pandas/_libs/tslibs/nattype.cpython-38-darwin.so::pandas._libs.tslibs.nattype.NaTType.utcnow
FAILED pandas/_libs/tslibs/period.cpython-38-darwin.so::pandas._libs.tslibs.period._Period.strftime
FAILED pandas/_libs/tslibs/period.cpython-38-darwin.so::pandas._libs.tslibs.period.__test__._Period.strftime (line 2243)
FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.Timestamp.now
FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.Timestamp.today
FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.Timestamp.utcfromtimestamp
FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.Timestamp.utcnow
FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps._Timestamp.to_period
FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.__test__.Timestamp.now (line 1048)
FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.__test__.Timestamp.today (line 1075)
FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.__test__.Timestamp.utcfromtimestamp (line 1115)
FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.__test__.Timestamp.utcnow (line 1101)
FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.__test__._Timestamp.to_period (line 905)
FAILED pandas/_libs/tslibs/timezones.cpython-38-darwin.so::pandas._libs.tslibs.timezones.__test__.tz_standardize (line 375)
FAILED pandas/_libs/tslibs/timezones.cpython-38-darwin.so::pandas._libs.tslibs.timezones.tz_standardize
================================================================================= 34 failed, 106 passed in 1.10s ==================================================================================
|
1.0
|
TST: Add doctests on cython files - Adding `--doctest-cython` to `ci/run_tests.sh` to cover doctests in cython files.
Lots of them are failing at the moment. But I think few of them (like now, today), can be skipped using `# doctest: +SKIP`.
Below is the current list of failures atm.
```python
===================================================================================== short test summary info =====================================================================================
FAILED pandas/_libs/hashtable.cpython-38-darwin.so::pandas._libs.hashtable.Int64Factorizer.factorize
FAILED pandas/_libs/hashtable.cpython-38-darwin.so::pandas._libs.hashtable.ObjectFactorizer.factorize
FAILED pandas/_libs/hashtable.cpython-38-darwin.so::pandas._libs.hashtable.__test__.Int64Factorizer.factorize (line 134)
FAILED pandas/_libs/hashtable.cpython-38-darwin.so::pandas._libs.hashtable.__test__.ObjectFactorizer.factorize (line 88)
FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.__test__.infer_dtype (line 1299)
FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.__test__.is_iterator (line 242)
FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.__test__.is_list_like (line 1051)
FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.__test__.is_scalar (line 149)
FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.infer_dtype
FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.is_iterator
FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.is_list_like
FAILED pandas/_libs/lib.cpython-38-darwin.so::pandas._libs.lib.is_scalar
FAILED pandas/_libs/tslibs/dtypes.cpython-38-darwin.so::pandas._libs.tslibs.dtypes.Resolution.from_attrname
FAILED pandas/_libs/tslibs/dtypes.cpython-38-darwin.so::pandas._libs.tslibs.dtypes.Resolution.get_reso_from_freq
FAILED pandas/_libs/tslibs/dtypes.cpython-38-darwin.so::pandas._libs.tslibs.dtypes.__test__.Resolution.from_attrname (line 223)
FAILED pandas/_libs/tslibs/dtypes.cpython-38-darwin.so::pandas._libs.tslibs.dtypes.__test__.Resolution.get_reso_from_freq (line 238)
FAILED pandas/_libs/tslibs/nattype.cpython-38-darwin.so::pandas._libs.tslibs.nattype.NaTType.now
FAILED pandas/_libs/tslibs/nattype.cpython-38-darwin.so::pandas._libs.tslibs.nattype.NaTType.today
FAILED pandas/_libs/tslibs/nattype.cpython-38-darwin.so::pandas._libs.tslibs.nattype.NaTType.utcfromtimestamp
FAILED pandas/_libs/tslibs/nattype.cpython-38-darwin.so::pandas._libs.tslibs.nattype.NaTType.utcnow
FAILED pandas/_libs/tslibs/period.cpython-38-darwin.so::pandas._libs.tslibs.period._Period.strftime
FAILED pandas/_libs/tslibs/period.cpython-38-darwin.so::pandas._libs.tslibs.period.__test__._Period.strftime (line 2243)
FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.Timestamp.now
FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.Timestamp.today
FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.Timestamp.utcfromtimestamp
FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.Timestamp.utcnow
FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps._Timestamp.to_period
FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.__test__.Timestamp.now (line 1048)
FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.__test__.Timestamp.today (line 1075)
FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.__test__.Timestamp.utcfromtimestamp (line 1115)
FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.__test__.Timestamp.utcnow (line 1101)
FAILED pandas/_libs/tslibs/timestamps.cpython-38-darwin.so::pandas._libs.tslibs.timestamps.__test__._Timestamp.to_period (line 905)
FAILED pandas/_libs/tslibs/timezones.cpython-38-darwin.so::pandas._libs.tslibs.timezones.__test__.tz_standardize (line 375)
FAILED pandas/_libs/tslibs/timezones.cpython-38-darwin.so::pandas._libs.tslibs.timezones.tz_standardize
================================================================================= 34 failed, 106 passed in 1.10s ==================================================================================
|
test
|
tst add doctests on cython files adding doctest cython to ci run tests sh to cover doctests in cython files lots of them are failing at the moment but i think few of them like now today can be skipped using doctest skip below is the current list of failures atm python short test summary info failed pandas libs hashtable cpython darwin so pandas libs hashtable factorize failed pandas libs hashtable cpython darwin so pandas libs hashtable objectfactorizer factorize failed pandas libs hashtable cpython darwin so pandas libs hashtable test factorize line failed pandas libs hashtable cpython darwin so pandas libs hashtable test objectfactorizer factorize line failed pandas libs lib cpython darwin so pandas libs lib test infer dtype line failed pandas libs lib cpython darwin so pandas libs lib test is iterator line failed pandas libs lib cpython darwin so pandas libs lib test is list like line failed pandas libs lib cpython darwin so pandas libs lib test is scalar line failed pandas libs lib cpython darwin so pandas libs lib infer dtype failed pandas libs lib cpython darwin so pandas libs lib is iterator failed pandas libs lib cpython darwin so pandas libs lib is list like failed pandas libs lib cpython darwin so pandas libs lib is scalar failed pandas libs tslibs dtypes cpython darwin so pandas libs tslibs dtypes resolution from attrname failed pandas libs tslibs dtypes cpython darwin so pandas libs tslibs dtypes resolution get reso from freq failed pandas libs tslibs dtypes cpython darwin so pandas libs tslibs dtypes test resolution from attrname line failed pandas libs tslibs dtypes cpython darwin so pandas libs tslibs dtypes test resolution get reso from freq line failed pandas libs tslibs nattype cpython darwin so pandas libs tslibs nattype nattype now failed pandas libs tslibs nattype cpython darwin so pandas libs tslibs nattype nattype today failed pandas libs tslibs nattype cpython darwin so pandas libs tslibs nattype nattype utcfromtimestamp failed pandas libs tslibs nattype cpython darwin so pandas libs tslibs nattype nattype utcnow failed pandas libs tslibs period cpython darwin so pandas libs tslibs period period strftime failed pandas libs tslibs period cpython darwin so pandas libs tslibs period test period strftime line failed pandas libs tslibs timestamps cpython darwin so pandas libs tslibs timestamps timestamp now failed pandas libs tslibs timestamps cpython darwin so pandas libs tslibs timestamps timestamp today failed pandas libs tslibs timestamps cpython darwin so pandas libs tslibs timestamps timestamp utcfromtimestamp failed pandas libs tslibs timestamps cpython darwin so pandas libs tslibs timestamps timestamp utcnow failed pandas libs tslibs timestamps cpython darwin so pandas libs tslibs timestamps timestamp to period failed pandas libs tslibs timestamps cpython darwin so pandas libs tslibs timestamps test timestamp now line failed pandas libs tslibs timestamps cpython darwin so pandas libs tslibs timestamps test timestamp today line failed pandas libs tslibs timestamps cpython darwin so pandas libs tslibs timestamps test timestamp utcfromtimestamp line failed pandas libs tslibs timestamps cpython darwin so pandas libs tslibs timestamps test timestamp utcnow line failed pandas libs tslibs timestamps cpython darwin so pandas libs tslibs timestamps test timestamp to period line failed pandas libs tslibs timezones cpython darwin so pandas libs tslibs timezones test tz standardize line failed pandas libs tslibs timezones cpython darwin so pandas libs tslibs timezones tz standardize failed passed in
| 1
|
114,732
| 9,753,401,510
|
IssuesEvent
|
2019-06-04 09:03:24
|
spacemeshos/go-spacemesh
|
https://api.github.com/repos/spacemeshos/go-spacemesh
|
closed
|
Create query for logs per node after test ends
|
automation tests
|
# Overview / Motivation
we want developers to easily debug logs from a test run, usually, this involves watching logs from different parts of the test isolated but simultaneously.
# The Task
create a query that gets all the relevant logs for that test grouped by the client, node or pod that wrote them.
# Implementation Notes
TODO: Add links to relevant resources, specs, related issues, etc...
# Contribution Guidelines
Important: Issue assignment to developers will be by the order of their application and proficiency level according to the tasks complexity. We will not assign tasks to developers who have'nt introduced themselves on our Gitter [dev channel](https://gitter.im/spacemesh-os/Lobby)
1. Introduce yourself on go-spacemesh [dev chat channel](https://gitter.im/spacemesh-os/Lobby) - ask our team any question you may have about this task
2. Fork branch `develop` to your own repo and work in your repo
3. You must document all methods, enums and types with [godoc comments](https://blog.golang.org/godoc-documenting-go-code)
4. You must write go unit tests for all types and methods when submitting a component, and integration tests if you submit a feature
5. When ready for code review, submit a PR from your repo back to branch `develop`
6. Attach relevant issue to PR
|
1.0
|
Create query for logs per node after test ends - # Overview / Motivation
we want developers to easily debug logs from a test run, usually, this involves watching logs from different parts of the test isolated but simultaneously.
# The Task
create a query that gets all the relevant logs for that test grouped by the client, node or pod that wrote them.
# Implementation Notes
TODO: Add links to relevant resources, specs, related issues, etc...
# Contribution Guidelines
Important: Issue assignment to developers will be by the order of their application and proficiency level according to the tasks complexity. We will not assign tasks to developers who have'nt introduced themselves on our Gitter [dev channel](https://gitter.im/spacemesh-os/Lobby)
1. Introduce yourself on go-spacemesh [dev chat channel](https://gitter.im/spacemesh-os/Lobby) - ask our team any question you may have about this task
2. Fork branch `develop` to your own repo and work in your repo
3. You must document all methods, enums and types with [godoc comments](https://blog.golang.org/godoc-documenting-go-code)
4. You must write go unit tests for all types and methods when submitting a component, and integration tests if you submit a feature
5. When ready for code review, submit a PR from your repo back to branch `develop`
6. Attach relevant issue to PR
|
test
|
create query for logs per node after test ends overview motivation we want developers to easily debug logs from a test run usually this involves watching logs from different parts of the test isolated but simultaneously the task create a query that gets all the relevant logs for that test grouped by the client node or pod that wrote them implementation notes todo add links to relevant resources specs related issues etc contribution guidelines important issue assignment to developers will be by the order of their application and proficiency level according to the tasks complexity we will not assign tasks to developers who have nt introduced themselves on our gitter introduce yourself on go spacemesh ask our team any question you may have about this task fork branch develop to your own repo and work in your repo you must document all methods enums and types with you must write go unit tests for all types and methods when submitting a component and integration tests if you submit a feature when ready for code review submit a pr from your repo back to branch develop attach relevant issue to pr
| 1
|
76,562
| 7,539,839,902
|
IssuesEvent
|
2018-04-17 02:51:22
|
sudar/bulk-delete
|
https://api.github.com/repos/sudar/bulk-delete
|
opened
|
User Post Meta - Migrate to new architecture
|
1 - Ready task tests
|
- [ ] Migrate User Post Meta
- [ ] Manually Test the changes
- [ ] Manually Test that the User User Meta Pro add-on works
## Tests
- [ ] Add tests for deleting single user meta fields from 'admin' user role
- [ ] Add tests for deleting single user meta fields from 'subscriber' user role
- [ ] Add tests for deleting multiple user meta fields from 'admin' user role
- [ ] Add tests for deleting multiple user meta fields from 'subscriber' user role
- [ ] Add tests for deleting user meta from users in batches
|
1.0
|
User Post Meta - Migrate to new architecture - - [ ] Migrate User Post Meta
- [ ] Manually Test the changes
- [ ] Manually Test that the User User Meta Pro add-on works
## Tests
- [ ] Add tests for deleting single user meta fields from 'admin' user role
- [ ] Add tests for deleting single user meta fields from 'subscriber' user role
- [ ] Add tests for deleting multiple user meta fields from 'admin' user role
- [ ] Add tests for deleting multiple user meta fields from 'subscriber' user role
- [ ] Add tests for deleting user meta from users in batches
|
test
|
user post meta migrate to new architecture migrate user post meta manually test the changes manually test that the user user meta pro add on works tests add tests for deleting single user meta fields from admin user role add tests for deleting single user meta fields from subscriber user role add tests for deleting multiple user meta fields from admin user role add tests for deleting multiple user meta fields from subscriber user role add tests for deleting user meta from users in batches
| 1
|
299,847
| 25,930,862,464
|
IssuesEvent
|
2022-12-16 09:53:37
|
parca-dev/parca-agent
|
https://api.github.com/repos/parca-dev/parca-agent
|
opened
|
Testing the Agent
|
stack-quality testing
|
## The state of testing affairs
At the moment we have:
- **unit-tests**: to ensure the correctness of small well-contained pieces of code, such as our [Kernel symbolizer](https://github.com/parca-dev/parca-agent/blob/main/pkg/ksym/ksym_test.go). This helps us ensure that the behaviour is correct, but not less importantly has everything in place for us to add regression tests if we encounter bugs that we then fix;
- **e2e tests**: running on Minikube that have the potential of testing our Kubernetes / systems metadata collection;
These two are great and are very necessary. However, we have some blindspots, that are only going to get worse once we start shipping more advanced features, such as the [DWARF-based unwinder](https://github.com/parca-dev/parca-agent/issues/768). In my opinion, we are missing three main things:
1. A way to ensure that the **kernels we want to support work** (it's easy to introduce regressions, such as a feature that's not supported in a kernel we care about, or some code pattern that the verifier in some kernels rejects. These two are things I've spotted before, not fabricated examples)
1. Ensuring that resource usage, e.g. **memory locking etc is within what we expect** (we had to recently [remediate this](https://github.com/parca-dev/parca-agent/pull/1074))
1. **Correctness testing**: on any kernel we support, running test programs and our unwinder and verifying that we get correct stack traces
## Propose additions
### VM testing
For both 1) and - partially - 2) we need to add some testing in VMs with different kernels. Instead of running all of our integration tests (which may be flaky) or unit tests (which don't add much value in different kernels), I would like to propose the following for the kernels we support:
- Can create the maps, so we ensure that our logic to create a map of size set at runtime is correct (note: we'll make this logic smarter to load as much as we can for a given host with a ceiling of 400-500MB);
- Load our BPF program, to ensure that the verifier is happy with our code;
This will also test that the BTF-pulling system for hosts that don't have it locally works, too (https://github.com/parca-dev/parca-agent/pull/1131).
Ideally, we'll publish a prototype of this by the end of the day with the following goals:
- quick iteration times: each VM should take less than 1min from spinning up to reporting results
- simplicity: simple to understand, and hopefully simple(r) to debug
- can be run locally, but also in CI. We would need to decide if we always want to run it in CI or only in certain occasions as it might slow things down (e.g. once a day, or just if bpf/ gets modified)
### Stack correctness tests
(This will be something to tackle early next year)
Running on the current machine, asserting that stacks are what we expect. Something similar to [this](https://github.com/javierhonduco/rbperf/blob/main/src/rbperf.rs#L597-L624). They can run on the host's kernel as we'll be testing that the programs can still be loaded in all the kernels we support with the VM testing infra
cc/ @v-thakkar @Sylfrena @kakkoyun @brancz
|
1.0
|
Testing the Agent - ## The state of testing affairs
At the moment we have:
- **unit-tests**: to ensure the correctness of small well-contained pieces of code, such as our [Kernel symbolizer](https://github.com/parca-dev/parca-agent/blob/main/pkg/ksym/ksym_test.go). This helps us ensure that the behaviour is correct, but not less importantly has everything in place for us to add regression tests if we encounter bugs that we then fix;
- **e2e tests**: running on Minikube that have the potential of testing our Kubernetes / systems metadata collection;
These two are great and are very necessary. However, we have some blindspots, that are only going to get worse once we start shipping more advanced features, such as the [DWARF-based unwinder](https://github.com/parca-dev/parca-agent/issues/768). In my opinion, we are missing three main things:
1. A way to ensure that the **kernels we want to support work** (it's easy to introduce regressions, such as a feature that's not supported in a kernel we care about, or some code pattern that the verifier in some kernels rejects. These two are things I've spotted before, not fabricated examples)
1. Ensuring that resource usage, e.g. **memory locking etc is within what we expect** (we had to recently [remediate this](https://github.com/parca-dev/parca-agent/pull/1074))
1. **Correctness testing**: on any kernel we support, running test programs and our unwinder and verifying that we get correct stack traces
## Propose additions
### VM testing
For both 1) and - partially - 2) we need to add some testing in VMs with different kernels. Instead of running all of our integration tests (which may be flaky) or unit tests (which don't add much value in different kernels), I would like to propose the following for the kernels we support:
- Can create the maps, so we ensure that our logic to create a map of size set at runtime is correct (note: we'll make this logic smarter to load as much as we can for a given host with a ceiling of 400-500MB);
- Load our BPF program, to ensure that the verifier is happy with our code;
This will also test that the BTF-pulling system for hosts that don't have it locally works, too (https://github.com/parca-dev/parca-agent/pull/1131).
Ideally, we'll publish a prototype of this by the end of the day with the following goals:
- quick iteration times: each VM should take less than 1min from spinning up to reporting results
- simplicity: simple to understand, and hopefully simple(r) to debug
- can be run locally, but also in CI. We would need to decide if we always want to run it in CI or only in certain occasions as it might slow things down (e.g. once a day, or just if bpf/ gets modified)
### Stack correctness tests
(This will be something to tackle early next year)
Running on the current machine, asserting that stacks are what we expect. Something similar to [this](https://github.com/javierhonduco/rbperf/blob/main/src/rbperf.rs#L597-L624). They can run on the host's kernel as we'll be testing that the programs can still be loaded in all the kernels we support with the VM testing infra
cc/ @v-thakkar @Sylfrena @kakkoyun @brancz
|
test
|
testing the agent the state of testing affairs at the moment we have unit tests to ensure the correctness of small well contained pieces of code such as our this helps us ensure that the behaviour is correct but not less importantly has everything in place for us to add regression tests if we encounter bugs that we then fix tests running on minikube that have the potential of testing our kubernetes systems metadata collection these two are great and are very necessary however we have some blindspots that are only going to get worse once we start shipping more advanced features such as the in my opinion we are missing three main things a way to ensure that the kernels we want to support work it s easy to introduce regressions such as a feature that s not supported in a kernel we care about or some code pattern that the verifier in some kernels rejects these two are things i ve spotted before not fabricated examples ensuring that resource usage e g memory locking etc is within what we expect we had to recently correctness testing on any kernel we support running test programs and our unwinder and verifying that we get correct stack traces propose additions vm testing for both and partially we need to add some testing in vms with different kernels instead of running all of our integration tests which may be flaky or unit tests which don t add much value in different kernels i would like to propose the following for the kernels we support can create the maps so we ensure that our logic to create a map of size set at runtime is correct note we ll make this logic smarter to load as much as we can for a given host with a ceiling of load our bpf program to ensure that the verifier is happy with our code this will also test that the btf pulling system for hosts that don t have it locally works too ideally we ll publish a prototype of this by the end of the day with the following goals quick iteration times each vm should take less than from spinning up to reporting results simplicity simple to understand and hopefully simple r to debug can be run locally but also in ci we would need to decide if we always want to run it in ci or only in certain occasions as it might slow things down e g once a day or just if bpf gets modified stack correctness tests this will be something to tackle early next year running on the current machine asserting that stacks are what we expect something similar to they can run on the host s kernel as we ll be testing that the programs can still be loaded in all the kernels we support with the vm testing infra cc v thakkar sylfrena kakkoyun brancz
| 1
|
2,216
| 4,613,932,496
|
IssuesEvent
|
2016-09-25 09:08:09
|
OpenSRP/opensrp-client
|
https://api.github.com/repos/OpenSRP/opensrp-client
|
closed
|
handling cursor close events in register fragments
|
BANGLADESH enhancement FATAL Functional Requirement High Priority
|
currently cursor closing is not handled properly for cursor adaptors and hence this is required when exiting a register in order to save memory
|
1.0
|
handling cursor close events in register fragments - currently cursor closing is not handled properly for cursor adaptors and hence this is required when exiting a register in order to save memory
|
non_test
|
handling cursor close events in register fragments currently cursor closing is not handled properly for cursor adaptors and hence this is required when exiting a register in order to save memory
| 0
|
311,349
| 23,383,260,307
|
IssuesEvent
|
2022-08-11 11:32:59
|
alphagov/govuk-prototype-kit
|
https://api.github.com/repos/alphagov/govuk-prototype-kit
|
opened
|
Document how to show parts of a page or component based on data
|
documentation awaiting triage
|
From support:
Need to know how to show parts of a page or component based on data, for example:
```
{% if data['over-18'] %}
[ Over 18 content here ]
{% else %}
[ Under 18 content here ]
{% endif %}
```
and
```
{{ govukRadios({
idPrefix: "where-do-you-live",
name: "where-do-you-live",
fieldset: {
legend: {
text: "Where do you live?",
isPageHeading: true,
classes: "govuk-fieldset__legend--l"
}
},
items: [
{
value: "england",
text: "England"
},
{
value: "scotland",
text: "Scotland"
},
{
value: "wales",
text: "Wales"
},
{
value: "northern-ireland",
text: "Northern Ireland"
} if (data['example'] == true)
]
}) }}
```
with the key part being `if (data['example'] == true)`
## Related documentation
Not sure we have any
## Suggestion
None yet apart from covering the topics as above
## Evidence (where applicable)
From multiple support requests
|
1.0
|
Document how to show parts of a page or component based on data - From support:
Need to know how to show parts of a page or component based on data, for example:
```
{% if data['over-18'] %}
[ Over 18 content here ]
{% else %}
[ Under 18 content here ]
{% endif %}
```
and
```
{{ govukRadios({
idPrefix: "where-do-you-live",
name: "where-do-you-live",
fieldset: {
legend: {
text: "Where do you live?",
isPageHeading: true,
classes: "govuk-fieldset__legend--l"
}
},
items: [
{
value: "england",
text: "England"
},
{
value: "scotland",
text: "Scotland"
},
{
value: "wales",
text: "Wales"
},
{
value: "northern-ireland",
text: "Northern Ireland"
} if (data['example'] == true)
]
}) }}
```
with the key part being `if (data['example'] == true)`
## Related documentation
Not sure we have any
## Suggestion
None yet apart from covering the topics as above
## Evidence (where applicable)
From multiple support requests
|
non_test
|
document how to show parts of a page or component based on data from support need to know how to show parts of a page or component based on data for example if data else endif and govukradios idprefix where do you live name where do you live fieldset legend text where do you live ispageheading true classes govuk fieldset legend l items value england text england value scotland text scotland value wales text wales value northern ireland text northern ireland if data true with the key part being if data true related documentation not sure we have any suggestion none yet apart from covering the topics as above evidence where applicable from multiple support requests
| 0
|
248,844
| 7,936,846,663
|
IssuesEvent
|
2018-07-09 10:47:38
|
status-im/status-react
|
https://api.github.com/repos/status-im/status-react
|
closed
|
Implement User Profile screen
|
desktop feature high-priority
|
[comment]: # (Please replace ... with your information. Remove < and >)
### User Story
As a user, I want to see nice profile screen with photo, options, etc
### Description
*Type*: Feature
*Summary*: implement user profile screen according to design zpl.io/25qZqB3
#### Expected behavior
User profile looks like here: zpl.io/25qZqB3
#### Actual behavior
![Uploading Prof.png…]()
### Reproduction
[comment]: # (Describe how we can replicate the bug step by step.)
- Open Status
- Create account
- Go to "Profile"
*Summary*:
- [ ] Add profile icon
- [ ] Left panel width should be 340px (fixed size), now is 280px.
- [ ] User name should be centered in left panel (now is aligned to the left)
### Additional Information
* Status version: [desktop build 29.06/2018](https://jenkins.status.im/job/status-react/job/desktop/job/manual/2/)
|
1.0
|
Implement User Profile screen - [comment]: # (Please replace ... with your information. Remove < and >)
### User Story
As a user, I want to see nice profile screen with photo, options, etc
### Description
*Type*: Feature
*Summary*: implement user profile screen according to design zpl.io/25qZqB3
#### Expected behavior
User profile looks like here: zpl.io/25qZqB3
#### Actual behavior
![Uploading Prof.png…]()
### Reproduction
[comment]: # (Describe how we can replicate the bug step by step.)
- Open Status
- Create account
- Go to "Profile"
*Summary*:
- [ ] Add profile icon
- [ ] Left panel width should be 340px (fixed size), now is 280px.
- [ ] User name should be centered in left panel (now is aligned to the left)
### Additional Information
* Status version: [desktop build 29.06/2018](https://jenkins.status.im/job/status-react/job/desktop/job/manual/2/)
|
non_test
|
implement user profile screen please replace with your information remove user story as a user i want to see nice profile screen with photo options etc description type feature summary implement user profile screen according to design zpl io expected behavior user profile looks like here zpl io actual behavior reproduction describe how we can replicate the bug step by step open status create account go to profile summary add profile icon left panel width should be fixed size now is user name should be centered in left panel now is aligned to the left additional information status version
| 0
|
747,098
| 26,073,171,664
|
IssuesEvent
|
2022-12-24 04:35:17
|
tinkerbell/boots
|
https://api.github.com/repos/tinkerbell/boots
|
closed
|
Make use of params command optional for phone-home
|
kind/feature priority/backlog
|
<!--- Provide a general summary of the issue in the Title above -->
Working through getting my physical servers working with tinkerbell to start doing some experimentation.. they fail saying the “params” command is not found.
Leased from datacenter so flashing or updating ipxe might not be possible or ideal.
## Expected Behaviour
<!--- If you're describing a bug, tell us what should happen -->
<!--- If you're suggesting a change/improvement, tell us how it should work -->
Boots
## Current Behaviour
<!--- If describing a bug, tell us what happens instead of the expected behavior -->
<!--- If suggesting a change/improvement, explain the difference from current behavior -->
Errors out at params not found
## Possible Solution
<!--- Not obligatory, but suggest a fix/reason for the bug, -->
<!--- or ideas how to implement the addition or change -->
Maybe hardware option? Or use query string
## Steps to Reproduce (for bugs)
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug. Include code to reproduce, if relevant -->
1. Run ipxe with out params built in
2. Try to boot
## Context
<!--- How has this issue affected you? What are you trying to accomplish? -->
<!--- Providing context helps us come up with a solution that is most useful in the real world -->
My solution was just to remove the params here:
https://github.com/tinkerbell/boots/blob/master/ipxe/script.go#L34
Then just let it phone home even with out params. Seems like using ?body=${body}&type=${body} might could work. I’m also not even sure if needed? Looking at the phone home code seems like it’s mostly ignored?
## Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* Operating System and version (e.g. Linux, Windows, MacOS):
* How are you running Tinkerbell? Using Vagrant & VirtualBox, Vagrant & Libvirt, on Packet using Terraform, or give details: tinkerbell it’s self kvm on the network
* Link to your project or a code example to reproduce issue:
|
1.0
|
Make use of params command optional for phone-home - <!--- Provide a general summary of the issue in the Title above -->
Working through getting my physical servers working with tinkerbell to start doing some experimentation.. they fail saying the “params” command is not found.
Leased from datacenter so flashing or updating ipxe might not be possible or ideal.
## Expected Behaviour
<!--- If you're describing a bug, tell us what should happen -->
<!--- If you're suggesting a change/improvement, tell us how it should work -->
Boots
## Current Behaviour
<!--- If describing a bug, tell us what happens instead of the expected behavior -->
<!--- If suggesting a change/improvement, explain the difference from current behavior -->
Errors out at params not found
## Possible Solution
<!--- Not obligatory, but suggest a fix/reason for the bug, -->
<!--- or ideas how to implement the addition or change -->
Maybe hardware option? Or use query string
## Steps to Reproduce (for bugs)
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug. Include code to reproduce, if relevant -->
1. Run ipxe with out params built in
2. Try to boot
## Context
<!--- How has this issue affected you? What are you trying to accomplish? -->
<!--- Providing context helps us come up with a solution that is most useful in the real world -->
My solution was just to remove the params here:
https://github.com/tinkerbell/boots/blob/master/ipxe/script.go#L34
Then just let it phone home even with out params. Seems like using ?body=${body}&type=${body} might could work. I’m also not even sure if needed? Looking at the phone home code seems like it’s mostly ignored?
## Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* Operating System and version (e.g. Linux, Windows, MacOS):
* How are you running Tinkerbell? Using Vagrant & VirtualBox, Vagrant & Libvirt, on Packet using Terraform, or give details: tinkerbell it’s self kvm on the network
* Link to your project or a code example to reproduce issue:
|
non_test
|
make use of params command optional for phone home working through getting my physical servers working with tinkerbell to start doing some experimentation they fail saying the “params” command is not found leased from datacenter so flashing or updating ipxe might not be possible or ideal expected behaviour boots current behaviour errors out at params not found possible solution maybe hardware option or use query string steps to reproduce for bugs run ipxe with out params built in try to boot context my solution was just to remove the params here then just let it phone home even with out params seems like using body body type body might could work i’m also not even sure if needed looking at the phone home code seems like it’s mostly ignored your environment operating system and version e g linux windows macos how are you running tinkerbell using vagrant virtualbox vagrant libvirt on packet using terraform or give details tinkerbell it’s self kvm on the network link to your project or a code example to reproduce issue
| 0
|
43,372
| 11,209,803,538
|
IssuesEvent
|
2020-01-06 11:27:02
|
go-gitea/gitea
|
https://api.github.com/repos/go-gitea/gitea
|
opened
|
Use nfpm for package release
|
kind/build
|
In https://github.com/go-gitea/gitea/pull/6671, it was discussed to use nfpm via https://github.com/techknowlogick/drone-nfpm
I don't think this has move forward.
I just open this issue to track the progress and keep a track of this choice since the last issue is closed.
I re-think of it because I stumble upon a good tools to track packages in various distrib.
[](https://repology.org/project/gitea/versions)
|
1.0
|
Use nfpm for package release - In https://github.com/go-gitea/gitea/pull/6671, it was discussed to use nfpm via https://github.com/techknowlogick/drone-nfpm
I don't think this has move forward.
I just open this issue to track the progress and keep a track of this choice since the last issue is closed.
I re-think of it because I stumble upon a good tools to track packages in various distrib.
[](https://repology.org/project/gitea/versions)
|
non_test
|
use nfpm for package release in it was discussed to use nfpm via i don t think this has move forward i just open this issue to track the progress and keep a track of this choice since the last issue is closed i re think of it because i stumble upon a good tools to track packages in various distrib
| 0
|
113,554
| 11,808,508,387
|
IssuesEvent
|
2020-03-19 13:30:35
|
nestauk/covres
|
https://api.github.com/repos/nestauk/covres
|
opened
|
Exploratory data analysis
|
documentation eda
|
Exploratory analysis of various CORD-19 datasets. Includes:
* Data structure (sp for full-text)
* Variables
* Missing values
* Duplicates
* Sources
* Time coverage
* Status of identifiers (researcher names and institutions)
* Salient terms
Produces a tidy dataset for additional enrichment and analysis.
|
1.0
|
Exploratory data analysis - Exploratory analysis of various CORD-19 datasets. Includes:
* Data structure (sp for full-text)
* Variables
* Missing values
* Duplicates
* Sources
* Time coverage
* Status of identifiers (researcher names and institutions)
* Salient terms
Produces a tidy dataset for additional enrichment and analysis.
|
non_test
|
exploratory data analysis exploratory analysis of various cord datasets includes data structure sp for full text variables missing values duplicates sources time coverage status of identifiers researcher names and institutions salient terms produces a tidy dataset for additional enrichment and analysis
| 0
|
41,080
| 5,331,941,748
|
IssuesEvent
|
2017-02-15 20:47:23
|
UCHIC/iUTAHData
|
https://api.github.com/repos/UCHIC/iUTAHData
|
closed
|
Update Site because of transition to HydroShare (and other updates)
|
ReadyForTesting
|
Need to update:
Homepage:
- [x] Change text: https://docs.google.com/document/d/1m4E31ygXoFX7WWt_FTZ6jJESuaJcDhHXiApuTR1OrlU/edit
Software Development:
- [x] Remove the sentence under "Software Development.
- [x] Change text: https://docs.google.com/document/d/12LYwc4OkIcGI_Zb4H6h0Lasvgcomxuy4kjmShHB5Vhw/edit
Data Policy and Publication:
- [x] Change text: https://docs.google.com/document/d/1zPIV21-NX9wSMgP_pXBJ8D1XRLeNtxh84U4cKmml4RE/edit
- [x] Under the "Data" menu, change the "Dataset Repository" item to "iUTAH Data on HydroShare" and link to https://www.hydroshare.org/search/?q=&selected_facets=owners_names_exact:iUTAH%20Data%20Manager
- [x] Under the "Data" menu, add an item "Survey Data" and link to data.iutahepscor.org/surveys
- [x] Personnel: Remove Tony Castronova with (in two places).
- [x] Training Materials: Remove the last two videos (Search & Discovery and Submittal & Publication).
- [x] Documentation: Update text in several places with some new links, too: https://docs.google.com/document/d/10bjsKWS81JzLoFlJORltH_qQAzQcnLSooYEB3VAoMKc/edit#
|
1.0
|
Update Site because of transition to HydroShare (and other updates) - Need to update:
Homepage:
- [x] Change text: https://docs.google.com/document/d/1m4E31ygXoFX7WWt_FTZ6jJESuaJcDhHXiApuTR1OrlU/edit
Software Development:
- [x] Remove the sentence under "Software Development.
- [x] Change text: https://docs.google.com/document/d/12LYwc4OkIcGI_Zb4H6h0Lasvgcomxuy4kjmShHB5Vhw/edit
Data Policy and Publication:
- [x] Change text: https://docs.google.com/document/d/1zPIV21-NX9wSMgP_pXBJ8D1XRLeNtxh84U4cKmml4RE/edit
- [x] Under the "Data" menu, change the "Dataset Repository" item to "iUTAH Data on HydroShare" and link to https://www.hydroshare.org/search/?q=&selected_facets=owners_names_exact:iUTAH%20Data%20Manager
- [x] Under the "Data" menu, add an item "Survey Data" and link to data.iutahepscor.org/surveys
- [x] Personnel: Remove Tony Castronova with (in two places).
- [x] Training Materials: Remove the last two videos (Search & Discovery and Submittal & Publication).
- [x] Documentation: Update text in several places with some new links, too: https://docs.google.com/document/d/10bjsKWS81JzLoFlJORltH_qQAzQcnLSooYEB3VAoMKc/edit#
|
test
|
update site because of transition to hydroshare and other updates need to update homepage change text software development remove the sentence under software development change text data policy and publication change text under the data menu change the dataset repository item to iutah data on hydroshare and link to under the data menu add an item survey data and link to data iutahepscor org surveys personnel remove tony castronova with in two places training materials remove the last two videos search discovery and submittal publication documentation update text in several places with some new links too
| 1
|
130,565
| 27,722,257,181
|
IssuesEvent
|
2023-03-14 21:43:39
|
Rothamsted/knetminer
|
https://api.github.com/repos/Rothamsted/knetminer
|
opened
|
I the /qtl API still used? If yes, it needs review
|
code review code quality
|
I can't see any call in the UI/client to `/qtl`. On the server, it does almost the same things that `/genome` does, worse its implementation is conflated with `/genome`, within `OndexLocalDataSource.handleMainSearch()`.
Is this API used, within the UI or externally? If not, it's better to clean-up things by removing it (including `QtlResponse` and the conditional code which is run when the API itself is invoked.
I guess @KeywanHP is the best to know it.
|
2.0
|
I the /qtl API still used? If yes, it needs review - I can't see any call in the UI/client to `/qtl`. On the server, it does almost the same things that `/genome` does, worse its implementation is conflated with `/genome`, within `OndexLocalDataSource.handleMainSearch()`.
Is this API used, within the UI or externally? If not, it's better to clean-up things by removing it (including `QtlResponse` and the conditional code which is run when the API itself is invoked.
I guess @KeywanHP is the best to know it.
|
non_test
|
i the qtl api still used if yes it needs review i can t see any call in the ui client to qtl on the server it does almost the same things that genome does worse its implementation is conflated with genome within ondexlocaldatasource handlemainsearch is this api used within the ui or externally if not it s better to clean up things by removing it including qtlresponse and the conditional code which is run when the api itself is invoked i guess keywanhp is the best to know it
| 0
|
271,423
| 29,498,277,336
|
IssuesEvent
|
2023-06-02 19:01:55
|
bcgov/rsbc-digital-forms
|
https://api.github.com/repos/bcgov/rsbc-digital-forms
|
closed
|
Critical vulnerabilities found in rsbc-roadside-forms-backend-api on PR 60
|
security
|
# Report
## Results
- **[ERROR]** **[CVE-2022-29361]** [[HELP LINK](https://avd.aquasec.com/nvd/cve-2022-29361)] `Package: Werkzeug
Installed Version: 2.0.2
Vulnerability CVE-2022-29361
Severity: CRITICAL
Fixed Version: 2.1.1
Link: \[CVE-2022-29361\]\(https:\/\/avd.aquasec.com\/nvd\/cve-2022-29361\)`
- https://github.com/bcgov/rsbc-digital-forms/blob/feature/dockerscanv2/opt/app-root/lib/python3.6/site-packages/Werkzeug-2.0.2.dist-info/METADATA#L1-L1
## Suppressed Results
Nothing here.
## Rules information
<!-- Rule Info -->
<details><summary>Rules details</summary>
- CVE-2022-29361 [undefined]
> CVE-2022-29361
</details>
## Tool information
- Name: Trivy
- Organization: undefined
- Version: undefined
|
True
|
Critical vulnerabilities found in rsbc-roadside-forms-backend-api on PR 60 - # Report
## Results
- **[ERROR]** **[CVE-2022-29361]** [[HELP LINK](https://avd.aquasec.com/nvd/cve-2022-29361)] `Package: Werkzeug
Installed Version: 2.0.2
Vulnerability CVE-2022-29361
Severity: CRITICAL
Fixed Version: 2.1.1
Link: \[CVE-2022-29361\]\(https:\/\/avd.aquasec.com\/nvd\/cve-2022-29361\)`
- https://github.com/bcgov/rsbc-digital-forms/blob/feature/dockerscanv2/opt/app-root/lib/python3.6/site-packages/Werkzeug-2.0.2.dist-info/METADATA#L1-L1
## Suppressed Results
Nothing here.
## Rules information
<!-- Rule Info -->
<details><summary>Rules details</summary>
- CVE-2022-29361 [undefined]
> CVE-2022-29361
</details>
## Tool information
- Name: Trivy
- Organization: undefined
- Version: undefined
|
non_test
|
critical vulnerabilities found in rsbc roadside forms backend api on pr report results package werkzeug installed version vulnerability cve severity critical fixed version link https avd aquasec com nvd cve suppressed results nothing here rules information rules details cve cve tool information name trivy organization undefined version undefined
| 0
|
92,887
| 26,794,706,592
|
IssuesEvent
|
2023-02-01 11:01:10
|
zeek/zeek
|
https://api.github.com/repos/zeek/zeek
|
closed
|
libspicy and libhilti should build as shared objects
|
Area: Build/Platform
|
They're being built as static libraries now which means that the size of the following binaries is needlessly large...
```
-rwxr-xr-x 1 root root 439337760 Jan 6 05:33 hilti-config
-rwxr-xr-x 1 root root 442081168 Jan 6 05:33 hiltic
-rwxr-xr-x 1 root root 725743872 Jan 6 05:36 spicy-config
-rwxr-xr-x 1 root root 728677696 Jan 6 05:36 spicyc
-rwxr-xr-x 1 root root 728891008 Jan 6 05:36 spicy-driver
-rwxr-xr-x 1 root root 731454632 Jan 6 05:36 spicy-dump
-rwxr-xr-x 1 root root 751046864 Jan 6 05:39 spicyz
-rwxr-xr-x 1 root root 1048532000 Jan 6 05:40 zeek
```
Other than zeek, those binaries should all be very small but with a default build they're huge right now. Even if you strip them they're rather large due to the large amount of code that is in libhilti and libspicy.
|
1.0
|
libspicy and libhilti should build as shared objects - They're being built as static libraries now which means that the size of the following binaries is needlessly large...
```
-rwxr-xr-x 1 root root 439337760 Jan 6 05:33 hilti-config
-rwxr-xr-x 1 root root 442081168 Jan 6 05:33 hiltic
-rwxr-xr-x 1 root root 725743872 Jan 6 05:36 spicy-config
-rwxr-xr-x 1 root root 728677696 Jan 6 05:36 spicyc
-rwxr-xr-x 1 root root 728891008 Jan 6 05:36 spicy-driver
-rwxr-xr-x 1 root root 731454632 Jan 6 05:36 spicy-dump
-rwxr-xr-x 1 root root 751046864 Jan 6 05:39 spicyz
-rwxr-xr-x 1 root root 1048532000 Jan 6 05:40 zeek
```
Other than zeek, those binaries should all be very small but with a default build they're huge right now. Even if you strip them they're rather large due to the large amount of code that is in libhilti and libspicy.
|
non_test
|
libspicy and libhilti should build as shared objects they re being built as static libraries now which means that the size of the following binaries is needlessly large rwxr xr x root root jan hilti config rwxr xr x root root jan hiltic rwxr xr x root root jan spicy config rwxr xr x root root jan spicyc rwxr xr x root root jan spicy driver rwxr xr x root root jan spicy dump rwxr xr x root root jan spicyz rwxr xr x root root jan zeek other than zeek those binaries should all be very small but with a default build they re huge right now even if you strip them they re rather large due to the large amount of code that is in libhilti and libspicy
| 0
|
164,272
| 12,795,271,185
|
IssuesEvent
|
2020-07-02 08:28:41
|
rancher/rke2
|
https://api.github.com/repos/rancher/rke2
|
closed
|
nginx-ingress-controller service is in pending state
|
[zube]: To Test
|
Version:
rke2 v0.0.1-alpha.4
Issue:
nginx-ingress-controller service is in pending state. Since we dont have servicelb it should not expected to be of type LoadBalancer
```
kube-system nginx-ingress-controller LoadBalancer 10.43.5.161 <pending> 80:30782/TCP,443:30488/TCP 4h32m
```
|
1.0
|
nginx-ingress-controller service is in pending state - Version:
rke2 v0.0.1-alpha.4
Issue:
nginx-ingress-controller service is in pending state. Since we dont have servicelb it should not expected to be of type LoadBalancer
```
kube-system nginx-ingress-controller LoadBalancer 10.43.5.161 <pending> 80:30782/TCP,443:30488/TCP 4h32m
```
|
test
|
nginx ingress controller service is in pending state version alpha issue nginx ingress controller service is in pending state since we dont have servicelb it should not expected to be of type loadbalancer kube system nginx ingress controller loadbalancer tcp tcp
| 1
|
342,312
| 30,613,395,184
|
IssuesEvent
|
2023-07-23 22:14:28
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
pkg/sql/schemachanger/schemachanger_test: TestValidateMixedVersionElements_drop_column_computed_index failed
|
C-test-failure O-robot branch-master T-sql-foundations
|
pkg/sql/schemachanger/schemachanger_test.TestValidateMixedVersionElements_drop_column_computed_index [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/11029922?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/11029922?buildTab=artifacts#/) on master @ [d293bb3acd9b7f67ac793ddb6a0836e637864629](https://github.com/cockroachdb/cockroach/commits/d293bb3acd9b7f67ac793ddb6a0836e637864629):
```
panic.go:522: -- test log scope end --
test logs left over in: /artifacts/tmp/_tmp/2d9af299258c11d4d3ade298f6c5b40e/logTestValidateMixedVersionElements_drop_column_computed_index1756802228
--- FAIL: TestValidateMixedVersionElements_drop_column_computed_index (249.43s)
=== RUN TestValidateMixedVersionElements_drop_column_computed_index/pause_upgrade_and_resume_at_each_stage
--- FAIL: TestValidateMixedVersionElements_drop_column_computed_index/pause_upgrade_and_resume_at_each_stage (248.42s)
=== RUN TestValidateMixedVersionElements_drop_column_computed_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3
sql_runner.go:104:
Error Trace: github.com/cockroachdb/cockroach/pkg/testutils/sqlutils/sql_runner.go:117
github.com/cockroachdb/cockroach/pkg/testutils/sqlutils/sql_runner.go:312
github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/end_to_end.go:494
github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1629
github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1730
Error: Received unexpected error:
query 'SELECT status, job_type, description FROM [SHOW JOBS] WHERE job_type IN ('SCHEMA CHANGE', 'TYPEDESC SCHEMA CHANGE', 'NEW SCHEMA CHANGE') AND status NOT IN ('succeeded', 'failed')': expected:
(1) attached stack trace
-- stack trace:
| github.com/cockroachdb/cockroach/pkg/testutils/sqlutils.(*SQLRunner).CheckQueryResultsRetry.func1
| github.com/cockroachdb/cockroach/pkg/testutils/sqlutils/sql_runner.go:315
| github.com/cockroachdb/cockroach/pkg/testutils.SucceedsWithinError.func1
| github.com/cockroachdb/cockroach/pkg/testutils/soon.go:75
| github.com/cockroachdb/cockroach/pkg/util/retry.ForDuration
| github.com/cockroachdb/cockroach/pkg/util/retry/retry.go:213
| github.com/cockroachdb/cockroach/pkg/testutils.SucceedsWithinError
| github.com/cockroachdb/cockroach/pkg/testutils/soon.go:81
| github.com/cockroachdb/cockroach/pkg/testutils/sqlutils.(*SQLRunner).succeedsWithin
| github.com/cockroachdb/cockroach/pkg/testutils/sqlutils/sql_runner.go:117
| github.com/cockroachdb/cockroach/pkg/testutils/sqlutils.(*SQLRunner).CheckQueryResultsRetry
| github.com/cockroachdb/cockroach/pkg/testutils/sqlutils/sql_runner.go:312
| github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest.waitForSchemaChangesToFinish
| github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/end_to_end.go:494
| github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest.executeSchemaChangeTxn
| github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1629
| github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest.ValidateMixedVersionElements.func5.1
| github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1730
| testing.tRunner
| GOROOT/src/testing/testing.go:1446
| runtime.goexit
| GOROOT/src/runtime/asm_arm64.s:1172
Wraps: (2) query 'SELECT status, job_type, description FROM [SHOW JOBS] WHERE job_type IN ('SCHEMA CHANGE', 'TYPEDESC SCHEMA CHANGE', 'NEW SCHEMA CHANGE') AND status NOT IN ('succeeded', 'failed')': expected:
|
| got:
| paused, NEW SCHEMA CHANGE, ALTER TABLE defaultdb.public.t DROP COLUMN j CASCADE
Error types: (1) *withstack.withStack (2) *errutil.leafError
cumulative.go:1725:
Error Trace: github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1725
GOROOT/src/runtime/asm_arm64.s:1172
Error: Received unexpected error:
dial tcp 127.0.0.1:38973: connect: connection refused
Test: TestValidateMixedVersionElements_drop_column_computed_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3
--- FAIL: TestValidateMixedVersionElements_drop_column_computed_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3 (51.57s)
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/sql-foundations
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestValidateMixedVersionElements_drop_column_computed_index.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
1.0
|
pkg/sql/schemachanger/schemachanger_test: TestValidateMixedVersionElements_drop_column_computed_index failed - pkg/sql/schemachanger/schemachanger_test.TestValidateMixedVersionElements_drop_column_computed_index [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/11029922?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/11029922?buildTab=artifacts#/) on master @ [d293bb3acd9b7f67ac793ddb6a0836e637864629](https://github.com/cockroachdb/cockroach/commits/d293bb3acd9b7f67ac793ddb6a0836e637864629):
```
panic.go:522: -- test log scope end --
test logs left over in: /artifacts/tmp/_tmp/2d9af299258c11d4d3ade298f6c5b40e/logTestValidateMixedVersionElements_drop_column_computed_index1756802228
--- FAIL: TestValidateMixedVersionElements_drop_column_computed_index (249.43s)
=== RUN TestValidateMixedVersionElements_drop_column_computed_index/pause_upgrade_and_resume_at_each_stage
--- FAIL: TestValidateMixedVersionElements_drop_column_computed_index/pause_upgrade_and_resume_at_each_stage (248.42s)
=== RUN TestValidateMixedVersionElements_drop_column_computed_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3
sql_runner.go:104:
Error Trace: github.com/cockroachdb/cockroach/pkg/testutils/sqlutils/sql_runner.go:117
github.com/cockroachdb/cockroach/pkg/testutils/sqlutils/sql_runner.go:312
github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/end_to_end.go:494
github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1629
github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1730
Error: Received unexpected error:
query 'SELECT status, job_type, description FROM [SHOW JOBS] WHERE job_type IN ('SCHEMA CHANGE', 'TYPEDESC SCHEMA CHANGE', 'NEW SCHEMA CHANGE') AND status NOT IN ('succeeded', 'failed')': expected:
(1) attached stack trace
-- stack trace:
| github.com/cockroachdb/cockroach/pkg/testutils/sqlutils.(*SQLRunner).CheckQueryResultsRetry.func1
| github.com/cockroachdb/cockroach/pkg/testutils/sqlutils/sql_runner.go:315
| github.com/cockroachdb/cockroach/pkg/testutils.SucceedsWithinError.func1
| github.com/cockroachdb/cockroach/pkg/testutils/soon.go:75
| github.com/cockroachdb/cockroach/pkg/util/retry.ForDuration
| github.com/cockroachdb/cockroach/pkg/util/retry/retry.go:213
| github.com/cockroachdb/cockroach/pkg/testutils.SucceedsWithinError
| github.com/cockroachdb/cockroach/pkg/testutils/soon.go:81
| github.com/cockroachdb/cockroach/pkg/testutils/sqlutils.(*SQLRunner).succeedsWithin
| github.com/cockroachdb/cockroach/pkg/testutils/sqlutils/sql_runner.go:117
| github.com/cockroachdb/cockroach/pkg/testutils/sqlutils.(*SQLRunner).CheckQueryResultsRetry
| github.com/cockroachdb/cockroach/pkg/testutils/sqlutils/sql_runner.go:312
| github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest.waitForSchemaChangesToFinish
| github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/end_to_end.go:494
| github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest.executeSchemaChangeTxn
| github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1629
| github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest.ValidateMixedVersionElements.func5.1
| github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1730
| testing.tRunner
| GOROOT/src/testing/testing.go:1446
| runtime.goexit
| GOROOT/src/runtime/asm_arm64.s:1172
Wraps: (2) query 'SELECT status, job_type, description FROM [SHOW JOBS] WHERE job_type IN ('SCHEMA CHANGE', 'TYPEDESC SCHEMA CHANGE', 'NEW SCHEMA CHANGE') AND status NOT IN ('succeeded', 'failed')': expected:
|
| got:
| paused, NEW SCHEMA CHANGE, ALTER TABLE defaultdb.public.t DROP COLUMN j CASCADE
Error types: (1) *withstack.withStack (2) *errutil.leafError
cumulative.go:1725:
Error Trace: github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1725
GOROOT/src/runtime/asm_arm64.s:1172
Error: Received unexpected error:
dial tcp 127.0.0.1:38973: connect: connection refused
Test: TestValidateMixedVersionElements_drop_column_computed_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3
--- FAIL: TestValidateMixedVersionElements_drop_column_computed_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3 (51.57s)
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/sql-foundations
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestValidateMixedVersionElements_drop_column_computed_index.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
test
|
pkg sql schemachanger schemachanger test testvalidatemixedversionelements drop column computed index failed pkg sql schemachanger schemachanger test testvalidatemixedversionelements drop column computed index with on master panic go test log scope end test logs left over in artifacts tmp tmp logtestvalidatemixedversionelements drop column computed fail testvalidatemixedversionelements drop column computed index run testvalidatemixedversionelements drop column computed index pause upgrade and resume at each stage fail testvalidatemixedversionelements drop column computed index pause upgrade and resume at each stage run testvalidatemixedversionelements drop column computed index pause upgrade and resume at each stage postcommitnonrevertiblephase of sql runner go error trace github com cockroachdb cockroach pkg testutils sqlutils sql runner go github com cockroachdb cockroach pkg testutils sqlutils sql runner go github com cockroachdb cockroach pkg sql schemachanger sctest end to end go github com cockroachdb cockroach pkg sql schemachanger sctest cumulative go github com cockroachdb cockroach pkg sql schemachanger sctest cumulative go error received unexpected error query select status job type description from where job type in schema change typedesc schema change new schema change and status not in succeeded failed expected attached stack trace stack trace github com cockroachdb cockroach pkg testutils sqlutils sqlrunner checkqueryresultsretry github com cockroachdb cockroach pkg testutils sqlutils sql runner go github com cockroachdb cockroach pkg testutils succeedswithinerror github com cockroachdb cockroach pkg testutils soon go github com cockroachdb cockroach pkg util retry forduration github com cockroachdb cockroach pkg util retry retry go github com cockroachdb cockroach pkg testutils succeedswithinerror github com cockroachdb cockroach pkg testutils soon go github com cockroachdb cockroach pkg testutils sqlutils sqlrunner succeedswithin github com cockroachdb cockroach pkg testutils sqlutils sql runner go github com cockroachdb cockroach pkg testutils sqlutils sqlrunner checkqueryresultsretry github com cockroachdb cockroach pkg testutils sqlutils sql runner go github com cockroachdb cockroach pkg sql schemachanger sctest waitforschemachangestofinish github com cockroachdb cockroach pkg sql schemachanger sctest end to end go github com cockroachdb cockroach pkg sql schemachanger sctest executeschemachangetxn github com cockroachdb cockroach pkg sql schemachanger sctest cumulative go github com cockroachdb cockroach pkg sql schemachanger sctest validatemixedversionelements github com cockroachdb cockroach pkg sql schemachanger sctest cumulative go testing trunner goroot src testing testing go runtime goexit goroot src runtime asm s wraps query select status job type description from where job type in schema change typedesc schema change new schema change and status not in succeeded failed expected got paused new schema change alter table defaultdb public t drop column j cascade error types withstack withstack errutil leaferror cumulative go error trace github com cockroachdb cockroach pkg sql schemachanger sctest cumulative go goroot src runtime asm s error received unexpected error dial tcp connect connection refused test testvalidatemixedversionelements drop column computed index pause upgrade and resume at each stage postcommitnonrevertiblephase of fail testvalidatemixedversionelements drop column computed index pause upgrade and resume at each stage postcommitnonrevertiblephase of help see also cc cockroachdb sql foundations
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.