Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
490,428
14,121,065,105
IssuesEvent
2020-11-09 00:41:24
kubeflow/kubeflow
https://api.github.com/repos/kubeflow/kubeflow
closed
Installation guide for GCP is incomplete and it's impossible to deploy new Kubeflow cluster
area/docs kind/bug platform/gcp priority/p1
/kind bug **What steps did you take and what happened:** I've carefully followed the GCP Kubeflow deployment guide **twice** (starting from scratch), without success. https://www.kubeflow.org/docs/gke/deploy/ My `set-values` from Makefile (the one from https://www.kubeflow.org/docs/gke/deploy/deploy-cli/#fetch-packages-using-kpt): ```makefile set-values: kpt cfg set ./instance gke.private false kpt cfg set ./instance mgmt-ctxt kf-mgmnt kpt cfg set ./upstream/manifests/gcp name kf kpt cfg set ./upstream/manifests/gcp gcloud.core.project XXX kpt cfg set ./upstream/manifests/gcp gcloud.compute.zone europe-west4 kpt cfg set ./upstream/manifests/gcp location europe-west4-b kpt cfg set ./upstream/manifests/gcp log-firewalls false kpt cfg set ./upstream/manifests/stacks/gcp name kf kpt cfg set ./upstream/manifests/stacks/gcp gcloud.core.project XXX kpt cfg set ./instance name kf kpt cfg set ./instance location europe-west4-b kpt cfg set ./instance gcloud.core.project XXX kpt cfg set ./instance email X@XX ``` Final error I get after running `make apply` in step https://www.kubeflow.org/docs/gke/deploy/deploy-cli/#deploy-kubeflow is: ```bash kubectl --context=kf-mgmnt wait --for=condition=Ready --timeout=600s containercluster kf error: timed out waiting for the condition on containerclusters/kf ``` **What did you expect to happen:** Installation is seamless and completes without errors, installation guide is complete. **Anything else you would like to add:** Current deployment documentation seems to be neglected and there are a lot of missing pieces there that waste people's time on digging through many GitHub issues. **Issues I've had to deal with manually / questions:** 1. When management cluster has long name, like **kubeflow-management**, the setup https://www.kubeflow.org/docs/gke/deploy/management-setup/ cannot be completed due to violation of some max characters limit for service account names 1. The guide https://www.kubeflow.org/docs/gke/deploy/ does not mention that you need kustomize IN SPECIFIC VERSION (https://github.com/kubeflow/manifests/issues/1490#issuecomment-674092679), istioctl IN GOOGLE-PROVIDED version (https://github.com/kubeflow/manifests/issues/1490#issuecomment-673729937) 1. There is a strong assumption that cluster needs to have the namespace exactly the same as the project? Why is that? 1. Why Kubeflow needs a separate "management" cluster now? **Environment:** - Kubeflow version: n/a - kfctl version: (use `kfctl version`): not installed - Kubernetes platform: GCP - Kubernetes version: ``` Client Version: version.Info{Major:"1", Minor:"16+", GitVersion:"v1.16.13-dispatcher", GitCommit:"fd22db44e150011eccc8729db223945384460143", GitTreeState:"clean", BuildDate:"2020-07-24T07:48:37Z", GoVersion:"go1.13.9", Compiler:"gc", Platform:"darwin/amd64"} Server Version: version.Info{Major:"1", Minor:"16+", GitVersion:"v1.16.13-gke.1", GitCommit:"688c6543aa4b285355723f100302d80431e411cc", GitTreeState:"clean", BuildDate:"2020-07-21T02:37:26Z", GoVersion:"go1.13.9b4", Compiler:"gc", Platform:"linux/amd64"} ``` - OS: macOS 10.15.1
1.0
Installation guide for GCP is incomplete and it's impossible to deploy new Kubeflow cluster - /kind bug **What steps did you take and what happened:** I've carefully followed the GCP Kubeflow deployment guide **twice** (starting from scratch), without success. https://www.kubeflow.org/docs/gke/deploy/ My `set-values` from Makefile (the one from https://www.kubeflow.org/docs/gke/deploy/deploy-cli/#fetch-packages-using-kpt): ```makefile set-values: kpt cfg set ./instance gke.private false kpt cfg set ./instance mgmt-ctxt kf-mgmnt kpt cfg set ./upstream/manifests/gcp name kf kpt cfg set ./upstream/manifests/gcp gcloud.core.project XXX kpt cfg set ./upstream/manifests/gcp gcloud.compute.zone europe-west4 kpt cfg set ./upstream/manifests/gcp location europe-west4-b kpt cfg set ./upstream/manifests/gcp log-firewalls false kpt cfg set ./upstream/manifests/stacks/gcp name kf kpt cfg set ./upstream/manifests/stacks/gcp gcloud.core.project XXX kpt cfg set ./instance name kf kpt cfg set ./instance location europe-west4-b kpt cfg set ./instance gcloud.core.project XXX kpt cfg set ./instance email X@XX ``` Final error I get after running `make apply` in step https://www.kubeflow.org/docs/gke/deploy/deploy-cli/#deploy-kubeflow is: ```bash kubectl --context=kf-mgmnt wait --for=condition=Ready --timeout=600s containercluster kf error: timed out waiting for the condition on containerclusters/kf ``` **What did you expect to happen:** Installation is seamless and completes without errors, installation guide is complete. **Anything else you would like to add:** Current deployment documentation seems to be neglected and there are a lot of missing pieces there that waste people's time on digging through many GitHub issues. **Issues I've had to deal with manually / questions:** 1. When management cluster has long name, like **kubeflow-management**, the setup https://www.kubeflow.org/docs/gke/deploy/management-setup/ cannot be completed due to violation of some max characters limit for service account names 1. The guide https://www.kubeflow.org/docs/gke/deploy/ does not mention that you need kustomize IN SPECIFIC VERSION (https://github.com/kubeflow/manifests/issues/1490#issuecomment-674092679), istioctl IN GOOGLE-PROVIDED version (https://github.com/kubeflow/manifests/issues/1490#issuecomment-673729937) 1. There is a strong assumption that cluster needs to have the namespace exactly the same as the project? Why is that? 1. Why Kubeflow needs a separate "management" cluster now? **Environment:** - Kubeflow version: n/a - kfctl version: (use `kfctl version`): not installed - Kubernetes platform: GCP - Kubernetes version: ``` Client Version: version.Info{Major:"1", Minor:"16+", GitVersion:"v1.16.13-dispatcher", GitCommit:"fd22db44e150011eccc8729db223945384460143", GitTreeState:"clean", BuildDate:"2020-07-24T07:48:37Z", GoVersion:"go1.13.9", Compiler:"gc", Platform:"darwin/amd64"} Server Version: version.Info{Major:"1", Minor:"16+", GitVersion:"v1.16.13-gke.1", GitCommit:"688c6543aa4b285355723f100302d80431e411cc", GitTreeState:"clean", BuildDate:"2020-07-21T02:37:26Z", GoVersion:"go1.13.9b4", Compiler:"gc", Platform:"linux/amd64"} ``` - OS: macOS 10.15.1
non_test
installation guide for gcp is incomplete and it s impossible to deploy new kubeflow cluster kind bug what steps did you take and what happened i ve carefully followed the gcp kubeflow deployment guide twice starting from scratch without success my set values from makefile the one from makefile set values kpt cfg set instance gke private false kpt cfg set instance mgmt ctxt kf mgmnt kpt cfg set upstream manifests gcp name kf kpt cfg set upstream manifests gcp gcloud core project xxx kpt cfg set upstream manifests gcp gcloud compute zone europe kpt cfg set upstream manifests gcp location europe b kpt cfg set upstream manifests gcp log firewalls false kpt cfg set upstream manifests stacks gcp name kf kpt cfg set upstream manifests stacks gcp gcloud core project xxx kpt cfg set instance name kf kpt cfg set instance location europe b kpt cfg set instance gcloud core project xxx kpt cfg set instance email x xx final error i get after running make apply in step is bash kubectl context kf mgmnt wait for condition ready timeout containercluster kf error timed out waiting for the condition on containerclusters kf what did you expect to happen installation is seamless and completes without errors installation guide is complete anything else you would like to add current deployment documentation seems to be neglected and there are a lot of missing pieces there that waste people s time on digging through many github issues issues i ve had to deal with manually questions when management cluster has long name like kubeflow management the setup cannot be completed due to violation of some max characters limit for service account names the guide does not mention that you need kustomize in specific version istioctl in google provided version there is a strong assumption that cluster needs to have the namespace exactly the same as the project why is that why kubeflow needs a separate management cluster now environment kubeflow version n a kfctl version use kfctl version not installed kubernetes platform gcp kubernetes version client version version info major minor gitversion dispatcher gitcommit gittreestate clean builddate goversion compiler gc platform darwin server version version info major minor gitversion gke gitcommit gittreestate clean builddate goversion compiler gc platform linux os macos
0
145,326
11,685,310,415
IssuesEvent
2020-03-05 08:51:12
MorphCast/Studio
https://api.github.com/repos/MorphCast/Studio
closed
Deleting a video timeline item, exiting timeline, reentering and undoing the delete will crash
bug ready for testing
The "recursive" application of the locallydeleted commands for the timeline item and its "childrens" will create issues both when the timeline item is firstly locallydeleted or lastly. It's not a matter of that change, but instead a matter of creating a TimelineBox for each media item inside the timeline item instead of linking it to the front one.
1.0
Deleting a video timeline item, exiting timeline, reentering and undoing the delete will crash - The "recursive" application of the locallydeleted commands for the timeline item and its "childrens" will create issues both when the timeline item is firstly locallydeleted or lastly. It's not a matter of that change, but instead a matter of creating a TimelineBox for each media item inside the timeline item instead of linking it to the front one.
test
deleting a video timeline item exiting timeline reentering and undoing the delete will crash the recursive application of the locallydeleted commands for the timeline item and its childrens will create issues both when the timeline item is firstly locallydeleted or lastly it s not a matter of that change but instead a matter of creating a timelinebox for each media item inside the timeline item instead of linking it to the front one
1
27,850
5,114,074,401
IssuesEvent
2017-01-06 17:16:04
edno/kleis
https://api.github.com/repos/edno/kleis
closed
Field Group displayed whilst creating Super Admin
defect
When a field validation error occurs, then the field Group is displayed whilst creating Super Admin account. The field is empty (no values available), and it does not block the creation. The field Group should be displayed only for Local Admin. Steps to reproduce: - Create a new Super Admin user with incorrect password format
1.0
Field Group displayed whilst creating Super Admin - When a field validation error occurs, then the field Group is displayed whilst creating Super Admin account. The field is empty (no values available), and it does not block the creation. The field Group should be displayed only for Local Admin. Steps to reproduce: - Create a new Super Admin user with incorrect password format
non_test
field group displayed whilst creating super admin when a field validation error occurs then the field group is displayed whilst creating super admin account the field is empty no values available and it does not block the creation the field group should be displayed only for local admin steps to reproduce create a new super admin user with incorrect password format
0
126,763
10,434,829,342
IssuesEvent
2019-09-17 15:57:54
Submitty/Submitty
https://api.github.com/repos/Submitty/Submitty
closed
Add testing for late days
Testing / Continuous Integration (CI)
**What problem are you trying to solve with Submitty** Add unit tests for the different types of late day states ![image](https://user-images.githubusercontent.com/12129065/64914490-f2909e00-d720-11e9-96a7-a7bda293af16.png)
1.0
Add testing for late days - **What problem are you trying to solve with Submitty** Add unit tests for the different types of late day states ![image](https://user-images.githubusercontent.com/12129065/64914490-f2909e00-d720-11e9-96a7-a7bda293af16.png)
test
add testing for late days what problem are you trying to solve with submitty add unit tests for the different types of late day states
1
348,118
31,468,224,035
IssuesEvent
2023-08-30 05:01:53
gear-tech/gear
https://api.github.com/repos/gear-tech/gear
closed
Write tests for each send signal case (with asserting appropriate signal code)
C1-feature D4-test
### Problem to Solve / .. / ### Possible Solution / .. / ### Notes _No response_
1.0
Write tests for each send signal case (with asserting appropriate signal code) - ### Problem to Solve / .. / ### Possible Solution / .. / ### Notes _No response_
test
write tests for each send signal case with asserting appropriate signal code problem to solve possible solution notes no response
1
252,515
21,582,207,046
IssuesEvent
2022-05-02 20:02:21
damccorm/test-migration-target
https://api.github.com/repos/damccorm/test-migration-target
opened
beam_PostCommit_XVR_GoUsingJava_Dataflow fails on some test transforms
bug test-failures cross-language sdk-go P2
Example failure: https://ci-beam.apache.org/job/beam_PostCommit_XVR_GoUsingJava_Dataflow/7/ I couldn't find accurate details about why the tests are failing, but TestXLang_Prefix, TestXLang_Multi, and TestXLang_Partition are failing while running for some reason. Investigating the Dataflow logs, we can see SDK harnesses are failing to connect for some reason. For example: `noformat` "getPodContainerStatuses for pod "df-go-testxlang-multi-03300551-62xv-harness-3msv_default(a7f1d8dfb2c3d2b4e80f5d92c1728787)" failed: rpc error: code = Unknown desc = Error: No such container: bea0d9bde42bf890f6fe1d4f589932471037a5948fb9588d01a06425cd14c177" `noformat` However I haven't been able to find any further details showing why the harness fails, and the tests keep running beyond that for a while with other errors that are also pretty inscrutable. Imported from Jira [BEAM-14214](https://issues.apache.org/jira/browse/BEAM-14214). Original Jira may contain additional context. Reported by: danoliveira.
1.0
beam_PostCommit_XVR_GoUsingJava_Dataflow fails on some test transforms - Example failure: https://ci-beam.apache.org/job/beam_PostCommit_XVR_GoUsingJava_Dataflow/7/ I couldn't find accurate details about why the tests are failing, but TestXLang_Prefix, TestXLang_Multi, and TestXLang_Partition are failing while running for some reason. Investigating the Dataflow logs, we can see SDK harnesses are failing to connect for some reason. For example: `noformat` "getPodContainerStatuses for pod "df-go-testxlang-multi-03300551-62xv-harness-3msv_default(a7f1d8dfb2c3d2b4e80f5d92c1728787)" failed: rpc error: code = Unknown desc = Error: No such container: bea0d9bde42bf890f6fe1d4f589932471037a5948fb9588d01a06425cd14c177" `noformat` However I haven't been able to find any further details showing why the harness fails, and the tests keep running beyond that for a while with other errors that are also pretty inscrutable. Imported from Jira [BEAM-14214](https://issues.apache.org/jira/browse/BEAM-14214). Original Jira may contain additional context. Reported by: danoliveira.
test
beam postcommit xvr gousingjava dataflow fails on some test transforms example failure i couldn t find accurate details about why the tests are failing but testxlang prefix testxlang multi and testxlang partition are failing while running for some reason investigating the dataflow logs we can see sdk harnesses are failing to connect for some reason for example noformat getpodcontainerstatuses for pod df go testxlang multi harness default failed rpc error code unknown desc error no such container noformat however i haven t been able to find any further details showing why the harness fails and the tests keep running beyond that for a while with other errors that are also pretty inscrutable imported from jira original jira may contain additional context reported by danoliveira
1
627,472
19,905,632,970
IssuesEvent
2022-01-25 12:30:21
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.derstandard.at - site is not usable
browser-firefox-mobile priority-normal type-tracking-protection-strict engine-gecko
<!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/98670 --> <!-- @extra_labels: type-tracking-protection-strict --> **URL**: https://www.derstandard.at/story/2000132199542/indische-regierung-nimmt-mutter-teresa-in-der-mangel?ref=rss **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android 9 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Missing items **Steps to Reproduce**: The community comments are not displayed. <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2022/1/bd1e895c-5908-4752-a4f0-6474f3df8776.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200204012329</li><li>channel: default</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: true (strict)</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2022/1/dd9dfc87-cfaa-4e00-b8d0-e184b5249f92) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.derstandard.at - site is not usable - <!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/98670 --> <!-- @extra_labels: type-tracking-protection-strict --> **URL**: https://www.derstandard.at/story/2000132199542/indische-regierung-nimmt-mutter-teresa-in-der-mangel?ref=rss **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android 9 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Missing items **Steps to Reproduce**: The community comments are not displayed. <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2022/1/bd1e895c-5908-4752-a4f0-6474f3df8776.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200204012329</li><li>channel: default</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: true (strict)</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2022/1/dd9dfc87-cfaa-4e00-b8d0-e184b5249f92) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_test
site is not usable url browser version firefox mobile operating system android tested another browser yes chrome problem type site is not usable description missing items steps to reproduce the community comments are not displayed view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel default hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked true strict from with ❤️
0
164,746
12,812,898,687
IssuesEvent
2020-07-04 09:30:29
aliasrobotics/RVD
https://api.github.com/repos/aliasrobotics/RVD
closed
RVD#2918: CWE-120 (buffer), Does not check for buffer overflows when concatenating to destination ... @ tforms/posix/src/main.cpp:500
CWE-120 bug components software flawfinder flawfinder_level_4 mitigated robot component: PX4 static analysis testing triage version: v1.8.0
```yaml id: 2918 title: 'RVD#2918: CWE-120 (buffer), Does not check for buffer overflows when concatenating to destination ... @ tforms/posix/src/main.cpp:500' type: bug description: 'Does not check for buffer overflows when concatenating to destination [MS-banned] (CWE-120). Consider using strcat_s, strncat, strlcat, or snprintf (warning: strncat is easily misused). . Happening @ ...tforms/posix/src/main.cpp:500' cwe: - CWE-120 cve: None keywords: - flawfinder - flawfinder_level_4 - static analysis - testing - triage - CWE-120 - bug - 'version: v1.8.0' - 'robot component: PX4' - components software system: ./Firmware/platforms/posix/src/main.cpp:500:18 vendor: null severity: rvss-score: 0 rvss-vector: '' severity-description: '' cvss-score: 0 cvss-vector: '' links: - https://github.com/aliasrobotics/RVD/issues/2918 flaw: phase: testing specificity: subject-specific architectural-location: application-specific application: N/A subsystem: N/A package: N/A languages: None date-detected: 2020-06-29 (15:49) detected-by: Alias Robotics detected-by-method: testing static date-reported: 2020-06-29 (15:49) reported-by: Alias Robotics reported-by-relationship: automatic issue: https://github.com/aliasrobotics/RVD/issues/2918 reproducibility: always trace: (context) \t\tif (nullptr == strcat(pwd_path, folderpath)) { reproduction: See artifacts below (if available) reproduction-image: gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_px4/-/jobs/615986299/artifacts/download exploitation: description: '' exploitation-image: '' exploitation-vector: '' exploitation-recipe: '' mitigation: description: 'Consider using strcat_s, strncat, strlcat, or snprintf (warning: strncat is easily misused)' pull-request: '' date-mitigation: '' ```
1.0
RVD#2918: CWE-120 (buffer), Does not check for buffer overflows when concatenating to destination ... @ tforms/posix/src/main.cpp:500 - ```yaml id: 2918 title: 'RVD#2918: CWE-120 (buffer), Does not check for buffer overflows when concatenating to destination ... @ tforms/posix/src/main.cpp:500' type: bug description: 'Does not check for buffer overflows when concatenating to destination [MS-banned] (CWE-120). Consider using strcat_s, strncat, strlcat, or snprintf (warning: strncat is easily misused). . Happening @ ...tforms/posix/src/main.cpp:500' cwe: - CWE-120 cve: None keywords: - flawfinder - flawfinder_level_4 - static analysis - testing - triage - CWE-120 - bug - 'version: v1.8.0' - 'robot component: PX4' - components software system: ./Firmware/platforms/posix/src/main.cpp:500:18 vendor: null severity: rvss-score: 0 rvss-vector: '' severity-description: '' cvss-score: 0 cvss-vector: '' links: - https://github.com/aliasrobotics/RVD/issues/2918 flaw: phase: testing specificity: subject-specific architectural-location: application-specific application: N/A subsystem: N/A package: N/A languages: None date-detected: 2020-06-29 (15:49) detected-by: Alias Robotics detected-by-method: testing static date-reported: 2020-06-29 (15:49) reported-by: Alias Robotics reported-by-relationship: automatic issue: https://github.com/aliasrobotics/RVD/issues/2918 reproducibility: always trace: (context) \t\tif (nullptr == strcat(pwd_path, folderpath)) { reproduction: See artifacts below (if available) reproduction-image: gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_px4/-/jobs/615986299/artifacts/download exploitation: description: '' exploitation-image: '' exploitation-vector: '' exploitation-recipe: '' mitigation: description: 'Consider using strcat_s, strncat, strlcat, or snprintf (warning: strncat is easily misused)' pull-request: '' date-mitigation: '' ```
test
rvd cwe buffer does not check for buffer overflows when concatenating to destination tforms posix src main cpp yaml id title rvd cwe buffer does not check for buffer overflows when concatenating to destination tforms posix src main cpp type bug description does not check for buffer overflows when concatenating to destination cwe consider using strcat s strncat strlcat or snprintf warning strncat is easily misused happening tforms posix src main cpp cwe cwe cve none keywords flawfinder flawfinder level static analysis testing triage cwe bug version robot component components software system firmware platforms posix src main cpp vendor null severity rvss score rvss vector severity description cvss score cvss vector links flaw phase testing specificity subject specific architectural location application specific application n a subsystem n a package n a languages none date detected detected by alias robotics detected by method testing static date reported reported by alias robotics reported by relationship automatic issue reproducibility always trace context t tif nullptr strcat pwd path folderpath reproduction see artifacts below if available reproduction image gitlab com aliasrobotics offensive alurity pipelines active pipeline jobs artifacts download exploitation description exploitation image exploitation vector exploitation recipe mitigation description consider using strcat s strncat strlcat or snprintf warning strncat is easily misused pull request date mitigation
1
318,293
27,297,075,801
IssuesEvent
2023-02-23 21:21:23
nucleus-security/Test-repo
https://api.github.com/repos/nucleus-security/Test-repo
closed
Nucleus - [High] - 440057
Test
Source: QUALYS Finding Description: CentOS has released security update for kernel to fix the vulnerabilities. Affected Products: centos 6 Impact: Successful exploitation allows attacker to compromise the system. Target(s): Asset name: 192.168.56.103 Solution: To resolve this issue, upgrade to the latest packages which contain a patch. Refer to CentOS advisory centos 6 (https://lists.centos.org/pipermail/centos-announce/2016-July/021977.html) for updates and patch information. Patch: Following are links for downloading patches to fix the vulnerabilities: CESA-2016:1406: centos 6 (https://lists.centos.org/pipermail/centos-announce/2016-July/021977.html) References: QID:440057 CVE:CVE-2016-4565 Category:CentOS PCI Flagged:yes Vendor References:CESA-2016:1406 centos 6 Bugtraq IDs:90301 Severity: High Date Discovered: 2022-11-12 08:04:44 Nucleus Notification Rules Triggered: Rule GitHub Project Name: 6716 Please see Nucleus for more information on these vulnerabilities:https://192.168.56.101/nucleus/public/app/index.html#vuln/201000007/NDQwMDU3/UVVBTFlT/VnVsbg--/false/MjAxMDAwMDA3/c3VtbWFyeQ--/false
1.0
Nucleus - [High] - 440057 - Source: QUALYS Finding Description: CentOS has released security update for kernel to fix the vulnerabilities. Affected Products: centos 6 Impact: Successful exploitation allows attacker to compromise the system. Target(s): Asset name: 192.168.56.103 Solution: To resolve this issue, upgrade to the latest packages which contain a patch. Refer to CentOS advisory centos 6 (https://lists.centos.org/pipermail/centos-announce/2016-July/021977.html) for updates and patch information. Patch: Following are links for downloading patches to fix the vulnerabilities: CESA-2016:1406: centos 6 (https://lists.centos.org/pipermail/centos-announce/2016-July/021977.html) References: QID:440057 CVE:CVE-2016-4565 Category:CentOS PCI Flagged:yes Vendor References:CESA-2016:1406 centos 6 Bugtraq IDs:90301 Severity: High Date Discovered: 2022-11-12 08:04:44 Nucleus Notification Rules Triggered: Rule GitHub Project Name: 6716 Please see Nucleus for more information on these vulnerabilities:https://192.168.56.101/nucleus/public/app/index.html#vuln/201000007/NDQwMDU3/UVVBTFlT/VnVsbg--/false/MjAxMDAwMDA3/c3VtbWFyeQ--/false
test
nucleus source qualys finding description centos has released security update for kernel to fix the vulnerabilities affected products centos impact successful exploitation allows attacker to compromise the system target s asset name solution to resolve this issue upgrade to the latest packages which contain a patch refer to centos advisory centos for updates and patch information patch following are links for downloading patches to fix the vulnerabilities cesa centos references qid cve cve category centos pci flagged yes vendor references cesa centos bugtraq ids severity high date discovered nucleus notification rules triggered rule github project name please see nucleus for more information on these vulnerabilities
1
65,486
16,371,664,725
IssuesEvent
2021-05-15 08:43:09
lettre/lettre
https://api.github.com/repos/lettre/lettre
closed
High-level methods for attachments and embedded files
component:builder component:docs status:in progress type:bug type:feature
Looks like the only way to add an attachment in v.0.10 at the moment is to manually create a multipart message. 1. It's too much hassle 2. I couldn't even make it work with the examples provided I noticed this comment `// TODO: High-level methods for attachments and embedded files` in https://github.com/lettre/lettre/blob/13b48b656d916308c26680db896695034fdec1d9/src/message/mod.rs#L374 Since I now have to make it work for our project I'm happy to make an extra effort and implement adding attachments as high level methods. #### Questions: 1. Can you make a simple example of having a plain text part with an attachment using multipart or singlepart to help me get started? Both https://github.com/lettre/lettre/blob/master/src/message/mod.rs#L85 and https://github.com/lettre/lettre/issues/492#issuecomment-716111178 examples panic over a non-existing boundary with `called Option::unwrap() on a None value, /home/ubuntu/.cargo/git/checkouts/lettre-53652803723a9045/86763cc/src/message/mimebody.rs:402:63` 2. Did you plan to replicate the file attaching interface from the previous version or do something completely different?
1.0
High-level methods for attachments and embedded files - Looks like the only way to add an attachment in v.0.10 at the moment is to manually create a multipart message. 1. It's too much hassle 2. I couldn't even make it work with the examples provided I noticed this comment `// TODO: High-level methods for attachments and embedded files` in https://github.com/lettre/lettre/blob/13b48b656d916308c26680db896695034fdec1d9/src/message/mod.rs#L374 Since I now have to make it work for our project I'm happy to make an extra effort and implement adding attachments as high level methods. #### Questions: 1. Can you make a simple example of having a plain text part with an attachment using multipart or singlepart to help me get started? Both https://github.com/lettre/lettre/blob/master/src/message/mod.rs#L85 and https://github.com/lettre/lettre/issues/492#issuecomment-716111178 examples panic over a non-existing boundary with `called Option::unwrap() on a None value, /home/ubuntu/.cargo/git/checkouts/lettre-53652803723a9045/86763cc/src/message/mimebody.rs:402:63` 2. Did you plan to replicate the file attaching interface from the previous version or do something completely different?
non_test
high level methods for attachments and embedded files looks like the only way to add an attachment in v at the moment is to manually create a multipart message it s too much hassle i couldn t even make it work with the examples provided i noticed this comment todo high level methods for attachments and embedded files in since i now have to make it work for our project i m happy to make an extra effort and implement adding attachments as high level methods questions can you make a simple example of having a plain text part with an attachment using multipart or singlepart to help me get started both and examples panic over a non existing boundary with called option unwrap on a none value home ubuntu cargo git checkouts lettre src message mimebody rs did you plan to replicate the file attaching interface from the previous version or do something completely different
0
305,968
26,423,890,835
IssuesEvent
2023-01-14 00:27:59
cosmos/cosmos-sdk
https://api.github.com/repos/cosmos/cosmos-sdk
closed
e2e has flaky CLI tests (TestCLIMultisignSortSignatures, TestEditValidatorMoniker and TestCLISignBatch)
good first issue T: Tests
<!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺ v ✰ Thanks for opening an issue! ✰ v Before smashing the submit button please review the template. v Please also ensure that this is not a duplicate issue :) ☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > --> <!-- IMPORTANT: Prior to opening a bug report, check if it affects one of the core modules and if its elegible for a bug bounty on `SECURITY.md`. Bugs that are not submitted through the appropriate channels won't receive any bounty. --> ## Summary of Bug I've seen 3 new flaky tests: ``` --- FAIL: TestIntegrationTestSuite/TestCLIMultisignSortSignatures (3.37s) suite.go:1024: Error Trace: /home/ubuntu/actions-runner/_work/cosmos-sdk/cosmos-sdk/tests/e2e/auth/suite.go:1024 Error: Not equal: expected: math.Int{i:(*big.Int)(0xc0043df100)} actual : math.Int{i:(*big.Int)(0xc002243280)} Diff: --- Expected +++ Actual @@ -3,5 +3,3 @@ neg: (bool) false, - abs: (big.nat) (len=1) { - (big.Word) 10 - } + abs: (big.nat) <nil> }) Test: TestIntegrationTestSuite/TestCLIMultisignSortSignatures ``` and ``` --- FAIL: TestIntegrationTestSuite (109.64s) suite.go:42: setting up integration test suite network.go:289: acquiring test network lock network.go:299: preparing test network with chain-id "chain-dzqeIU" network.go:568: starting test network... network.go:574: started validator 0 network.go:574: started validator 1 network.go:582: started test network at height: 0 --- FAIL: TestIntegrationTestSuite/TestEditValidatorMoniker (2.19s) suite.go:1538: Error Trace: /home/ubuntu/actions-runner/_work/cosmos-sdk/cosmos-sdk/tests/e2e/staking/client/testutil/suite.go:1538 Error: Not equal: expected: "node0" actual : "testing" Diff: --- Expected +++ Actual @@ -1 +1 @@ -node0 +testing Test: TestIntegrationTestSuite/TestEditValidatorMoniker ``` and ``` --- FAIL: TestIntegrationTestSuite/TestCLISignBatch (4.69s) suite.go:320: Error Trace: /home/runner/work/cosmos-sdk/cosmos-sdk/tests/e2e/auth/suite.go:320 Error: Not equal: expected: 0x9 actual : 0x8 Test: TestIntegrationTestSuite/TestCLISignBatch ```
1.0
e2e has flaky CLI tests (TestCLIMultisignSortSignatures, TestEditValidatorMoniker and TestCLISignBatch) - <!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺ v ✰ Thanks for opening an issue! ✰ v Before smashing the submit button please review the template. v Please also ensure that this is not a duplicate issue :) ☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > --> <!-- IMPORTANT: Prior to opening a bug report, check if it affects one of the core modules and if its elegible for a bug bounty on `SECURITY.md`. Bugs that are not submitted through the appropriate channels won't receive any bounty. --> ## Summary of Bug I've seen 3 new flaky tests: ``` --- FAIL: TestIntegrationTestSuite/TestCLIMultisignSortSignatures (3.37s) suite.go:1024: Error Trace: /home/ubuntu/actions-runner/_work/cosmos-sdk/cosmos-sdk/tests/e2e/auth/suite.go:1024 Error: Not equal: expected: math.Int{i:(*big.Int)(0xc0043df100)} actual : math.Int{i:(*big.Int)(0xc002243280)} Diff: --- Expected +++ Actual @@ -3,5 +3,3 @@ neg: (bool) false, - abs: (big.nat) (len=1) { - (big.Word) 10 - } + abs: (big.nat) <nil> }) Test: TestIntegrationTestSuite/TestCLIMultisignSortSignatures ``` and ``` --- FAIL: TestIntegrationTestSuite (109.64s) suite.go:42: setting up integration test suite network.go:289: acquiring test network lock network.go:299: preparing test network with chain-id "chain-dzqeIU" network.go:568: starting test network... network.go:574: started validator 0 network.go:574: started validator 1 network.go:582: started test network at height: 0 --- FAIL: TestIntegrationTestSuite/TestEditValidatorMoniker (2.19s) suite.go:1538: Error Trace: /home/ubuntu/actions-runner/_work/cosmos-sdk/cosmos-sdk/tests/e2e/staking/client/testutil/suite.go:1538 Error: Not equal: expected: "node0" actual : "testing" Diff: --- Expected +++ Actual @@ -1 +1 @@ -node0 +testing Test: TestIntegrationTestSuite/TestEditValidatorMoniker ``` and ``` --- FAIL: TestIntegrationTestSuite/TestCLISignBatch (4.69s) suite.go:320: Error Trace: /home/runner/work/cosmos-sdk/cosmos-sdk/tests/e2e/auth/suite.go:320 Error: Not equal: expected: 0x9 actual : 0x8 Test: TestIntegrationTestSuite/TestCLISignBatch ```
test
has flaky cli tests testclimultisignsortsignatures testeditvalidatormoniker and testclisignbatch ☺ v ✰ thanks for opening an issue ✰ v before smashing the submit button please review the template v please also ensure that this is not a duplicate issue ☺ important prior to opening a bug report check if it affects one of the core modules and if its elegible for a bug bounty on security md bugs that are not submitted through the appropriate channels won t receive any bounty summary of bug i ve seen new flaky tests fail testintegrationtestsuite testclimultisignsortsignatures suite go error trace home ubuntu actions runner work cosmos sdk cosmos sdk tests auth suite go error not equal expected math int i big int actual math int i big int diff expected actual neg bool false abs big nat len big word abs big nat test testintegrationtestsuite testclimultisignsortsignatures and fail testintegrationtestsuite suite go setting up integration test suite network go acquiring test network lock network go preparing test network with chain id chain dzqeiu network go starting test network network go started validator network go started validator network go started test network at height fail testintegrationtestsuite testeditvalidatormoniker suite go error trace home ubuntu actions runner work cosmos sdk cosmos sdk tests staking client testutil suite go error not equal expected actual testing diff expected actual testing test testintegrationtestsuite testeditvalidatormoniker and fail testintegrationtestsuite testclisignbatch suite go error trace home runner work cosmos sdk cosmos sdk tests auth suite go error not equal expected actual test testintegrationtestsuite testclisignbatch
1
64,174
15,818,670,032
IssuesEvent
2021-04-05 16:21:27
ns1labs/pktvisor
https://api.github.com/repos/ns1labs/pktvisor
opened
statically linked binaries
build-system
generate statically linked versions of `pktvisord`, `pktvisor-pcap` for various platforms
1.0
statically linked binaries - generate statically linked versions of `pktvisord`, `pktvisor-pcap` for various platforms
non_test
statically linked binaries generate statically linked versions of pktvisord pktvisor pcap for various platforms
0
12,227
7,811,372,643
IssuesEvent
2018-06-12 09:54:09
crate/crate
https://api.github.com/repos/crate/crate
closed
Bad join performance with no possible matches
feature: performance feature: sql: joins
Hello, I have create 2 "clones" tables named people3 and people 4: cr> select table_name, column_name, data_type from information_schema.columns where table_name like 'people%'; +------------+-------------+-----------+ | table_name | column_name | data_type | +------------+-------------+-----------+ | people3 | birthdate | timestamp | | people3 | name | string | | people4 | birthdate | timestamp | | people4 | name | string | +------------+-------------+-----------+ And add many rows: cr> select count(*) from people3; +----------+ | count(*) | +----------+ | 3449535 | +----------+
True
Bad join performance with no possible matches - Hello, I have create 2 "clones" tables named people3 and people 4: cr> select table_name, column_name, data_type from information_schema.columns where table_name like 'people%'; +------------+-------------+-----------+ | table_name | column_name | data_type | +------------+-------------+-----------+ | people3 | birthdate | timestamp | | people3 | name | string | | people4 | birthdate | timestamp | | people4 | name | string | +------------+-------------+-----------+ And add many rows: cr> select count(*) from people3; +----------+ | count(*) | +----------+ | 3449535 | +----------+
non_test
bad join performance with no possible matches hello i have create clones tables named and people cr select table name column name data type from information schema columns where table name like people table name column name data type birthdate timestamp name string birthdate timestamp name string and add many rows cr select count from count
0
194,681
15,437,620,646
IssuesEvent
2021-03-07 17:22:21
microsoft/azure-pipelines-task-lib
https://api.github.com/repos/microsoft/azure-pipelines-task-lib
closed
Documentation
documentation stale
### Environment azure-pipelines-task-lib version: current ### Issue Description There is no documentation on how to use the node lib for common scenarios. ### Expected behaviour There is clear documentation with examples for the most common (or all) commands. 1. for Example the `tl.exec()` command 2. for getting `connectedService` secrets, handling and using them 3. general construction of the UI and its features Just pointing to working code or config files is not documentation. I am sorry if something like this already exists, but I was unable to find it. It's late over here so please point to the right links.
1.0
Documentation - ### Environment azure-pipelines-task-lib version: current ### Issue Description There is no documentation on how to use the node lib for common scenarios. ### Expected behaviour There is clear documentation with examples for the most common (or all) commands. 1. for Example the `tl.exec()` command 2. for getting `connectedService` secrets, handling and using them 3. general construction of the UI and its features Just pointing to working code or config files is not documentation. I am sorry if something like this already exists, but I was unable to find it. It's late over here so please point to the right links.
non_test
documentation environment azure pipelines task lib version current issue description there is no documentation on how to use the node lib for common scenarios expected behaviour there is clear documentation with examples for the most common or all commands for example the tl exec command for getting connectedservice secrets handling and using them general construction of the ui and its features just pointing to working code or config files is not documentation i am sorry if something like this already exists but i was unable to find it it s late over here so please point to the right links
0
38,116
12,528,264,883
IssuesEvent
2020-06-04 09:17:44
ckauhaus/nixpkgs
https://api.github.com/repos/ckauhaus/nixpkgs
opened
Vulnerability roundup 4: advancecomp-2.1: 1 advisory
1.severity: security
[search](https://search.nix.gsc.io/?q=advancecomp&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=advancecomp+in%3Apath&type=Code) * [ ] [CVE-2019-9210](https://nvd.nist.gov/vuln/detail/CVE-2019-9210) CVSSv3=7.8 (nixos-19.03) Scanned versions: nixos-19.03: 34c7eb7545d. May contain false positives.
True
Vulnerability roundup 4: advancecomp-2.1: 1 advisory - [search](https://search.nix.gsc.io/?q=advancecomp&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=advancecomp+in%3Apath&type=Code) * [ ] [CVE-2019-9210](https://nvd.nist.gov/vuln/detail/CVE-2019-9210) CVSSv3=7.8 (nixos-19.03) Scanned versions: nixos-19.03: 34c7eb7545d. May contain false positives.
non_test
vulnerability roundup advancecomp advisory nixos scanned versions nixos may contain false positives
0
244,992
20,736,789,504
IssuesEvent
2022-03-14 14:23:00
dnd-side-project/dnd-6th-5-backend
https://api.github.com/repos/dnd-side-project/dnd-6th-5-backend
closed
test: patchUserNickname api에 대한 통합테스트, 유닛테스트 작성
test
- [x] node-mocks-http, supertest, jest 모듈 설치 - [x] tscongfig에 json파일을 import하기 위한 설정 추가 - [x] 에러 핸들링 미들웨어 추가 - [x] test script 수정 - [x] ormconfig.js test DB설정 추가 - [x] jset.config.js 설정중 transform에서 js확장자도 추가 - [x] tsconfig.json에서 js확장자 파일인 jest설정파일을 읽을 수 있도록 allowJs 설정 true로 설정 - [x] patchUserNickname 컨트롤러에 대한 unittest code작성 - [x] patchUserNickname 컨트롤러에 대한 통합테스트 code작성
1.0
test: patchUserNickname api에 대한 통합테스트, 유닛테스트 작성 - - [x] node-mocks-http, supertest, jest 모듈 설치 - [x] tscongfig에 json파일을 import하기 위한 설정 추가 - [x] 에러 핸들링 미들웨어 추가 - [x] test script 수정 - [x] ormconfig.js test DB설정 추가 - [x] jset.config.js 설정중 transform에서 js확장자도 추가 - [x] tsconfig.json에서 js확장자 파일인 jest설정파일을 읽을 수 있도록 allowJs 설정 true로 설정 - [x] patchUserNickname 컨트롤러에 대한 unittest code작성 - [x] patchUserNickname 컨트롤러에 대한 통합테스트 code작성
test
test patchusernickname api에 대한 통합테스트 유닛테스트 작성 node mocks http supertest jest 모듈 설치 tscongfig에 json파일을 import하기 위한 설정 추가 에러 핸들링 미들웨어 추가 test script 수정 ormconfig js test db설정 추가 jset config js 설정중 transform에서 js확장자도 추가 tsconfig json에서 js확장자 파일인 jest설정파일을 읽을 수 있도록 allowjs 설정 true로 설정 patchusernickname 컨트롤러에 대한 unittest code작성 patchusernickname 컨트롤러에 대한 통합테스트 code작성
1
315,804
27,107,822,811
IssuesEvent
2023-02-15 13:23:40
ventoy/Ventoy
https://api.github.com/repos/ventoy/Ventoy
closed
[Success Image Report]: Tiny 11 ( by NTDEV ) based on Windows 11 Pro 22H2
【Tested Image Report】
### Official Website List - [X] I have checked the list in official website and the image file is not listed there. ### Ventoy Version 1.0.88 ### BIOS Mode Both ### Partition Style MBR ### Image file name tiny11 b1.iso ### Image file checksum type MD5 ### Image file checksum value efd53d1bd51854ee57391ea3a4700cbf ### Image file download link (if applicable) https://archive.org/download/tiny-11-NTDEV/tiny11%20b1.iso ### Test environment Lenovo ThinkPad X230 Laptop ### More Details? This image file booted successfully in Ventoy and no errors were seen. Laptop Specs: - Intel(R) Core(TM) i5-3320M CPU @ 2.60GHz, 2601 Mhz, 2 Core(s), 4 Logical Processor(s) x64-based PC 8.00 GB RAM 500 GB SSD
1.0
[Success Image Report]: Tiny 11 ( by NTDEV ) based on Windows 11 Pro 22H2 - ### Official Website List - [X] I have checked the list in official website and the image file is not listed there. ### Ventoy Version 1.0.88 ### BIOS Mode Both ### Partition Style MBR ### Image file name tiny11 b1.iso ### Image file checksum type MD5 ### Image file checksum value efd53d1bd51854ee57391ea3a4700cbf ### Image file download link (if applicable) https://archive.org/download/tiny-11-NTDEV/tiny11%20b1.iso ### Test environment Lenovo ThinkPad X230 Laptop ### More Details? This image file booted successfully in Ventoy and no errors were seen. Laptop Specs: - Intel(R) Core(TM) i5-3320M CPU @ 2.60GHz, 2601 Mhz, 2 Core(s), 4 Logical Processor(s) x64-based PC 8.00 GB RAM 500 GB SSD
test
tiny by ntdev based on windows pro official website list i have checked the list in official website and the image file is not listed there ventoy version bios mode both partition style mbr image file name iso image file checksum type image file checksum value image file download link if applicable test environment lenovo thinkpad laptop more details this image file booted successfully in ventoy and no errors were seen laptop specs intel r core tm cpu mhz core s logical processor s based pc gb ram gb ssd
1
219,049
24,436,717,553
IssuesEvent
2022-10-06 12:02:06
pazhanivel07/tcpdump-4.9.2
https://api.github.com/repos/pazhanivel07/tcpdump-4.9.2
opened
CVE-2018-14470 (High) detected in tcpdumptcpdump-4.9.2
security vulnerability
## CVE-2018-14470 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tcpdumptcpdump-4.9.2</b></p></summary> <p> <p>the TCPdump network dissector</p> <p>Library home page: <a href=https://github.com/the-tcpdump-group/tcpdump.git>https://github.com/the-tcpdump-group/tcpdump.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/tcpdump-4.9.2/commit/761aa8f39eabb1228c1d03e3a55c861d76c46817">761aa8f39eabb1228c1d03e3a55c861d76c46817</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/print-babel.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The Babel parser in tcpdump before 4.9.3 has a buffer over-read in print-babel.c:babel_print_v2(). <p>Publish Date: 2019-10-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14470>CVE-2018-14470</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-14470">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-14470</a></p> <p>Release Date: 2019-10-03</p> <p>Fix Resolution: 4.9.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-14470 (High) detected in tcpdumptcpdump-4.9.2 - ## CVE-2018-14470 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tcpdumptcpdump-4.9.2</b></p></summary> <p> <p>the TCPdump network dissector</p> <p>Library home page: <a href=https://github.com/the-tcpdump-group/tcpdump.git>https://github.com/the-tcpdump-group/tcpdump.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/tcpdump-4.9.2/commit/761aa8f39eabb1228c1d03e3a55c861d76c46817">761aa8f39eabb1228c1d03e3a55c861d76c46817</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/print-babel.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The Babel parser in tcpdump before 4.9.3 has a buffer over-read in print-babel.c:babel_print_v2(). <p>Publish Date: 2019-10-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14470>CVE-2018-14470</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-14470">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-14470</a></p> <p>Release Date: 2019-10-03</p> <p>Fix Resolution: 4.9.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in tcpdumptcpdump cve high severity vulnerability vulnerable library tcpdumptcpdump the tcpdump network dissector library home page a href found in head commit a href found in base branch master vulnerable source files print babel c vulnerability details the babel parser in tcpdump before has a buffer over read in print babel c babel print publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
199,387
15,036,594,904
IssuesEvent
2021-02-02 15:25:25
apache/buildstream
https://api.github.com/repos/apache/buildstream
closed
Run CI as non-root user
enhancement tests
[See original issue on GitLab](https://gitlab.com/BuildStream/buildstream/-/issues/102) In GitLab by [[Gitlab user @tlater]](https://gitlab.com/tlater) on Oct 3, 2017, 14:31 Currently all tests are run as root, which is not likely to reflect the actual user environment, and may mask permission issues. CI tests should run as a separate user for the linux platform. Simple `su` calls don't appear to work in the container.
1.0
Run CI as non-root user - [See original issue on GitLab](https://gitlab.com/BuildStream/buildstream/-/issues/102) In GitLab by [[Gitlab user @tlater]](https://gitlab.com/tlater) on Oct 3, 2017, 14:31 Currently all tests are run as root, which is not likely to reflect the actual user environment, and may mask permission issues. CI tests should run as a separate user for the linux platform. Simple `su` calls don't appear to work in the container.
test
run ci as non root user in gitlab by on oct currently all tests are run as root which is not likely to reflect the actual user environment and may mask permission issues ci tests should run as a separate user for the linux platform simple su calls don t appear to work in the container
1
130,471
10,617,051,333
IssuesEvent
2019-10-12 16:16:32
unsuitable001/memoryJS
https://api.github.com/repos/unsuitable001/memoryJS
opened
[unit testing] Add coverage check using CODEBEAT
enhancement hacktoberfest help wanted testing up-for-grabs
[unit testing] Add coverage check using CODEBEAT. You have to write the tests & try to fix if any error occurs. Refer - https://hub.codebeat.co/docs/test-coverage-reports#section-setting-up-for-javascript-typescript
1.0
[unit testing] Add coverage check using CODEBEAT - [unit testing] Add coverage check using CODEBEAT. You have to write the tests & try to fix if any error occurs. Refer - https://hub.codebeat.co/docs/test-coverage-reports#section-setting-up-for-javascript-typescript
test
add coverage check using codebeat add coverage check using codebeat you have to write the tests try to fix if any error occurs refer
1
158,072
12,401,453,792
IssuesEvent
2020-05-21 09:54:17
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
opened
An invalid time warning still occurs when switching to 'None' in add message dialog
:gear: queues 🧪 testing
**Storage Explorer Version:** 1.14.0-dev **Build**: [20200521.3](https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=3746638&view=results) **Branch:** master **Platform/OS:** Windows 10/ Linux Ubuntu 16.04/ macOS Mojave **Architecture:** ia32/x64 **Regression From:** Not a regression **Steps to reproduce:** 1. Expand one storage account -> Queues. 2. Create a new queue -> Click 'Add Message'. 3. Check 'Expire in:' -> Input an invalid string -> Switch to 'None'. 4. Check the result. **Expect Experience:** No warning occurs. **Actual Experience:** The warning still occurs. ![image](https://user-images.githubusercontent.com/54055206/82547171-de445100-9b8b-11ea-9788-569d22baf35d.png)
1.0
An invalid time warning still occurs when switching to 'None' in add message dialog - **Storage Explorer Version:** 1.14.0-dev **Build**: [20200521.3](https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=3746638&view=results) **Branch:** master **Platform/OS:** Windows 10/ Linux Ubuntu 16.04/ macOS Mojave **Architecture:** ia32/x64 **Regression From:** Not a regression **Steps to reproduce:** 1. Expand one storage account -> Queues. 2. Create a new queue -> Click 'Add Message'. 3. Check 'Expire in:' -> Input an invalid string -> Switch to 'None'. 4. Check the result. **Expect Experience:** No warning occurs. **Actual Experience:** The warning still occurs. ![image](https://user-images.githubusercontent.com/54055206/82547171-de445100-9b8b-11ea-9788-569d22baf35d.png)
test
an invalid time warning still occurs when switching to none in add message dialog storage explorer version dev build branch master platform os windows linux ubuntu macos mojave architecture regression from not a regression steps to reproduce expand one storage account queues create a new queue click add message check expire in input an invalid string switch to none check the result expect experience no warning occurs actual experience the warning still occurs
1
22,342
2,648,808,102
IssuesEvent
2015-03-14 08:37:31
gunmetalbackupgooglecode/opendbg
https://api.github.com/repos/gunmetalbackupgooglecode/opendbg
closed
tracer doesn't have some functions for work with GUI
imported OpSys-Windows Priority-Critical Type-Review wontfix
_From [d1mk4nah@gmail.com](https://code.google.com/u/d1mk4nah@gmail.com/) on September 13, 2008 06:07:57_ We should have thunks for some functions (you can look at it when you try compile tracer) at tracer or kernel side. _Original issue: http://code.google.com/p/opendbg/issues/detail?id=1_
1.0
tracer doesn't have some functions for work with GUI - _From [d1mk4nah@gmail.com](https://code.google.com/u/d1mk4nah@gmail.com/) on September 13, 2008 06:07:57_ We should have thunks for some functions (you can look at it when you try compile tracer) at tracer or kernel side. _Original issue: http://code.google.com/p/opendbg/issues/detail?id=1_
non_test
tracer doesn t have some functions for work with gui from on september we should have thunks for some functions you can look at it when you try compile tracer at tracer or kernel side original issue
0
53,098
6,300,485,965
IssuesEvent
2017-07-21 03:58:41
yarnpkg/yarn
https://api.github.com/repos/yarnpkg/yarn
closed
yarn-e2e/label=docker,os=ubuntu-14.04 #258 failed
failure test
Build 'yarn-e2e/label=docker,os=ubuntu-14.04' is failing! Last 50 lines of build output: ``` Started by upstream project "yarn-e2e" build number 258 originally caused by: Started by timer Building on master in workspace /var/lib/jenkins/workspace/yarn-e2e/label/docker/os/ubuntu-14.04 Cloning the remote Git repository Cloning repository https://github.com/yarnpkg/yarn.git > git init /var/lib/jenkins/workspace/yarn-e2e/label/docker/os/ubuntu-14.04 # timeout=10 Fetching upstream changes from https://github.com/yarnpkg/yarn.git > git --version # timeout=10 > git fetch --tags --progress https://github.com/yarnpkg/yarn.git +refs/heads/*:refs/remotes/origin/* ERROR: Error cloning remote repo 'origin' hudson.plugins.git.GitException: Command "git fetch --tags --progress https://github.com/yarnpkg/yarn.git +refs/heads/*:refs/remotes/origin/*" returned status code 128: stdout: stderr: remote: Counting objects: 45503, done. remote: Compressing objects: 1% (1/76) remote: Compressing objects: 2% (2/76) remote: Compressing objects: 3% (3/76) remote: Compressing objects: 5% (4/76) remote: Compressing objects: 6% (5/76) remote: Compressing objects: 7% (6/76) remote: Compressing objects: 9% (7/76) remote: Compressing objects: 10% (8/76) remote: Compressing objects: 11% (9/76) remote: Compressing objects: 13% (10/76) remote: Compressing objects: 14% (11/76) remote: Compressing objects: 15% (12/76) remote: Compressing objects: 17% (13/76) remote: Compressing objects: 18% (14/76) remote: Compressing objects: 19% (15/76) remote: Compressing objects: 21% (16/76) remote: Compressing objects: 22% (17/76) remote: Compressing objects: 23% (18/76) remote: Compressing objects: 25% (19/76) remote: Compressing objects: 26% (20/76) remote: Compressing objects: 27% (21/76) remote: Compressing objects: 28% (22/76) remote: Compressing objects: 30% (23/76) remote: Compressing objects: 31% (24/76) remote: Compressing objects: 32% (25/76) remote: Compressing objects: 34% (26/76) remote: Compressing objects: 35% (27/76) remote: Compressing objects: 36% (28/76) remote: Compressing objects: 38% (29/76) remote: Compressing objects: 39% (30/76) remote: Compressing objects: 40% (31/76) remote: Compressing objects: 42% (32/76) remote: Compressing objects: 43% (33/76) remote: Compressing objects: 44% (34/76) remote: Compressing objects: 46% (35/76) remote: Compressing objects: 47% (36/76) remote: Compressing objects: 48% (37/76) remote: Compressing objects: 50% (38/76) remote: Compressing objects: 51% (39/76) remote: Compressing objects: 52% (40/76) remote: Compressing objects: 53% (41/76) remote: Compressing objects: 55% (42/76) remote: Compressing objects: 56% (43/76) remote: Compressing objects: 57% (44/76) remote: Compressing objects: 59% (45/76) remote: Compressing objects: 60% (46/76) remote: Compressing objects: 61% (47/76) remote: Compressing objects: 63% (48/76) remote: Compressing objects: 64% (49/76) remote: Compressing objects: 65% (50/76) remote: Compressing objects: 67% (51/76) remote: Compressing objects: 68% (52/76) remote: Compressing objects: 69% (53/76) remote: Compressing objects: 71% (54/76) remote: Compressing objects: 72% (55/76) remote: Compressing objects: 73% (56/76) remote: Compressing objects: 75% (57/76) remote: Compressing objects: 76% (58/76) remote: Compressing objects: 77% (59/76) remote: Compressing objects: 78% (60/76) remote: Compressing objects: 80% (61/76) remote: Compressing objects: 81% (62/76) remote: Compressing objects: 82% (63/76) remote: Compressing objects: 84% (64/76) remote: Compressing objects: 85% (65/76) remote: Compressing objects: 86% (66/76) remote: Compressing objects: 88% (67/76) remote: Compressing objects: 89% (68/76) remote: Compressing objects: 90% (69/76) remote: Compressing objects: 92% (70/76) remote: Compressing objects: 93% (71/76) remote: Compressing objects: 94% (72/76) remote: Compressing objects: 96% (73/76) remote: Compressing objects: 97% (74/76) remote: Compressing objects: 98% (75/76) remote: Compressing objects: 100% (76/76) remote: Compressing objects: 100% (76/76), done. Receiving objects: 0% (1/45503) Receiving objects: 1% (456/45503) Receiving objects: 2% (911/45503) Receiving objects: 3% (1366/45503) Receiving objects: 4% (1821/45503) Receiving objects: 5% (2276/45503) Receiving objects: 6% (2731/45503) Receiving objects: 7% (3186/45503) error: RPC failed; curl 56 GnuTLS recv error (-54): Error in the pull function. fatal: The remote end hung up unexpectedly fatal: early EOF fatal: index-pack failed at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommandIn(CliGitAPIImpl.java:1903) at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommandWithCredentials(CliGitAPIImpl.java:1622) at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.access$300(CliGitAPIImpl.java:71) at org.jenkinsci.plugins.gitclient.CliGitAPIImpl$1.execute(CliGitAPIImpl.java:348) at org.jenkinsci.plugins.gitclient.CliGitAPIImpl$2.execute(CliGitAPIImpl.java:545) at hudson.plugins.git.GitSCM.retrieveChanges(GitSCM.java:1070) at hudson.plugins.git.GitSCM.checkout(GitSCM.java:1110) at hudson.scm.SCM.checkout(SCM.java:495) at hudson.model.AbstractProject.checkout(AbstractProject.java:1276) at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:560) at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:86) at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:485) at hudson.model.Run.execute(Run.java:1735) at hudson.matrix.MatrixRun.run(MatrixRun.java:146) at hudson.model.ResourceController.execute(ResourceController.java:97) at hudson.model.Executor.run(Executor.java:405) ERROR: Error cloning remote repo 'origin' ``` Changes since last successful build: No changes [View full output](https://build.dan.cx/job/yarn-e2e/label=docker,os=ubuntu-14.04/258/) cc @Daniel15
1.0
yarn-e2e/label=docker,os=ubuntu-14.04 #258 failed - Build 'yarn-e2e/label=docker,os=ubuntu-14.04' is failing! Last 50 lines of build output: ``` Started by upstream project "yarn-e2e" build number 258 originally caused by: Started by timer Building on master in workspace /var/lib/jenkins/workspace/yarn-e2e/label/docker/os/ubuntu-14.04 Cloning the remote Git repository Cloning repository https://github.com/yarnpkg/yarn.git > git init /var/lib/jenkins/workspace/yarn-e2e/label/docker/os/ubuntu-14.04 # timeout=10 Fetching upstream changes from https://github.com/yarnpkg/yarn.git > git --version # timeout=10 > git fetch --tags --progress https://github.com/yarnpkg/yarn.git +refs/heads/*:refs/remotes/origin/* ERROR: Error cloning remote repo 'origin' hudson.plugins.git.GitException: Command "git fetch --tags --progress https://github.com/yarnpkg/yarn.git +refs/heads/*:refs/remotes/origin/*" returned status code 128: stdout: stderr: remote: Counting objects: 45503, done. remote: Compressing objects: 1% (1/76) remote: Compressing objects: 2% (2/76) remote: Compressing objects: 3% (3/76) remote: Compressing objects: 5% (4/76) remote: Compressing objects: 6% (5/76) remote: Compressing objects: 7% (6/76) remote: Compressing objects: 9% (7/76) remote: Compressing objects: 10% (8/76) remote: Compressing objects: 11% (9/76) remote: Compressing objects: 13% (10/76) remote: Compressing objects: 14% (11/76) remote: Compressing objects: 15% (12/76) remote: Compressing objects: 17% (13/76) remote: Compressing objects: 18% (14/76) remote: Compressing objects: 19% (15/76) remote: Compressing objects: 21% (16/76) remote: Compressing objects: 22% (17/76) remote: Compressing objects: 23% (18/76) remote: Compressing objects: 25% (19/76) remote: Compressing objects: 26% (20/76) remote: Compressing objects: 27% (21/76) remote: Compressing objects: 28% (22/76) remote: Compressing objects: 30% (23/76) remote: Compressing objects: 31% (24/76) remote: Compressing objects: 32% (25/76) remote: Compressing objects: 34% (26/76) remote: Compressing objects: 35% (27/76) remote: Compressing objects: 36% (28/76) remote: Compressing objects: 38% (29/76) remote: Compressing objects: 39% (30/76) remote: Compressing objects: 40% (31/76) remote: Compressing objects: 42% (32/76) remote: Compressing objects: 43% (33/76) remote: Compressing objects: 44% (34/76) remote: Compressing objects: 46% (35/76) remote: Compressing objects: 47% (36/76) remote: Compressing objects: 48% (37/76) remote: Compressing objects: 50% (38/76) remote: Compressing objects: 51% (39/76) remote: Compressing objects: 52% (40/76) remote: Compressing objects: 53% (41/76) remote: Compressing objects: 55% (42/76) remote: Compressing objects: 56% (43/76) remote: Compressing objects: 57% (44/76) remote: Compressing objects: 59% (45/76) remote: Compressing objects: 60% (46/76) remote: Compressing objects: 61% (47/76) remote: Compressing objects: 63% (48/76) remote: Compressing objects: 64% (49/76) remote: Compressing objects: 65% (50/76) remote: Compressing objects: 67% (51/76) remote: Compressing objects: 68% (52/76) remote: Compressing objects: 69% (53/76) remote: Compressing objects: 71% (54/76) remote: Compressing objects: 72% (55/76) remote: Compressing objects: 73% (56/76) remote: Compressing objects: 75% (57/76) remote: Compressing objects: 76% (58/76) remote: Compressing objects: 77% (59/76) remote: Compressing objects: 78% (60/76) remote: Compressing objects: 80% (61/76) remote: Compressing objects: 81% (62/76) remote: Compressing objects: 82% (63/76) remote: Compressing objects: 84% (64/76) remote: Compressing objects: 85% (65/76) remote: Compressing objects: 86% (66/76) remote: Compressing objects: 88% (67/76) remote: Compressing objects: 89% (68/76) remote: Compressing objects: 90% (69/76) remote: Compressing objects: 92% (70/76) remote: Compressing objects: 93% (71/76) remote: Compressing objects: 94% (72/76) remote: Compressing objects: 96% (73/76) remote: Compressing objects: 97% (74/76) remote: Compressing objects: 98% (75/76) remote: Compressing objects: 100% (76/76) remote: Compressing objects: 100% (76/76), done. Receiving objects: 0% (1/45503) Receiving objects: 1% (456/45503) Receiving objects: 2% (911/45503) Receiving objects: 3% (1366/45503) Receiving objects: 4% (1821/45503) Receiving objects: 5% (2276/45503) Receiving objects: 6% (2731/45503) Receiving objects: 7% (3186/45503) error: RPC failed; curl 56 GnuTLS recv error (-54): Error in the pull function. fatal: The remote end hung up unexpectedly fatal: early EOF fatal: index-pack failed at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommandIn(CliGitAPIImpl.java:1903) at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommandWithCredentials(CliGitAPIImpl.java:1622) at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.access$300(CliGitAPIImpl.java:71) at org.jenkinsci.plugins.gitclient.CliGitAPIImpl$1.execute(CliGitAPIImpl.java:348) at org.jenkinsci.plugins.gitclient.CliGitAPIImpl$2.execute(CliGitAPIImpl.java:545) at hudson.plugins.git.GitSCM.retrieveChanges(GitSCM.java:1070) at hudson.plugins.git.GitSCM.checkout(GitSCM.java:1110) at hudson.scm.SCM.checkout(SCM.java:495) at hudson.model.AbstractProject.checkout(AbstractProject.java:1276) at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:560) at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:86) at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:485) at hudson.model.Run.execute(Run.java:1735) at hudson.matrix.MatrixRun.run(MatrixRun.java:146) at hudson.model.ResourceController.execute(ResourceController.java:97) at hudson.model.Executor.run(Executor.java:405) ERROR: Error cloning remote repo 'origin' ``` Changes since last successful build: No changes [View full output](https://build.dan.cx/job/yarn-e2e/label=docker,os=ubuntu-14.04/258/) cc @Daniel15
test
yarn label docker os ubuntu failed build yarn label docker os ubuntu is failing last lines of build output started by upstream project yarn build number originally caused by started by timer building on master in workspace var lib jenkins workspace yarn label docker os ubuntu cloning the remote git repository cloning repository git init var lib jenkins workspace yarn label docker os ubuntu timeout fetching upstream changes from git version timeout git fetch tags progress refs heads refs remotes origin error error cloning remote repo origin hudson plugins git gitexception command git fetch tags progress refs heads refs remotes origin returned status code stdout stderr remote counting objects done remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects remote compressing objects done receiving objects receiving objects receiving objects receiving objects receiving objects receiving objects receiving objects receiving objects error rpc failed curl gnutls recv error error in the pull function fatal the remote end hung up unexpectedly fatal early eof fatal index pack failed at org jenkinsci plugins gitclient cligitapiimpl launchcommandin cligitapiimpl java at org jenkinsci plugins gitclient cligitapiimpl launchcommandwithcredentials cligitapiimpl java at org jenkinsci plugins gitclient cligitapiimpl access cligitapiimpl java at org jenkinsci plugins gitclient cligitapiimpl execute cligitapiimpl java at org jenkinsci plugins gitclient cligitapiimpl execute cligitapiimpl java at hudson plugins git gitscm retrievechanges gitscm java at hudson plugins git gitscm checkout gitscm java at hudson scm scm checkout scm java at hudson model abstractproject checkout abstractproject java at hudson model abstractbuild abstractbuildexecution defaultcheckout abstractbuild java at jenkins scm scmcheckoutstrategy checkout scmcheckoutstrategy java at hudson model abstractbuild abstractbuildexecution run abstractbuild java at hudson model run execute run java at hudson matrix matrixrun run matrixrun java at hudson model resourcecontroller execute resourcecontroller java at hudson model executor run executor java error error cloning remote repo origin changes since last successful build no changes cc
1
215,940
16,621,585,807
IssuesEvent
2021-06-03 02:23:56
JustinCally/VicmapR
https://api.github.com/repos/JustinCally/VicmapR
closed
Improve Documentation And Package Aesthetics
documentation
Tasks to do for improving the ability for users to navigate and use the package: - [x] Write a vignetter on querying data - [x] Review the manual for the functions - [x] Add badges to Readme
1.0
Improve Documentation And Package Aesthetics - Tasks to do for improving the ability for users to navigate and use the package: - [x] Write a vignetter on querying data - [x] Review the manual for the functions - [x] Add badges to Readme
non_test
improve documentation and package aesthetics tasks to do for improving the ability for users to navigate and use the package write a vignetter on querying data review the manual for the functions add badges to readme
0
264,382
23,115,665,265
IssuesEvent
2022-07-27 16:26:10
gravitational/teleport
https://api.github.com/repos/gravitational/teleport
closed
`TestFnCacheCancellation` flakiness
flaky tests
## Failure #### Link(s) to logs - https://console.cloud.google.com/cloud-build/builds;region=us-west1/1ddfaee7-0787-49ed-a3a6-0b236d5ba3f2?project=ci-account #### Relevant snippet ``` === CONT TestFnCacheCancellation fncache_test.go:189: Error Trace: fncache_test.go:189 Error: Received unexpected error: context deadline exceeded Test: TestFnCacheCancellation {"caller":"utils/loadbalancer.go:272","component":"loadbalancer","dest":{"IP":"127.0.0.1","Port":34899,"Zone":""},"level":"debug","message":"forward","source":{"IP":"127.0.0.1","Port":43326,"Zone":""},"timestamp":"2022-07-16T01:01:59Z","trace.fields":{"listen":"127.0.0.1:0"}} {"caller":"utils/loadbalancer.go:201","component":"loadbalancer","level":"debug","message":"created listening socket","timestamp":"2022-07-16T01:01:59Z","trace.fields":{"listen":"127.0.0.1:0"}} --- FAIL: TestFnCacheCancellation (0.21s) ```
1.0
`TestFnCacheCancellation` flakiness - ## Failure #### Link(s) to logs - https://console.cloud.google.com/cloud-build/builds;region=us-west1/1ddfaee7-0787-49ed-a3a6-0b236d5ba3f2?project=ci-account #### Relevant snippet ``` === CONT TestFnCacheCancellation fncache_test.go:189: Error Trace: fncache_test.go:189 Error: Received unexpected error: context deadline exceeded Test: TestFnCacheCancellation {"caller":"utils/loadbalancer.go:272","component":"loadbalancer","dest":{"IP":"127.0.0.1","Port":34899,"Zone":""},"level":"debug","message":"forward","source":{"IP":"127.0.0.1","Port":43326,"Zone":""},"timestamp":"2022-07-16T01:01:59Z","trace.fields":{"listen":"127.0.0.1:0"}} {"caller":"utils/loadbalancer.go:201","component":"loadbalancer","level":"debug","message":"created listening socket","timestamp":"2022-07-16T01:01:59Z","trace.fields":{"listen":"127.0.0.1:0"}} --- FAIL: TestFnCacheCancellation (0.21s) ```
test
testfncachecancellation flakiness failure link s to logs relevant snippet cont testfncachecancellation fncache test go error trace fncache test go error received unexpected error context deadline exceeded test testfncachecancellation caller utils loadbalancer go component loadbalancer dest ip port zone level debug message forward source ip port zone timestamp trace fields listen caller utils loadbalancer go component loadbalancer level debug message created listening socket timestamp trace fields listen fail testfncachecancellation
1
133,545
18,298,940,829
IssuesEvent
2021-10-05 23:47:43
bsbtd/Teste
https://api.github.com/repos/bsbtd/Teste
opened
CVE-2020-14061 (High) detected in jackson-databind-2.9.5.jar, jackson-databind-2.6.7.3.jar
security vulnerability
## CVE-2020-14061 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.5.jar</b>, <b>jackson-databind-2.6.7.3.jar</b></p></summary> <p> <details><summary><b>jackson-databind-2.9.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: Teste/liferay-portal/modules/etl/talend/talend-runtime/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p> <p> Dependency Hierarchy: - components-api-0.25.3.jar (Root Library) - daikon-0.27.0.jar - :x: **jackson-databind-2.9.5.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.6.7.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: Teste/opentelemetry-java/sdk_contrib/aws_v1_support/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.7.3/c8bad9f57a5d298280f8bc0efcb7bf8393a41534/jackson-databind-2.6.7.3.jar</p> <p> Dependency Hierarchy: - aws-java-sdk-ec2-1.11.701.jar (Root Library) - jmespath-java-1.11.701.jar - :x: **jackson-databind-2.6.7.3.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/bsbtd/Teste/commit/64dde89c50c07496423c4d4a865f2e16b92399ad">64dde89c50c07496423c4d4a865f2e16b92399ad</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to oracle.jms.AQjmsQueueConnectionFactory, oracle.jms.AQjmsXATopicConnectionFactory, oracle.jms.AQjmsTopicConnectionFactory, oracle.jms.AQjmsXAQueueConnectionFactory, and oracle.jms.AQjmsXAConnectionFactory (aka weblogic/oracle-aqjms). <p>Publish Date: 2020-06-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14061>CVE-2020-14061</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14061">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14061</a></p> <p>Release Date: 2020-06-14</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-14061 (High) detected in jackson-databind-2.9.5.jar, jackson-databind-2.6.7.3.jar - ## CVE-2020-14061 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.5.jar</b>, <b>jackson-databind-2.6.7.3.jar</b></p></summary> <p> <details><summary><b>jackson-databind-2.9.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: Teste/liferay-portal/modules/etl/talend/talend-runtime/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p> <p> Dependency Hierarchy: - components-api-0.25.3.jar (Root Library) - daikon-0.27.0.jar - :x: **jackson-databind-2.9.5.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.6.7.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: Teste/opentelemetry-java/sdk_contrib/aws_v1_support/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.6.7.3/c8bad9f57a5d298280f8bc0efcb7bf8393a41534/jackson-databind-2.6.7.3.jar</p> <p> Dependency Hierarchy: - aws-java-sdk-ec2-1.11.701.jar (Root Library) - jmespath-java-1.11.701.jar - :x: **jackson-databind-2.6.7.3.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/bsbtd/Teste/commit/64dde89c50c07496423c4d4a865f2e16b92399ad">64dde89c50c07496423c4d4a865f2e16b92399ad</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to oracle.jms.AQjmsQueueConnectionFactory, oracle.jms.AQjmsXATopicConnectionFactory, oracle.jms.AQjmsTopicConnectionFactory, oracle.jms.AQjmsXAQueueConnectionFactory, and oracle.jms.AQjmsXAConnectionFactory (aka weblogic/oracle-aqjms). <p>Publish Date: 2020-06-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14061>CVE-2020-14061</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14061">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14061</a></p> <p>Release Date: 2020-06-14</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in jackson databind jar jackson databind jar cve high severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file teste liferay portal modules etl talend talend runtime pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy components api jar root library daikon jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file teste opentelemetry java sdk contrib aws support build gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy aws java sdk jar root library jmespath java jar x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to oracle jms aqjmsqueueconnectionfactory oracle jms aqjmsxatopicconnectionfactory oracle jms aqjmstopicconnectionfactory oracle jms aqjmsxaqueueconnectionfactory and oracle jms aqjmsxaconnectionfactory aka weblogic oracle aqjms publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource
0
141,978
11,450,171,652
IssuesEvent
2020-02-06 09:04:46
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
closed
Fail to upload an empty folder to one file share
:gear: files 🧪 testing
**Storage Explorer Version:** 1.11.1 **Build:** [20191210.3](https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=3309673) **Branch:** master **Platform/OS:** Windows 10/ Linux Ubuntu 18.04/macOS High Sierra **Architecture:** ia32/x64 **Regression From:** Not a regression **Steps to reproduce:** 1. Expand one storage account -> File Shares. 2. Create one file share -> Upload one empty folder. 3. Check the result. **Expect Experience:** The uploaded empty folder shows. **Actual Experience:** 1. The uploaded empty folder doesn't show. 2. The activity log shows 0 completed. ![image](https://user-images.githubusercontent.com/41351993/70612883-b054fb00-1c42-11ea-9c30-a7979021c91f.png) **More Info:** Succeed to download one empty folder. ![image](https://user-images.githubusercontent.com/41351993/70610549-bb0d9100-1c3e-11ea-94d4-cb87d8d0db9c.png)
1.0
Fail to upload an empty folder to one file share - **Storage Explorer Version:** 1.11.1 **Build:** [20191210.3](https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=3309673) **Branch:** master **Platform/OS:** Windows 10/ Linux Ubuntu 18.04/macOS High Sierra **Architecture:** ia32/x64 **Regression From:** Not a regression **Steps to reproduce:** 1. Expand one storage account -> File Shares. 2. Create one file share -> Upload one empty folder. 3. Check the result. **Expect Experience:** The uploaded empty folder shows. **Actual Experience:** 1. The uploaded empty folder doesn't show. 2. The activity log shows 0 completed. ![image](https://user-images.githubusercontent.com/41351993/70612883-b054fb00-1c42-11ea-9c30-a7979021c91f.png) **More Info:** Succeed to download one empty folder. ![image](https://user-images.githubusercontent.com/41351993/70610549-bb0d9100-1c3e-11ea-94d4-cb87d8d0db9c.png)
test
fail to upload an empty folder to one file share storage explorer version build branch master platform os windows linux ubuntu macos high sierra architecture regression from not a regression steps to reproduce expand one storage account file shares create one file share upload one empty folder check the result expect experience the uploaded empty folder shows actual experience the uploaded empty folder doesn t show the activity log shows completed more info succeed to download one empty folder
1
319,202
27,355,906,736
IssuesEvent
2023-02-27 12:51:48
QubesOS/updates-status
https://api.github.com/repos/QubesOS/updates-status
closed
desktop-linux-menu v1.0.2 (r4.2)
r4.2-host-cur-test r4.2-vm-bullseye-cur-test r4.2-vm-bookworm-cur-test r4.2-vm-fc37-cur-test r4.2-vm-fc36-cur-test r4.2-vm-centos-stream8-cur-test
Update of desktop-linux-menu to v1.0.2 for Qubes r4.2, see comments below for details and build status. From commit: https://github.com/QubesOS/qubes-desktop-linux-menu/commit/d47acae5d80ae6bbac4b6c111281c7ed243bfe40 [Changes since previous version](https://github.com/QubesOS/qubes-desktop-linux-menu/compare/v1.0.1...v1.0.2): QubesOS/qubes-desktop-linux-menu@d47acae version 1.0.2 QubesOS/qubes-desktop-linux-menu@ab9c386 Merge remote-tracking branch 'origin/pr/24' QubesOS/qubes-desktop-linux-menu@bcbb629 Try to wrangle menu into having correct env vars set QubesOS/qubes-desktop-linux-menu@78983ec Merge remote-tracking branch 'origin/pr/23' QubesOS/qubes-desktop-linux-menu@c3f1c80 Merge remote-tracking branch 'origin/pr/22' QubesOS/qubes-desktop-linux-menu@a9471e5 Improve readability on normal DPI systems QubesOS/qubes-desktop-linux-menu@4477562 Set KillMode on menu to process Referenced issues: If you're release manager, you can issue GPG-inline signed command: * `Upload-component r4.2 desktop-linux-menu d47acae5d80ae6bbac4b6c111281c7ed243bfe40 current all` (available 5 days from now) * `Upload-component r4.2 desktop-linux-menu d47acae5d80ae6bbac4b6c111281c7ed243bfe40 security-testing` You can choose subset of distributions like: * `Upload-component r4.2 desktop-linux-menu d47acae5d80ae6bbac4b6c111281c7ed243bfe40 current vm-bookworm,vm-fc37` (available 5 days from now) Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it). For more information on how to test this update, please take a look at https://www.qubes-os.org/doc/testing/#updates.
6.0
desktop-linux-menu v1.0.2 (r4.2) - Update of desktop-linux-menu to v1.0.2 for Qubes r4.2, see comments below for details and build status. From commit: https://github.com/QubesOS/qubes-desktop-linux-menu/commit/d47acae5d80ae6bbac4b6c111281c7ed243bfe40 [Changes since previous version](https://github.com/QubesOS/qubes-desktop-linux-menu/compare/v1.0.1...v1.0.2): QubesOS/qubes-desktop-linux-menu@d47acae version 1.0.2 QubesOS/qubes-desktop-linux-menu@ab9c386 Merge remote-tracking branch 'origin/pr/24' QubesOS/qubes-desktop-linux-menu@bcbb629 Try to wrangle menu into having correct env vars set QubesOS/qubes-desktop-linux-menu@78983ec Merge remote-tracking branch 'origin/pr/23' QubesOS/qubes-desktop-linux-menu@c3f1c80 Merge remote-tracking branch 'origin/pr/22' QubesOS/qubes-desktop-linux-menu@a9471e5 Improve readability on normal DPI systems QubesOS/qubes-desktop-linux-menu@4477562 Set KillMode on menu to process Referenced issues: If you're release manager, you can issue GPG-inline signed command: * `Upload-component r4.2 desktop-linux-menu d47acae5d80ae6bbac4b6c111281c7ed243bfe40 current all` (available 5 days from now) * `Upload-component r4.2 desktop-linux-menu d47acae5d80ae6bbac4b6c111281c7ed243bfe40 security-testing` You can choose subset of distributions like: * `Upload-component r4.2 desktop-linux-menu d47acae5d80ae6bbac4b6c111281c7ed243bfe40 current vm-bookworm,vm-fc37` (available 5 days from now) Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it). For more information on how to test this update, please take a look at https://www.qubes-os.org/doc/testing/#updates.
test
desktop linux menu update of desktop linux menu to for qubes see comments below for details and build status from commit qubesos qubes desktop linux menu version qubesos qubes desktop linux menu merge remote tracking branch origin pr qubesos qubes desktop linux menu try to wrangle menu into having correct env vars set qubesos qubes desktop linux menu merge remote tracking branch origin pr qubesos qubes desktop linux menu merge remote tracking branch origin pr qubesos qubes desktop linux menu improve readability on normal dpi systems qubesos qubes desktop linux menu set killmode on menu to process referenced issues if you re release manager you can issue gpg inline signed command upload component desktop linux menu current all available days from now upload component desktop linux menu security testing you can choose subset of distributions like upload component desktop linux menu current vm bookworm vm available days from now above commands will work only if packages in current testing repository were built from given commit i e no new version superseded it for more information on how to test this update please take a look at
1
155,830
12,279,239,728
IssuesEvent
2020-05-08 11:45:43
ufcg-lsd/saps-engine
https://api.github.com/repos/ufcg-lsd/saps-engine
opened
Perform endtoend testing with real algorithms
test
The entire environment was created to perform the endtoend test described in [endtoend doc](/docs/end-to-end-test.md#end-to-end-test-using-real-algorithms), but the current deployment of test does not have workers with sufficient capacity to result in successful processing. It must be done after allocating workers with the minimum settings of 16GB ram and 2 vcpus with 50GB host size.
1.0
Perform endtoend testing with real algorithms - The entire environment was created to perform the endtoend test described in [endtoend doc](/docs/end-to-end-test.md#end-to-end-test-using-real-algorithms), but the current deployment of test does not have workers with sufficient capacity to result in successful processing. It must be done after allocating workers with the minimum settings of 16GB ram and 2 vcpus with 50GB host size.
test
perform endtoend testing with real algorithms the entire environment was created to perform the endtoend test described in docs end to end test md end to end test using real algorithms but the current deployment of test does not have workers with sufficient capacity to result in successful processing it must be done after allocating workers with the minimum settings of ram and vcpus with host size
1
3,991
2,699,306,245
IssuesEvent
2015-04-03 16:00:29
redmatrix/redmatrix
https://api.github.com/repos/redmatrix/redmatrix
closed
Events page shows code at the top
fix-applied-needs-test
_From @HaakonME on May 9, 2014 20:20_ Revision: 7302a36 option:selected').each( function() { selstr = $(this).text(); $('#jot-public').hide(); }); if(selstr == null) { $('#jot-public').show(); } }).trigger('change'); }); _Copied from original issue: friendica/red#452_
1.0
Events page shows code at the top - _From @HaakonME on May 9, 2014 20:20_ Revision: 7302a36 option:selected').each( function() { selstr = $(this).text(); $('#jot-public').hide(); }); if(selstr == null) { $('#jot-public').show(); } }).trigger('change'); }); _Copied from original issue: friendica/red#452_
test
events page shows code at the top from haakonme on may revision option selected each function selstr this text jot public hide if selstr null jot public show trigger change copied from original issue friendica red
1
324,128
9,884,348,860
IssuesEvent
2019-06-24 21:52:14
jenkins-x/jx
https://api.github.com/repos/jenkins-x/jx
closed
Jenkins X stuck on installing jenkins-x-platform
priority/awaiting-more-evidence
### Summary I have tried to create a new kubernetes cluster and split the create and install process into two steps, but both fail with the following error: error: installing the Jenkins X platform: failed to wait for jenkins-x chart installation to be ready: Unauthorized ### Steps to reproduce the behavior Method 1: $jx create cluster eks --cluster-name jx-eks --region us-east-2 --node-type t2.small Method 2: $jx create cluster eks --cluster-name jx-eks --region us-east-2 --node-type=t2.small --skip-installation $jx install --provider=eks --cluster-name jx-eks --no-tiller ### Expected behavior 1. EKS cluster is created 2. Jenkins Platform created ### Actual behavior 1. EKS cluster is created 2. Jenkins Platform installation hangs and eventually produces the following error: error: installing the Jenkins X platform: failed to wait for jenkins-x chart installation to be ready: Unauthorized ### Jx version The output of `jx version` is: NAME VERSION jx 2.0.317 jenkins x platform 2.0.672 Kubernetes cluster v1.12.6-eks-d69f1b kubectl v1.14.0 helm client Client: v2.14.1+g5270352 git git version 2.21.0.windows.1 Operating System Windows 10 Home 1809 build 17763 ``` COPY OUTPUT HERE ``` ### Jenkins type <!-- Select which installation type are you using. --> - [ X] Next Generation (Tekton + Prow) - [ ] Classic Jenkins - [ ] Serverless Jenkins (JenkinsFileRunner + Prow) ### Kubernetes cluster <!-- What kind of Kubernetes cluster are you using & how did you create it? --> I created the EKS cluster with the following command via jx: $jx create cluster eks --cluster-name jx-eks --region us-east-2 --node-type t2.small ### Operating system / Environment <!-- In which environment are you running the jx CLI? --> Windows 10 Home
1.0
Jenkins X stuck on installing jenkins-x-platform - ### Summary I have tried to create a new kubernetes cluster and split the create and install process into two steps, but both fail with the following error: error: installing the Jenkins X platform: failed to wait for jenkins-x chart installation to be ready: Unauthorized ### Steps to reproduce the behavior Method 1: $jx create cluster eks --cluster-name jx-eks --region us-east-2 --node-type t2.small Method 2: $jx create cluster eks --cluster-name jx-eks --region us-east-2 --node-type=t2.small --skip-installation $jx install --provider=eks --cluster-name jx-eks --no-tiller ### Expected behavior 1. EKS cluster is created 2. Jenkins Platform created ### Actual behavior 1. EKS cluster is created 2. Jenkins Platform installation hangs and eventually produces the following error: error: installing the Jenkins X platform: failed to wait for jenkins-x chart installation to be ready: Unauthorized ### Jx version The output of `jx version` is: NAME VERSION jx 2.0.317 jenkins x platform 2.0.672 Kubernetes cluster v1.12.6-eks-d69f1b kubectl v1.14.0 helm client Client: v2.14.1+g5270352 git git version 2.21.0.windows.1 Operating System Windows 10 Home 1809 build 17763 ``` COPY OUTPUT HERE ``` ### Jenkins type <!-- Select which installation type are you using. --> - [ X] Next Generation (Tekton + Prow) - [ ] Classic Jenkins - [ ] Serverless Jenkins (JenkinsFileRunner + Prow) ### Kubernetes cluster <!-- What kind of Kubernetes cluster are you using & how did you create it? --> I created the EKS cluster with the following command via jx: $jx create cluster eks --cluster-name jx-eks --region us-east-2 --node-type t2.small ### Operating system / Environment <!-- In which environment are you running the jx CLI? --> Windows 10 Home
non_test
jenkins x stuck on installing jenkins x platform summary i have tried to create a new kubernetes cluster and split the create and install process into two steps but both fail with the following error error installing the jenkins x platform failed to wait for jenkins x chart installation to be ready unauthorized steps to reproduce the behavior method jx create cluster eks cluster name jx eks region us east node type small method jx create cluster eks cluster name jx eks region us east node type small skip installation jx install provider eks cluster name jx eks no tiller expected behavior eks cluster is created jenkins platform created actual behavior eks cluster is created jenkins platform installation hangs and eventually produces the following error error installing the jenkins x platform failed to wait for jenkins x chart installation to be ready unauthorized jx version the output of jx version is name version jx jenkins x platform kubernetes cluster eks kubectl helm client client git git version windows operating system windows home build copy output here jenkins type select which installation type are you using next generation tekton prow classic jenkins serverless jenkins jenkinsfilerunner prow kubernetes cluster what kind of kubernetes cluster are you using how did you create it i created the eks cluster with the following command via jx jx create cluster eks cluster name jx eks region us east node type small operating system environment in which environment are you running the jx cli windows home
0
231,126
18,740,354,588
IssuesEvent
2021-11-04 12:56:54
Mexahoid/ESTC-project
https://api.github.com/repos/Mexahoid/ESTC-project
reopened
Use GPIO to work with button
intest implemented
The task is: In your current implementation, you always iterate through all LEDs. Now change this logic. Iterate only when a button is pressed. If the button is pressed, released, and pressed again continue from the previous active LED. Example for the sequence: RRGGGB Button released => nothing Button pressed for a 3 sec => RRG Button released => nothing Button pressed for a 2 sec => GG Button released => nothing Button pressed for a g sec => BRRGG Use GPIO HAL (modules/nrfx/hal/nrf_gpio.h). Useful functions and macro: NRF_GPIO_PIN_MAP nrf_gpio_cfg_input nrf_gpio_pin_read
1.0
Use GPIO to work with button - The task is: In your current implementation, you always iterate through all LEDs. Now change this logic. Iterate only when a button is pressed. If the button is pressed, released, and pressed again continue from the previous active LED. Example for the sequence: RRGGGB Button released => nothing Button pressed for a 3 sec => RRG Button released => nothing Button pressed for a 2 sec => GG Button released => nothing Button pressed for a g sec => BRRGG Use GPIO HAL (modules/nrfx/hal/nrf_gpio.h). Useful functions and macro: NRF_GPIO_PIN_MAP nrf_gpio_cfg_input nrf_gpio_pin_read
test
use gpio to work with button the task is in your current implementation you always iterate through all leds now change this logic iterate only when a button is pressed if the button is pressed released and pressed again continue from the previous active led example for the sequence rrgggb button released nothing button pressed for a sec rrg button released nothing button pressed for a sec gg button released nothing button pressed for a g sec brrgg use gpio hal modules nrfx hal nrf gpio h useful functions and macro nrf gpio pin map nrf gpio cfg input nrf gpio pin read
1
201,097
15,801,905,970
IssuesEvent
2021-04-03 07:05:48
litone01/ped
https://api.github.com/repos/litone01/ped
opened
[UG] Find command is by name or?
severity.Low type.DocumentationBug
In the UG, 'find' is locating persons by name, however, in the content, it can be searching for other fields as well. This may be a typo. In the UG: ![image.png](https://raw.githubusercontent.com/litone01/ped/main/files/ea6e93ef-08d9-4773-8f28-9c45d4566b91.png) ![image.png](https://raw.githubusercontent.com/litone01/ped/main/files/cc68aee9-3fda-49cc-8c1e-4f9dbdd5f2c1.png) <!--session: 1617429541871-a01e6ec5-0ac4-4b7e-978d-0edd3f0e9e2f-->
1.0
[UG] Find command is by name or? - In the UG, 'find' is locating persons by name, however, in the content, it can be searching for other fields as well. This may be a typo. In the UG: ![image.png](https://raw.githubusercontent.com/litone01/ped/main/files/ea6e93ef-08d9-4773-8f28-9c45d4566b91.png) ![image.png](https://raw.githubusercontent.com/litone01/ped/main/files/cc68aee9-3fda-49cc-8c1e-4f9dbdd5f2c1.png) <!--session: 1617429541871-a01e6ec5-0ac4-4b7e-978d-0edd3f0e9e2f-->
non_test
find command is by name or in the ug find is locating persons by name however in the content it can be searching for other fields as well this may be a typo in the ug
0
369,470
25,850,315,613
IssuesEvent
2022-12-13 09:56:53
maxfischer2781/asyncstdlib
https://api.github.com/repos/maxfischer2781/asyncstdlib
closed
clearing cache before it exists raises AttributeError
documentation
Where `self.site_map` is a `@cached_property`, I use the following to work around the issue: ```python def reset_cache(self) -> None: if "site_map" in self.__dict__: del self.__dict__["site_map"] ``` Perhaps some checks could be added in the library so this isn't necessary in my code? FYI Calling `del self.site_map` gives me a PyCharm linter error that I can override with: `# noinspection PyPropertyAccess`. The code above works around both the linter error and the AttributeError issues. Perhaps providing this alternate approach to clear cache could be provided in the documentation? The storage of the cached value in `self.__dict__` is an implementation detail that my code shouldn't know about and may change in the future. I am thus of the strong belief that this should be done upstream (in the library) so that this implementation detail need not be coupled in downstream code. Thanks for the great library. I love it and use it often.
1.0
clearing cache before it exists raises AttributeError - Where `self.site_map` is a `@cached_property`, I use the following to work around the issue: ```python def reset_cache(self) -> None: if "site_map" in self.__dict__: del self.__dict__["site_map"] ``` Perhaps some checks could be added in the library so this isn't necessary in my code? FYI Calling `del self.site_map` gives me a PyCharm linter error that I can override with: `# noinspection PyPropertyAccess`. The code above works around both the linter error and the AttributeError issues. Perhaps providing this alternate approach to clear cache could be provided in the documentation? The storage of the cached value in `self.__dict__` is an implementation detail that my code shouldn't know about and may change in the future. I am thus of the strong belief that this should be done upstream (in the library) so that this implementation detail need not be coupled in downstream code. Thanks for the great library. I love it and use it often.
non_test
clearing cache before it exists raises attributeerror where self site map is a cached property i use the following to work around the issue python def reset cache self none if site map in self dict del self dict perhaps some checks could be added in the library so this isn t necessary in my code fyi calling del self site map gives me a pycharm linter error that i can override with noinspection pypropertyaccess the code above works around both the linter error and the attributeerror issues perhaps providing this alternate approach to clear cache could be provided in the documentation the storage of the cached value in self dict is an implementation detail that my code shouldn t know about and may change in the future i am thus of the strong belief that this should be done upstream in the library so that this implementation detail need not be coupled in downstream code thanks for the great library i love it and use it often
0
179,622
13,891,591,017
IssuesEvent
2020-10-19 10:55:13
SEAI-H-2020/Software
https://api.github.com/repos/SEAI-H-2020/Software
opened
Test deep sleep mode for hw team
High Priority Testing
- Minimum time to wake up, connect to network, send a message and sleep again
1.0
Test deep sleep mode for hw team - - Minimum time to wake up, connect to network, send a message and sleep again
test
test deep sleep mode for hw team minimum time to wake up connect to network send a message and sleep again
1
293,162
25,274,764,684
IssuesEvent
2022-11-16 11:50:57
epiphany-platform/epiphany
https://api.github.com/repos/epiphany-platform/epiphany
closed
[FEATURE REQUEST] Verify components backup/recovery data
area/testing
**Is your feature request related to a problem? Please describe.** We have pipelines for running the backup and recovery of components, however we don't verify the data after running this. **Describe the solution you'd like** Add tests to verify the data after running the backup and recovery of components. **Describe alternatives you've considered** No alternatives. **Additional context** No additional context. --- **DoD checklist** - Changelog - [ ] updated - [ ] not needed - COMPONENTS.md - [ ] updated - [ ] not needed - Schema - [ ] updated - [ ] not needed - Backport tasks - [ ] created - [ ] not needed - Documentation - [ ] added - [ ] updated - [ ] not needed - [ ] Feature has automated tests - [ ] Automated tests passed (QA pipelines) - [ ] apply - [ ] upgrade - [ ] backup/restore - [ ] Idempotency tested - [ ] All conversations in PR resolved - [ ] Solution meets requirements and is done according to design doc - [ ] Usage compliant with license
1.0
[FEATURE REQUEST] Verify components backup/recovery data - **Is your feature request related to a problem? Please describe.** We have pipelines for running the backup and recovery of components, however we don't verify the data after running this. **Describe the solution you'd like** Add tests to verify the data after running the backup and recovery of components. **Describe alternatives you've considered** No alternatives. **Additional context** No additional context. --- **DoD checklist** - Changelog - [ ] updated - [ ] not needed - COMPONENTS.md - [ ] updated - [ ] not needed - Schema - [ ] updated - [ ] not needed - Backport tasks - [ ] created - [ ] not needed - Documentation - [ ] added - [ ] updated - [ ] not needed - [ ] Feature has automated tests - [ ] Automated tests passed (QA pipelines) - [ ] apply - [ ] upgrade - [ ] backup/restore - [ ] Idempotency tested - [ ] All conversations in PR resolved - [ ] Solution meets requirements and is done according to design doc - [ ] Usage compliant with license
test
verify components backup recovery data is your feature request related to a problem please describe we have pipelines for running the backup and recovery of components however we don t verify the data after running this describe the solution you d like add tests to verify the data after running the backup and recovery of components describe alternatives you ve considered no alternatives additional context no additional context dod checklist changelog updated not needed components md updated not needed schema updated not needed backport tasks created not needed documentation added updated not needed feature has automated tests automated tests passed qa pipelines apply upgrade backup restore idempotency tested all conversations in pr resolved solution meets requirements and is done according to design doc usage compliant with license
1
101,868
8,806,555,202
IssuesEvent
2018-12-27 04:51:37
drussell1974/schemeofwork_web2py_app
https://api.github.com/repos/drussell1974/schemeofwork_web2py_app
closed
Selenium - Create new LearningEpisode
test
**Create New** _check page elements_ - [x] title and headings - [x] navigation _edit_ - [x] submit invalid - [x] submit valid
1.0
Selenium - Create new LearningEpisode - **Create New** _check page elements_ - [x] title and headings - [x] navigation _edit_ - [x] submit invalid - [x] submit valid
test
selenium create new learningepisode create new check page elements title and headings navigation edit submit invalid submit valid
1
277,452
24,073,198,434
IssuesEvent
2022-09-18 13:01:03
GiveUsMoney/FanFixiv-auth
https://api.github.com/repos/GiveUsMoney/FanFixiv-auth
opened
로그인, 회원가입 e2e 테스트 작성
B-E TEST
`/login`, `/register` API의 e2e 테스트를 작성합니다. 인증과 밀접한 관련이 있는 API므로 신중하게 작성합시다. 만약 API상의 문제점을 발견한다면 @Roharui 에게 DM을 날려주세요. - [ ] Github Secret 등록 - [ ] h2 DB 환경 세팅 - [ ] 테스트시 h2 연결 로직 설정 > 회원가입 이후 로그인 테스트를 진행합니다. **/register** - [ ] 잘못된 값 입력시 에러 호출 확인 - [ ] 값 누락시 에러 호출 확인 - [ ] 정상적인 값 입력시 DB에 입력되는 값 확인 **/login** - [ ] 잘못된 값 입력시 에러 호출 확인 - [ ] 값 누락시 에러 호출 확인 - [ ] 정상적인 값 입력시 200 호출
1.0
로그인, 회원가입 e2e 테스트 작성 - `/login`, `/register` API의 e2e 테스트를 작성합니다. 인증과 밀접한 관련이 있는 API므로 신중하게 작성합시다. 만약 API상의 문제점을 발견한다면 @Roharui 에게 DM을 날려주세요. - [ ] Github Secret 등록 - [ ] h2 DB 환경 세팅 - [ ] 테스트시 h2 연결 로직 설정 > 회원가입 이후 로그인 테스트를 진행합니다. **/register** - [ ] 잘못된 값 입력시 에러 호출 확인 - [ ] 값 누락시 에러 호출 확인 - [ ] 정상적인 값 입력시 DB에 입력되는 값 확인 **/login** - [ ] 잘못된 값 입력시 에러 호출 확인 - [ ] 값 누락시 에러 호출 확인 - [ ] 정상적인 값 입력시 200 호출
test
로그인 회원가입 테스트 작성 login register api의 테스트를 작성합니다 인증과 밀접한 관련이 있는 api므로 신중하게 작성합시다 만약 api상의 문제점을 발견한다면 roharui 에게 dm을 날려주세요 github secret 등록 db 환경 세팅 테스트시 연결 로직 설정 회원가입 이후 로그인 테스트를 진행합니다 register 잘못된 값 입력시 에러 호출 확인 값 누락시 에러 호출 확인 정상적인 값 입력시 db에 입력되는 값 확인 login 잘못된 값 입력시 에러 호출 확인 값 누락시 에러 호출 확인 정상적인 값 입력시 호출
1
168,226
20,747,153,053
IssuesEvent
2022-03-15 01:06:22
Tim-sandbox/WebGoat-8.1
https://api.github.com/repos/Tim-sandbox/WebGoat-8.1
opened
CVE-2020-36518 (High) detected in jackson-databind-2.12.4.jar
security vulnerability
## CVE-2020-36518 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.12.4.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /webgoat-lessons/jwt/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.12.4/jackson-databind-2.12.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.12.4/jackson-databind-2.12.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.12.4/jackson-databind-2.12.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.12.4/jackson-databind-2.12.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.12.4/jackson-databind-2.12.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.12.4/jackson-databind-2.12.4.jar</p> <p> Dependency Hierarchy: - jjwt-0.9.1.jar (Root Library) - :x: **jackson-databind-2.12.4.jar** (Vulnerable Library) <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jackson-databind before 2.13.0 allows a Java StackOverflow exception and denial of service via a large depth of nested objects. WhiteSource Note: After conducting further research, WhiteSource has determined that all versions of com.fasterxml.jackson.core:jackson-databind up to version 2.13.2 are vulnerable to CVE-2020-36518. <p>Publish Date: 2022-03-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36518>CVE-2020-36518</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-36518">https://nvd.nist.gov/vuln/detail/CVE-2020-36518</a></p> <p>Release Date: 2022-03-11</p> <p>Fix Resolution: jackson-databind-2.10 - 2.10.1;com.fasterxml.jackson.core.jackson-databind - 2.6.2.v20161117-2150</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.12.4","packageFilePaths":["/webgoat-lessons/jwt/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"io.jsonwebtoken:jjwt:0.9.1;com.fasterxml.jackson.core:jackson-databind:2.12.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jackson-databind-2.10 - 2.10.1;com.fasterxml.jackson.core.jackson-databind - 2.6.2.v20161117-2150","isBinary":false}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2020-36518","vulnerabilityDetails":"jackson-databind before 2.13.0 allows a Java StackOverflow exception and denial of service via a large depth of nested objects.\n WhiteSource Note: After conducting further research, WhiteSource has determined that all versions of com.fasterxml.jackson.core:jackson-databind up to version 2.13.2 are vulnerable to CVE-2020-36518.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36518","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-36518 (High) detected in jackson-databind-2.12.4.jar - ## CVE-2020-36518 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.12.4.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /webgoat-lessons/jwt/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.12.4/jackson-databind-2.12.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.12.4/jackson-databind-2.12.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.12.4/jackson-databind-2.12.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.12.4/jackson-databind-2.12.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.12.4/jackson-databind-2.12.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.12.4/jackson-databind-2.12.4.jar</p> <p> Dependency Hierarchy: - jjwt-0.9.1.jar (Root Library) - :x: **jackson-databind-2.12.4.jar** (Vulnerable Library) <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jackson-databind before 2.13.0 allows a Java StackOverflow exception and denial of service via a large depth of nested objects. WhiteSource Note: After conducting further research, WhiteSource has determined that all versions of com.fasterxml.jackson.core:jackson-databind up to version 2.13.2 are vulnerable to CVE-2020-36518. <p>Publish Date: 2022-03-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36518>CVE-2020-36518</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-36518">https://nvd.nist.gov/vuln/detail/CVE-2020-36518</a></p> <p>Release Date: 2022-03-11</p> <p>Fix Resolution: jackson-databind-2.10 - 2.10.1;com.fasterxml.jackson.core.jackson-databind - 2.6.2.v20161117-2150</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.12.4","packageFilePaths":["/webgoat-lessons/jwt/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"io.jsonwebtoken:jjwt:0.9.1;com.fasterxml.jackson.core:jackson-databind:2.12.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jackson-databind-2.10 - 2.10.1;com.fasterxml.jackson.core.jackson-databind - 2.6.2.v20161117-2150","isBinary":false}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2020-36518","vulnerabilityDetails":"jackson-databind before 2.13.0 allows a Java StackOverflow exception and denial of service via a large depth of nested objects.\n WhiteSource Note: After conducting further research, WhiteSource has determined that all versions of com.fasterxml.jackson.core:jackson-databind up to version 2.13.2 are vulnerable to CVE-2020-36518.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36518","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_test
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file webgoat lessons jwt pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy jjwt jar root library x jackson databind jar vulnerable library found in base branch develop vulnerability details jackson databind before allows a java stackoverflow exception and denial of service via a large depth of nested objects whitesource note after conducting further research whitesource has determined that all versions of com fasterxml jackson core jackson databind up to version are vulnerable to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jackson databind com fasterxml jackson core jackson databind isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree io jsonwebtoken jjwt com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion jackson databind com fasterxml jackson core jackson databind isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails jackson databind before allows a java stackoverflow exception and denial of service via a large depth of nested objects n whitesource note after conducting further research whitesource has determined that all versions of com fasterxml jackson core jackson databind up to version are vulnerable to cve vulnerabilityurl
0
654,609
21,657,722,260
IssuesEvent
2022-05-06 15:40:28
googleapis/python-aiplatform
https://api.github.com/repos/googleapis/python-aiplatform
closed
tests.system.aiplatform.test_featurestore.TestFeaturestore: test_batch_serve_to_bq failed
type: bug priority: p1 flakybot: issue api: vertex-ai
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 5e5aa558b65de102c39ba9c8e3dbdf6444c05071 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/6f5d6d39-780b-478d-8204-eb3ba9f06069), [Sponge](http://sponge2/6f5d6d39-780b-478d-8204-eb3ba9f06069) status: failed <details><summary>Test output</summary><br><pre>self = <tests.system.aiplatform.test_featurestore.TestFeaturestore object at 0x7f09bd29e760> shared_state = {'bigquery_client': <google.cloud.bigquery.client.Client object at 0x7f09da6e6550>, 'bigquery_dataset': Dataset(Datase...2549_46b2_ac57_e14319b0eb73', 'bucket': <Bucket: temp_vertex_sdk_e2e_featurestore_test-638afb3f-61ae-47e8-bcf2-0>, ...} caplog = <_pytest.logging.LogCaptureFixture object at 0x7f09d251e700> def test_batch_serve_to_bq(self, shared_state, caplog): assert shared_state["featurestore"] assert shared_state["bigquery_dataset"] assert shared_state["user_age_feature_resource_name"] assert shared_state["user_gender_feature_resource_name"] assert shared_state["user_liked_genres_feature_resource_name"] featurestore = shared_state["featurestore"] bigquery_dataset_id = shared_state["bigquery_dataset_id"] user_age_feature_resource_name = shared_state["user_age_feature_resource_name"] user_gender_feature_resource_name = shared_state[ "user_gender_feature_resource_name" ] user_liked_genres_feature_resource_name = shared_state[ "user_liked_genres_feature_resource_name" ] aiplatform.init( project=e2e_base._PROJECT, location=e2e_base._LOCATION, ) caplog.set_level(logging.INFO) > featurestore.batch_serve_to_bq( serving_feature_ids={ _TEST_USER_ENTITY_TYPE_ID: [ _TEST_USER_AGE_FEATURE_ID, _TEST_USER_GENDER_FEATURE_ID, _TEST_USER_LIKED_GENRES_FEATURE_ID, ], _TEST_MOVIE_ENTITY_TYPE_ID: [ _TEST_MOVIE_TITLE_FEATURE_ID, _TEST_MOVIE_GENRES_FEATURE_ID, _TEST_MOVIE_AVERAGE_RATING_FEATURE_ID, ], }, read_instances_uri=_TEST_READ_INSTANCE_SRC, feature_destination_fields={ user_age_feature_resource_name: "user_age_dest", user_gender_feature_resource_name: "user_gender_dest", user_liked_genres_feature_resource_name: "user_liked_genres_dest", }, bq_destination_output_uri=f"bq://{bigquery_dataset_id}.test_table", ) tests/system/aiplatform/test_featurestore.py:599: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ google/cloud/aiplatform/base.py:750: in wrapper return method(*args, **kwargs) google/cloud/aiplatform/featurestore/featurestore.py:930: in batch_serve_to_bq return self._batch_read_feature_values( google/cloud/aiplatform/featurestore/featurestore.py:647: in _batch_read_feature_values batch_read_lro.result() _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <google.api_core.operation.Operation object at 0x7f09d251a0a0> timeout = None, retry = <google.api_core.retry.Retry object at 0x7f09be630b20> def result(self, timeout=None, retry=DEFAULT_RETRY): """Get the result of the operation, blocking if necessary. Args: timeout (int): How long (in seconds) to wait for the operation to complete. If None, wait indefinitely. Returns: google.protobuf.Message: The Operation's result. Raises: google.api_core.GoogleAPICallError: If the operation errors or if the timeout is reached before the operation completes. """ kwargs = {} if retry is DEFAULT_RETRY else {"retry": retry} self._blocking_poll(timeout=timeout, **kwargs) if self._exception is not None: # pylint: disable=raising-bad-type # Pylint doesn't recognize that this is valid in this case. > raise self._exception E google.api_core.exceptions.InternalServerError: 500 Failed to process query file: gs://cloud-samples-data-us-central1/vertex-ai/feature-store/datasets/movie_prediction.csv. An internal error occurred and the request could not be completed. This is usually caused by a transient issue. Retrying the job with back-off as described in the BigQuery SLA should solve the problem: https://cloud.google.com/bigquery/sla. If the error continues to occur please contact support at https://cloud.google.com/support. Error: 3750392 .nox/system-3-8/lib/python3.8/site-packages/google/api_core/future/polling.py:137: InternalServerError</pre></details>
1.0
tests.system.aiplatform.test_featurestore.TestFeaturestore: test_batch_serve_to_bq failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 5e5aa558b65de102c39ba9c8e3dbdf6444c05071 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/6f5d6d39-780b-478d-8204-eb3ba9f06069), [Sponge](http://sponge2/6f5d6d39-780b-478d-8204-eb3ba9f06069) status: failed <details><summary>Test output</summary><br><pre>self = <tests.system.aiplatform.test_featurestore.TestFeaturestore object at 0x7f09bd29e760> shared_state = {'bigquery_client': <google.cloud.bigquery.client.Client object at 0x7f09da6e6550>, 'bigquery_dataset': Dataset(Datase...2549_46b2_ac57_e14319b0eb73', 'bucket': <Bucket: temp_vertex_sdk_e2e_featurestore_test-638afb3f-61ae-47e8-bcf2-0>, ...} caplog = <_pytest.logging.LogCaptureFixture object at 0x7f09d251e700> def test_batch_serve_to_bq(self, shared_state, caplog): assert shared_state["featurestore"] assert shared_state["bigquery_dataset"] assert shared_state["user_age_feature_resource_name"] assert shared_state["user_gender_feature_resource_name"] assert shared_state["user_liked_genres_feature_resource_name"] featurestore = shared_state["featurestore"] bigquery_dataset_id = shared_state["bigquery_dataset_id"] user_age_feature_resource_name = shared_state["user_age_feature_resource_name"] user_gender_feature_resource_name = shared_state[ "user_gender_feature_resource_name" ] user_liked_genres_feature_resource_name = shared_state[ "user_liked_genres_feature_resource_name" ] aiplatform.init( project=e2e_base._PROJECT, location=e2e_base._LOCATION, ) caplog.set_level(logging.INFO) > featurestore.batch_serve_to_bq( serving_feature_ids={ _TEST_USER_ENTITY_TYPE_ID: [ _TEST_USER_AGE_FEATURE_ID, _TEST_USER_GENDER_FEATURE_ID, _TEST_USER_LIKED_GENRES_FEATURE_ID, ], _TEST_MOVIE_ENTITY_TYPE_ID: [ _TEST_MOVIE_TITLE_FEATURE_ID, _TEST_MOVIE_GENRES_FEATURE_ID, _TEST_MOVIE_AVERAGE_RATING_FEATURE_ID, ], }, read_instances_uri=_TEST_READ_INSTANCE_SRC, feature_destination_fields={ user_age_feature_resource_name: "user_age_dest", user_gender_feature_resource_name: "user_gender_dest", user_liked_genres_feature_resource_name: "user_liked_genres_dest", }, bq_destination_output_uri=f"bq://{bigquery_dataset_id}.test_table", ) tests/system/aiplatform/test_featurestore.py:599: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ google/cloud/aiplatform/base.py:750: in wrapper return method(*args, **kwargs) google/cloud/aiplatform/featurestore/featurestore.py:930: in batch_serve_to_bq return self._batch_read_feature_values( google/cloud/aiplatform/featurestore/featurestore.py:647: in _batch_read_feature_values batch_read_lro.result() _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <google.api_core.operation.Operation object at 0x7f09d251a0a0> timeout = None, retry = <google.api_core.retry.Retry object at 0x7f09be630b20> def result(self, timeout=None, retry=DEFAULT_RETRY): """Get the result of the operation, blocking if necessary. Args: timeout (int): How long (in seconds) to wait for the operation to complete. If None, wait indefinitely. Returns: google.protobuf.Message: The Operation's result. Raises: google.api_core.GoogleAPICallError: If the operation errors or if the timeout is reached before the operation completes. """ kwargs = {} if retry is DEFAULT_RETRY else {"retry": retry} self._blocking_poll(timeout=timeout, **kwargs) if self._exception is not None: # pylint: disable=raising-bad-type # Pylint doesn't recognize that this is valid in this case. > raise self._exception E google.api_core.exceptions.InternalServerError: 500 Failed to process query file: gs://cloud-samples-data-us-central1/vertex-ai/feature-store/datasets/movie_prediction.csv. An internal error occurred and the request could not be completed. This is usually caused by a transient issue. Retrying the job with back-off as described in the BigQuery SLA should solve the problem: https://cloud.google.com/bigquery/sla. If the error continues to occur please contact support at https://cloud.google.com/support. Error: 3750392 .nox/system-3-8/lib/python3.8/site-packages/google/api_core/future/polling.py:137: InternalServerError</pre></details>
non_test
tests system aiplatform test featurestore testfeaturestore test batch serve to bq failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output self shared state bigquery client bigquery dataset dataset datase bucket caplog def test batch serve to bq self shared state caplog assert shared state assert shared state assert shared state assert shared state assert shared state featurestore shared state bigquery dataset id shared state user age feature resource name shared state user gender feature resource name shared state user gender feature resource name user liked genres feature resource name shared state user liked genres feature resource name aiplatform init project base project location base location caplog set level logging info featurestore batch serve to bq serving feature ids test user entity type id test user age feature id test user gender feature id test user liked genres feature id test movie entity type id test movie title feature id test movie genres feature id test movie average rating feature id read instances uri test read instance src feature destination fields user age feature resource name user age dest user gender feature resource name user gender dest user liked genres feature resource name user liked genres dest bq destination output uri f bq bigquery dataset id test table tests system aiplatform test featurestore py google cloud aiplatform base py in wrapper return method args kwargs google cloud aiplatform featurestore featurestore py in batch serve to bq return self batch read feature values google cloud aiplatform featurestore featurestore py in batch read feature values batch read lro result self timeout none retry def result self timeout none retry default retry get the result of the operation blocking if necessary args timeout int how long in seconds to wait for the operation to complete if none wait indefinitely returns google protobuf message the operation s result raises google api core googleapicallerror if the operation errors or if the timeout is reached before the operation completes kwargs if retry is default retry else retry retry self blocking poll timeout timeout kwargs if self exception is not none pylint disable raising bad type pylint doesn t recognize that this is valid in this case raise self exception e google api core exceptions internalservererror failed to process query file gs cloud samples data us vertex ai feature store datasets movie prediction csv an internal error occurred and the request could not be completed this is usually caused by a transient issue retrying the job with back off as described in the bigquery sla should solve the problem if the error continues to occur please contact support at error nox system lib site packages google api core future polling py internalservererror
0
93,371
8,411,289,822
IssuesEvent
2018-10-12 13:29:25
sherpa/sherpa
https://api.github.com/repos/sherpa/sherpa
opened
Repeated test code related to ARF/RMF creation
area:tests
It looks like #482 ended up repeating some tests - e.g. `test_rspmodelnopha_delta_call` is in both `sherpa/astro/tests/test_astro_diag_arfrmf.py` and `sherpa/astro/tests/test_instrument.py`. I haven't checked closely, so there may be differences, but I note that they both fail the same way in https://travis-ci.org/sherpa/sherpa/jobs/440614460#L4148 which suggests that they are the same. I think we should also update `sherpa/astro/tests/test_instrument.py` so that it uses the "top-level" create-an-arf / rmf functions added in #482 rather than having its own version (which predates #482), to reduce code duplication (and maybe improve test coverage).
1.0
Repeated test code related to ARF/RMF creation - It looks like #482 ended up repeating some tests - e.g. `test_rspmodelnopha_delta_call` is in both `sherpa/astro/tests/test_astro_diag_arfrmf.py` and `sherpa/astro/tests/test_instrument.py`. I haven't checked closely, so there may be differences, but I note that they both fail the same way in https://travis-ci.org/sherpa/sherpa/jobs/440614460#L4148 which suggests that they are the same. I think we should also update `sherpa/astro/tests/test_instrument.py` so that it uses the "top-level" create-an-arf / rmf functions added in #482 rather than having its own version (which predates #482), to reduce code duplication (and maybe improve test coverage).
test
repeated test code related to arf rmf creation it looks like ended up repeating some tests e g test rspmodelnopha delta call is in both sherpa astro tests test astro diag arfrmf py and sherpa astro tests test instrument py i haven t checked closely so there may be differences but i note that they both fail the same way in which suggests that they are the same i think we should also update sherpa astro tests test instrument py so that it uses the top level create an arf rmf functions added in rather than having its own version which predates to reduce code duplication and maybe improve test coverage
1
399,428
11,748,088,224
IssuesEvent
2020-03-12 14:40:37
oslc-op/oslc-specs
https://api.github.com/repos/oslc-op/oslc-specs
closed
Reference in old query spec does not exist in core 3.0
Core: Query Priority: Medium Xtra: Jira
In [https://open-services.net/bin/view/Main/OSLCCoreSpecQuery](https://open-services.net/bin/view/Main/OSLCCoreSpecQuery) under Graph Patterns: A Query Capability defines a starting subject resource, namely the base URI itself (see oslc:queryBase in the OSLC Core Specification - Query Capability). However, there is no corresponding section in core 3.0 to reference. The link was to [https://open-services.net/bin/view/Main/OslcCoreSpecification#Query\_Capabilities](https://open-services.net/bin/view/Main/OslcCoreSpecification#Query\_Capabilities). However, there is [https://tools.oasis-open.org/version-control/svn/oslc-core/trunk/specs/discovery.html#queryCapabilityShape](https://tools.oasis-open.org/version-control/svn/oslc-core/trunk/specs/discovery.html#queryCapabilityShape). Should we reference that instead, or as well as some new core section? --- _Migrated from https://issues.oasis-open.org/browse/OSLCCORE-100 (opened by @DavidJHoney; previously assigned to @jamsden)_
1.0
Reference in old query spec does not exist in core 3.0 - In [https://open-services.net/bin/view/Main/OSLCCoreSpecQuery](https://open-services.net/bin/view/Main/OSLCCoreSpecQuery) under Graph Patterns: A Query Capability defines a starting subject resource, namely the base URI itself (see oslc:queryBase in the OSLC Core Specification - Query Capability). However, there is no corresponding section in core 3.0 to reference. The link was to [https://open-services.net/bin/view/Main/OslcCoreSpecification#Query\_Capabilities](https://open-services.net/bin/view/Main/OslcCoreSpecification#Query\_Capabilities). However, there is [https://tools.oasis-open.org/version-control/svn/oslc-core/trunk/specs/discovery.html#queryCapabilityShape](https://tools.oasis-open.org/version-control/svn/oslc-core/trunk/specs/discovery.html#queryCapabilityShape). Should we reference that instead, or as well as some new core section? --- _Migrated from https://issues.oasis-open.org/browse/OSLCCORE-100 (opened by @DavidJHoney; previously assigned to @jamsden)_
non_test
reference in old query spec does not exist in core in under graph patterns a query capability defines a starting subject resource namely the base uri itself see oslc querybase in the oslc core specification query capability however there is no corresponding section in core to reference the link was to however there is should we reference that instead or as well as some new core section migrated from opened by davidjhoney previously assigned to jamsden
0
97,336
28,180,521,125
IssuesEvent
2023-04-04 01:48:53
cake-build/cake
https://api.github.com/repos/cake-build/cake
closed
Update Azure Pipelines build to use Ubuntu 22.04 now that Ubuntu 18.04 has been deprecated
Build
Azure builds are failing with errors: ``` The ubuntu-18.04 environment is deprecated, consider switching to ubuntu-20.04 or ubuntu-22.04 (ubuntu-latest). For more details, see https://github.com/actions/runner-images/issues/6002 The remote provider was unable to process the request. ``` <img width="1052" alt="image" src="https://user-images.githubusercontent.com/177608/229654686-2cd173fc-9c96-418e-9b36-93bb208d0c32.png"> --- - [Azure Pipelines - Build Cake Debian Stretch](https://dev.azure.com/cake-build/Cake/_build/results?buildId=29485&view=results) - [Azure Pipelines - Build Cake Fedora 28](https://dev.azure.com/cake-build/Cake/_build/results?buildId=29484&view=results)
1.0
Update Azure Pipelines build to use Ubuntu 22.04 now that Ubuntu 18.04 has been deprecated - Azure builds are failing with errors: ``` The ubuntu-18.04 environment is deprecated, consider switching to ubuntu-20.04 or ubuntu-22.04 (ubuntu-latest). For more details, see https://github.com/actions/runner-images/issues/6002 The remote provider was unable to process the request. ``` <img width="1052" alt="image" src="https://user-images.githubusercontent.com/177608/229654686-2cd173fc-9c96-418e-9b36-93bb208d0c32.png"> --- - [Azure Pipelines - Build Cake Debian Stretch](https://dev.azure.com/cake-build/Cake/_build/results?buildId=29485&view=results) - [Azure Pipelines - Build Cake Fedora 28](https://dev.azure.com/cake-build/Cake/_build/results?buildId=29484&view=results)
non_test
update azure pipelines build to use ubuntu now that ubuntu has been deprecated azure builds are failing with errors the ubuntu environment is deprecated consider switching to ubuntu or ubuntu ubuntu latest for more details see the remote provider was unable to process the request img width alt image src
0
138,833
5,348,421,607
IssuesEvent
2017-02-18 04:46:34
ACEmulator/ACE
https://api.github.com/repos/ACEmulator/ACE
closed
Friends list functionality
In Progress priority:medium type:Feature
+ Persist/read friends from db table + Send full friends list on login + Respond to Add friend client message + Respond to Remove friend client message Likely not able to add yet until multiplayer support added: + When player logs on or off, send friend status update to their friends.
1.0
Friends list functionality - + Persist/read friends from db table + Send full friends list on login + Respond to Add friend client message + Respond to Remove friend client message Likely not able to add yet until multiplayer support added: + When player logs on or off, send friend status update to their friends.
non_test
friends list functionality persist read friends from db table send full friends list on login respond to add friend client message respond to remove friend client message likely not able to add yet until multiplayer support added when player logs on or off send friend status update to their friends
0
268,669
20,359,010,569
IssuesEvent
2022-02-20 12:03:07
Tweetmarked/Tweetmarked
https://api.github.com/repos/Tweetmarked/Tweetmarked
closed
Definition of Done: Review in Progess
documentation
What should be the requirements for the review in progress to consider it as done? AB#8
1.0
Definition of Done: Review in Progess - What should be the requirements for the review in progress to consider it as done? AB#8
non_test
definition of done review in progess what should be the requirements for the review in progress to consider it as done ab
0
74,802
9,116,751,454
IssuesEvent
2019-02-22 09:52:19
ByteClubGames/YumiAndTheYokai
https://api.github.com/repos/ByteClubGames/YumiAndTheYokai
reopened
Level Spacing
Done Level Design MEDIUM PRIORITY
Test size and spacing of level and elements implemented into the level -- Make corresponding adjustments to geometry / kick it to art to do so
1.0
Level Spacing - Test size and spacing of level and elements implemented into the level -- Make corresponding adjustments to geometry / kick it to art to do so
non_test
level spacing test size and spacing of level and elements implemented into the level make corresponding adjustments to geometry kick it to art to do so
0
253,875
21,711,179,786
IssuesEvent
2022-05-10 14:00:10
Azordev/various-tasks
https://api.github.com/repos/Azordev/various-tasks
closed
Component Unit tests #2
EE-1 Under Review Test
## Explanation In SP we need to create unit tests for some components, in this task you will see collected 5 components that do not have unit tests and need to be added, each component must meet the following tests: - Snapshot - Is render correctly - Functionality tests if apply - Negative tests if apply The components that are being mentioned are inside **/src/components**, first you should go to **src/components,** and then follow the path that the checklist says: for example **src/components/Buttons/ArrowButton** ### Checklist - [x] FeatureGroupBanner has the tests mentioned in the explanation - [x] Headers/TrialSignupHeader has the tests mentioned in the explanation - [x] Loadings/ButtonLoading has the tests mentioned in the explanation - [x] Loadings/ButtonLoading2 has the tests mentioned in the explanation - [x] Loadings/SignUpLoading has the tests mentioned in the explanation
1.0
Component Unit tests #2 - ## Explanation In SP we need to create unit tests for some components, in this task you will see collected 5 components that do not have unit tests and need to be added, each component must meet the following tests: - Snapshot - Is render correctly - Functionality tests if apply - Negative tests if apply The components that are being mentioned are inside **/src/components**, first you should go to **src/components,** and then follow the path that the checklist says: for example **src/components/Buttons/ArrowButton** ### Checklist - [x] FeatureGroupBanner has the tests mentioned in the explanation - [x] Headers/TrialSignupHeader has the tests mentioned in the explanation - [x] Loadings/ButtonLoading has the tests mentioned in the explanation - [x] Loadings/ButtonLoading2 has the tests mentioned in the explanation - [x] Loadings/SignUpLoading has the tests mentioned in the explanation
test
component unit tests explanation in sp we need to create unit tests for some components in this task you will see collected components that do not have unit tests and need to be added each component must meet the following tests snapshot is render correctly functionality tests if apply negative tests if apply the components that are being mentioned are inside src components first you should go to src components and then follow the path that the checklist says for example src components buttons arrowbutton checklist featuregroupbanner has the tests mentioned in the explanation headers trialsignupheader has the tests mentioned in the explanation loadings buttonloading has the tests mentioned in the explanation loadings has the tests mentioned in the explanation loadings signuploading has the tests mentioned in the explanation
1
93,986
8,461,297,537
IssuesEvent
2018-10-22 21:22:29
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
teamcity: failed test: TestDropIndexWithZoneConfigCCL
C-test-failure O-robot
The following tests appear to have failed on master (testrace): TestDropIndexWithZoneConfigCCL You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+TestDropIndexWithZoneConfigCCL). [#978156](https://teamcity.cockroachdb.com/viewLog.html?buildId=978156): ``` TestDropIndexWithZoneConfigCCL ...log.go:126 [n1,client=127.0.0.1:49890,user=root] Event: "set_zone_config", target: 53, info: {Target:t.kv@primary Config: Options: User:root} I181021 09:04:07.236813 786 storage/replica_command.go:300 [n1,split,s1,r22/1:/{Table/53/1-Max}] initiating a split of this range at key /Table/53/2 [r23] I181021 09:04:07.398343 599 sql/event_log.go:126 [n1,client=127.0.0.1:49890,user=root] Event: "set_zone_config", target: 53, info: {Target:t.kv@i Config:gc: {ttlseconds: 1} Options: User:root} I181021 09:04:07.412181 741 storage/replica_command.go:300 [n1,split,s1,r23/1:/{Table/53/2-Max}] initiating a split of this range at key /Table/53/3 [r24] I181021 09:04:07.451894 769 storage/replica_command.go:300 [n1,split,s1,r24/1:/{Table/53/3-Max}] initiating a split of this range at key /Table/53/4 [r25] I181021 09:04:07.540919 599 sql/event_log.go:126 [n1,client=127.0.0.1:49890,user=root] Event: "set_zone_config", target: 53, info: {Target:t.kv.p2 Config:gc: {ttlseconds: 1} Options: User:root} I181021 09:04:07.550891 807 storage/replica_command.go:300 [n1,split,s1,r24/1:/Table/53/{3-4}] initiating a split of this range at key /Table/53/3/2 [r26] I181021 09:04:07.582036 599 sql/lease.go:345 [n1,client=127.0.0.1:49890,user=root] publish: descID=53 (kv) version=8 mtime=2018-10-21 09:04:07.54189151 +0000 UTC I181021 09:04:07.592865 657 storage/replica_command.go:300 [n1,split,s1,r26/1:/Table/53/{3/2-4}] initiating a split of this range at key /Table/53/3/3 [r27] I181021 09:04:07.644884 599 sql/event_log.go:126 [n1,client=127.0.0.1:49890,user=root] Event: "drop_index", target: 53, info: {TableName:t.public.kv IndexName:i Statement:DROP INDEX t.public.kv@i User:root MutationID:3 CascadeDroppedViews:[]} I181021 09:04:07.694218 812 sql/lease.go:877 new lease: 53("kv") ver=8:1540112945.550088829,0, refcount=0 I181021 09:04:07.776949 599 sql/lease.go:345 [n1,client=127.0.0.1:49890,user=root,scExec] publish: descID=53 (kv) version=9 mtime=2018-10-21 09:04:07.775894162 +0000 UTC I181021 09:04:07.809634 852 sql/lease.go:877 new lease: 53("kv") ver=9:1540112918.080928968,0, refcount=0 I181021 09:04:07.817096 599 sql/lease.go:315 publish (1 leases): desc=[{kv 53 8}] I181021 09:04:07.854840 599 sql/backfill.go:132 [n1,client=127.0.0.1:49890,user=root,scExec] Running backfill for "kv", v=9, m=3 I181021 09:04:07.875949 599 sql/lease.go:345 [n1,client=127.0.0.1:49890,user=root,scExec] publish: descID=53 (kv) version=10 mtime=2018-10-21 09:04:07.868268247 +0000 UTC I181021 09:04:07.975329 599 sql/event_log.go:126 [n1,client=127.0.0.1:49890,user=root,scExec] Event: "finish_schema_change", target: 53, info: {MutationID:3} I181021 09:04:08.018651 857 sql/lease.go:877 new lease: 53("kv") ver=10:1540112933.419412470,0, refcount=0 I181021 09:04:08.042729 599 sql/lease.go:315 publish (1 leases): desc=[{kv 53 9}] I181021 09:04:08.944519 51 util/stop/stopper.go:537 quiescing; tasks left: 1 node.Node: batch 1 [async] transport racer 1 [async] kv.TxnCoordSender: heartbeat loop 1 [async] closedts-rangefeed-subscriber I181021 09:04:08.945098 324 kv/transport_race.go:113 transport race promotion: ran 20 iterations on up to 2051 requests I181021 09:04:08.945326 51 util/stop/stopper.go:537 quiescing; tasks left: 1 node.Node: batch 1 [async] kv.TxnCoordSender: heartbeat loop 1 [async] closedts-rangefeed-subscriber W181021 09:04:08.947065 118 internal/client/txn.go:532 [n1,scExec] failure aborting transaction: node unavailable; try another peer; abort caused by: kv/txn_interceptor_heartbeat.go:405: node already quiescing W181021 09:04:08.947471 118 sql/schema_changer.go:757 [n1,scExec] node unavailable; try another peer W181021 09:04:08.947621 118 sql/schema_changer.go:1522 [n1] Error executing schema change: kv/txn_interceptor_heartbeat.go:405: node already quiescing I181021 09:04:08.949558 51 util/stop/stopper.go:537 quiescing; tasks left: 1 [async] closedts-rangefeed-subscriber TestDropIndexWithZoneConfigCCL ...oot,scExec] Event: "finish_schema_change", target: 53, info: {MutationID:2} I181021 09:05:43.934654 587 sql/event_log.go:126 [n1,client=127.0.0.1:43650,user=root] Event: "set_zone_config", target: 53, info: {Target:t.kv@primary Config: Options: User:root} I181021 09:05:43.935946 520 storage/replica_command.go:300 [n1,split,s1,r21/1:/{Table/53-Max}] initiating a split of this range at key /Table/53/1 [r22] I181021 09:05:43.940784 772 storage/replica_command.go:300 [n1,split,s1,r22/1:/{Table/53/1-Max}] initiating a split of this range at key /Table/53/2 [r23] I181021 09:05:43.941669 587 sql/event_log.go:126 [n1,client=127.0.0.1:43650,user=root] Event: "set_zone_config", target: 53, info: {Target:t.kv@i Config:gc: {ttlseconds: 1} Options: User:root} I181021 09:05:43.946175 792 storage/replica_command.go:300 [n1,split,s1,r23/1:/{Table/53/2-Max}] initiating a split of this range at key /Table/53/3 [r24] I181021 09:05:43.950366 587 sql/event_log.go:126 [n1,client=127.0.0.1:43650,user=root] Event: "set_zone_config", target: 53, info: {Target:t.kv.p2 Config:gc: {ttlseconds: 1} Options: User:root} I181021 09:05:43.953748 587 sql/lease.go:345 [n1,client=127.0.0.1:43650,user=root] publish: descID=53 (kv) version=8 mtime=2018-10-21 09:05:43.950650091 +0000 UTC I181021 09:05:43.953751 727 storage/replica_command.go:300 [n1,split,s1,r24/1:/{Table/53/3-Max}] initiating a split of this range at key /Table/53/3/2 [r25] I181021 09:05:43.959069 737 storage/replica_command.go:300 [n1,split,s1,r25/1:/{Table/53/3/2-Max}] initiating a split of this range at key /Table/53/3/3 [r26] I181021 09:05:43.965772 587 sql/event_log.go:126 [n1,client=127.0.0.1:43650,user=root] Event: "drop_index", target: 53, info: {TableName:t.public.kv IndexName:i Statement:DROP INDEX t.public.kv@i User:root MutationID:3 CascadeDroppedViews:[]} I181021 09:05:43.969696 769 storage/replica_command.go:300 [n1,split,s1,r26/1:/{Table/53/3/3-Max}] initiating a split of this range at key /Table/53/4 [r27] I181021 09:05:43.977449 765 sql/lease.go:877 new lease: 53("kv") ver=8:1540113060.293186899,0, refcount=0 I181021 09:05:44.009719 587 sql/lease.go:345 [n1,client=127.0.0.1:43650,user=root,scExec] publish: descID=53 (kv) version=9 mtime=2018-10-21 09:05:44.009574047 +0000 UTC I181021 09:05:44.014061 815 sql/lease.go:877 new lease: 53("kv") ver=9:1540113071.562652268,0, refcount=0 I181021 09:05:44.019925 587 sql/backfill.go:132 [n1,client=127.0.0.1:43650,user=root,scExec] Running backfill for "kv", v=9, m=3 I181021 09:05:44.022276 587 sql/lease.go:345 [n1,client=127.0.0.1:43650,user=root,scExec] publish: descID=53 (kv) version=10 mtime=2018-10-21 09:05:44.022123018 +0000 UTC I181021 09:05:44.031979 587 sql/event_log.go:126 [n1,client=127.0.0.1:43650,user=root,scExec] Event: "finish_schema_change", target: 53, info: {MutationID:3} I181021 09:05:44.034514 587 sql/lease.go:315 publish (1 leases): desc=[{kv 53 9}] I181021 09:05:44.040260 855 sql/lease.go:877 new lease: 53("kv") ver=10:1540113024.443665630,0, refcount=0 I181021 09:05:44.957520 85 gossip/gossip.go:1510 [n1] node has connected to cluster via gossip I181021 09:05:44.957723 85 storage/stores.go:261 [n1] wrote 0 node addresses to persistent storage I181021 09:05:45.048871 266 sql/lease.go:345 [n1,scExec] publish: descID=53 (kv) version=11 mtime=2018-10-21 09:05:45.048718965 +0000 UTC I181021 09:05:45.049923 688 storage/replica_command.go:432 [n1,merge,s1,r25/1:/Table/53/3/{2-3}] initiating a merge of r26:/Table/53/{3/3-4} [(n1,s1):1, next=2, gen=1] into this range I181021 09:05:45.095447 266 sql/lease.go:315 publish (1 leases): desc=[{kv 53 10}] I181021 09:05:45.097409 221 storage/store.go:2580 [n1,s1,r25/1:/Table/53/3/{2-3}] removing replica r26/1 I181021 09:05:45.101033 998 sql/lease.go:877 new lease: 53("kv") ver=11:1540113014.088685589,0, refcount=0 I181021 09:05:45.234851 29 util/stop/stopper.go:537 quiescing; tasks left: 1 [async] closedts-rangefeed-subscriber ``` Please assign, take a look and update the issue accordingly.
1.0
teamcity: failed test: TestDropIndexWithZoneConfigCCL - The following tests appear to have failed on master (testrace): TestDropIndexWithZoneConfigCCL You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+TestDropIndexWithZoneConfigCCL). [#978156](https://teamcity.cockroachdb.com/viewLog.html?buildId=978156): ``` TestDropIndexWithZoneConfigCCL ...log.go:126 [n1,client=127.0.0.1:49890,user=root] Event: "set_zone_config", target: 53, info: {Target:t.kv@primary Config: Options: User:root} I181021 09:04:07.236813 786 storage/replica_command.go:300 [n1,split,s1,r22/1:/{Table/53/1-Max}] initiating a split of this range at key /Table/53/2 [r23] I181021 09:04:07.398343 599 sql/event_log.go:126 [n1,client=127.0.0.1:49890,user=root] Event: "set_zone_config", target: 53, info: {Target:t.kv@i Config:gc: {ttlseconds: 1} Options: User:root} I181021 09:04:07.412181 741 storage/replica_command.go:300 [n1,split,s1,r23/1:/{Table/53/2-Max}] initiating a split of this range at key /Table/53/3 [r24] I181021 09:04:07.451894 769 storage/replica_command.go:300 [n1,split,s1,r24/1:/{Table/53/3-Max}] initiating a split of this range at key /Table/53/4 [r25] I181021 09:04:07.540919 599 sql/event_log.go:126 [n1,client=127.0.0.1:49890,user=root] Event: "set_zone_config", target: 53, info: {Target:t.kv.p2 Config:gc: {ttlseconds: 1} Options: User:root} I181021 09:04:07.550891 807 storage/replica_command.go:300 [n1,split,s1,r24/1:/Table/53/{3-4}] initiating a split of this range at key /Table/53/3/2 [r26] I181021 09:04:07.582036 599 sql/lease.go:345 [n1,client=127.0.0.1:49890,user=root] publish: descID=53 (kv) version=8 mtime=2018-10-21 09:04:07.54189151 +0000 UTC I181021 09:04:07.592865 657 storage/replica_command.go:300 [n1,split,s1,r26/1:/Table/53/{3/2-4}] initiating a split of this range at key /Table/53/3/3 [r27] I181021 09:04:07.644884 599 sql/event_log.go:126 [n1,client=127.0.0.1:49890,user=root] Event: "drop_index", target: 53, info: {TableName:t.public.kv IndexName:i Statement:DROP INDEX t.public.kv@i User:root MutationID:3 CascadeDroppedViews:[]} I181021 09:04:07.694218 812 sql/lease.go:877 new lease: 53("kv") ver=8:1540112945.550088829,0, refcount=0 I181021 09:04:07.776949 599 sql/lease.go:345 [n1,client=127.0.0.1:49890,user=root,scExec] publish: descID=53 (kv) version=9 mtime=2018-10-21 09:04:07.775894162 +0000 UTC I181021 09:04:07.809634 852 sql/lease.go:877 new lease: 53("kv") ver=9:1540112918.080928968,0, refcount=0 I181021 09:04:07.817096 599 sql/lease.go:315 publish (1 leases): desc=[{kv 53 8}] I181021 09:04:07.854840 599 sql/backfill.go:132 [n1,client=127.0.0.1:49890,user=root,scExec] Running backfill for "kv", v=9, m=3 I181021 09:04:07.875949 599 sql/lease.go:345 [n1,client=127.0.0.1:49890,user=root,scExec] publish: descID=53 (kv) version=10 mtime=2018-10-21 09:04:07.868268247 +0000 UTC I181021 09:04:07.975329 599 sql/event_log.go:126 [n1,client=127.0.0.1:49890,user=root,scExec] Event: "finish_schema_change", target: 53, info: {MutationID:3} I181021 09:04:08.018651 857 sql/lease.go:877 new lease: 53("kv") ver=10:1540112933.419412470,0, refcount=0 I181021 09:04:08.042729 599 sql/lease.go:315 publish (1 leases): desc=[{kv 53 9}] I181021 09:04:08.944519 51 util/stop/stopper.go:537 quiescing; tasks left: 1 node.Node: batch 1 [async] transport racer 1 [async] kv.TxnCoordSender: heartbeat loop 1 [async] closedts-rangefeed-subscriber I181021 09:04:08.945098 324 kv/transport_race.go:113 transport race promotion: ran 20 iterations on up to 2051 requests I181021 09:04:08.945326 51 util/stop/stopper.go:537 quiescing; tasks left: 1 node.Node: batch 1 [async] kv.TxnCoordSender: heartbeat loop 1 [async] closedts-rangefeed-subscriber W181021 09:04:08.947065 118 internal/client/txn.go:532 [n1,scExec] failure aborting transaction: node unavailable; try another peer; abort caused by: kv/txn_interceptor_heartbeat.go:405: node already quiescing W181021 09:04:08.947471 118 sql/schema_changer.go:757 [n1,scExec] node unavailable; try another peer W181021 09:04:08.947621 118 sql/schema_changer.go:1522 [n1] Error executing schema change: kv/txn_interceptor_heartbeat.go:405: node already quiescing I181021 09:04:08.949558 51 util/stop/stopper.go:537 quiescing; tasks left: 1 [async] closedts-rangefeed-subscriber TestDropIndexWithZoneConfigCCL ...oot,scExec] Event: "finish_schema_change", target: 53, info: {MutationID:2} I181021 09:05:43.934654 587 sql/event_log.go:126 [n1,client=127.0.0.1:43650,user=root] Event: "set_zone_config", target: 53, info: {Target:t.kv@primary Config: Options: User:root} I181021 09:05:43.935946 520 storage/replica_command.go:300 [n1,split,s1,r21/1:/{Table/53-Max}] initiating a split of this range at key /Table/53/1 [r22] I181021 09:05:43.940784 772 storage/replica_command.go:300 [n1,split,s1,r22/1:/{Table/53/1-Max}] initiating a split of this range at key /Table/53/2 [r23] I181021 09:05:43.941669 587 sql/event_log.go:126 [n1,client=127.0.0.1:43650,user=root] Event: "set_zone_config", target: 53, info: {Target:t.kv@i Config:gc: {ttlseconds: 1} Options: User:root} I181021 09:05:43.946175 792 storage/replica_command.go:300 [n1,split,s1,r23/1:/{Table/53/2-Max}] initiating a split of this range at key /Table/53/3 [r24] I181021 09:05:43.950366 587 sql/event_log.go:126 [n1,client=127.0.0.1:43650,user=root] Event: "set_zone_config", target: 53, info: {Target:t.kv.p2 Config:gc: {ttlseconds: 1} Options: User:root} I181021 09:05:43.953748 587 sql/lease.go:345 [n1,client=127.0.0.1:43650,user=root] publish: descID=53 (kv) version=8 mtime=2018-10-21 09:05:43.950650091 +0000 UTC I181021 09:05:43.953751 727 storage/replica_command.go:300 [n1,split,s1,r24/1:/{Table/53/3-Max}] initiating a split of this range at key /Table/53/3/2 [r25] I181021 09:05:43.959069 737 storage/replica_command.go:300 [n1,split,s1,r25/1:/{Table/53/3/2-Max}] initiating a split of this range at key /Table/53/3/3 [r26] I181021 09:05:43.965772 587 sql/event_log.go:126 [n1,client=127.0.0.1:43650,user=root] Event: "drop_index", target: 53, info: {TableName:t.public.kv IndexName:i Statement:DROP INDEX t.public.kv@i User:root MutationID:3 CascadeDroppedViews:[]} I181021 09:05:43.969696 769 storage/replica_command.go:300 [n1,split,s1,r26/1:/{Table/53/3/3-Max}] initiating a split of this range at key /Table/53/4 [r27] I181021 09:05:43.977449 765 sql/lease.go:877 new lease: 53("kv") ver=8:1540113060.293186899,0, refcount=0 I181021 09:05:44.009719 587 sql/lease.go:345 [n1,client=127.0.0.1:43650,user=root,scExec] publish: descID=53 (kv) version=9 mtime=2018-10-21 09:05:44.009574047 +0000 UTC I181021 09:05:44.014061 815 sql/lease.go:877 new lease: 53("kv") ver=9:1540113071.562652268,0, refcount=0 I181021 09:05:44.019925 587 sql/backfill.go:132 [n1,client=127.0.0.1:43650,user=root,scExec] Running backfill for "kv", v=9, m=3 I181021 09:05:44.022276 587 sql/lease.go:345 [n1,client=127.0.0.1:43650,user=root,scExec] publish: descID=53 (kv) version=10 mtime=2018-10-21 09:05:44.022123018 +0000 UTC I181021 09:05:44.031979 587 sql/event_log.go:126 [n1,client=127.0.0.1:43650,user=root,scExec] Event: "finish_schema_change", target: 53, info: {MutationID:3} I181021 09:05:44.034514 587 sql/lease.go:315 publish (1 leases): desc=[{kv 53 9}] I181021 09:05:44.040260 855 sql/lease.go:877 new lease: 53("kv") ver=10:1540113024.443665630,0, refcount=0 I181021 09:05:44.957520 85 gossip/gossip.go:1510 [n1] node has connected to cluster via gossip I181021 09:05:44.957723 85 storage/stores.go:261 [n1] wrote 0 node addresses to persistent storage I181021 09:05:45.048871 266 sql/lease.go:345 [n1,scExec] publish: descID=53 (kv) version=11 mtime=2018-10-21 09:05:45.048718965 +0000 UTC I181021 09:05:45.049923 688 storage/replica_command.go:432 [n1,merge,s1,r25/1:/Table/53/3/{2-3}] initiating a merge of r26:/Table/53/{3/3-4} [(n1,s1):1, next=2, gen=1] into this range I181021 09:05:45.095447 266 sql/lease.go:315 publish (1 leases): desc=[{kv 53 10}] I181021 09:05:45.097409 221 storage/store.go:2580 [n1,s1,r25/1:/Table/53/3/{2-3}] removing replica r26/1 I181021 09:05:45.101033 998 sql/lease.go:877 new lease: 53("kv") ver=11:1540113014.088685589,0, refcount=0 I181021 09:05:45.234851 29 util/stop/stopper.go:537 quiescing; tasks left: 1 [async] closedts-rangefeed-subscriber ``` Please assign, take a look and update the issue accordingly.
test
teamcity failed test testdropindexwithzoneconfigccl the following tests appear to have failed on master testrace testdropindexwithzoneconfigccl you may want to check testdropindexwithzoneconfigccl log go event set zone config target info target t kv primary config options user root storage replica command go initiating a split of this range at key table sql event log go event set zone config target info target t kv i config gc ttlseconds options user root storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event set zone config target info target t kv config gc ttlseconds options user root storage replica command go initiating a split of this range at key table sql lease go publish descid kv version mtime utc storage replica command go initiating a split of this range at key table sql event log go event drop index target info tablename t public kv indexname i statement drop index t public kv i user root mutationid cascadedroppedviews sql lease go new lease kv ver refcount sql lease go publish descid kv version mtime utc sql lease go new lease kv ver refcount sql lease go publish leases desc sql backfill go running backfill for kv v m sql lease go publish descid kv version mtime utc sql event log go event finish schema change target info mutationid sql lease go new lease kv ver refcount sql lease go publish leases desc util stop stopper go quiescing tasks left node node batch transport racer kv txncoordsender heartbeat loop closedts rangefeed subscriber kv transport race go transport race promotion ran iterations on up to requests util stop stopper go quiescing tasks left node node batch kv txncoordsender heartbeat loop closedts rangefeed subscriber internal client txn go failure aborting transaction node unavailable try another peer abort caused by kv txn interceptor heartbeat go node already quiescing sql schema changer go node unavailable try another peer sql schema changer go error executing schema change kv txn interceptor heartbeat go node already quiescing util stop stopper go quiescing tasks left closedts rangefeed subscriber testdropindexwithzoneconfigccl oot scexec event finish schema change target info mutationid sql event log go event set zone config target info target t kv primary config options user root storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event set zone config target info target t kv i config gc ttlseconds options user root storage replica command go initiating a split of this range at key table sql event log go event set zone config target info target t kv config gc ttlseconds options user root sql lease go publish descid kv version mtime utc storage replica command go initiating a split of this range at key table storage replica command go initiating a split of this range at key table sql event log go event drop index target info tablename t public kv indexname i statement drop index t public kv i user root mutationid cascadedroppedviews storage replica command go initiating a split of this range at key table sql lease go new lease kv ver refcount sql lease go publish descid kv version mtime utc sql lease go new lease kv ver refcount sql backfill go running backfill for kv v m sql lease go publish descid kv version mtime utc sql event log go event finish schema change target info mutationid sql lease go publish leases desc sql lease go new lease kv ver refcount gossip gossip go node has connected to cluster via gossip storage stores go wrote node addresses to persistent storage sql lease go publish descid kv version mtime utc storage replica command go initiating a merge of table into this range sql lease go publish leases desc storage store go removing replica sql lease go new lease kv ver refcount util stop stopper go quiescing tasks left closedts rangefeed subscriber please assign take a look and update the issue accordingly
1
418,428
28,116,105,266
IssuesEvent
2023-03-31 10:52:05
2sic/2sxc
https://api.github.com/repos/2sic/2sxc
closed
Feat: Enable 2sxc Modules in Skin to also register js/css assets
documentation
**I'm submitting a ...** <!-- HINT: remove lines which don't apply --> [x] bug report => search github for a similar issue before submitting [x] feature request **...about** <!-- HINT: remove lines which don't apply --> [x] Razor templating [x] DNN parts **Current behavior** <!-- HINT: Describe how the bug manifests. --> When including a 2sxc module in the theme using the IRenderService, it does not auto-include the JS/CSS assets in the page. But it should... * [x] fix * [ ] document fixed in 9e508ad056972c65f9f8035265b5b5f91610d19b and d0d70c568eedbbe10661a57c261e259c8d65e8a3 **Your environment** <!-- HINT: Operating system, IDE, package manager, HTTP server, ... --> * **2sxc version(s):** 14.11
1.0
Feat: Enable 2sxc Modules in Skin to also register js/css assets - **I'm submitting a ...** <!-- HINT: remove lines which don't apply --> [x] bug report => search github for a similar issue before submitting [x] feature request **...about** <!-- HINT: remove lines which don't apply --> [x] Razor templating [x] DNN parts **Current behavior** <!-- HINT: Describe how the bug manifests. --> When including a 2sxc module in the theme using the IRenderService, it does not auto-include the JS/CSS assets in the page. But it should... * [x] fix * [ ] document fixed in 9e508ad056972c65f9f8035265b5b5f91610d19b and d0d70c568eedbbe10661a57c261e259c8d65e8a3 **Your environment** <!-- HINT: Operating system, IDE, package manager, HTTP server, ... --> * **2sxc version(s):** 14.11
non_test
feat enable modules in skin to also register js css assets i m submitting a bug report search github for a similar issue before submitting feature request about razor templating dnn parts current behavior when including a module in the theme using the irenderservice it does not auto include the js css assets in the page but it should fix document fixed in and your environment version s
0
118,026
9,968,267,884
IssuesEvent
2019-07-08 15:14:17
XanaduAI/SFOpenBoson
https://api.github.com/repos/XanaduAI/SFOpenBoson
closed
Bose-Hubbard tests currently fail on Python 3.5 due to quirk in how Python changed dictionary sorting in Python 3.6
bug good first issue tests
Note that this does not affect the library itself, just the way some of the tests perform assertions on Python 3.5. This can likely be fixed by modifying the tests to compare the keys/values individually, instead of considering the entire object.
1.0
Bose-Hubbard tests currently fail on Python 3.5 due to quirk in how Python changed dictionary sorting in Python 3.6 - Note that this does not affect the library itself, just the way some of the tests perform assertions on Python 3.5. This can likely be fixed by modifying the tests to compare the keys/values individually, instead of considering the entire object.
test
bose hubbard tests currently fail on python due to quirk in how python changed dictionary sorting in python note that this does not affect the library itself just the way some of the tests perform assertions on python this can likely be fixed by modifying the tests to compare the keys values individually instead of considering the entire object
1
58,782
3,091,365,575
IssuesEvent
2015-08-26 12:48:54
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
Expose EfficientLength interface
Area-Library Library-Collection Priority-Unassigned Triaged Type-Enhancement
*This issue was originally filed by davidm...&#064;google.com* _____ EfficientLength is currently part of dart._internal, meaning third party collection libraries can't advertise their efficientness and performance suffers as a result. Please share ;)
1.0
Expose EfficientLength interface - *This issue was originally filed by davidm...&#064;google.com* _____ EfficientLength is currently part of dart._internal, meaning third party collection libraries can't advertise their efficientness and performance suffers as a result. Please share ;)
non_test
expose efficientlength interface this issue was originally filed by davidm google com efficientlength is currently part of dart internal meaning third party collection libraries can t advertise their efficientness and performance suffers as a result please share
0
260,406
22,618,075,362
IssuesEvent
2022-06-30 01:39:54
thexerteproject/xerteonlinetoolkits
https://api.github.com/repos/thexerteproject/xerteonlinetoolkits
reopened
New Audio Recorder button in WYSIWYG editor requires testing
New feature Needs testing Editor
Hi @ronm123, @torinfo, @FayCross this has been a long time idea and I've finally got something up and running and committed this morning. This relates to several issues #23 #24 #25 #447 #525 The functionality isn't finalised yet - I just wanted to get it up there so that it can be tested in practice and can get some feedback but it seems pretty robust to my testing (in Chrome only at present) I also still need to do some work on the language file and the sources dropdown - at present it might only record from the default audio source. I need to get a few other devices hooked up to my laptop so that I have some to choose from but if anyone else has multiple things populated in that dropdown then i'd welcome any feedback on whether it works or not. The button is after the Audio icon in the editor: ![image](https://user-images.githubusercontent.com/3907362/173029081-fb90fab5-b868-4d2b-b64b-bccd60017280.png) It transcodes into mp3 by default but can also do ogg and wav if required. Possibly I can have it write out the default Opus/.webm format also, if anyone thinks that is a good idea. Any suggestions greatly received!
1.0
New Audio Recorder button in WYSIWYG editor requires testing - Hi @ronm123, @torinfo, @FayCross this has been a long time idea and I've finally got something up and running and committed this morning. This relates to several issues #23 #24 #25 #447 #525 The functionality isn't finalised yet - I just wanted to get it up there so that it can be tested in practice and can get some feedback but it seems pretty robust to my testing (in Chrome only at present) I also still need to do some work on the language file and the sources dropdown - at present it might only record from the default audio source. I need to get a few other devices hooked up to my laptop so that I have some to choose from but if anyone else has multiple things populated in that dropdown then i'd welcome any feedback on whether it works or not. The button is after the Audio icon in the editor: ![image](https://user-images.githubusercontent.com/3907362/173029081-fb90fab5-b868-4d2b-b64b-bccd60017280.png) It transcodes into mp3 by default but can also do ogg and wav if required. Possibly I can have it write out the default Opus/.webm format also, if anyone thinks that is a good idea. Any suggestions greatly received!
test
new audio recorder button in wysiwyg editor requires testing hi torinfo faycross this has been a long time idea and i ve finally got something up and running and committed this morning this relates to several issues the functionality isn t finalised yet i just wanted to get it up there so that it can be tested in practice and can get some feedback but it seems pretty robust to my testing in chrome only at present i also still need to do some work on the language file and the sources dropdown at present it might only record from the default audio source i need to get a few other devices hooked up to my laptop so that i have some to choose from but if anyone else has multiple things populated in that dropdown then i d welcome any feedback on whether it works or not the button is after the audio icon in the editor it transcodes into by default but can also do ogg and wav if required possibly i can have it write out the default opus webm format also if anyone thinks that is a good idea any suggestions greatly received
1
217,951
16,891,853,375
IssuesEvent
2021-06-23 10:11:36
IntellectualSites/PlotSquared
https://api.github.com/repos/IntellectualSites/PlotSquared
opened
Moving between plots causes console error
Requires Testing
### Server Implementation Paper ### Server Version 1.17.1 ### Describe the bug Going to the builders world that plot6 handles, moving between plots owned by different people spits out console errors. I also noticed the title msg isn't showing up ### To Reproduce move between plots owned by different people ### Expected behaviour no errors, smooth transition, and titlemsg to show saying who owns the plot. ### Screenshots / Videos <img width="1551" alt="Screenshot 2021-06-23 at 12 11 22" src="https://user-images.githubusercontent.com/28841349/123079324-20d5fe00-d41c-11eb-8ee3-8b863903c57d.png"> ### Error log (if applicable) _No response_ ### Plot Debugpaste https://athion.net/ISPaster/paste/view/a2cf784cdc414e66a8aa68e27783f167 ### PlotSquared Version PlotSquared-6.0.0-Premium ### Checklist - [X] I have included a Plot debugpaste. - [X] I am using the newest build from https://www.spigotmc.org/resources/77506/ and the issue still persists. ### Anything else? ``` > plotsquared version [12:10:29 INFO]: >> PlotSquared (Version: PlotSquared-6.0.0-Premium) [12:10:29 INFO]: >> Authors: Citymonstret & Empire92 & MattBDev & dordsor21 & NotMyFault & SirYwell [12:10:29 INFO]: >> Wiki: https://github.com/IntellectualSites/PlotSquared-Documentation/wiki [12:10:29 INFO]: >> Discord: https://discord.gg/intellectualsites [12:10:29 INFO]: >> Premium: true > ``` ``` [12:07:22 ERROR]: Could not pass event PlayerMoveEvent to PlotSquared v6.0.0-Premium java.lang.NoSuchMethodError: 'com.plotsquared.core.player.PlotPlayer com.plotsquared.core.player.PlotPlayer.wrap(java.lang.Object)' at com.Zrips.sv.commands.list.plot.showCurrentPlot(plot.java:55) ~[?:?] at com.Zrips.sv.AllListeners.onPlotSquared5Enter.onPlayerEnterPlotEvent(onPlotSquared5Enter.java:22) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:78) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:567) ~[?:?] at com.google.common.eventbus.Subscriber.invokeSubscriberMethod(Subscriber.java:91) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at com.google.common.eventbus.Subscriber$SynchronizedSubscriber.invokeSubscriberMethod(Subscriber.java:150) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at com.google.common.eventbus.Subscriber$1.run(Subscriber.java:76) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at com.google.common.util.concurrent.MoreExecutors$DirectExecutor.execute(MoreExecutors.java:399) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at com.google.common.eventbus.Subscriber.dispatchEvent(Subscriber.java:71) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at com.google.common.eventbus.Dispatcher$PerThreadQueuedDispatcher.dispatch(Dispatcher.java:116) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at com.google.common.eventbus.EventBus.post(EventBus.java:217) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at com.plotsquared.core.util.EventDispatcher.callEvent(EventDispatcher.java:114) ~[?:?] at com.plotsquared.core.util.EventDispatcher.callEntry(EventDispatcher.java:192) ~[?:?] at com.plotsquared.core.listener.PlotListener.plotEntry(PlotListener.java:161) ~[?:?] at com.plotsquared.bukkit.listener.PlayerEventListener.playerMove(PlayerEventListener.java:627) ~[?:?] at com.destroystokyo.paper.event.executor.asm.generated.GeneratedEventExecutor150.execute(Unknown Source) ~[?:?] at org.bukkit.plugin.EventExecutor.lambda$create$1(EventExecutor.java:69) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at co.aikar.timings.TimedEventExecutor.execute(TimedEventExecutor.java:80) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at org.bukkit.plugin.RegisteredListener.callEvent(RegisteredListener.java:70) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at org.bukkit.plugin.SimplePluginManager.callEvent(SimplePluginManager.java:624) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at net.minecraft.server.network.ServerGamePacketListenerImpl.handleMovePlayer(ServerGamePacketListenerImpl.java:1440) ~[?:?] at net.minecraft.network.protocol.game.ServerboundMovePlayerPacket.handle(ServerboundMovePlayerPacket.java:114) ~[?:?] at net.minecraft.network.protocol.game.ServerboundMovePlayerPacket$Pos.handle(ServerboundMovePlayerPacket.java:42) ~[?:?] at net.minecraft.network.protocol.PacketUtils.lambda$ensureRunningOnSameThread$1(PacketUtils.java:36) ~[?:?] at net.minecraft.server.TickTask.run(TickTask.java:18) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at net.minecraft.util.thread.BlockableEventLoop.doRunTask(BlockableEventLoop.java:149) ~[?:?] at net.minecraft.util.thread.ReentrantBlockableEventLoop.doRunTask(ReentrantBlockableEventLoop.java:23) ~[?:?] at net.minecraft.server.MinecraftServer.doRunTask(MinecraftServer.java:1340) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at net.minecraft.server.MinecraftServer.shouldRun(MinecraftServer.java:193) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at net.minecraft.util.thread.BlockableEventLoop.pollTask(BlockableEventLoop.java:122) ~[?:?] at net.minecraft.server.MinecraftServer.pollTaskInternal(MinecraftServer.java:1319) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at net.minecraft.server.MinecraftServer.pollTask(MinecraftServer.java:1312) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at net.minecraft.util.thread.BlockableEventLoop.runAllTasks(BlockableEventLoop.java:110) ~[?:?] at net.minecraft.server.MinecraftServer.tickServer(MinecraftServer.java:1456) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at net.minecraft.server.MinecraftServer.runServer(MinecraftServer.java:1180) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at net.minecraft.server.MinecraftServer.lambda$spin$0(MinecraftServer.java:320) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at java.lang.Thread.run(Thread.java:831) [?:?] [12:07:28 INFO]: mrfloris issued server command: /plot info ```
1.0
Moving between plots causes console error - ### Server Implementation Paper ### Server Version 1.17.1 ### Describe the bug Going to the builders world that plot6 handles, moving between plots owned by different people spits out console errors. I also noticed the title msg isn't showing up ### To Reproduce move between plots owned by different people ### Expected behaviour no errors, smooth transition, and titlemsg to show saying who owns the plot. ### Screenshots / Videos <img width="1551" alt="Screenshot 2021-06-23 at 12 11 22" src="https://user-images.githubusercontent.com/28841349/123079324-20d5fe00-d41c-11eb-8ee3-8b863903c57d.png"> ### Error log (if applicable) _No response_ ### Plot Debugpaste https://athion.net/ISPaster/paste/view/a2cf784cdc414e66a8aa68e27783f167 ### PlotSquared Version PlotSquared-6.0.0-Premium ### Checklist - [X] I have included a Plot debugpaste. - [X] I am using the newest build from https://www.spigotmc.org/resources/77506/ and the issue still persists. ### Anything else? ``` > plotsquared version [12:10:29 INFO]: >> PlotSquared (Version: PlotSquared-6.0.0-Premium) [12:10:29 INFO]: >> Authors: Citymonstret & Empire92 & MattBDev & dordsor21 & NotMyFault & SirYwell [12:10:29 INFO]: >> Wiki: https://github.com/IntellectualSites/PlotSquared-Documentation/wiki [12:10:29 INFO]: >> Discord: https://discord.gg/intellectualsites [12:10:29 INFO]: >> Premium: true > ``` ``` [12:07:22 ERROR]: Could not pass event PlayerMoveEvent to PlotSquared v6.0.0-Premium java.lang.NoSuchMethodError: 'com.plotsquared.core.player.PlotPlayer com.plotsquared.core.player.PlotPlayer.wrap(java.lang.Object)' at com.Zrips.sv.commands.list.plot.showCurrentPlot(plot.java:55) ~[?:?] at com.Zrips.sv.AllListeners.onPlotSquared5Enter.onPlayerEnterPlotEvent(onPlotSquared5Enter.java:22) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:78) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:567) ~[?:?] at com.google.common.eventbus.Subscriber.invokeSubscriberMethod(Subscriber.java:91) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at com.google.common.eventbus.Subscriber$SynchronizedSubscriber.invokeSubscriberMethod(Subscriber.java:150) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at com.google.common.eventbus.Subscriber$1.run(Subscriber.java:76) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at com.google.common.util.concurrent.MoreExecutors$DirectExecutor.execute(MoreExecutors.java:399) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at com.google.common.eventbus.Subscriber.dispatchEvent(Subscriber.java:71) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at com.google.common.eventbus.Dispatcher$PerThreadQueuedDispatcher.dispatch(Dispatcher.java:116) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at com.google.common.eventbus.EventBus.post(EventBus.java:217) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at com.plotsquared.core.util.EventDispatcher.callEvent(EventDispatcher.java:114) ~[?:?] at com.plotsquared.core.util.EventDispatcher.callEntry(EventDispatcher.java:192) ~[?:?] at com.plotsquared.core.listener.PlotListener.plotEntry(PlotListener.java:161) ~[?:?] at com.plotsquared.bukkit.listener.PlayerEventListener.playerMove(PlayerEventListener.java:627) ~[?:?] at com.destroystokyo.paper.event.executor.asm.generated.GeneratedEventExecutor150.execute(Unknown Source) ~[?:?] at org.bukkit.plugin.EventExecutor.lambda$create$1(EventExecutor.java:69) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at co.aikar.timings.TimedEventExecutor.execute(TimedEventExecutor.java:80) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at org.bukkit.plugin.RegisteredListener.callEvent(RegisteredListener.java:70) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at org.bukkit.plugin.SimplePluginManager.callEvent(SimplePluginManager.java:624) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at net.minecraft.server.network.ServerGamePacketListenerImpl.handleMovePlayer(ServerGamePacketListenerImpl.java:1440) ~[?:?] at net.minecraft.network.protocol.game.ServerboundMovePlayerPacket.handle(ServerboundMovePlayerPacket.java:114) ~[?:?] at net.minecraft.network.protocol.game.ServerboundMovePlayerPacket$Pos.handle(ServerboundMovePlayerPacket.java:42) ~[?:?] at net.minecraft.network.protocol.PacketUtils.lambda$ensureRunningOnSameThread$1(PacketUtils.java:36) ~[?:?] at net.minecraft.server.TickTask.run(TickTask.java:18) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at net.minecraft.util.thread.BlockableEventLoop.doRunTask(BlockableEventLoop.java:149) ~[?:?] at net.minecraft.util.thread.ReentrantBlockableEventLoop.doRunTask(ReentrantBlockableEventLoop.java:23) ~[?:?] at net.minecraft.server.MinecraftServer.doRunTask(MinecraftServer.java:1340) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at net.minecraft.server.MinecraftServer.shouldRun(MinecraftServer.java:193) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at net.minecraft.util.thread.BlockableEventLoop.pollTask(BlockableEventLoop.java:122) ~[?:?] at net.minecraft.server.MinecraftServer.pollTaskInternal(MinecraftServer.java:1319) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at net.minecraft.server.MinecraftServer.pollTask(MinecraftServer.java:1312) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at net.minecraft.util.thread.BlockableEventLoop.runAllTasks(BlockableEventLoop.java:110) ~[?:?] at net.minecraft.server.MinecraftServer.tickServer(MinecraftServer.java:1456) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at net.minecraft.server.MinecraftServer.runServer(MinecraftServer.java:1180) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at net.minecraft.server.MinecraftServer.lambda$spin$0(MinecraftServer.java:320) ~[patched_1.17.jar:git-Paper-"4e2f0be"] at java.lang.Thread.run(Thread.java:831) [?:?] [12:07:28 INFO]: mrfloris issued server command: /plot info ```
test
moving between plots causes console error server implementation paper server version describe the bug going to the builders world that handles moving between plots owned by different people spits out console errors i also noticed the title msg isn t showing up to reproduce move between plots owned by different people expected behaviour no errors smooth transition and titlemsg to show saying who owns the plot screenshots videos img width alt screenshot at src error log if applicable no response plot debugpaste plotsquared version plotsquared premium checklist i have included a plot debugpaste i am using the newest build from and the issue still persists anything else plotsquared version plotsquared version plotsquared premium authors citymonstret mattbdev notmyfault sirywell wiki discord premium true could not pass event playermoveevent to plotsquared premium java lang nosuchmethoderror com plotsquared core player plotplayer com plotsquared core player plotplayer wrap java lang object at com zrips sv commands list plot showcurrentplot plot java at com zrips sv alllisteners onplayerenterplotevent java at jdk internal reflect nativemethodaccessorimpl native method at jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com google common eventbus subscriber invokesubscribermethod subscriber java at com google common eventbus subscriber synchronizedsubscriber invokesubscribermethod subscriber java at com google common eventbus subscriber run subscriber java at com google common util concurrent moreexecutors directexecutor execute moreexecutors java at com google common eventbus subscriber dispatchevent subscriber java at com google common eventbus dispatcher perthreadqueueddispatcher dispatch dispatcher java at com google common eventbus eventbus post eventbus java at com plotsquared core util eventdispatcher callevent eventdispatcher java at com plotsquared core util eventdispatcher callentry eventdispatcher java at com plotsquared core listener plotlistener plotentry plotlistener java at com plotsquared bukkit listener playereventlistener playermove playereventlistener java at com destroystokyo paper event executor asm generated execute unknown source at org bukkit plugin eventexecutor lambda create eventexecutor java at co aikar timings timedeventexecutor execute timedeventexecutor java at org bukkit plugin registeredlistener callevent registeredlistener java at org bukkit plugin simplepluginmanager callevent simplepluginmanager java at net minecraft server network servergamepacketlistenerimpl handlemoveplayer servergamepacketlistenerimpl java at net minecraft network protocol game serverboundmoveplayerpacket handle serverboundmoveplayerpacket java at net minecraft network protocol game serverboundmoveplayerpacket pos handle serverboundmoveplayerpacket java at net minecraft network protocol packetutils lambda ensurerunningonsamethread packetutils java at net minecraft server ticktask run ticktask java at net minecraft util thread blockableeventloop doruntask blockableeventloop java at net minecraft util thread reentrantblockableeventloop doruntask reentrantblockableeventloop java at net minecraft server minecraftserver doruntask minecraftserver java at net minecraft server minecraftserver shouldrun minecraftserver java at net minecraft util thread blockableeventloop polltask blockableeventloop java at net minecraft server minecraftserver polltaskinternal minecraftserver java at net minecraft server minecraftserver polltask minecraftserver java at net minecraft util thread blockableeventloop runalltasks blockableeventloop java at net minecraft server minecraftserver tickserver minecraftserver java at net minecraft server minecraftserver runserver minecraftserver java at net minecraft server minecraftserver lambda spin minecraftserver java at java lang thread run thread java mrfloris issued server command plot info
1
29,633
24,126,371,843
IssuesEvent
2022-09-21 01:04:01
antlr/grammars-v4
https://api.github.com/repos/antlr/grammars-v4
closed
[GrammarTest] Grammar tester doesn't support parsers with `options { superClass=...;}`
infrastructure
There doesn't seem to be a way to make the [grammar tester](https://github.com/antlr/grammars-v4/tree/master/_grammar-test) work for a grammar that uses `options {superClass=...;}`. Specifically, there doesn't appear to be a way to feed the GenericParser a set of grammar files that includes either the Java source file for the superclass or a compiled class from it. Or, for that matter, superclass files for any language. I'd love to be wrong, because I'm trying to use `GrammarTest` to test my changes for [[rexx] Hexstrings and binarystrings aren't parsed correctly](https://github.com/antlr/grammars-v4/issues/2741). @julianthome - Am I missing something key here?
1.0
[GrammarTest] Grammar tester doesn't support parsers with `options { superClass=...;}` - There doesn't seem to be a way to make the [grammar tester](https://github.com/antlr/grammars-v4/tree/master/_grammar-test) work for a grammar that uses `options {superClass=...;}`. Specifically, there doesn't appear to be a way to feed the GenericParser a set of grammar files that includes either the Java source file for the superclass or a compiled class from it. Or, for that matter, superclass files for any language. I'd love to be wrong, because I'm trying to use `GrammarTest` to test my changes for [[rexx] Hexstrings and binarystrings aren't parsed correctly](https://github.com/antlr/grammars-v4/issues/2741). @julianthome - Am I missing something key here?
non_test
grammar tester doesn t support parsers with options superclass there doesn t seem to be a way to make the work for a grammar that uses options superclass specifically there doesn t appear to be a way to feed the genericparser a set of grammar files that includes either the java source file for the superclass or a compiled class from it or for that matter superclass files for any language i d love to be wrong because i m trying to use grammartest to test my changes for hexstrings and binarystrings aren t parsed correctly julianthome am i missing something key here
0
252,407
21,576,191,165
IssuesEvent
2022-05-02 13:59:20
rstudio/rstudio
https://api.github.com/repos/rstudio/rstudio
closed
Feature request) Refresh Find in Files
enhancement find in files test
I often use Find in Files to remove or alter instances where I use a function across my package. It would be neat if there was a Refresh button in the Find in Files pane rather than having to Edit->Find in Files->Find over and over (the shortcut helps too). That way I could remove a few of them, refresh, see what I have left, and iteratively do this until I change them all.
1.0
Feature request) Refresh Find in Files - I often use Find in Files to remove or alter instances where I use a function across my package. It would be neat if there was a Refresh button in the Find in Files pane rather than having to Edit->Find in Files->Find over and over (the shortcut helps too). That way I could remove a few of them, refresh, see what I have left, and iteratively do this until I change them all.
test
feature request refresh find in files i often use find in files to remove or alter instances where i use a function across my package it would be neat if there was a refresh button in the find in files pane rather than having to edit find in files find over and over the shortcut helps too that way i could remove a few of them refresh see what i have left and iteratively do this until i change them all
1
78,589
15,031,236,725
IssuesEvent
2021-02-02 08:40:05
edgi-govdata-archiving/web-monitoring-db
https://api.github.com/repos/edgi-govdata-archiving/web-monitoring-db
opened
Remove/replace deprecated google-api-client dependency
Code Quality
The `google-api-client` gem has been deprecated. On install, it now prints: ``` ******************************************************************************* The google-api-client gem is deprecated and will likely not be updated further. Instead, please install the gem corresponding to the specific service to use. For example, to use the Google Drive V3 client, install google-apis-drive_v3. For more information, see the FAQ in the OVERVIEW.md file or the YARD docs. ******************************************************************************* ``` I believe (but we need to check) our only usage is in [`lib/tasks/import_from_sheets.rake`](https://github.com/edgi-govdata-archiving/web-monitoring-db/blob/main/lib/tasks/import_from_sheets.rake), which we don’t really actively use. We should replace it with the sheets-specific gem or remove the rake task.
1.0
Remove/replace deprecated google-api-client dependency - The `google-api-client` gem has been deprecated. On install, it now prints: ``` ******************************************************************************* The google-api-client gem is deprecated and will likely not be updated further. Instead, please install the gem corresponding to the specific service to use. For example, to use the Google Drive V3 client, install google-apis-drive_v3. For more information, see the FAQ in the OVERVIEW.md file or the YARD docs. ******************************************************************************* ``` I believe (but we need to check) our only usage is in [`lib/tasks/import_from_sheets.rake`](https://github.com/edgi-govdata-archiving/web-monitoring-db/blob/main/lib/tasks/import_from_sheets.rake), which we don’t really actively use. We should replace it with the sheets-specific gem or remove the rake task.
non_test
remove replace deprecated google api client dependency the google api client gem has been deprecated on install it now prints the google api client gem is deprecated and will likely not be updated further instead please install the gem corresponding to the specific service to use for example to use the google drive client install google apis drive for more information see the faq in the overview md file or the yard docs i believe but we need to check our only usage is in which we don’t really actively use we should replace it with the sheets specific gem or remove the rake task
0
233,182
7,695,121,797
IssuesEvent
2018-05-18 11:06:05
fac-13/stackMatch
https://api.github.com/repos/fac-13/stackMatch
closed
Front end - tech_stack to be integrated into my details form
T1h-4h in-progress priority-2
User Story #117 ## Front End - [x] On my profile, I can view my current tech stack - [x] On click of a tech stack 'edit' button - it opens up another pop up window: ![edit tech stack screen](https://i.imgur.com/79vXLej.png) -- or this could be within the my detail form
1.0
Front end - tech_stack to be integrated into my details form - User Story #117 ## Front End - [x] On my profile, I can view my current tech stack - [x] On click of a tech stack 'edit' button - it opens up another pop up window: ![edit tech stack screen](https://i.imgur.com/79vXLej.png) -- or this could be within the my detail form
non_test
front end tech stack to be integrated into my details form user story front end on my profile i can view my current tech stack on click of a tech stack edit button it opens up another pop up window or this could be within the my detail form
0
272,922
20,763,433,696
IssuesEvent
2022-03-15 18:15:52
awslabs/aws-ddk
https://api.github.com/repos/awslabs/aws-ddk
closed
(docs) How-to-guide: Create your own DDK data stage
documentation enhancement effort/small p1
**Is your idea related to a problem? Please describe.** How-to Guide for adding custom DDK data stages. *P.S. Don't attach files. Please, prefer add code snippets directly in the message body.*
1.0
(docs) How-to-guide: Create your own DDK data stage - **Is your idea related to a problem? Please describe.** How-to Guide for adding custom DDK data stages. *P.S. Don't attach files. Please, prefer add code snippets directly in the message body.*
non_test
docs how to guide create your own ddk data stage is your idea related to a problem please describe how to guide for adding custom ddk data stages p s don t attach files please prefer add code snippets directly in the message body
0
128,310
10,524,743,737
IssuesEvent
2019-09-30 13:52:00
operator-framework/operator-sdk
https://api.github.com/repos/operator-framework/operator-sdk
closed
Generating tests from `operator-sdk generate`
kind/feature lifecycle/rotten testing
## Feature Request **Is your feature request related to a problem? Please describe.** When using `operator-sdk generate`, I would ideally like to have had the build also create minimal tests. Ideally, it could run e2e as well. **Describe the solution you'd like** [This document](https://kubernetes.io/blog/2019/03/22/kubernetes-end-to-end-testing-for-everyone/) seems to be quite current and I'm going to implement the tests for my controller this way, mostly because TDD is much faster when the cost to bring up a test environment is high. It would also help lower the bar for fully tested operators, so hopefully more of them would emerge into the community that way.
1.0
Generating tests from `operator-sdk generate` - ## Feature Request **Is your feature request related to a problem? Please describe.** When using `operator-sdk generate`, I would ideally like to have had the build also create minimal tests. Ideally, it could run e2e as well. **Describe the solution you'd like** [This document](https://kubernetes.io/blog/2019/03/22/kubernetes-end-to-end-testing-for-everyone/) seems to be quite current and I'm going to implement the tests for my controller this way, mostly because TDD is much faster when the cost to bring up a test environment is high. It would also help lower the bar for fully tested operators, so hopefully more of them would emerge into the community that way.
test
generating tests from operator sdk generate feature request is your feature request related to a problem please describe when using operator sdk generate i would ideally like to have had the build also create minimal tests ideally it could run as well describe the solution you d like seems to be quite current and i m going to implement the tests for my controller this way mostly because tdd is much faster when the cost to bring up a test environment is high it would also help lower the bar for fully tested operators so hopefully more of them would emerge into the community that way
1
287,151
24,811,714,645
IssuesEvent
2022-10-25 09:56:09
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
opened
There is no warning message in the 'Rename Connection' dialog when first opening it
🧪 testing
**Storage Explorer Version**: 1.27.0-dev **Build Number**: 20221025.2 **Branch**: main **Platform/OS**: Windows 10/Linux Ubuntu 22.04/MacOS Monterey 12.6 (Apple M1 Pro) **Architecture**: ia32/x64 **How Found**: Ad-hoc testing **Regression From**: Not a regression ## Steps to Reproduce ## 1. Expand one storage account -> Tables. 2. Select one table -> Attach it. 3. Right click the attached table -> Click the action 'Rename connection...'. 4. Check whether there is a warning message. ## Expected Experience ## There is a warning message indicating 'the name should be different'. ![image](https://user-images.githubusercontent.com/41351993/197740301-db138bfd-d5d9-409a-b539-5027cd67b48a.png) ## Actual Experience ## There is no warning message. ![image](https://user-images.githubusercontent.com/41351993/197740314-f908455c-4d80-45b8-b15c-de24b3b8aa7e.png) ## Additional Connext ## This issue also reproduces for one attached file share.
1.0
There is no warning message in the 'Rename Connection' dialog when first opening it - **Storage Explorer Version**: 1.27.0-dev **Build Number**: 20221025.2 **Branch**: main **Platform/OS**: Windows 10/Linux Ubuntu 22.04/MacOS Monterey 12.6 (Apple M1 Pro) **Architecture**: ia32/x64 **How Found**: Ad-hoc testing **Regression From**: Not a regression ## Steps to Reproduce ## 1. Expand one storage account -> Tables. 2. Select one table -> Attach it. 3. Right click the attached table -> Click the action 'Rename connection...'. 4. Check whether there is a warning message. ## Expected Experience ## There is a warning message indicating 'the name should be different'. ![image](https://user-images.githubusercontent.com/41351993/197740301-db138bfd-d5d9-409a-b539-5027cd67b48a.png) ## Actual Experience ## There is no warning message. ![image](https://user-images.githubusercontent.com/41351993/197740314-f908455c-4d80-45b8-b15c-de24b3b8aa7e.png) ## Additional Connext ## This issue also reproduces for one attached file share.
test
there is no warning message in the rename connection dialog when first opening it storage explorer version dev build number branch main platform os windows linux ubuntu macos monterey apple pro architecture how found ad hoc testing regression from not a regression steps to reproduce expand one storage account tables select one table attach it right click the attached table click the action rename connection check whether there is a warning message expected experience there is a warning message indicating the name should be different actual experience there is no warning message additional connext this issue also reproduces for one attached file share
1
138,870
11,220,501,419
IssuesEvent
2020-01-07 15:55:44
spacemeshos/go-spacemesh
https://api.github.com/repos/spacemeshos/go-spacemesh
closed
poet keeps running after app test
PoET bug tests
# Overview / Motivation on app test we deploy a local poet service to all of our nodes that run through the test. we want to deploy a new poet for each test and clean it up when the test ends. poet takes quite a lot of resources this why leaking the poet is really bad. though right now running the app test (which starts a poet in a separate process) sometimes leaves poet running on your machine. this might also cause problems when tests might try to use that poet instead of a new one. # The Task make sure to stop poet after the test has finished. # Implementation Notes TODO: Add links to relevant resources, specs, related issues, etc... # Contribution Guidelines Important: Issue assignment to developers will be by the order of their application and proficiency level according to the tasks complexity. We will not assign tasks to developers who have'nt introduced themselves on our Gitter [dev channel](https://gitter.im/spacemesh-os/Lobby) 1. Introduce yourself on go-spacemesh [dev chat channel](https://gitter.im/spacemesh-os/Lobby) - ask our team any question you may have about this task 2. Fork branch `develop` to your own repo and work in your repo 3. You must document all methods, enums and types with [godoc comments](https://blog.golang.org/godoc-documenting-go-code) 4. You must write go unit tests for all types and methods when submitting a component, and integration tests if you submit a feature 5. When ready for code review, submit a PR from your repo back to branch `develop` 6. Attach relevant issue to PR
1.0
poet keeps running after app test - # Overview / Motivation on app test we deploy a local poet service to all of our nodes that run through the test. we want to deploy a new poet for each test and clean it up when the test ends. poet takes quite a lot of resources this why leaking the poet is really bad. though right now running the app test (which starts a poet in a separate process) sometimes leaves poet running on your machine. this might also cause problems when tests might try to use that poet instead of a new one. # The Task make sure to stop poet after the test has finished. # Implementation Notes TODO: Add links to relevant resources, specs, related issues, etc... # Contribution Guidelines Important: Issue assignment to developers will be by the order of their application and proficiency level according to the tasks complexity. We will not assign tasks to developers who have'nt introduced themselves on our Gitter [dev channel](https://gitter.im/spacemesh-os/Lobby) 1. Introduce yourself on go-spacemesh [dev chat channel](https://gitter.im/spacemesh-os/Lobby) - ask our team any question you may have about this task 2. Fork branch `develop` to your own repo and work in your repo 3. You must document all methods, enums and types with [godoc comments](https://blog.golang.org/godoc-documenting-go-code) 4. You must write go unit tests for all types and methods when submitting a component, and integration tests if you submit a feature 5. When ready for code review, submit a PR from your repo back to branch `develop` 6. Attach relevant issue to PR
test
poet keeps running after app test overview motivation on app test we deploy a local poet service to all of our nodes that run through the test we want to deploy a new poet for each test and clean it up when the test ends poet takes quite a lot of resources this why leaking the poet is really bad though right now running the app test which starts a poet in a separate process sometimes leaves poet running on your machine this might also cause problems when tests might try to use that poet instead of a new one the task make sure to stop poet after the test has finished implementation notes todo add links to relevant resources specs related issues etc contribution guidelines important issue assignment to developers will be by the order of their application and proficiency level according to the tasks complexity we will not assign tasks to developers who have nt introduced themselves on our gitter introduce yourself on go spacemesh ask our team any question you may have about this task fork branch develop to your own repo and work in your repo you must document all methods enums and types with you must write go unit tests for all types and methods when submitting a component and integration tests if you submit a feature when ready for code review submit a pr from your repo back to branch develop attach relevant issue to pr
1
70,043
7,175,597,342
IssuesEvent
2018-01-31 06:29:18
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Add Node title text missing from Add Node page/modal
area/host area/ui kind/bug status/resolved status/to-test version/2.0
**Rancher versions:** 2.0 master 1/18 **Steps to Reproduce:** 1. Go to the Add node page and click Configure 2. Go to the Add RKE cluster and click on Add Node then Configure **Results:** The page/modal does have the Add Node title text at the top ![image](https://user-images.githubusercontent.com/11514927/35124382-241376b8-fc63-11e7-834d-a2ef30b79dd6.png) ![image](https://user-images.githubusercontent.com/11514927/35124587-d8bc4464-fc63-11e7-8b54-cf26cc45a2b9.png)
1.0
Add Node title text missing from Add Node page/modal - **Rancher versions:** 2.0 master 1/18 **Steps to Reproduce:** 1. Go to the Add node page and click Configure 2. Go to the Add RKE cluster and click on Add Node then Configure **Results:** The page/modal does have the Add Node title text at the top ![image](https://user-images.githubusercontent.com/11514927/35124382-241376b8-fc63-11e7-834d-a2ef30b79dd6.png) ![image](https://user-images.githubusercontent.com/11514927/35124587-d8bc4464-fc63-11e7-8b54-cf26cc45a2b9.png)
test
add node title text missing from add node page modal rancher versions master steps to reproduce go to the add node page and click configure go to the add rke cluster and click on add node then configure results the page modal does have the add node title text at the top
1
109,590
9,395,130,243
IssuesEvent
2019-04-08 00:45:48
Travmatth/21sh
https://api.github.com/repos/Travmatth/21sh
closed
2.6.1 Tilde Expansion
enhancement needs test case
- Fifth step of post lexer processing - Need to implement > A "tilde-prefix" consists of an unquoted <tilde> character at the beginning of a word, followed by all of the characters preceding the first unquoted <slash> in the word, or all the characters in the word if there is no <slash>. In an assignment (see XBD Variable Assignment), multiple tilde-prefixes can be used: at the beginning of the word (that is, following the <equals-sign> of the assignment), following any unquoted <colon>, or both. A tilde-prefix in an assignment is terminated by the first unquoted <colon> or <slash>. If none of the characters in the tilde-prefix are quoted, the characters in the tilde-prefix following the <tilde> are treated as a possible login name from the user database. A portable login name cannot contain characters outside the set given in the description of the LOGNAME environment variable in XBD Other Environment Variables. If the login name is null (that is, the tilde-prefix contains only the tilde), the tilde-prefix is replaced by the value of the variable HOME. If HOME is unset, the results are unspecified. Otherwise, the tilde-prefix shall be replaced by a pathname of the initial working directory associated with the login name obtained using the getpwnam() function as defined in the System Interfaces volume of POSIX.1-2017. If the system does not recognize the login name, the results are undefined. > The pathname resulting from tilde expansion shall be treated as if quoted to prevent it being altered by field splitting and pathname expansion.
1.0
2.6.1 Tilde Expansion - - Fifth step of post lexer processing - Need to implement > A "tilde-prefix" consists of an unquoted <tilde> character at the beginning of a word, followed by all of the characters preceding the first unquoted <slash> in the word, or all the characters in the word if there is no <slash>. In an assignment (see XBD Variable Assignment), multiple tilde-prefixes can be used: at the beginning of the word (that is, following the <equals-sign> of the assignment), following any unquoted <colon>, or both. A tilde-prefix in an assignment is terminated by the first unquoted <colon> or <slash>. If none of the characters in the tilde-prefix are quoted, the characters in the tilde-prefix following the <tilde> are treated as a possible login name from the user database. A portable login name cannot contain characters outside the set given in the description of the LOGNAME environment variable in XBD Other Environment Variables. If the login name is null (that is, the tilde-prefix contains only the tilde), the tilde-prefix is replaced by the value of the variable HOME. If HOME is unset, the results are unspecified. Otherwise, the tilde-prefix shall be replaced by a pathname of the initial working directory associated with the login name obtained using the getpwnam() function as defined in the System Interfaces volume of POSIX.1-2017. If the system does not recognize the login name, the results are undefined. > The pathname resulting from tilde expansion shall be treated as if quoted to prevent it being altered by field splitting and pathname expansion.
test
tilde expansion fifth step of post lexer processing need to implement a tilde prefix consists of an unquoted character at the beginning of a word followed by all of the characters preceding the first unquoted in the word or all the characters in the word if there is no in an assignment see xbd variable assignment multiple tilde prefixes can be used at the beginning of the word that is following the of the assignment following any unquoted or both a tilde prefix in an assignment is terminated by the first unquoted or if none of the characters in the tilde prefix are quoted the characters in the tilde prefix following the are treated as a possible login name from the user database a portable login name cannot contain characters outside the set given in the description of the logname environment variable in xbd other environment variables if the login name is null that is the tilde prefix contains only the tilde the tilde prefix is replaced by the value of the variable home if home is unset the results are unspecified otherwise the tilde prefix shall be replaced by a pathname of the initial working directory associated with the login name obtained using the getpwnam function as defined in the system interfaces volume of posix if the system does not recognize the login name the results are undefined the pathname resulting from tilde expansion shall be treated as if quoted to prevent it being altered by field splitting and pathname expansion
1
109,270
9,376,489,944
IssuesEvent
2019-04-04 08:05:08
italia/spid
https://api.github.com/repos/italia/spid
closed
Validazione metadata Comune di Agropoli
metadata nuovo md test
Buongiorno, per conto del Comune di Agropoli abbiamo predisposto metadata e pubblicati all'URL https://agropoli.comune-online.it/serviziSPID/metadata.xml cordiali saluti Facondini Stefano Maggioli Spa [metadata_agropoli-signed.zip](https://github.com/italia/spid/files/3028767/metadata_agropoli-signed.zip)
1.0
Validazione metadata Comune di Agropoli - Buongiorno, per conto del Comune di Agropoli abbiamo predisposto metadata e pubblicati all'URL https://agropoli.comune-online.it/serviziSPID/metadata.xml cordiali saluti Facondini Stefano Maggioli Spa [metadata_agropoli-signed.zip](https://github.com/italia/spid/files/3028767/metadata_agropoli-signed.zip)
test
validazione metadata comune di agropoli buongiorno per conto del comune di agropoli abbiamo predisposto metadata e pubblicati all url cordiali saluti facondini stefano maggioli spa
1
275,113
23,893,381,441
IssuesEvent
2022-09-08 13:11:27
ARUP-CAS/aiscr-webamcr
https://api.github.com/repos/ARUP-CAS/aiscr-webamcr
closed
Prázdný konec seznamů se jmény v Mozille
bug / maintanance TESTED
Vyčleněno z #338 a #327 V Mozille při výběru z drop-down listu osob (bez ohledu na umístění v aplikaci) při pokusu dojet na konec seznamu přeskočí seznam na prázdnou “stránku”. Při probliknutí před prázdnou stránkou je však vidět, že za “Žohová, Martina” jsou i další jména. týká se např.: http://192.168.254.30:8080/dokument/edit/X-C-TX-201904031 http://192.168.254.30:8080/dokument/model/edit/C-3D-202000004 http://192.168.254.30:8080/pas/edit/M-202101534-N00282 http://192.168.254.30:8080/arch-z/akce/detail/C-202002083A
1.0
Prázdný konec seznamů se jmény v Mozille - Vyčleněno z #338 a #327 V Mozille při výběru z drop-down listu osob (bez ohledu na umístění v aplikaci) při pokusu dojet na konec seznamu přeskočí seznam na prázdnou “stránku”. Při probliknutí před prázdnou stránkou je však vidět, že za “Žohová, Martina” jsou i další jména. týká se např.: http://192.168.254.30:8080/dokument/edit/X-C-TX-201904031 http://192.168.254.30:8080/dokument/model/edit/C-3D-202000004 http://192.168.254.30:8080/pas/edit/M-202101534-N00282 http://192.168.254.30:8080/arch-z/akce/detail/C-202002083A
test
prázdný konec seznamů se jmény v mozille vyčleněno z a v mozille při výběru z drop down listu osob bez ohledu na umístění v aplikaci při pokusu dojet na konec seznamu přeskočí seznam na prázdnou “stránku” při probliknutí před prázdnou stránkou je však vidět že za “žohová martina” jsou i další jména týká se např
1
152,783
12,126,583,871
IssuesEvent
2020-04-22 17:16:07
sparklyr/sparklyr
https://api.github.com/repos/sparklyr/sparklyr
closed
make test infra work better with Databricks
tests
https://github.com/sparklyr/sparklyr/pull/2421#issuecomment-614618682 test cases should use dbfs:// URLs to locate any required input file when running on Databricks connection, and there is a helper function that does exactly that already, so, we should use it whenever applicable
1.0
make test infra work better with Databricks - https://github.com/sparklyr/sparklyr/pull/2421#issuecomment-614618682 test cases should use dbfs:// URLs to locate any required input file when running on Databricks connection, and there is a helper function that does exactly that already, so, we should use it whenever applicable
test
make test infra work better with databricks test cases should use dbfs urls to locate any required input file when running on databricks connection and there is a helper function that does exactly that already so we should use it whenever applicable
1
292,163
25,204,719,168
IssuesEvent
2022-11-13 14:44:04
go-gitea/gitea
https://api.github.com/repos/go-gitea/gitea
closed
Unit tests with mysql/postgres/mssql
kind/testing reviewed/confirmed
Currently unit tests only with sqlite, so we needs run unit tests with mysql/postgres/mssql like we did on migration tests.
1.0
Unit tests with mysql/postgres/mssql - Currently unit tests only with sqlite, so we needs run unit tests with mysql/postgres/mssql like we did on migration tests.
test
unit tests with mysql postgres mssql currently unit tests only with sqlite so we needs run unit tests with mysql postgres mssql like we did on migration tests
1
177,470
13,725,556,318
IssuesEvent
2020-10-03 19:01:45
roderik/unfurl-dropbox-screenshots
https://api.github.com/repos/roderik/unfurl-dropbox-screenshots
closed
Automated issue test Thursday, October 1st 2020, 5:38:37 pm
Stale test
This is an automated test of the unfurling action # Regular link https://vanderveer.be # Dropbox link ![](https://www.dropbox.com/s/7e7qjbmlahvvy09/Screenshot%202020-04-04%2010.25.23.png?raw=1) # Markdown image (dl=0) ![DL0](https://www.dropbox.com/s/0698tr08sxvk3o7/Screenshot%202020-04-04%2010.26.56.png?raw=1) # Markdown image (dl=1) ![DL1](https://www.dropbox.com/s/0698tr08sxvk3o7/Screenshot%202020-04-04%2010.26.56.png?raw=1) # Markdown image () ![none](https://www.dropbox.com/s/0698tr08sxvk3o7/Screenshot%202020-04-04%2010.26.56.png?raw=1) # Markdown image (raw=1) ![raw](https://www.dropbox.com/s/0698tr08sxvk3o7/Screenshot%202020-04-04%2010.26.56.png?raw=1) # Manually uploaded image <img width="316" alt="Screenshot 2020-04-04 10 25 41" src="https://user-images.githubusercontent.com/16780/78422312-ad9d6c00-765e-11ea-8414-da0b05478372.png">
1.0
Automated issue test Thursday, October 1st 2020, 5:38:37 pm - This is an automated test of the unfurling action # Regular link https://vanderveer.be # Dropbox link ![](https://www.dropbox.com/s/7e7qjbmlahvvy09/Screenshot%202020-04-04%2010.25.23.png?raw=1) # Markdown image (dl=0) ![DL0](https://www.dropbox.com/s/0698tr08sxvk3o7/Screenshot%202020-04-04%2010.26.56.png?raw=1) # Markdown image (dl=1) ![DL1](https://www.dropbox.com/s/0698tr08sxvk3o7/Screenshot%202020-04-04%2010.26.56.png?raw=1) # Markdown image () ![none](https://www.dropbox.com/s/0698tr08sxvk3o7/Screenshot%202020-04-04%2010.26.56.png?raw=1) # Markdown image (raw=1) ![raw](https://www.dropbox.com/s/0698tr08sxvk3o7/Screenshot%202020-04-04%2010.26.56.png?raw=1) # Manually uploaded image <img width="316" alt="Screenshot 2020-04-04 10 25 41" src="https://user-images.githubusercontent.com/16780/78422312-ad9d6c00-765e-11ea-8414-da0b05478372.png">
test
automated issue test thursday october pm this is an automated test of the unfurling action regular link dropbox link markdown image dl markdown image dl markdown image markdown image raw manually uploaded image img width alt screenshot src
1
22,638
2,649,608,497
IssuesEvent
2015-03-15 03:01:32
chharvey/chharvey.github.io
https://api.github.com/repos/chharvey/chharvey.github.io
opened
html to jade
feature-major layer-HTML priority-high
Use jade instead of html for all blog and page source. For now, compile locally then upload, but later find a way to do this on the server? Helps the closure of #9 because Jade has a template system. Also helps any issue tagged `layer-HTML`, especially #40 and #22 Pages to convert: - [ ] all blog posts - [ ] all home pages including index, home/, resume - [ ] core/doc/ if applicable (might be removing later---see #32 and #38 ) - [ ] lesson plans - [ ] higher ed pages
1.0
html to jade - Use jade instead of html for all blog and page source. For now, compile locally then upload, but later find a way to do this on the server? Helps the closure of #9 because Jade has a template system. Also helps any issue tagged `layer-HTML`, especially #40 and #22 Pages to convert: - [ ] all blog posts - [ ] all home pages including index, home/, resume - [ ] core/doc/ if applicable (might be removing later---see #32 and #38 ) - [ ] lesson plans - [ ] higher ed pages
non_test
html to jade use jade instead of html for all blog and page source for now compile locally then upload but later find a way to do this on the server helps the closure of because jade has a template system also helps any issue tagged layer html especially and pages to convert all blog posts all home pages including index home resume core doc if applicable might be removing later see and lesson plans higher ed pages
0
144,870
13,128,935,066
IssuesEvent
2020-08-06 13:08:37
blackstream-x/smallparts
https://api.github.com/repos/blackstream-x/smallparts
opened
Enhance the HtmlTagStripper class to make it reusable
documentation enhancement quality assurance
Implement the https://docs.python.org/3/library/html.parser.html#html.parser.HTMLParser.close and https://docs.python.org/3/library/html.parser.html#html.parser.HTMLParser.reset methods. Provide an additional method to parse a complete HTML document.
1.0
Enhance the HtmlTagStripper class to make it reusable - Implement the https://docs.python.org/3/library/html.parser.html#html.parser.HTMLParser.close and https://docs.python.org/3/library/html.parser.html#html.parser.HTMLParser.reset methods. Provide an additional method to parse a complete HTML document.
non_test
enhance the htmltagstripper class to make it reusable implement the and methods provide an additional method to parse a complete html document
0
178,826
13,796,711,446
IssuesEvent
2020-10-09 20:20:54
WarcraftPriests/sl-shadow-priest
https://api.github.com/repos/WarcraftPriests/sl-shadow-priest
opened
Shadow Crash has fixed travel time again
needs-testing
Need to fix this, although i dont think i ever removed it
1.0
Shadow Crash has fixed travel time again - Need to fix this, although i dont think i ever removed it
test
shadow crash has fixed travel time again need to fix this although i dont think i ever removed it
1
123,344
10,264,811,090
IssuesEvent
2019-08-22 17:17:46
wprig/wprig
https://api.github.com/repos/wprig/wprig
closed
assets directory structure could be improved.
css enhancement needs-testing wontfix
## Issue Overview When viewing the wprig src theme, relative src css references fail with current assets directory structure e.g. `background-image: url('../images/image.jpg');` ## Steps to Reproduce 1. Set theme to wprig source theme 2. Set a relative path in your css file. Something like this: `background-image: url('../images/image.jpg');` 3. The image can not be found because the reference in the wprig src theme would need to be: `background-image: url('../../images/src/image.jpg');` But in the dist theme it would need to be: `background-image: url('../images/src/image.jpg');` ## Possible Solution The src css references will work perfectly if we move the src assets into their own src assets directory and change the assets tree to look something like this: ``` \assets\css \assets\images \assets\js \assets-src\css \assets-src\images \assets-src\js ``` Another advantage is that it simplifies the .gitignore file a little :) ``` !/assets/ ``` vs ``` !/assets/ /assets/* !/assets/css/ /assets/css/* !/assets/css/src/ !/assets/js/ /assets/js/* !/assets/js/src/ !/assets/images/ /assets/images/* !/assets/images/src/ ```
1.0
assets directory structure could be improved. - ## Issue Overview When viewing the wprig src theme, relative src css references fail with current assets directory structure e.g. `background-image: url('../images/image.jpg');` ## Steps to Reproduce 1. Set theme to wprig source theme 2. Set a relative path in your css file. Something like this: `background-image: url('../images/image.jpg');` 3. The image can not be found because the reference in the wprig src theme would need to be: `background-image: url('../../images/src/image.jpg');` But in the dist theme it would need to be: `background-image: url('../images/src/image.jpg');` ## Possible Solution The src css references will work perfectly if we move the src assets into their own src assets directory and change the assets tree to look something like this: ``` \assets\css \assets\images \assets\js \assets-src\css \assets-src\images \assets-src\js ``` Another advantage is that it simplifies the .gitignore file a little :) ``` !/assets/ ``` vs ``` !/assets/ /assets/* !/assets/css/ /assets/css/* !/assets/css/src/ !/assets/js/ /assets/js/* !/assets/js/src/ !/assets/images/ /assets/images/* !/assets/images/src/ ```
test
assets directory structure could be improved issue overview when viewing the wprig src theme relative src css references fail with current assets directory structure e g background image url images image jpg steps to reproduce set theme to wprig source theme set a relative path in your css file something like this background image url images image jpg the image can not be found because the reference in the wprig src theme would need to be background image url images src image jpg but in the dist theme it would need to be background image url images src image jpg possible solution the src css references will work perfectly if we move the src assets into their own src assets directory and change the assets tree to look something like this assets css assets images assets js assets src css assets src images assets src js another advantage is that it simplifies the gitignore file a little assets vs assets assets assets css assets css assets css src assets js assets js assets js src assets images assets images assets images src
1
486,433
14,008,975,882
IssuesEvent
2020-10-29 01:07:12
woocommerce/woocommerce-gateway-stripe
https://api.github.com/repos/woocommerce/woocommerce-gateway-stripe
closed
Copy WCS Stripe OAuth code into this plugin
Priority: High
Right now, to use Stripe Connect (Standard) to manage things like automated key copying, you have to have the WooCommerce Services (WCS) plugin installed. To make connecting to our Stripe Connect Standard platform an option for more merchants, let's add that code here. I say "copy/add" (and not "move") because we probably shouldn't remove that code from WooCommerce Services without first considering any impact it has to Stores on WordPress.com merchants. Testing needs to include - what happens if a merchant already has WCS and then OAuths through this code (i.e. in the Stripe extension) - what happens if a merchant later activates WCS (whether or not they OAuth through this code) - what happens if a merchant deactivates WCS after OAuthing through WCS - will the the connection still work through this code? cc @thenbrent @chickenn00dle @bmccotter
1.0
Copy WCS Stripe OAuth code into this plugin - Right now, to use Stripe Connect (Standard) to manage things like automated key copying, you have to have the WooCommerce Services (WCS) plugin installed. To make connecting to our Stripe Connect Standard platform an option for more merchants, let's add that code here. I say "copy/add" (and not "move") because we probably shouldn't remove that code from WooCommerce Services without first considering any impact it has to Stores on WordPress.com merchants. Testing needs to include - what happens if a merchant already has WCS and then OAuths through this code (i.e. in the Stripe extension) - what happens if a merchant later activates WCS (whether or not they OAuth through this code) - what happens if a merchant deactivates WCS after OAuthing through WCS - will the the connection still work through this code? cc @thenbrent @chickenn00dle @bmccotter
non_test
copy wcs stripe oauth code into this plugin right now to use stripe connect standard to manage things like automated key copying you have to have the woocommerce services wcs plugin installed to make connecting to our stripe connect standard platform an option for more merchants let s add that code here i say copy add and not move because we probably shouldn t remove that code from woocommerce services without first considering any impact it has to stores on wordpress com merchants testing needs to include what happens if a merchant already has wcs and then oauths through this code i e in the stripe extension what happens if a merchant later activates wcs whether or not they oauth through this code what happens if a merchant deactivates wcs after oauthing through wcs will the the connection still work through this code cc thenbrent bmccotter
0
13,484
23,200,811,393
IssuesEvent
2022-08-01 21:15:13
vectordotdev/vector
https://api.github.com/repos/vectordotdev/vector
closed
Add fuzzing tests to Vector
domain: tests needs: requirements domain: security have: nice
In order to improve Vector's security and stability, we can add fuzzing tests to Vector. As @lukesteensen [mentioned](https://github.com/timberio/vector/issues/843#issuecomment-532408827), fuzzing was already used to harden individual components of Vector, such as the `tokenizer` transform. However, it might be interesting to invest more in this. For example we can start from defining our general approach to fuzzing which can then be applied to different components. Such approach is to be specified and details can change, but I have in mind something like this: * Agree which fuzzer(s) to use and how to call them (on entire Vector binary or on individual functions). * Define inputs which are to be fuzzed for different components. These definitions can be stored either in TOML files or as Rust code. * Create a Dockerfile which would setup everything necessary for reproducible running the fuzzying process. Also, having either a service in the `docker-compose` file (see https://github.com/timberio/vector/issues/1613#issuecomment-579268809 for details) or a wrapping shell script in `scripts` could help with reproducible runs of the fuzzing tests. Some notes: I think it might be worth to focus on making it easy to maintain these tests and add fuzzing definitions for new components. If it comes come at cost of having not the most computationally efficient fuzzing process, it could be fine as long as we are not going to run these tests in CI.
1.0
Add fuzzing tests to Vector - In order to improve Vector's security and stability, we can add fuzzing tests to Vector. As @lukesteensen [mentioned](https://github.com/timberio/vector/issues/843#issuecomment-532408827), fuzzing was already used to harden individual components of Vector, such as the `tokenizer` transform. However, it might be interesting to invest more in this. For example we can start from defining our general approach to fuzzing which can then be applied to different components. Such approach is to be specified and details can change, but I have in mind something like this: * Agree which fuzzer(s) to use and how to call them (on entire Vector binary or on individual functions). * Define inputs which are to be fuzzed for different components. These definitions can be stored either in TOML files or as Rust code. * Create a Dockerfile which would setup everything necessary for reproducible running the fuzzying process. Also, having either a service in the `docker-compose` file (see https://github.com/timberio/vector/issues/1613#issuecomment-579268809 for details) or a wrapping shell script in `scripts` could help with reproducible runs of the fuzzing tests. Some notes: I think it might be worth to focus on making it easy to maintain these tests and add fuzzing definitions for new components. If it comes come at cost of having not the most computationally efficient fuzzing process, it could be fine as long as we are not going to run these tests in CI.
non_test
add fuzzing tests to vector in order to improve vector s security and stability we can add fuzzing tests to vector as lukesteensen fuzzing was already used to harden individual components of vector such as the tokenizer transform however it might be interesting to invest more in this for example we can start from defining our general approach to fuzzing which can then be applied to different components such approach is to be specified and details can change but i have in mind something like this agree which fuzzer s to use and how to call them on entire vector binary or on individual functions define inputs which are to be fuzzed for different components these definitions can be stored either in toml files or as rust code create a dockerfile which would setup everything necessary for reproducible running the fuzzying process also having either a service in the docker compose file see for details or a wrapping shell script in scripts could help with reproducible runs of the fuzzing tests some notes i think it might be worth to focus on making it easy to maintain these tests and add fuzzing definitions for new components if it comes come at cost of having not the most computationally efficient fuzzing process it could be fine as long as we are not going to run these tests in ci
0
26,627
5,277,659,538
IssuesEvent
2017-02-07 04:20:07
allinurl/goaccess
https://api.github.com/repos/allinurl/goaccess
closed
How to do realtime graphs with history saving option.
documentation question
Hi, i tried my best level to figureout how to implement realtime graph. i used below command :+1: # goaccess -f /var/log/httpd/access_log -o /usr/local/share/piwik/report.html --real-time-html --ws-url=212.43.0.66 but the page is not refreshing. i have environment which runs centos 7 with firewall and selinux enabled. what all things i need to do to get live repots ? >Regards, Ben
1.0
How to do realtime graphs with history saving option. - Hi, i tried my best level to figureout how to implement realtime graph. i used below command :+1: # goaccess -f /var/log/httpd/access_log -o /usr/local/share/piwik/report.html --real-time-html --ws-url=212.43.0.66 but the page is not refreshing. i have environment which runs centos 7 with firewall and selinux enabled. what all things i need to do to get live repots ? >Regards, Ben
non_test
how to do realtime graphs with history saving option hi i tried my best level to figureout how to implement realtime graph i used below command goaccess f var log httpd access log o usr local share piwik report html real time html ws url but the page is not refreshing i have environment which runs centos with firewall and selinux enabled what all things i need to do to get live repots regards ben
0
328,838
28,137,579,491
IssuesEvent
2023-04-01 15:08:56
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
closed
Fix raw_ops.test_tensorflow_Minimum
TensorFlow Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4548616395/jobs/8019827411" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4548616395/jobs/8019827411" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4548616395/jobs/8019827411" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4548616395/jobs/8019827411" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_raw_ops.py::test_tensorflow_Minimum[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T23:40:26.8904414Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8907790Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8912234Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8912694Z E Falsifying example: test_tensorflow_Minimum( 2023-03-28T23:40:26.8913069Z E dtype_and_x=(['bfloat16', 'bfloat16'], 2023-03-28T23:40:26.8913459Z E [array(-1, dtype=bfloat16), array(-1, dtype=bfloat16)]), 2023-03-28T23:40:26.8913794Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T23:40:26.8914073Z E num_positional_args=0, 2023-03-28T23:40:26.8914316Z E with_out=False, 2023-03-28T23:40:26.8914544Z E inplace=False, 2023-03-28T23:40:26.8914780Z E as_variable=[False], 2023-03-28T23:40:26.8915027Z E native_arrays=[False], 2023-03-28T23:40:26.8915294Z E generate_frontend_arrays=False, 2023-03-28T23:40:26.8915528Z E ), 2023-03-28T23:40:26.8916035Z E fn_tree='ivy.functional.frontends.tensorflow.raw_ops.Minimum', 2023-03-28T23:40:26.8916520Z E frontend='tensorflow', 2023-03-28T23:40:26.8916877Z E on_device='cpu', 2023-03-28T23:40:26.8917258Z E ) 2023-03-28T23:40:26.8917514Z E 2023-03-28T23:40:26.8918153Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2AAAkYGCGBEYzMwAAAAXwAF') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_raw_ops.py::test_tensorflow_Minimum[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T23:40:26.8904414Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8907790Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8912234Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8912694Z E Falsifying example: test_tensorflow_Minimum( 2023-03-28T23:40:26.8913069Z E dtype_and_x=(['bfloat16', 'bfloat16'], 2023-03-28T23:40:26.8913459Z E [array(-1, dtype=bfloat16), array(-1, dtype=bfloat16)]), 2023-03-28T23:40:26.8913794Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T23:40:26.8914073Z E num_positional_args=0, 2023-03-28T23:40:26.8914316Z E with_out=False, 2023-03-28T23:40:26.8914544Z E inplace=False, 2023-03-28T23:40:26.8914780Z E as_variable=[False], 2023-03-28T23:40:26.8915027Z E native_arrays=[False], 2023-03-28T23:40:26.8915294Z E generate_frontend_arrays=False, 2023-03-28T23:40:26.8915528Z E ), 2023-03-28T23:40:26.8916035Z E fn_tree='ivy.functional.frontends.tensorflow.raw_ops.Minimum', 2023-03-28T23:40:26.8916520Z E frontend='tensorflow', 2023-03-28T23:40:26.8916877Z E on_device='cpu', 2023-03-28T23:40:26.8917258Z E ) 2023-03-28T23:40:26.8917514Z E 2023-03-28T23:40:26.8918153Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2AAAkYGCGBEYzMwAAAAXwAF') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_raw_ops.py::test_tensorflow_Minimum[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T23:40:26.8904414Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8907790Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8912234Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8912694Z E Falsifying example: test_tensorflow_Minimum( 2023-03-28T23:40:26.8913069Z E dtype_and_x=(['bfloat16', 'bfloat16'], 2023-03-28T23:40:26.8913459Z E [array(-1, dtype=bfloat16), array(-1, dtype=bfloat16)]), 2023-03-28T23:40:26.8913794Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T23:40:26.8914073Z E num_positional_args=0, 2023-03-28T23:40:26.8914316Z E with_out=False, 2023-03-28T23:40:26.8914544Z E inplace=False, 2023-03-28T23:40:26.8914780Z E as_variable=[False], 2023-03-28T23:40:26.8915027Z E native_arrays=[False], 2023-03-28T23:40:26.8915294Z E generate_frontend_arrays=False, 2023-03-28T23:40:26.8915528Z E ), 2023-03-28T23:40:26.8916035Z E fn_tree='ivy.functional.frontends.tensorflow.raw_ops.Minimum', 2023-03-28T23:40:26.8916520Z E frontend='tensorflow', 2023-03-28T23:40:26.8916877Z E on_device='cpu', 2023-03-28T23:40:26.8917258Z E ) 2023-03-28T23:40:26.8917514Z E 2023-03-28T23:40:26.8918153Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2AAAkYGCGBEYzMwAAAAXwAF') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_raw_ops.py::test_tensorflow_Minimum[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T23:40:26.8904414Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8907790Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8912234Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8912694Z E Falsifying example: test_tensorflow_Minimum( 2023-03-28T23:40:26.8913069Z E dtype_and_x=(['bfloat16', 'bfloat16'], 2023-03-28T23:40:26.8913459Z E [array(-1, dtype=bfloat16), array(-1, dtype=bfloat16)]), 2023-03-28T23:40:26.8913794Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T23:40:26.8914073Z E num_positional_args=0, 2023-03-28T23:40:26.8914316Z E with_out=False, 2023-03-28T23:40:26.8914544Z E inplace=False, 2023-03-28T23:40:26.8914780Z E as_variable=[False], 2023-03-28T23:40:26.8915027Z E native_arrays=[False], 2023-03-28T23:40:26.8915294Z E generate_frontend_arrays=False, 2023-03-28T23:40:26.8915528Z E ), 2023-03-28T23:40:26.8916035Z E fn_tree='ivy.functional.frontends.tensorflow.raw_ops.Minimum', 2023-03-28T23:40:26.8916520Z E frontend='tensorflow', 2023-03-28T23:40:26.8916877Z E on_device='cpu', 2023-03-28T23:40:26.8917258Z E ) 2023-03-28T23:40:26.8917514Z E 2023-03-28T23:40:26.8918153Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2AAAkYGCGBEYzMwAAAAXwAF') as a decorator on your test case </details>
1.0
Fix raw_ops.test_tensorflow_Minimum - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4548616395/jobs/8019827411" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4548616395/jobs/8019827411" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4548616395/jobs/8019827411" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4548616395/jobs/8019827411" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_raw_ops.py::test_tensorflow_Minimum[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T23:40:26.8904414Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8907790Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8912234Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8912694Z E Falsifying example: test_tensorflow_Minimum( 2023-03-28T23:40:26.8913069Z E dtype_and_x=(['bfloat16', 'bfloat16'], 2023-03-28T23:40:26.8913459Z E [array(-1, dtype=bfloat16), array(-1, dtype=bfloat16)]), 2023-03-28T23:40:26.8913794Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T23:40:26.8914073Z E num_positional_args=0, 2023-03-28T23:40:26.8914316Z E with_out=False, 2023-03-28T23:40:26.8914544Z E inplace=False, 2023-03-28T23:40:26.8914780Z E as_variable=[False], 2023-03-28T23:40:26.8915027Z E native_arrays=[False], 2023-03-28T23:40:26.8915294Z E generate_frontend_arrays=False, 2023-03-28T23:40:26.8915528Z E ), 2023-03-28T23:40:26.8916035Z E fn_tree='ivy.functional.frontends.tensorflow.raw_ops.Minimum', 2023-03-28T23:40:26.8916520Z E frontend='tensorflow', 2023-03-28T23:40:26.8916877Z E on_device='cpu', 2023-03-28T23:40:26.8917258Z E ) 2023-03-28T23:40:26.8917514Z E 2023-03-28T23:40:26.8918153Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2AAAkYGCGBEYzMwAAAAXwAF') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_raw_ops.py::test_tensorflow_Minimum[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T23:40:26.8904414Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8907790Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8912234Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8912694Z E Falsifying example: test_tensorflow_Minimum( 2023-03-28T23:40:26.8913069Z E dtype_and_x=(['bfloat16', 'bfloat16'], 2023-03-28T23:40:26.8913459Z E [array(-1, dtype=bfloat16), array(-1, dtype=bfloat16)]), 2023-03-28T23:40:26.8913794Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T23:40:26.8914073Z E num_positional_args=0, 2023-03-28T23:40:26.8914316Z E with_out=False, 2023-03-28T23:40:26.8914544Z E inplace=False, 2023-03-28T23:40:26.8914780Z E as_variable=[False], 2023-03-28T23:40:26.8915027Z E native_arrays=[False], 2023-03-28T23:40:26.8915294Z E generate_frontend_arrays=False, 2023-03-28T23:40:26.8915528Z E ), 2023-03-28T23:40:26.8916035Z E fn_tree='ivy.functional.frontends.tensorflow.raw_ops.Minimum', 2023-03-28T23:40:26.8916520Z E frontend='tensorflow', 2023-03-28T23:40:26.8916877Z E on_device='cpu', 2023-03-28T23:40:26.8917258Z E ) 2023-03-28T23:40:26.8917514Z E 2023-03-28T23:40:26.8918153Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2AAAkYGCGBEYzMwAAAAXwAF') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_raw_ops.py::test_tensorflow_Minimum[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T23:40:26.8904414Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8907790Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8912234Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8912694Z E Falsifying example: test_tensorflow_Minimum( 2023-03-28T23:40:26.8913069Z E dtype_and_x=(['bfloat16', 'bfloat16'], 2023-03-28T23:40:26.8913459Z E [array(-1, dtype=bfloat16), array(-1, dtype=bfloat16)]), 2023-03-28T23:40:26.8913794Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T23:40:26.8914073Z E num_positional_args=0, 2023-03-28T23:40:26.8914316Z E with_out=False, 2023-03-28T23:40:26.8914544Z E inplace=False, 2023-03-28T23:40:26.8914780Z E as_variable=[False], 2023-03-28T23:40:26.8915027Z E native_arrays=[False], 2023-03-28T23:40:26.8915294Z E generate_frontend_arrays=False, 2023-03-28T23:40:26.8915528Z E ), 2023-03-28T23:40:26.8916035Z E fn_tree='ivy.functional.frontends.tensorflow.raw_ops.Minimum', 2023-03-28T23:40:26.8916520Z E frontend='tensorflow', 2023-03-28T23:40:26.8916877Z E on_device='cpu', 2023-03-28T23:40:26.8917258Z E ) 2023-03-28T23:40:26.8917514Z E 2023-03-28T23:40:26.8918153Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2AAAkYGCGBEYzMwAAAAXwAF') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_raw_ops.py::test_tensorflow_Minimum[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T23:40:26.8904414Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8907790Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8912234Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T23:40:26.8912694Z E Falsifying example: test_tensorflow_Minimum( 2023-03-28T23:40:26.8913069Z E dtype_and_x=(['bfloat16', 'bfloat16'], 2023-03-28T23:40:26.8913459Z E [array(-1, dtype=bfloat16), array(-1, dtype=bfloat16)]), 2023-03-28T23:40:26.8913794Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T23:40:26.8914073Z E num_positional_args=0, 2023-03-28T23:40:26.8914316Z E with_out=False, 2023-03-28T23:40:26.8914544Z E inplace=False, 2023-03-28T23:40:26.8914780Z E as_variable=[False], 2023-03-28T23:40:26.8915027Z E native_arrays=[False], 2023-03-28T23:40:26.8915294Z E generate_frontend_arrays=False, 2023-03-28T23:40:26.8915528Z E ), 2023-03-28T23:40:26.8916035Z E fn_tree='ivy.functional.frontends.tensorflow.raw_ops.Minimum', 2023-03-28T23:40:26.8916520Z E frontend='tensorflow', 2023-03-28T23:40:26.8916877Z E on_device='cpu', 2023-03-28T23:40:26.8917258Z E ) 2023-03-28T23:40:26.8917514Z E 2023-03-28T23:40:26.8918153Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2AAAkYGCGBEYzMwAAAAXwAF') as a decorator on your test case </details>
test
fix raw ops test tensorflow minimum tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test frontends test tensorflow test raw ops py test tensorflow minimum e typeerror to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map tensorflow nested map to native got an unexpected keyword argument dtype e falsifying example test tensorflow minimum e dtype and x e e test flags frontendfunctiontestflags e num positional args e with out false e inplace false e as variable e native arrays e generate frontend arrays false e e fn tree ivy functional frontends tensorflow raw ops minimum e frontend tensorflow e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test frontends test tensorflow test raw ops py test tensorflow minimum e typeerror to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map tensorflow nested map to native got an unexpected keyword argument dtype e falsifying example test tensorflow minimum e dtype and x e e test flags frontendfunctiontestflags e num positional args e with out false e inplace false e as variable e native arrays e generate frontend arrays false e e fn tree ivy functional frontends tensorflow raw ops minimum e frontend tensorflow e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test frontends test tensorflow test raw ops py test tensorflow minimum e typeerror to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map tensorflow nested map to native got an unexpected keyword argument dtype e falsifying example test tensorflow minimum e dtype and x e e test flags frontendfunctiontestflags e num positional args e with out false e inplace false e as variable e native arrays e generate frontend arrays false e e fn tree ivy functional frontends tensorflow raw ops minimum e frontend tensorflow e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test frontends test tensorflow test raw ops py test tensorflow minimum e typeerror to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map tensorflow nested map to native got an unexpected keyword argument dtype e falsifying example test tensorflow minimum e dtype and x e e test flags frontendfunctiontestflags e num positional args e with out false e inplace false e as variable e native arrays e generate frontend arrays false e e fn tree ivy functional frontends tensorflow raw ops minimum e frontend tensorflow e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case
1
38,402
5,185,240,111
IssuesEvent
2017-01-20 09:46:05
elastic/logstash
https://api.github.com/repos/elastic/logstash
closed
auto_reload test failed
test failure
https://logstash-ci.elastic.co/job/elastic+logstash+5.0+multijob-os-compatibility/os=opensuse/233/console ``` Failures: 1) LogStash::Agent#execute when auto_reload is true if state is clean should periodically reload_state Failure/Error: expect(subject).to receive(:reload_state!).at_least(3).times (#<LogStash::Agent:0x458babc3>).reload_state!(any args) expected: at least 3 times with any arguments received: 1 time with any arguments # /var/lib/jenkins/workspace/elastic+logstash+5.0+multijob-os-compatibility/os/opensuse/logstash-core/spec/logstash/agent_spec.rb:158:in `(root)' ```
1.0
auto_reload test failed - https://logstash-ci.elastic.co/job/elastic+logstash+5.0+multijob-os-compatibility/os=opensuse/233/console ``` Failures: 1) LogStash::Agent#execute when auto_reload is true if state is clean should periodically reload_state Failure/Error: expect(subject).to receive(:reload_state!).at_least(3).times (#<LogStash::Agent:0x458babc3>).reload_state!(any args) expected: at least 3 times with any arguments received: 1 time with any arguments # /var/lib/jenkins/workspace/elastic+logstash+5.0+multijob-os-compatibility/os/opensuse/logstash-core/spec/logstash/agent_spec.rb:158:in `(root)' ```
test
auto reload test failed failures logstash agent execute when auto reload is true if state is clean should periodically reload state failure error expect subject to receive reload state at least times reload state any args expected at least times with any arguments received time with any arguments var lib jenkins workspace elastic logstash multijob os compatibility os opensuse logstash core spec logstash agent spec rb in root
1
251,873
21,526,771,267
IssuesEvent
2022-04-28 19:17:54
dagster-io/dagster
https://api.github.com/repos/dagster-io/dagster
closed
Default values for configs are not surface anywhere in dagit
good second issue page:job-overview page:job-launchpad flow:dev+test flow:execute-jobs dagit
Right now if you specify a default value, it is not surface anywhere in dagit. We should display config both in right hand bar of the definition viewer as well as in the config editor.
1.0
Default values for configs are not surface anywhere in dagit - Right now if you specify a default value, it is not surface anywhere in dagit. We should display config both in right hand bar of the definition viewer as well as in the config editor.
test
default values for configs are not surface anywhere in dagit right now if you specify a default value it is not surface anywhere in dagit we should display config both in right hand bar of the definition viewer as well as in the config editor
1
287,526
8,816,464,025
IssuesEvent
2018-12-30 11:08:15
azerothcore/azerothcore-wotlk
https://api.github.com/repos/azerothcore/azerothcore-wotlk
opened
DB restructuring - world - creature
Priority - High
This is part of: https://github.com/azerothcore/azerothcore-wotlk/issues/1118 **General note**: not all structure differences can or should be imported. Every change must be first verified. If a structure change is not suitable, we just close the issue. --- ```diff --- Dumping structure for table acore_world.creature +-- Dumping structure for table tc_world.creature DROP TABLE IF EXISTS `creature`; CREATE TABLE IF NOT EXISTS `creature` ( `guid` int(10) unsigned NOT NULL AUTO_INCREMENT COMMENT 'Global Unique Identifier', `id` mediumint(8) unsigned NOT NULL DEFAULT '0' COMMENT 'Creature Identifier', `map` smallint(5) unsigned NOT NULL DEFAULT '0' COMMENT 'Map Identifier', + `zoneId` smallint(5) unsigned NOT NULL DEFAULT '0' COMMENT 'Zone Identifier', + `areaId` smallint(5) unsigned NOT NULL DEFAULT '0' COMMENT 'Area Identifier', `spawnMask` tinyint(3) unsigned NOT NULL DEFAULT '1', `phaseMask` int(10) unsigned NOT NULL DEFAULT '1', `modelid` mediumint(8) unsigned NOT NULL DEFAULT '0', @@ -235,13 +244,15 @@ CREATE TABLE IF NOT EXISTS `creature` ( `npcflag` int(10) unsigned NOT NULL DEFAULT '0', `unit_flags` int(10) unsigned NOT NULL DEFAULT '0', `dynamicflags` int(10) unsigned NOT NULL DEFAULT '0', + `ScriptName` char(64) DEFAULT '', + `VerifiedBuild` smallint(5) DEFAULT '0', PRIMARY KEY (`guid`), KEY `idx_map` (`map`), KEY `idx_id` (`id`) ```
1.0
DB restructuring - world - creature - This is part of: https://github.com/azerothcore/azerothcore-wotlk/issues/1118 **General note**: not all structure differences can or should be imported. Every change must be first verified. If a structure change is not suitable, we just close the issue. --- ```diff --- Dumping structure for table acore_world.creature +-- Dumping structure for table tc_world.creature DROP TABLE IF EXISTS `creature`; CREATE TABLE IF NOT EXISTS `creature` ( `guid` int(10) unsigned NOT NULL AUTO_INCREMENT COMMENT 'Global Unique Identifier', `id` mediumint(8) unsigned NOT NULL DEFAULT '0' COMMENT 'Creature Identifier', `map` smallint(5) unsigned NOT NULL DEFAULT '0' COMMENT 'Map Identifier', + `zoneId` smallint(5) unsigned NOT NULL DEFAULT '0' COMMENT 'Zone Identifier', + `areaId` smallint(5) unsigned NOT NULL DEFAULT '0' COMMENT 'Area Identifier', `spawnMask` tinyint(3) unsigned NOT NULL DEFAULT '1', `phaseMask` int(10) unsigned NOT NULL DEFAULT '1', `modelid` mediumint(8) unsigned NOT NULL DEFAULT '0', @@ -235,13 +244,15 @@ CREATE TABLE IF NOT EXISTS `creature` ( `npcflag` int(10) unsigned NOT NULL DEFAULT '0', `unit_flags` int(10) unsigned NOT NULL DEFAULT '0', `dynamicflags` int(10) unsigned NOT NULL DEFAULT '0', + `ScriptName` char(64) DEFAULT '', + `VerifiedBuild` smallint(5) DEFAULT '0', PRIMARY KEY (`guid`), KEY `idx_map` (`map`), KEY `idx_id` (`id`) ```
non_test
db restructuring world creature this is part of general note not all structure differences can or should be imported every change must be first verified if a structure change is not suitable we just close the issue diff dumping structure for table acore world creature dumping structure for table tc world creature drop table if exists creature create table if not exists creature guid int unsigned not null auto increment comment global unique identifier id mediumint unsigned not null default comment creature identifier map smallint unsigned not null default comment map identifier zoneid smallint unsigned not null default comment zone identifier areaid smallint unsigned not null default comment area identifier spawnmask tinyint unsigned not null default phasemask int unsigned not null default modelid mediumint unsigned not null default create table if not exists creature npcflag int unsigned not null default unit flags int unsigned not null default dynamicflags int unsigned not null default scriptname char default verifiedbuild smallint default primary key guid key idx map map key idx id id
0
120,038
10,090,919,785
IssuesEvent
2019-07-26 13:02:00
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Failing test: Chrome UI Functional Tests.test/functional/apps/dashboard/dashboard_time·js - dashboard app using legacy data dashboard time time changes preserved during navigation
Feature:Dashboard Team:KibanaApp failed-test
A test failed on a tracked branch ``` Error: retry.try timeout: TimeoutError: Waiting for element to be located By(css selector, [data-test-subj~="superDatePickerendDatePopoverButton"]) Wait timed out after 10016ms at node_modules/selenium-webdriver/lib/webdriver.js:841:17 at process._tickCallback (internal/process/next_tick.js:68:7) at lastError (test/common/services/retry/retry_for_success.ts:28:9) at onFailure (test/common/services/retry/retry_for_success.ts:68:13) ``` First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/JOB=kibana-ciGroup5,node=linux-immutable/1143/) <!-- kibanaCiData = {"failed-test":{"test.class":"Chrome UI Functional Tests.test/functional/apps/dashboard/dashboard_time·js","test.name":"dashboard app using legacy data dashboard time time changes preserved during navigation","test.failCount":7}} -->
1.0
Failing test: Chrome UI Functional Tests.test/functional/apps/dashboard/dashboard_time·js - dashboard app using legacy data dashboard time time changes preserved during navigation - A test failed on a tracked branch ``` Error: retry.try timeout: TimeoutError: Waiting for element to be located By(css selector, [data-test-subj~="superDatePickerendDatePopoverButton"]) Wait timed out after 10016ms at node_modules/selenium-webdriver/lib/webdriver.js:841:17 at process._tickCallback (internal/process/next_tick.js:68:7) at lastError (test/common/services/retry/retry_for_success.ts:28:9) at onFailure (test/common/services/retry/retry_for_success.ts:68:13) ``` First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/JOB=kibana-ciGroup5,node=linux-immutable/1143/) <!-- kibanaCiData = {"failed-test":{"test.class":"Chrome UI Functional Tests.test/functional/apps/dashboard/dashboard_time·js","test.name":"dashboard app using legacy data dashboard time time changes preserved during navigation","test.failCount":7}} -->
test
failing test chrome ui functional tests test functional apps dashboard dashboard time·js dashboard app using legacy data dashboard time time changes preserved during navigation a test failed on a tracked branch error retry try timeout timeouterror waiting for element to be located by css selector wait timed out after at node modules selenium webdriver lib webdriver js at process tickcallback internal process next tick js at lasterror test common services retry retry for success ts at onfailure test common services retry retry for success ts first failure
1
229,356
18,293,059,485
IssuesEvent
2021-10-05 17:19:23
FTBTeam/FTB-Quests
https://api.github.com/repos/FTBTeam/FTB-Quests
closed
[bug] bad server performance
bug priority: high needs testing
To help us help you, please provide as much information as you can, this helps us with debugging and reproducting the issue. **Mod & Pack version** ftb-quests-forge-1605.3.3-build.49 **Forge / Fabric version, if applicable** ftb-quests-forge-1605.3.3-build.49 **How to reproduce the issue** 1) Install the autoclicker 2) click on quests where there are tasks for crafting or for finding an item in your inventory 3) get a drop in tps on the server to 2-3 Simply put, launch about 30 players on a server with quests and get terrible profiler results the same was on version 1.12.2 if there are at least 100 quests on the server **A screenshot, if applicable** https://github.com/FTBTeam/FTB-Quests/blob/main/common/src/main/java/dev/ftb/mods/ftbquests/util/FTBQuestsInventoryListener.java#L27 ![image](https://user-images.githubusercontent.com/17977139/130367946-f58aa4bd-82db-4ab8-8670-bd10f0dde65f.png) ![image](https://user-images.githubusercontent.com/17977139/130367972-f4ab2d27-231b-4551-a565-58657d0c89d2.png) Spark profiler https://spark.lucko.me/zNFHVFFCDz
1.0
[bug] bad server performance - To help us help you, please provide as much information as you can, this helps us with debugging and reproducting the issue. **Mod & Pack version** ftb-quests-forge-1605.3.3-build.49 **Forge / Fabric version, if applicable** ftb-quests-forge-1605.3.3-build.49 **How to reproduce the issue** 1) Install the autoclicker 2) click on quests where there are tasks for crafting or for finding an item in your inventory 3) get a drop in tps on the server to 2-3 Simply put, launch about 30 players on a server with quests and get terrible profiler results the same was on version 1.12.2 if there are at least 100 quests on the server **A screenshot, if applicable** https://github.com/FTBTeam/FTB-Quests/blob/main/common/src/main/java/dev/ftb/mods/ftbquests/util/FTBQuestsInventoryListener.java#L27 ![image](https://user-images.githubusercontent.com/17977139/130367946-f58aa4bd-82db-4ab8-8670-bd10f0dde65f.png) ![image](https://user-images.githubusercontent.com/17977139/130367972-f4ab2d27-231b-4551-a565-58657d0c89d2.png) Spark profiler https://spark.lucko.me/zNFHVFFCDz
test
bad server performance to help us help you please provide as much information as you can this helps us with debugging and reproducting the issue mod pack version ftb quests forge build forge fabric version if applicable ftb quests forge build how to reproduce the issue install the autoclicker click on quests where there are tasks for crafting or for finding an item in your inventory get a drop in tps on the server to simply put launch about players on a server with quests and get terrible profiler results the same was on version if there are at least quests on the server a screenshot if applicable spark profiler
1
746,781
26,045,402,061
IssuesEvent
2022-12-22 14:00:34
Laravel-Backpack/Generators
https://api.github.com/repos/Laravel-Backpack/Generators
closed
[bug] generating nested controllers and models
bug Priority: SHOULD
If I write in small letters: `php artisan backpack:crud post/badge` Creates classes with small letters (badgeCrudController). If I write in capital letters: `php artisan backpack:crud Post/Badge` Cannot add CrudTrait to model. And an error occurs: ``` Illuminate\Contracts\Container\BindingResolutionException Target class [App\Http\Controllers\Admin\Post/BadgeCrudController] does not exist. ``` Because of the slash in the 'Post/BadgeCrudController' routes. Also, does not find the model in the controller due to the wrong namespace. `CRUD::setModel(\App\Models\Badge::class);` Although the model is located here \App\Models\Post\Badge::class Now I checked it - it does not display entries for models in folders, it gives a 404 error. ### Backpack, Laravel, PHP, DB version When I run ```php artisan backpack:version``` the output is: ### PHP VERSION: PHP 8.1.10 (cli) (built: Aug 30 2022 18:05:49) (ZTS Visual C++ 2019 x64) Copyright (c) The PHP Group Zend Engine v4.1.10, Copyright (c) Zend Technologies ### LARAVEL VERSION: v8.83.25@b77b908a9426efa41d6286a2ef4c3adbf5398ca1 ### BACKPACK VERSION: 5.4.7@6e1bb116de9f3091530cb9b11edcceb4d252daa8
1.0
[bug] generating nested controllers and models - If I write in small letters: `php artisan backpack:crud post/badge` Creates classes with small letters (badgeCrudController). If I write in capital letters: `php artisan backpack:crud Post/Badge` Cannot add CrudTrait to model. And an error occurs: ``` Illuminate\Contracts\Container\BindingResolutionException Target class [App\Http\Controllers\Admin\Post/BadgeCrudController] does not exist. ``` Because of the slash in the 'Post/BadgeCrudController' routes. Also, does not find the model in the controller due to the wrong namespace. `CRUD::setModel(\App\Models\Badge::class);` Although the model is located here \App\Models\Post\Badge::class Now I checked it - it does not display entries for models in folders, it gives a 404 error. ### Backpack, Laravel, PHP, DB version When I run ```php artisan backpack:version``` the output is: ### PHP VERSION: PHP 8.1.10 (cli) (built: Aug 30 2022 18:05:49) (ZTS Visual C++ 2019 x64) Copyright (c) The PHP Group Zend Engine v4.1.10, Copyright (c) Zend Technologies ### LARAVEL VERSION: v8.83.25@b77b908a9426efa41d6286a2ef4c3adbf5398ca1 ### BACKPACK VERSION: 5.4.7@6e1bb116de9f3091530cb9b11edcceb4d252daa8
non_test
generating nested controllers and models if i write in small letters php artisan backpack crud post badge creates classes with small letters badgecrudcontroller if i write in capital letters php artisan backpack crud post badge cannot add crudtrait to model and an error occurs illuminate contracts container bindingresolutionexception target class does not exist because of the slash in the post badgecrudcontroller routes also does not find the model in the controller due to the wrong namespace crud setmodel app models badge class although the model is located here app models post badge class now i checked it it does not display entries for models in folders it gives a error backpack laravel php db version when i run php artisan backpack version the output is php version php cli built aug zts visual c copyright c the php group zend engine copyright c zend technologies laravel version backpack version
0
329,763
28,305,782,965
IssuesEvent
2023-04-10 10:51:47
wazuh/wazuh
https://api.github.com/repos/wazuh/wazuh
closed
Release 4.4.1 - Release Candidate 1 - Python unit tests
module/framework type/test module/api level/task release test/4.4.1
The following issue aims to run all `python unit tests` for the current release candidate, report the results, and open new issues for any encountered errors. ## Python unit tests information | | | |--------------------------------------|--------------------------------------------| | **Main release candidate issue** |https://github.com/wazuh/wazuh/issues/16620| | **Version** | 4.4.1| | **Release candidate #** |1| | **Tag** |[v4.4.1-rc1](https://github.com/wazuh/wazuh/tree/v4.4.1-rc1)| | **Previous python unit tests issue** |[#16426](https://github.com/wazuh/wazuh/issues/16426)| ## Test report procedure All individual test checks must be marked as: | | | |---------------------------------|--------------------------------------------| | Pass | The test ran successfully. | | Xfail | The test was expected to fail and it failed. It must be properly justified and reported in an issue. | | Skip | The test was not run. It must be properly justified and reported in an issue. | | Fail | The test failed. A new issue must be opened to evaluate and address the problem. | All test results must have one the following statuses: | | | |---------------------------------|--------------------------------------------| | :green_circle: | All checks passed. | | :red_circle: | There is at least one failed check. | | :yellow_circle: | There is at least one expected fail or skipped test and no failures. | Any failing test must be properly addressed with a new issue, detailing the error and the possible cause. It must be included in the `Fixes` section of the current release candidate main issue. Any expected fail or skipped test must have an issue justifying the reason. All auditors must validate the justification for an expected fail or skipped test. An extended report of the test results must be attached as a zip or txt. This report can be used by the auditors to dig deeper into any possible failures and details. As for the coverage results, they must be added split by modules. Test files must not be included in the coverage report. All test coverage results must have one of the following statuses: | | | |-----------------|--------| | :green_square: | >= 75% | | :yellow_square: | >= 50% | | :orange_square: | >= 25% | | :red_square: | < 25% | ## Conclusions All tests have been executed and the results can be found [here](https://github.com/wazuh/wazuh/issues/16596#issuecomment-1494314131). All tests have passed and the fails have been reported or justified. I therefore conclude that this issue is finished and OK for this release candidate. | | | | |----------------------------------------|--------------|------------| | | **Coverage** | **Status** | | **Overall python unit tests coverage** |85|🟩| ## Auditors validation The definition of done for this one is the validation of the conclusions and the test results from all auditors. All checks from below must be accepted in order to close this issue. - [x] @davidjiglesias - [x] @Selutario
2.0
Release 4.4.1 - Release Candidate 1 - Python unit tests - The following issue aims to run all `python unit tests` for the current release candidate, report the results, and open new issues for any encountered errors. ## Python unit tests information | | | |--------------------------------------|--------------------------------------------| | **Main release candidate issue** |https://github.com/wazuh/wazuh/issues/16620| | **Version** | 4.4.1| | **Release candidate #** |1| | **Tag** |[v4.4.1-rc1](https://github.com/wazuh/wazuh/tree/v4.4.1-rc1)| | **Previous python unit tests issue** |[#16426](https://github.com/wazuh/wazuh/issues/16426)| ## Test report procedure All individual test checks must be marked as: | | | |---------------------------------|--------------------------------------------| | Pass | The test ran successfully. | | Xfail | The test was expected to fail and it failed. It must be properly justified and reported in an issue. | | Skip | The test was not run. It must be properly justified and reported in an issue. | | Fail | The test failed. A new issue must be opened to evaluate and address the problem. | All test results must have one the following statuses: | | | |---------------------------------|--------------------------------------------| | :green_circle: | All checks passed. | | :red_circle: | There is at least one failed check. | | :yellow_circle: | There is at least one expected fail or skipped test and no failures. | Any failing test must be properly addressed with a new issue, detailing the error and the possible cause. It must be included in the `Fixes` section of the current release candidate main issue. Any expected fail or skipped test must have an issue justifying the reason. All auditors must validate the justification for an expected fail or skipped test. An extended report of the test results must be attached as a zip or txt. This report can be used by the auditors to dig deeper into any possible failures and details. As for the coverage results, they must be added split by modules. Test files must not be included in the coverage report. All test coverage results must have one of the following statuses: | | | |-----------------|--------| | :green_square: | >= 75% | | :yellow_square: | >= 50% | | :orange_square: | >= 25% | | :red_square: | < 25% | ## Conclusions All tests have been executed and the results can be found [here](https://github.com/wazuh/wazuh/issues/16596#issuecomment-1494314131). All tests have passed and the fails have been reported or justified. I therefore conclude that this issue is finished and OK for this release candidate. | | | | |----------------------------------------|--------------|------------| | | **Coverage** | **Status** | | **Overall python unit tests coverage** |85|🟩| ## Auditors validation The definition of done for this one is the validation of the conclusions and the test results from all auditors. All checks from below must be accepted in order to close this issue. - [x] @davidjiglesias - [x] @Selutario
test
release release candidate python unit tests the following issue aims to run all python unit tests for the current release candidate report the results and open new issues for any encountered errors python unit tests information main release candidate issue version release candidate tag previous python unit tests issue test report procedure all individual test checks must be marked as pass the test ran successfully xfail the test was expected to fail and it failed it must be properly justified and reported in an issue skip the test was not run it must be properly justified and reported in an issue fail the test failed a new issue must be opened to evaluate and address the problem all test results must have one the following statuses green circle all checks passed red circle there is at least one failed check yellow circle there is at least one expected fail or skipped test and no failures any failing test must be properly addressed with a new issue detailing the error and the possible cause it must be included in the fixes section of the current release candidate main issue any expected fail or skipped test must have an issue justifying the reason all auditors must validate the justification for an expected fail or skipped test an extended report of the test results must be attached as a zip or txt this report can be used by the auditors to dig deeper into any possible failures and details as for the coverage results they must be added split by modules test files must not be included in the coverage report all test coverage results must have one of the following statuses green square yellow square orange square red square conclusions all tests have been executed and the results can be found all tests have passed and the fails have been reported or justified i therefore conclude that this issue is finished and ok for this release candidate coverage status overall python unit tests coverage 🟩 auditors validation the definition of done for this one is the validation of the conclusions and the test results from all auditors all checks from below must be accepted in order to close this issue davidjiglesias selutario
1
94,233
8,476,902,313
IssuesEvent
2018-10-25 00:03:19
Brycey92/Galaxy-Craft-Issues
https://api.github.com/repos/Brycey92/Galaxy-Craft-Issues
closed
Food sucks
fixed - needs testing priority
**Pack version** 1.0.3 **Describe the bug** All food has low hunger/saturation restoration. **Expected behavior** The values should be nerfed from vanilla, but not this ridiculously. It's time to muck around in the HungerOverhaul config.
1.0
Food sucks - **Pack version** 1.0.3 **Describe the bug** All food has low hunger/saturation restoration. **Expected behavior** The values should be nerfed from vanilla, but not this ridiculously. It's time to muck around in the HungerOverhaul config.
test
food sucks pack version describe the bug all food has low hunger saturation restoration expected behavior the values should be nerfed from vanilla but not this ridiculously it s time to muck around in the hungeroverhaul config
1
326,745
28,016,000,500
IssuesEvent
2023-03-27 22:46:42
iotaledger/firefly
https://api.github.com/repos/iotaledger/firefly
closed
Unit / integration test coverage
scope:ci type:tests
### Issue description Firefly has lots of source code files for both business logic and UI / component logic that remains untested (in a "unit test" sense; we have audits regularly and releases are tested by a group of community testers). It would be good to increase unit test coverage gradually in dedicated PRs that contain chunks of unit tests. This will most likely have to been done for `migration.ts` and `wallet.rs` as they contain so much functionality. #### Some thoughts - It may be worth thinking splitting up the large 1000+ line files into sub-directories with smaller source code files - As the collection of source code files in `packages/shared/lib` grows, it would be helpful for us to come up with some organizational system for them or group them differently... need to give this one some more thought 🧠 - i.e. `packages/shared/lib/app/` - `index.ts` - `settings.ts` - `updater.ts` - AFAIK TypeScript files are usually named in a `kebab-case` format rather than `camelCase`, how the Svelte component files named... should we switch to this convention? ### Firefly version `develop` ### OS details Catalina (10.15.7) ### Can the issue be reliably reproduced? Yes ### Steps to reproduce the issue 1. Write a test in `package/shared/lib/tests/` (refer to the ["Writing tests"](https://github.com/iotaledger/firefly/wiki/Testing-Guide#writing-tests) section of the testing guide wiki) 2. Run `yarn test` in the root directory of the repository ### Coverage The following files are `*.ts` files in the `packages/shared/lib/` directory. This is where most of the business logic of Firefly lives. #### Required The following functions and files __must__ have unit test coverage: - `app.ts` - [ ] `cleanupSignup` - [ ] `logout` - `ledger.ts` - [ ] `getLedgerDeviceStatus` - [ ] `isLedgerConnected` - [ ] `calculateLedgerDeviceState` - [ ] `displayNotificationForLedgerProfile` - [ ] `isLedgerError` - [ ] `getLegacyErrorMessage` - `migration.ts` - [ ] everything... will probably need a dedicated PR - `network.ts` - [x] everything, but depends on developer mode / network configuration feature because current network code has less logic / less strict typing - `profile.ts` - [ ] `saveProfile` - [ ] `createProfile` - [ ] `disposeNewProfile` - [ ] `updateProfile` - [ ] `removeProfileFolder` - Unsure of what testing may look for accessing dev / user machine storage - [ ] `cleanupEmptyProfiles` - `units.ts` - [ ] `formatUnitBestMatch` - [ ] `formatUnitPrecision` - [ ] `changeUnits` - `wallet.rs` - [ ] everything... although not sure how testing / mocking the wallet.rs `api` will work #### Optional The following functions and files __should__ have unit test coverage: - `bech32.ts` - [ ] `encode` - [ ] `encode5BitArray` - [ ] `decode` - [ ] `decodeTo5BitArray` - [ ] `to5Bit` - [ ] `from5Bit` - [ ] `matches` - `chart.ts` - [ ] `getChartDataFromBalanceHistory` - [ ] `getChartDataForTokenValue` - [ ] `getAccountActivityData` - `currency.ts` - [x] `convertToFiat` - [x] `convertFromFiat` - [x] `isFiatCurrency` - [ ] `formatCurrencyValue` - [ ] `getDecimalSeparator` - [ ] `getCurrencyPosition` - [ ] `getGroupSeparator` - [ ] `parseCurrency` - [ ] `formatCurrency` - [ ] `formatNumber` - [ ] `ensureZeros` - [ ] `replaceCurrencyDecimal` - `errors.ts` - [x] `addError` - `helpers.ts` - [ ] `getTrimmedLength` - [ ] `validateFilenameChars` - [ ] `getInitials` - `market.ts` - [ ] `addProfileCurrencyPriceData` - `notification.ts` - [ ] `isNewNotification` - [ ] `removeDisplayNotification` - `popup.ts` - [ ] `openPopup` - [ ] `closePopup` - `router.ts` - [ ] `routerNext` - [ ] `forceNextRoute` - [ ] `routerPrevious` - [ ] `resetRouter` - `utils.ts` - [ ] `validatePinFormat` - [ ] `generateRandomId` - [ ] `parseDeepLink`* - [ ] `parseAddress`* - [ ] `isValidUrl` - [ ] `isValidHttpsUrl` - [ ] `validateBech32Address` - [ ] `setClipboard` - Is this worth testing? Any edge cases? \* These might be best in a specific source code file so we can find them more easily. Or we could create a `packages/shared/lib/utils/` directory with `parse.ts` or something like that. For example: - `packages/shared/lib/utils/` - `parsing.ts` - `validation.ts` - ... It also might be better to stick to a standardized format for function names and ask ourselves a few questions: - Should we use `isValidNOUN` or `validateNOUN` as a validation format for data? If they return a `bool` then I think it makes sense to use `isValidNOUN` as it's like a yes/no question. If it returns `NOUN` (or maybe `void`) then `validateNOUN` is better as it's more descriptive. - ...
1.0
Unit / integration test coverage - ### Issue description Firefly has lots of source code files for both business logic and UI / component logic that remains untested (in a "unit test" sense; we have audits regularly and releases are tested by a group of community testers). It would be good to increase unit test coverage gradually in dedicated PRs that contain chunks of unit tests. This will most likely have to been done for `migration.ts` and `wallet.rs` as they contain so much functionality. #### Some thoughts - It may be worth thinking splitting up the large 1000+ line files into sub-directories with smaller source code files - As the collection of source code files in `packages/shared/lib` grows, it would be helpful for us to come up with some organizational system for them or group them differently... need to give this one some more thought 🧠 - i.e. `packages/shared/lib/app/` - `index.ts` - `settings.ts` - `updater.ts` - AFAIK TypeScript files are usually named in a `kebab-case` format rather than `camelCase`, how the Svelte component files named... should we switch to this convention? ### Firefly version `develop` ### OS details Catalina (10.15.7) ### Can the issue be reliably reproduced? Yes ### Steps to reproduce the issue 1. Write a test in `package/shared/lib/tests/` (refer to the ["Writing tests"](https://github.com/iotaledger/firefly/wiki/Testing-Guide#writing-tests) section of the testing guide wiki) 2. Run `yarn test` in the root directory of the repository ### Coverage The following files are `*.ts` files in the `packages/shared/lib/` directory. This is where most of the business logic of Firefly lives. #### Required The following functions and files __must__ have unit test coverage: - `app.ts` - [ ] `cleanupSignup` - [ ] `logout` - `ledger.ts` - [ ] `getLedgerDeviceStatus` - [ ] `isLedgerConnected` - [ ] `calculateLedgerDeviceState` - [ ] `displayNotificationForLedgerProfile` - [ ] `isLedgerError` - [ ] `getLegacyErrorMessage` - `migration.ts` - [ ] everything... will probably need a dedicated PR - `network.ts` - [x] everything, but depends on developer mode / network configuration feature because current network code has less logic / less strict typing - `profile.ts` - [ ] `saveProfile` - [ ] `createProfile` - [ ] `disposeNewProfile` - [ ] `updateProfile` - [ ] `removeProfileFolder` - Unsure of what testing may look for accessing dev / user machine storage - [ ] `cleanupEmptyProfiles` - `units.ts` - [ ] `formatUnitBestMatch` - [ ] `formatUnitPrecision` - [ ] `changeUnits` - `wallet.rs` - [ ] everything... although not sure how testing / mocking the wallet.rs `api` will work #### Optional The following functions and files __should__ have unit test coverage: - `bech32.ts` - [ ] `encode` - [ ] `encode5BitArray` - [ ] `decode` - [ ] `decodeTo5BitArray` - [ ] `to5Bit` - [ ] `from5Bit` - [ ] `matches` - `chart.ts` - [ ] `getChartDataFromBalanceHistory` - [ ] `getChartDataForTokenValue` - [ ] `getAccountActivityData` - `currency.ts` - [x] `convertToFiat` - [x] `convertFromFiat` - [x] `isFiatCurrency` - [ ] `formatCurrencyValue` - [ ] `getDecimalSeparator` - [ ] `getCurrencyPosition` - [ ] `getGroupSeparator` - [ ] `parseCurrency` - [ ] `formatCurrency` - [ ] `formatNumber` - [ ] `ensureZeros` - [ ] `replaceCurrencyDecimal` - `errors.ts` - [x] `addError` - `helpers.ts` - [ ] `getTrimmedLength` - [ ] `validateFilenameChars` - [ ] `getInitials` - `market.ts` - [ ] `addProfileCurrencyPriceData` - `notification.ts` - [ ] `isNewNotification` - [ ] `removeDisplayNotification` - `popup.ts` - [ ] `openPopup` - [ ] `closePopup` - `router.ts` - [ ] `routerNext` - [ ] `forceNextRoute` - [ ] `routerPrevious` - [ ] `resetRouter` - `utils.ts` - [ ] `validatePinFormat` - [ ] `generateRandomId` - [ ] `parseDeepLink`* - [ ] `parseAddress`* - [ ] `isValidUrl` - [ ] `isValidHttpsUrl` - [ ] `validateBech32Address` - [ ] `setClipboard` - Is this worth testing? Any edge cases? \* These might be best in a specific source code file so we can find them more easily. Or we could create a `packages/shared/lib/utils/` directory with `parse.ts` or something like that. For example: - `packages/shared/lib/utils/` - `parsing.ts` - `validation.ts` - ... It also might be better to stick to a standardized format for function names and ask ourselves a few questions: - Should we use `isValidNOUN` or `validateNOUN` as a validation format for data? If they return a `bool` then I think it makes sense to use `isValidNOUN` as it's like a yes/no question. If it returns `NOUN` (or maybe `void`) then `validateNOUN` is better as it's more descriptive. - ...
test
unit integration test coverage issue description firefly has lots of source code files for both business logic and ui component logic that remains untested in a unit test sense we have audits regularly and releases are tested by a group of community testers it would be good to increase unit test coverage gradually in dedicated prs that contain chunks of unit tests this will most likely have to been done for migration ts and wallet rs as they contain so much functionality some thoughts it may be worth thinking splitting up the large line files into sub directories with smaller source code files as the collection of source code files in packages shared lib grows it would be helpful for us to come up with some organizational system for them or group them differently need to give this one some more thought 🧠 i e packages shared lib app index ts settings ts updater ts afaik typescript files are usually named in a kebab case format rather than camelcase how the svelte component files named should we switch to this convention firefly version develop os details catalina can the issue be reliably reproduced yes steps to reproduce the issue write a test in package shared lib tests refer to the section of the testing guide wiki run yarn test in the root directory of the repository coverage the following files are ts files in the packages shared lib directory this is where most of the business logic of firefly lives required the following functions and files must have unit test coverage app ts cleanupsignup logout ledger ts getledgerdevicestatus isledgerconnected calculateledgerdevicestate displaynotificationforledgerprofile isledgererror getlegacyerrormessage migration ts everything will probably need a dedicated pr network ts everything but depends on developer mode network configuration feature because current network code has less logic less strict typing profile ts saveprofile createprofile disposenewprofile updateprofile removeprofilefolder unsure of what testing may look for accessing dev user machine storage cleanupemptyprofiles units ts formatunitbestmatch formatunitprecision changeunits wallet rs everything although not sure how testing mocking the wallet rs api will work optional the following functions and files should have unit test coverage ts encode decode matches chart ts getchartdatafrombalancehistory getchartdatafortokenvalue getaccountactivitydata currency ts converttofiat convertfromfiat isfiatcurrency formatcurrencyvalue getdecimalseparator getcurrencyposition getgroupseparator parsecurrency formatcurrency formatnumber ensurezeros replacecurrencydecimal errors ts adderror helpers ts gettrimmedlength validatefilenamechars getinitials market ts addprofilecurrencypricedata notification ts isnewnotification removedisplaynotification popup ts openpopup closepopup router ts routernext forcenextroute routerprevious resetrouter utils ts validatepinformat generaterandomid parsedeeplink parseaddress isvalidurl isvalidhttpsurl setclipboard is this worth testing any edge cases these might be best in a specific source code file so we can find them more easily or we could create a packages shared lib utils directory with parse ts or something like that for example packages shared lib utils parsing ts validation ts it also might be better to stick to a standardized format for function names and ask ourselves a few questions should we use isvalidnoun or validatenoun as a validation format for data if they return a bool then i think it makes sense to use isvalidnoun as it s like a yes no question if it returns noun or maybe void then validatenoun is better as it s more descriptive
1
346,440
30,914,013,449
IssuesEvent
2023-08-05 03:40:21
sphinx-doc/sphinx
https://api.github.com/repos/sphinx-doc/sphinx
closed
1.3.5 - make test fails with Python 2.7.10
type:tests
Tried to package 1.3.5 for Python 2.7.10. But "make test" fails: https://build.opensuse.org/build/home:stroeder:branches:devel:languages:python/openSUSE_Tumbleweed/x86_64/python-Sphinx/_log
1.0
1.3.5 - make test fails with Python 2.7.10 - Tried to package 1.3.5 for Python 2.7.10. But "make test" fails: https://build.opensuse.org/build/home:stroeder:branches:devel:languages:python/openSUSE_Tumbleweed/x86_64/python-Sphinx/_log
test
make test fails with python tried to package for python but make test fails
1
139,591
11,274,491,591
IssuesEvent
2020-01-14 18:41:12
aliasrobotics/RVD
https://api.github.com/repos/aliasrobotics/RVD
opened
(error) Buffer is accessed out of bounds
bug cppcheck static analysis testing triage
```yaml { "id": 1, "cve": "None", "description": "[src/industrial_calibration/industrial_extrinsic_cal/src/nodes/stereo_cal_srv.cpp:301]: (error) Buffer is accessed out of bounds: image_scene_chars", "system": "src/industrial_calibration/industrial_extrinsic_cal/src/nodes/stereo_cal_srv.cpp", "title": "(error) Buffer is accessed out of bounds", "severity": { "severity-description": "", "cvss-score": 0, "rvss-vector": "", "rvss-score": 0, "cvss-vector": "" }, "vendor": null, "exploitation": { "exploitation-image": "", "exploitation-vector": "", "description": "" }, "type": "bug", "mitigation": { "date-mitigation": "", "pull-request": "", "description": "" }, "links": "", "keywords": [ "cppcheck", "static analysis", "testing", "triage", "bug" ], "cwe": "None", "flaw": { "application": "N/A", "phase": "testing", "languages": "None", "reproduction": "See artifacts below (if available)", "specificity": "N/A", "date-detected": "2020-01-14 (18:41)", "architectural-location": "N/A", "subsystem": "N/A", "reproduction-image": "gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_ros_industrial/-/jobs/402982901/artifacts/download", "date-reported": "2020-01-14 (18:41)", "reproducibility": "always", "detected-by-method": "testing static", "reported-by-relationship": "automatic", "trace": "", "reported-by": "Alias Robotics", "issue": "", "package": "N/A", "detected-by": "Alias Robotics" } } ```
1.0
(error) Buffer is accessed out of bounds - ```yaml { "id": 1, "cve": "None", "description": "[src/industrial_calibration/industrial_extrinsic_cal/src/nodes/stereo_cal_srv.cpp:301]: (error) Buffer is accessed out of bounds: image_scene_chars", "system": "src/industrial_calibration/industrial_extrinsic_cal/src/nodes/stereo_cal_srv.cpp", "title": "(error) Buffer is accessed out of bounds", "severity": { "severity-description": "", "cvss-score": 0, "rvss-vector": "", "rvss-score": 0, "cvss-vector": "" }, "vendor": null, "exploitation": { "exploitation-image": "", "exploitation-vector": "", "description": "" }, "type": "bug", "mitigation": { "date-mitigation": "", "pull-request": "", "description": "" }, "links": "", "keywords": [ "cppcheck", "static analysis", "testing", "triage", "bug" ], "cwe": "None", "flaw": { "application": "N/A", "phase": "testing", "languages": "None", "reproduction": "See artifacts below (if available)", "specificity": "N/A", "date-detected": "2020-01-14 (18:41)", "architectural-location": "N/A", "subsystem": "N/A", "reproduction-image": "gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_ros_industrial/-/jobs/402982901/artifacts/download", "date-reported": "2020-01-14 (18:41)", "reproducibility": "always", "detected-by-method": "testing static", "reported-by-relationship": "automatic", "trace": "", "reported-by": "Alias Robotics", "issue": "", "package": "N/A", "detected-by": "Alias Robotics" } } ```
test
error buffer is accessed out of bounds yaml id cve none description error buffer is accessed out of bounds image scene chars system src industrial calibration industrial extrinsic cal src nodes stereo cal srv cpp title error buffer is accessed out of bounds severity severity description cvss score rvss vector rvss score cvss vector vendor null exploitation exploitation image exploitation vector description type bug mitigation date mitigation pull request description links keywords cppcheck static analysis testing triage bug cwe none flaw application n a phase testing languages none reproduction see artifacts below if available specificity n a date detected architectural location n a subsystem n a reproduction image gitlab com aliasrobotics offensive alurity pipelines active pipeline ros industrial jobs artifacts download date reported reproducibility always detected by method testing static reported by relationship automatic trace reported by alias robotics issue package n a detected by alias robotics
1
300,086
25,944,763,620
IssuesEvent
2022-12-16 22:44:25
hashicorp/terraform-provider-google
https://api.github.com/repos/hashicorp/terraform-provider-google
opened
Failing test(s): TestAccDataprocCluster_withLabels
test failure
<!--- This is a template for reporting test failures on nightly builds. It should only be used by core contributors who have access to our CI/CD results. ---> <!-- i.e. "Consistently since X date" or "X% failure in MONTH" --> Failure rate: 100% since 2022-12-09 <!-- List all impacted tests for searchability. The title of the issue can instead list one or more groups of tests, or describe the overall root cause. --> Impacted tests: - TestAccDataprocCluster_withLabels <!-- Link to the nightly build(s), ideally with one impacted test opened --> Nightly builds: - https://ci-oss.hashicorp.engineering/buildConfiguration/GoogleCloudBeta_ProviderGoogleCloudBetaGoogleProject/360401?buildTab=tests&expandedTest=-6423698170585265788 <!-- The error message that displays in the tests tab, for reference --> Message: ``` provider_test.go:315: Step 1/1 error: Check failed: Check 2/3 error: google_dataproc_cluster.with_labels: Attribute 'labels.%' expected "4", got "5" ```
1.0
Failing test(s): TestAccDataprocCluster_withLabels - <!--- This is a template for reporting test failures on nightly builds. It should only be used by core contributors who have access to our CI/CD results. ---> <!-- i.e. "Consistently since X date" or "X% failure in MONTH" --> Failure rate: 100% since 2022-12-09 <!-- List all impacted tests for searchability. The title of the issue can instead list one or more groups of tests, or describe the overall root cause. --> Impacted tests: - TestAccDataprocCluster_withLabels <!-- Link to the nightly build(s), ideally with one impacted test opened --> Nightly builds: - https://ci-oss.hashicorp.engineering/buildConfiguration/GoogleCloudBeta_ProviderGoogleCloudBetaGoogleProject/360401?buildTab=tests&expandedTest=-6423698170585265788 <!-- The error message that displays in the tests tab, for reference --> Message: ``` provider_test.go:315: Step 1/1 error: Check failed: Check 2/3 error: google_dataproc_cluster.with_labels: Attribute 'labels.%' expected "4", got "5" ```
test
failing test s testaccdataproccluster withlabels failure rate since impacted tests testaccdataproccluster withlabels nightly builds message provider test go step error check failed check error google dataproc cluster with labels attribute labels expected got
1
32,354
4,763,346,749
IssuesEvent
2016-10-25 14:14:24
Code-Poets/sandbox
https://api.github.com/repos/Code-Poets/sandbox
closed
Add a data migration to your scraper application
waiting for testing
Let's start with your application that contains the `Question` and `Category` models. - Add some questions and categories in your application. - Remove the `Category` model. - Add a category field to `Question` (as a `CharField`) - Create a migration that changes the structure of the database to match the new models but **without losing the data you entered**. - Apply the migration with `manage.py migrate` and verify that questions and categories you added have been preserved. You can read more about data migrations in Django documentation: [Data migrations](https://docs.djangoproject.com/en/1.10/topics/migrations/#data-migrations)
1.0
Add a data migration to your scraper application - Let's start with your application that contains the `Question` and `Category` models. - Add some questions and categories in your application. - Remove the `Category` model. - Add a category field to `Question` (as a `CharField`) - Create a migration that changes the structure of the database to match the new models but **without losing the data you entered**. - Apply the migration with `manage.py migrate` and verify that questions and categories you added have been preserved. You can read more about data migrations in Django documentation: [Data migrations](https://docs.djangoproject.com/en/1.10/topics/migrations/#data-migrations)
test
add a data migration to your scraper application let s start with your application that contains the question and category models add some questions and categories in your application remove the category model add a category field to question as a charfield create a migration that changes the structure of the database to match the new models but without losing the data you entered apply the migration with manage py migrate and verify that questions and categories you added have been preserved you can read more about data migrations in django documentation
1
702,212
24,120,740,507
IssuesEvent
2022-09-20 18:28:09
googleapis/nodejs-speech
https://api.github.com/repos/googleapis/nodejs-speech
closed
SpeechClient system test default: calls longRunningRecognize failed
priority: p1 type: bug api: speech flakybot: issue
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: ea654b5bf2e4067bf7682f9729cf8364ca4145a5 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/9663f5d7-8c67-4c75-9e46-ba62a7af791f), [Sponge](http://sponge2/9663f5d7-8c67-4c75-9e46-ba62a7af791f) status: failed <details><summary>Test output</summary><br><pre>16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. Error: 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. at Object.callErrorFromStatus (node_modules/@grpc/grpc-js/build/src/call.js:31:19) -> /workspace/node_modules/@grpc/grpc-js/src/call.ts:81:17 at Object.onReceiveStatus (node_modules/@grpc/grpc-js/build/src/client.js:190:52) -> /workspace/node_modules/@grpc/grpc-js/src/client.ts:352:36 at Object.onReceiveStatus (node_modules/@grpc/grpc-js/build/src/client-interceptors.js:365:141) -> /workspace/node_modules/@grpc/grpc-js/src/client-interceptors.ts:462:34 at Object.onReceiveStatus (node_modules/@grpc/grpc-js/build/src/client-interceptors.js:328:181) -> /workspace/node_modules/@grpc/grpc-js/src/client-interceptors.ts:424:48 at /workspace/node_modules/@grpc/grpc-js/build/src/call-stream.js:188:78 -> /workspace/node_modules/@grpc/grpc-js/src/call-stream.ts:330:24 at processTicksAndRejections (internal/process/task_queues.js:79:11) for call at at ServiceClientImpl.makeUnaryRequest (node_modules/@grpc/grpc-js/build/src/client.js:160:30) -> /workspace/node_modules/@grpc/grpc-js/src/client.ts:324:26 at ServiceClientImpl.<anonymous> (node_modules/@grpc/grpc-js/build/src/make-client.js:105:19) -> /workspace/node_modules/@grpc/grpc-js/src/make-client.ts:189:15 at /workspace/build/src/v1/speech_client.js:228:29 -> /workspace/src/v1/speech_client.ts:309:25 at /workspace/node_modules/google-gax/build/src/normalCalls/timeout.js:44:16 -> /workspace/node_modules/google-gax/src/normalCalls/timeout.ts:54:13 at LongrunningApiCaller._wrapOperation (node_modules/google-gax/build/src/longRunningCalls/longRunningApiCaller.js:55:16) -> /workspace/node_modules/google-gax/src/longRunningCalls/longRunningApiCaller.ts:80:12 at /workspace/node_modules/google-gax/build/src/longRunningCalls/longRunningApiCaller.js:46:25 -> /workspace/node_modules/google-gax/src/longRunningCalls/longRunningApiCaller.ts:64:19 at OngoingCallPromise.call (node_modules/google-gax/build/src/call.js:67:27) -> /workspace/node_modules/google-gax/src/call.ts:81:23 at LongrunningApiCaller.call (node_modules/google-gax/build/src/longRunningCalls/longRunningApiCaller.js:45:19) -> /workspace/node_modules/google-gax/src/longRunningCalls/longRunningApiCaller.ts:63:15 at /workspace/node_modules/google-gax/build/src/createApiCall.js:84:30 -> /workspace/node_modules/google-gax/src/createApiCall.ts:118:26 at processTicksAndRejections (internal/process/task_queues.js:97:5)</pre></details>
1.0
SpeechClient system test default: calls longRunningRecognize failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: ea654b5bf2e4067bf7682f9729cf8364ca4145a5 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/9663f5d7-8c67-4c75-9e46-ba62a7af791f), [Sponge](http://sponge2/9663f5d7-8c67-4c75-9e46-ba62a7af791f) status: failed <details><summary>Test output</summary><br><pre>16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. Error: 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. at Object.callErrorFromStatus (node_modules/@grpc/grpc-js/build/src/call.js:31:19) -> /workspace/node_modules/@grpc/grpc-js/src/call.ts:81:17 at Object.onReceiveStatus (node_modules/@grpc/grpc-js/build/src/client.js:190:52) -> /workspace/node_modules/@grpc/grpc-js/src/client.ts:352:36 at Object.onReceiveStatus (node_modules/@grpc/grpc-js/build/src/client-interceptors.js:365:141) -> /workspace/node_modules/@grpc/grpc-js/src/client-interceptors.ts:462:34 at Object.onReceiveStatus (node_modules/@grpc/grpc-js/build/src/client-interceptors.js:328:181) -> /workspace/node_modules/@grpc/grpc-js/src/client-interceptors.ts:424:48 at /workspace/node_modules/@grpc/grpc-js/build/src/call-stream.js:188:78 -> /workspace/node_modules/@grpc/grpc-js/src/call-stream.ts:330:24 at processTicksAndRejections (internal/process/task_queues.js:79:11) for call at at ServiceClientImpl.makeUnaryRequest (node_modules/@grpc/grpc-js/build/src/client.js:160:30) -> /workspace/node_modules/@grpc/grpc-js/src/client.ts:324:26 at ServiceClientImpl.<anonymous> (node_modules/@grpc/grpc-js/build/src/make-client.js:105:19) -> /workspace/node_modules/@grpc/grpc-js/src/make-client.ts:189:15 at /workspace/build/src/v1/speech_client.js:228:29 -> /workspace/src/v1/speech_client.ts:309:25 at /workspace/node_modules/google-gax/build/src/normalCalls/timeout.js:44:16 -> /workspace/node_modules/google-gax/src/normalCalls/timeout.ts:54:13 at LongrunningApiCaller._wrapOperation (node_modules/google-gax/build/src/longRunningCalls/longRunningApiCaller.js:55:16) -> /workspace/node_modules/google-gax/src/longRunningCalls/longRunningApiCaller.ts:80:12 at /workspace/node_modules/google-gax/build/src/longRunningCalls/longRunningApiCaller.js:46:25 -> /workspace/node_modules/google-gax/src/longRunningCalls/longRunningApiCaller.ts:64:19 at OngoingCallPromise.call (node_modules/google-gax/build/src/call.js:67:27) -> /workspace/node_modules/google-gax/src/call.ts:81:23 at LongrunningApiCaller.call (node_modules/google-gax/build/src/longRunningCalls/longRunningApiCaller.js:45:19) -> /workspace/node_modules/google-gax/src/longRunningCalls/longRunningApiCaller.ts:63:15 at /workspace/node_modules/google-gax/build/src/createApiCall.js:84:30 -> /workspace/node_modules/google-gax/src/createApiCall.ts:118:26 at processTicksAndRejections (internal/process/task_queues.js:97:5)</pre></details>
non_test
speechclient system test default calls longrunningrecognize failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see error unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see at object callerrorfromstatus node modules grpc grpc js build src call js workspace node modules grpc grpc js src call ts at object onreceivestatus node modules grpc grpc js build src client js workspace node modules grpc grpc js src client ts at object onreceivestatus node modules grpc grpc js build src client interceptors js workspace node modules grpc grpc js src client interceptors ts at object onreceivestatus node modules grpc grpc js build src client interceptors js workspace node modules grpc grpc js src client interceptors ts at workspace node modules grpc grpc js build src call stream js workspace node modules grpc grpc js src call stream ts at processticksandrejections internal process task queues js for call at at serviceclientimpl makeunaryrequest node modules grpc grpc js build src client js workspace node modules grpc grpc js src client ts at serviceclientimpl node modules grpc grpc js build src make client js workspace node modules grpc grpc js src make client ts at workspace build src speech client js workspace src speech client ts at workspace node modules google gax build src normalcalls timeout js workspace node modules google gax src normalcalls timeout ts at longrunningapicaller wrapoperation node modules google gax build src longrunningcalls longrunningapicaller js workspace node modules google gax src longrunningcalls longrunningapicaller ts at workspace node modules google gax build src longrunningcalls longrunningapicaller js workspace node modules google gax src longrunningcalls longrunningapicaller ts at ongoingcallpromise call node modules google gax build src call js workspace node modules google gax src call ts at longrunningapicaller call node modules google gax build src longrunningcalls longrunningapicaller js workspace node modules google gax src longrunningcalls longrunningapicaller ts at workspace node modules google gax build src createapicall js workspace node modules google gax src createapicall ts at processticksandrejections internal process task queues js
0
78,217
7,623,499,057
IssuesEvent
2018-05-03 15:10:53
jupyterhub/jupyterhub
https://api.github.com/repos/jupyterhub/jupyterhub
closed
jupyterhub upgrade-db fails due to SQLite limitations
needs: testing tests
The alembic migration 1cebaf56856c -> 56cc5a70207e fails due to the [create_foreign_key](https://github.com/jupyterhub/jupyterhub/blob/master/jupyterhub/alembic/versions/56cc5a70207e_token_tracking.py#L27) statements, and SQLite does not support adding a foreign key constraint to an existing column, apparently. ``` [I 2018-03-09 09:26:51.060 JupyterHub dbutil:125] Upgrading sqlite:///data/jupyterhub.sqlite [I 2018-03-09 09:26:51.061 JupyterHub dbutil:105] Backing up data/jupyterhub.sqlite => data/jupyterhub.sqlite.2018-03-09-092651 [I 2018-03-09 09:26:51.417 alembic.runtime.migration migration:117] Context impl SQLiteImpl. [I 2018-03-09 09:26:51.418 alembic.runtime.migration migration:122] Will assume non-transactional DDL. [I 2018-03-09 09:26:51.421 alembic.runtime.migration migration:327] Running upgrade 1cebaf56856c -> 56cc5a70207e, token tracking Traceback (most recent call last): File "/opt/conda/bin/alembic", line 11, in <module> ... File "/opt/conda/lib/python3.5/site-packages/jupyterhub/alembic/versions/56cc5a70207e_token_tracking.py", line 27, in upgrade op.create_foreign_key(None, 'oauth_access_tokens', 'oauth_clients', ['client_id'], ['identifier'], ondelete='CASCADE') ... File "/opt/conda/lib/python3.5/site-packages/alembic/ddl/sqlite.py", line 34, in add_constraint "No support for ALTER of constraints in SQLite dialect") NotImplementedError: No support for ALTER of constraints in SQLite dialect ```
2.0
jupyterhub upgrade-db fails due to SQLite limitations - The alembic migration 1cebaf56856c -> 56cc5a70207e fails due to the [create_foreign_key](https://github.com/jupyterhub/jupyterhub/blob/master/jupyterhub/alembic/versions/56cc5a70207e_token_tracking.py#L27) statements, and SQLite does not support adding a foreign key constraint to an existing column, apparently. ``` [I 2018-03-09 09:26:51.060 JupyterHub dbutil:125] Upgrading sqlite:///data/jupyterhub.sqlite [I 2018-03-09 09:26:51.061 JupyterHub dbutil:105] Backing up data/jupyterhub.sqlite => data/jupyterhub.sqlite.2018-03-09-092651 [I 2018-03-09 09:26:51.417 alembic.runtime.migration migration:117] Context impl SQLiteImpl. [I 2018-03-09 09:26:51.418 alembic.runtime.migration migration:122] Will assume non-transactional DDL. [I 2018-03-09 09:26:51.421 alembic.runtime.migration migration:327] Running upgrade 1cebaf56856c -> 56cc5a70207e, token tracking Traceback (most recent call last): File "/opt/conda/bin/alembic", line 11, in <module> ... File "/opt/conda/lib/python3.5/site-packages/jupyterhub/alembic/versions/56cc5a70207e_token_tracking.py", line 27, in upgrade op.create_foreign_key(None, 'oauth_access_tokens', 'oauth_clients', ['client_id'], ['identifier'], ondelete='CASCADE') ... File "/opt/conda/lib/python3.5/site-packages/alembic/ddl/sqlite.py", line 34, in add_constraint "No support for ALTER of constraints in SQLite dialect") NotImplementedError: No support for ALTER of constraints in SQLite dialect ```
test
jupyterhub upgrade db fails due to sqlite limitations the alembic migration fails due to the statements and sqlite does not support adding a foreign key constraint to an existing column apparently upgrading sqlite data jupyterhub sqlite backing up data jupyterhub sqlite data jupyterhub sqlite context impl sqliteimpl will assume non transactional ddl running upgrade token tracking traceback most recent call last file opt conda bin alembic line in file opt conda lib site packages jupyterhub alembic versions token tracking py line in upgrade op create foreign key none oauth access tokens oauth clients ondelete cascade file opt conda lib site packages alembic ddl sqlite py line in add constraint no support for alter of constraints in sqlite dialect notimplementederror no support for alter of constraints in sqlite dialect
1
418,025
28,113,287,172
IssuesEvent
2023-03-31 08:52:32
NivethaUdayakumar/ped
https://api.github.com/repos/NivethaUdayakumar/ped
opened
Could add more FAQ's to the UG Guide
type.DocumentationBug severity.VeryLow
Could add more FAQs commonly faced by users to the UG Guide. Suggestion: A possible FAQ could be: How can I transfer my data to another computer? ![image.png](https://raw.githubusercontent.com/NivethaUdayakumar/ped/main/files/d6a04fef-913a-4cd9-8c89-2edfd8c0d917.png) <!--session: 1680252441490-27f7ffdd-adc6-4773-a88f-a8f92ff55cc4--> <!--Version: Web v3.4.7-->
1.0
Could add more FAQ's to the UG Guide - Could add more FAQs commonly faced by users to the UG Guide. Suggestion: A possible FAQ could be: How can I transfer my data to another computer? ![image.png](https://raw.githubusercontent.com/NivethaUdayakumar/ped/main/files/d6a04fef-913a-4cd9-8c89-2edfd8c0d917.png) <!--session: 1680252441490-27f7ffdd-adc6-4773-a88f-a8f92ff55cc4--> <!--Version: Web v3.4.7-->
non_test
could add more faq s to the ug guide could add more faqs commonly faced by users to the ug guide suggestion a possible faq could be how can i transfer my data to another computer
0
143,145
13,055,502,129
IssuesEvent
2020-07-30 01:53:04
cbeleites/hyperSpec
https://api.github.com/repos/cbeleites/hyperSpec
closed
Add missing / update inappropriate @concepts/@keywords in documentation
blocked :no_entry: documentation :blue_book: priority low
Documentation should have either appropriate `@concepts` or `@keywords` (or both). They should be added or updated. But firstly: - [ ] #151 should be solved
1.0
Add missing / update inappropriate @concepts/@keywords in documentation - Documentation should have either appropriate `@concepts` or `@keywords` (or both). They should be added or updated. But firstly: - [ ] #151 should be solved
non_test
add missing update inappropriate concepts keywords in documentation documentation should have either appropriate concepts or keywords or both they should be added or updated but firstly should be solved
0
586
7,986,127,106
IssuesEvent
2018-07-19 00:10:25
rust-lang-nursery/stdsimd
https://api.github.com/repos/rust-lang-nursery/stdsimd
closed
simd_fmax is broken in LLVM6
A-portable
See https://github.com/rust-lang/rust/pull/49249, `simd_fmax` is broken in LLVM6, and the decision there was to work around this in `stdsimd` until rustc starts using LLVM7. This rust-lang/rust issue tracks that https://github.com/rust-lang/rust/issues/49261
True
simd_fmax is broken in LLVM6 - See https://github.com/rust-lang/rust/pull/49249, `simd_fmax` is broken in LLVM6, and the decision there was to work around this in `stdsimd` until rustc starts using LLVM7. This rust-lang/rust issue tracks that https://github.com/rust-lang/rust/issues/49261
non_test
simd fmax is broken in see simd fmax is broken in and the decision there was to work around this in stdsimd until rustc starts using this rust lang rust issue tracks that
0
1,559
6,335,238,819
IssuesEvent
2017-07-26 18:24:54
LearnersGuild/echo
https://api.github.com/repos/LearnersGuild/echo
reopened
Move changefeed listeners to web service
architecture chore
Currently, the workers set up listening to db changefeeds and, effectively, queue items for their _own_ work queues. This is problematic because if changes to the database occur that should be handled by a worker while the worker is not running, that change event is lost and the job is never processed. Instead of connecting the changefeed listeners in the worker process, we should do it in the standing web service. This also makes it easier to choose to alternative mechanisms for background task execution (instead of having always-running worker dynos).
1.0
Move changefeed listeners to web service - Currently, the workers set up listening to db changefeeds and, effectively, queue items for their _own_ work queues. This is problematic because if changes to the database occur that should be handled by a worker while the worker is not running, that change event is lost and the job is never processed. Instead of connecting the changefeed listeners in the worker process, we should do it in the standing web service. This also makes it easier to choose to alternative mechanisms for background task execution (instead of having always-running worker dynos).
non_test
move changefeed listeners to web service currently the workers set up listening to db changefeeds and effectively queue items for their own work queues this is problematic because if changes to the database occur that should be handled by a worker while the worker is not running that change event is lost and the job is never processed instead of connecting the changefeed listeners in the worker process we should do it in the standing web service this also makes it easier to choose to alternative mechanisms for background task execution instead of having always running worker dynos
0
109,121
9,368,868,954
IssuesEvent
2019-04-03 09:41:13
cjlee112/socraticqs2
https://api.github.com/repos/cjlee112/socraticqs2
closed
Add Courselets Assistant (Intercom) on old Instructor UI pages?
Enhancement Need to test
We are not yet in a position where we can get rid of the old Instructor UI entirely. And until we do so, it seems like it's even more important to offer instructors immediate help with their questions about the old UI, since it's more likely to cause them confusions!
1.0
Add Courselets Assistant (Intercom) on old Instructor UI pages? - We are not yet in a position where we can get rid of the old Instructor UI entirely. And until we do so, it seems like it's even more important to offer instructors immediate help with their questions about the old UI, since it's more likely to cause them confusions!
test
add courselets assistant intercom on old instructor ui pages we are not yet in a position where we can get rid of the old instructor ui entirely and until we do so it seems like it s even more important to offer instructors immediate help with their questions about the old ui since it s more likely to cause them confusions
1
26,674
4,238,699,160
IssuesEvent
2016-07-06 05:51:12
ausshir/MPU341
https://api.github.com/repos/ausshir/MPU341
reopened
Add Final Enhancements
5 - Testing
Misc. things like memory paging, large instructions and extra functionality. * Make it modular just in case! <!--- @huboard:{"order":0.9998000299960005,"milestone_order":1,"custom_state":""} -->
1.0
Add Final Enhancements - Misc. things like memory paging, large instructions and extra functionality. * Make it modular just in case! <!--- @huboard:{"order":0.9998000299960005,"milestone_order":1,"custom_state":""} -->
test
add final enhancements misc things like memory paging large instructions and extra functionality make it modular just in case huboard order milestone order custom state
1
119,328
10,039,061,897
IssuesEvent
2019-07-18 16:26:57
ValveSoftware/steam-for-linux
https://api.github.com/repos/ValveSoftware/steam-for-linux
closed
Steam Big Picture Mode sound settings segfault
Audio Big Picture Distro Family: Arch Need Retest reviewed
#### Your system information * Steam client version: * Distribution (e.g. Ubuntu): Arch * Opted into Steam client beta?: Yes * Have you checked for system updates?: Yes After clicking in settings on Interface or Voice audio settings, steam crashes with steam.sh: line 713: 10424 Segmentation fault (core dumped) $STEAM_DEBUGGER "$STEAMROOT/$STEAMEXEPATH" "$@" I've tried turning off beta and on. It was working like...half a year ago. I've got 5.1 audio setup. Can this be an issue? Thanks #### Steps for reproducing this issue: 1. Run Stam in BPM 2. Go to settings 3. Try Interface or Voice audio settings
1.0
Steam Big Picture Mode sound settings segfault - #### Your system information * Steam client version: * Distribution (e.g. Ubuntu): Arch * Opted into Steam client beta?: Yes * Have you checked for system updates?: Yes After clicking in settings on Interface or Voice audio settings, steam crashes with steam.sh: line 713: 10424 Segmentation fault (core dumped) $STEAM_DEBUGGER "$STEAMROOT/$STEAMEXEPATH" "$@" I've tried turning off beta and on. It was working like...half a year ago. I've got 5.1 audio setup. Can this be an issue? Thanks #### Steps for reproducing this issue: 1. Run Stam in BPM 2. Go to settings 3. Try Interface or Voice audio settings
test
steam big picture mode sound settings segfault your system information steam client version distribution e g ubuntu arch opted into steam client beta yes have you checked for system updates yes after clicking in settings on interface or voice audio settings steam crashes with steam sh line segmentation fault core dumped steam debugger steamroot steamexepath i ve tried turning off beta and on it was working like half a year ago i ve got audio setup can this be an issue thanks steps for reproducing this issue run stam in bpm go to settings try interface or voice audio settings
1
108,199
9,285,714,648
IssuesEvent
2019-03-21 08:15:50
KplusH/jsbach-timeline
https://api.github.com/repos/KplusH/jsbach-timeline
closed
Ladefehler Galerie bei Köthen/Alter Friedhof
ready for test
Wenn man das Zweite Mal eine Galeriebild bei Köthen/alter Friedhof anklickt, läd es nicht mehr ![Bildschirmfoto 2019-03-18 um 12 06 11](https://user-images.githubusercontent.com/23217462/54526521-56efc200-4977-11e9-8017-8c46c991dc10.png) ![Bildschirmfoto 2019-03-18 um 12 07 03](https://user-images.githubusercontent.com/23217462/54526522-56efc200-4977-11e9-9c10-3d995046aeaf.png)
1.0
Ladefehler Galerie bei Köthen/Alter Friedhof - Wenn man das Zweite Mal eine Galeriebild bei Köthen/alter Friedhof anklickt, läd es nicht mehr ![Bildschirmfoto 2019-03-18 um 12 06 11](https://user-images.githubusercontent.com/23217462/54526521-56efc200-4977-11e9-8017-8c46c991dc10.png) ![Bildschirmfoto 2019-03-18 um 12 07 03](https://user-images.githubusercontent.com/23217462/54526522-56efc200-4977-11e9-9c10-3d995046aeaf.png)
test
ladefehler galerie bei köthen alter friedhof wenn man das zweite mal eine galeriebild bei köthen alter friedhof anklickt läd es nicht mehr
1
728,169
25,069,211,099
IssuesEvent
2022-11-07 10:45:34
scylladb/scylla-operator
https://api.github.com/repos/scylladb/scylla-operator
closed
Flake: Scylla Manager integration [It] should discover cluster and sync tasks
priority/important-soon kind/flake
https://github.com/scylladb/scylla-operator/actions/runs/3067025626/jobs/4954445399#step:12:441 https://github.com/scylladb/scylla-operator/suites/8323775595/artifacts/365955898 ``` • [FAILED] [462.877 seconds] Scylla Manager integration github.com/scylladb/scylla-operator/test/e2e/set/scyllacluster/scyllamanager.go:18 [It] should discover cluster and sync tasks github.com/scylladb/scylla-operator/test/e2e/set/scyllacluster/scyllamanager.go:23 Begin Captured StdOut/StdErr Output >> I0916 12:29:55.912346 159 framework/framework.go:249] "Namespace removed." Namespace="e2e-test-scyllacluster-7t8gg-mxsmm" << End Captured StdOut/StdErr Output Begin Captured GinkgoWriter Output >> [BeforeEach] Scylla Manager integration github.com/scylladb/scylla-operator/test/e2e/framework/framework.go:57 STEP: Creating a new namespace 09/16/22 12:22:13.035 Sep 16 12:22:13.121: INFO: Created namespace "e2e-test-scyllacluster-7t8gg-mxsmm". STEP: Waiting for ServiceAccount "e2e-user" in namespace "e2e-test-scyllacluster-7t8gg-mxsmm". 09/16/22 12:22:13.216 STEP: Waiting for default ServiceAccount in namespace "e2e-test-scyllacluster-7t8gg-mxsmm". 09/16/22 12:22:15.18 [It] should discover cluster and sync tasks github.com/scylladb/scylla-operator/test/e2e/set/scyllacluster/scyllamanager.go:23 STEP: Creating a ScyllaCluster 09/16/22 12:22:15.184 STEP: Waiting for the ScyllaCluster to deploy 09/16/22 12:22:15.199 Sep 16 12:24:25.070: INFO: ScyllaCluster e2e-test-scyllacluster-7t8gg-mxsmm/basic-sxvd8 (RV=3430) is rolled out STEP: Inserting data with RF=1 09/16/22 12:24:30.151 STEP: Verifying the ScyllaCluster 09/16/22 12:24:30.729 Sep 16 12:24:30.751: INFO: Found 1 pvc(s) in namespace "e2e-test-scyllacluster-7t8gg-mxsmm" Sep 16 12:24:30.751: INFO: Found 1 pvc(s) for ScyllaCluster "e2e-test-scyllacluster-7t8gg-mxsmm/basic-sxvd8" STEP: Reading data with RF=1 09/16/22 12:24:30.807 STEP: Verifying data consistency 09/16/22 12:24:30.812 STEP: Waiting for the cluster sync with Scylla Manager 09/16/22 12:24:30.812 [AfterEach] Scylla Manager integration github.com/scylladb/scylla-operator/test/e2e/framework/framework.go:58 STEP: Collecting events from namespace "e2e-test-scyllacluster-7t8gg-mxsmm". 09/16/22 12:28:30.817 STEP: Found 30 events. 09/16/22 12:28:30.852 Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:41 +0000 UTC - event for basic-sxvd8-auth-token: {scyllacluster-controller } SecretCreated: Secret e2e-test-scyllacluster-7t8gg-mxsmm/basic-sxvd8-auth-token created Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:41 +0000 UTC - event for basic-sxvd8-member: {scyllacluster-controller } ServiceAccountCreated: ServiceAccount e2e-test-scyllacluster-7t8gg-mxsmm/basic-sxvd8-member created Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:41 +0000 UTC - event for basic-sxvd8-member: {scyllacluster-controller } RoleBindingCreated: RoleBinding e2e-test-scyllacluster-7t8gg-mxsmm/basic-sxvd8-member created Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:41 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a: {scyllacluster-controller } StatefulSetCreated: StatefulSet e2e-test-scyllacluster-7t8gg-mxsmm/basic-sxvd8-us-east-1-us-east-1a created Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:42 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a: {statefulset-controller } SuccessfulCreate: create Claim data-basic-sxvd8-us-east-1-us-east-1a-0 Pod basic-sxvd8-us-east-1-us-east-1a-0 in StatefulSet basic-sxvd8-us-east-1-us-east-1a success Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:42 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a: {statefulset-controller } SuccessfulCreate: create Pod basic-sxvd8-us-east-1-us-east-1a-0 in StatefulSet basic-sxvd8-us-east-1-us-east-1a successful Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:42 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {default-scheduler } FailedScheduling: 0/1 nodes are available: 1 pod has unbound immediate PersistentVolumeClaims. Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:42 +0000 UTC - event for data-basic-sxvd8-us-east-1-us-east-1a-0: {k8s.io/minikube-hostpath_fv-az136-60_065a1f85-feac-4ea1-b55a-0bda0d957ed6 } ProvisioningSucceeded: Successfully provisioned volume pvc-0e5f40ca-[443](https://github.com/scylladb/scylla-operator/actions/runs/3067025626/jobs/4954445399#step:12:444)e-49ed-bc13-c26[451](https://github.com/scylladb/scylla-operator/actions/runs/3067025626/jobs/4954445399#step:12:452)46d07a Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:42 +0000 UTC - event for data-basic-sxvd8-us-east-1-us-east-1a-0: {persistentvolume-controller } ExternalProvisioning: waiting for a volume to be created, either by external provisioner "k8s.io/minikube-hostpath" or manually created by system administrator Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:42 +0000 UTC - event for data-basic-sxvd8-us-east-1-us-east-1a-0: {k8s.io/minikube-hostpath_fv-az136-60_065a1f85-feac-4ea1-b55a-0bda0d957ed6 } Provisioning: External provisioner is provisioning volume for claim "e2e-test-scyllacluster-7t8gg-mxsmm/data-basic-sxvd8-us-east-1-us-east-1a-0" Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:43 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {default-scheduler } Scheduled: Successfully assigned e2e-test-scyllacluster-7t8gg-mxsmm/basic-sxvd8-us-east-1-us-east-1a-0 to fv-az136-60 Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:52 +0000 UTC - event for basic-sxvd8: {scyllacluster-controller } PodDisruptionBudgetCreated: PodDisruptionBudget e2e-test-scyllacluster-7t8gg-mxsmm/basic-sxvd8 created Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:52 +0000 UTC - event for basic-sxvd8-client: {scyllacluster-controller } ServiceCreated: Service e2e-test-scyllacluster-7t8gg-mxsmm/basic-sxvd8-client created Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:52 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {scyllacluster-controller } ServiceCreated: Service e2e-test-scyllacluster-7t8gg-mxsmm/basic-sxvd8-us-east-1-us-east-1a-0 created Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:58 +0000 UTC - event for nodeconfig-podinfo-70cb9e73-afd9-44ec-bbbf-cc8e41fdb038: {NodeConfigCM-controller } ConfigMapCreated: ConfigMap e2e-test-scyllacluster-7t8gg-mxsmm/nodeconfig-podinfo-70cb9e73-afd9-44ec-bbbf-cc8e41fdb038 created Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:59 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} FailedMount: MountVolume.SetUp failed for volume "scylla-client-config-volume" : failed to sync secret cache: timed out waiting for the condition Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:00 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} FailedMount: MountVolume.SetUp failed for volume "kube-api-access-bm7g6" : failed to sync configmap cache: timed out waiting for the condition Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:03 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Started: Started container sidecar-injection Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:03 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Pulled: Container image "docker.io/scylladb/scylla-operator:ci" already present on machine Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:03 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Created: Created container sidecar-injection Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:13 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Started: Started container scylla Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:13 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Pulled: Container image "docker.io/scylladb/scylla-manager-agent:2.6.3" already present on machine Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:13 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Created: Created container scylla-manager-agent Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:13 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Created: Created container scylla Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:13 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Pulled: Container image "docker.io/scylladb/scylla:4.6.3" already present on machine Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:14 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Started: Started container scylla-manager-agent Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:18 +0000 UTC - event for nodeconfig-podinfo-70cb9e73-afd9-44ec-bbbf-cc8e41fdb038: {NodeConfigCM-controller } ConfigMapUpdated: ConfigMap e2e-test-scyllacluster-7t8gg-mxsmm/nodeconfig-podinfo-70cb9e73-afd9-44ec-bbbf-cc8e41fdb038 updated Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:23 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Unhealthy: Startup probe failed: HTTP probe failed with statuscode: 503 Sep 16 12:28:30.854: INFO: At 2022-09-16 12:23:33 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Unhealthy: Readiness probe failed: HTTP probe failed with statuscode: 500 Sep 16 12:28:30.854: INFO: At 2022-09-16 12:23:43 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Unhealthy: Readiness probe failed: HTTP probe failed with statuscode: 503 STEP: Collecting dumps from namespace "e2e-test-scyllacluster-7t8gg-mxsmm". 09/16/22 12:28:30.854 STEP: Destroying namespace "e2e-test-scyllacluster-7t8gg-mxsmm". 09/16/22 12:28:31.094 STEP: Waiting for namespace "e2e-test-scyllacluster-7t8gg-mxsmm" to be removed. 09/16/22 12:28:31.106 << End Captured GinkgoWriter Output Unexpected error: <*errors.errorString | 0xc0002530b0>: { s: "timed out waiting for the condition", } timed out waiting for the condition occurred In [It] at: github.com/scylladb/scylla-operator/test/e2e/set/scyllacluster/scyllamanager.go:96 Full Stack Trace github.com/scylladb/scylla-operator/test/e2e/set/scyllacluster.glob..func15.1() github.com/scylladb/scylla-operator/test/e2e/set/scyllacluster/scyllamanager.go:96 +0x813 ```
1.0
Flake: Scylla Manager integration [It] should discover cluster and sync tasks - https://github.com/scylladb/scylla-operator/actions/runs/3067025626/jobs/4954445399#step:12:441 https://github.com/scylladb/scylla-operator/suites/8323775595/artifacts/365955898 ``` • [FAILED] [462.877 seconds] Scylla Manager integration github.com/scylladb/scylla-operator/test/e2e/set/scyllacluster/scyllamanager.go:18 [It] should discover cluster and sync tasks github.com/scylladb/scylla-operator/test/e2e/set/scyllacluster/scyllamanager.go:23 Begin Captured StdOut/StdErr Output >> I0916 12:29:55.912346 159 framework/framework.go:249] "Namespace removed." Namespace="e2e-test-scyllacluster-7t8gg-mxsmm" << End Captured StdOut/StdErr Output Begin Captured GinkgoWriter Output >> [BeforeEach] Scylla Manager integration github.com/scylladb/scylla-operator/test/e2e/framework/framework.go:57 STEP: Creating a new namespace 09/16/22 12:22:13.035 Sep 16 12:22:13.121: INFO: Created namespace "e2e-test-scyllacluster-7t8gg-mxsmm". STEP: Waiting for ServiceAccount "e2e-user" in namespace "e2e-test-scyllacluster-7t8gg-mxsmm". 09/16/22 12:22:13.216 STEP: Waiting for default ServiceAccount in namespace "e2e-test-scyllacluster-7t8gg-mxsmm". 09/16/22 12:22:15.18 [It] should discover cluster and sync tasks github.com/scylladb/scylla-operator/test/e2e/set/scyllacluster/scyllamanager.go:23 STEP: Creating a ScyllaCluster 09/16/22 12:22:15.184 STEP: Waiting for the ScyllaCluster to deploy 09/16/22 12:22:15.199 Sep 16 12:24:25.070: INFO: ScyllaCluster e2e-test-scyllacluster-7t8gg-mxsmm/basic-sxvd8 (RV=3430) is rolled out STEP: Inserting data with RF=1 09/16/22 12:24:30.151 STEP: Verifying the ScyllaCluster 09/16/22 12:24:30.729 Sep 16 12:24:30.751: INFO: Found 1 pvc(s) in namespace "e2e-test-scyllacluster-7t8gg-mxsmm" Sep 16 12:24:30.751: INFO: Found 1 pvc(s) for ScyllaCluster "e2e-test-scyllacluster-7t8gg-mxsmm/basic-sxvd8" STEP: Reading data with RF=1 09/16/22 12:24:30.807 STEP: Verifying data consistency 09/16/22 12:24:30.812 STEP: Waiting for the cluster sync with Scylla Manager 09/16/22 12:24:30.812 [AfterEach] Scylla Manager integration github.com/scylladb/scylla-operator/test/e2e/framework/framework.go:58 STEP: Collecting events from namespace "e2e-test-scyllacluster-7t8gg-mxsmm". 09/16/22 12:28:30.817 STEP: Found 30 events. 09/16/22 12:28:30.852 Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:41 +0000 UTC - event for basic-sxvd8-auth-token: {scyllacluster-controller } SecretCreated: Secret e2e-test-scyllacluster-7t8gg-mxsmm/basic-sxvd8-auth-token created Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:41 +0000 UTC - event for basic-sxvd8-member: {scyllacluster-controller } ServiceAccountCreated: ServiceAccount e2e-test-scyllacluster-7t8gg-mxsmm/basic-sxvd8-member created Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:41 +0000 UTC - event for basic-sxvd8-member: {scyllacluster-controller } RoleBindingCreated: RoleBinding e2e-test-scyllacluster-7t8gg-mxsmm/basic-sxvd8-member created Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:41 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a: {scyllacluster-controller } StatefulSetCreated: StatefulSet e2e-test-scyllacluster-7t8gg-mxsmm/basic-sxvd8-us-east-1-us-east-1a created Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:42 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a: {statefulset-controller } SuccessfulCreate: create Claim data-basic-sxvd8-us-east-1-us-east-1a-0 Pod basic-sxvd8-us-east-1-us-east-1a-0 in StatefulSet basic-sxvd8-us-east-1-us-east-1a success Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:42 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a: {statefulset-controller } SuccessfulCreate: create Pod basic-sxvd8-us-east-1-us-east-1a-0 in StatefulSet basic-sxvd8-us-east-1-us-east-1a successful Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:42 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {default-scheduler } FailedScheduling: 0/1 nodes are available: 1 pod has unbound immediate PersistentVolumeClaims. Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:42 +0000 UTC - event for data-basic-sxvd8-us-east-1-us-east-1a-0: {k8s.io/minikube-hostpath_fv-az136-60_065a1f85-feac-4ea1-b55a-0bda0d957ed6 } ProvisioningSucceeded: Successfully provisioned volume pvc-0e5f40ca-[443](https://github.com/scylladb/scylla-operator/actions/runs/3067025626/jobs/4954445399#step:12:444)e-49ed-bc13-c26[451](https://github.com/scylladb/scylla-operator/actions/runs/3067025626/jobs/4954445399#step:12:452)46d07a Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:42 +0000 UTC - event for data-basic-sxvd8-us-east-1-us-east-1a-0: {persistentvolume-controller } ExternalProvisioning: waiting for a volume to be created, either by external provisioner "k8s.io/minikube-hostpath" or manually created by system administrator Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:42 +0000 UTC - event for data-basic-sxvd8-us-east-1-us-east-1a-0: {k8s.io/minikube-hostpath_fv-az136-60_065a1f85-feac-4ea1-b55a-0bda0d957ed6 } Provisioning: External provisioner is provisioning volume for claim "e2e-test-scyllacluster-7t8gg-mxsmm/data-basic-sxvd8-us-east-1-us-east-1a-0" Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:43 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {default-scheduler } Scheduled: Successfully assigned e2e-test-scyllacluster-7t8gg-mxsmm/basic-sxvd8-us-east-1-us-east-1a-0 to fv-az136-60 Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:52 +0000 UTC - event for basic-sxvd8: {scyllacluster-controller } PodDisruptionBudgetCreated: PodDisruptionBudget e2e-test-scyllacluster-7t8gg-mxsmm/basic-sxvd8 created Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:52 +0000 UTC - event for basic-sxvd8-client: {scyllacluster-controller } ServiceCreated: Service e2e-test-scyllacluster-7t8gg-mxsmm/basic-sxvd8-client created Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:52 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {scyllacluster-controller } ServiceCreated: Service e2e-test-scyllacluster-7t8gg-mxsmm/basic-sxvd8-us-east-1-us-east-1a-0 created Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:58 +0000 UTC - event for nodeconfig-podinfo-70cb9e73-afd9-44ec-bbbf-cc8e41fdb038: {NodeConfigCM-controller } ConfigMapCreated: ConfigMap e2e-test-scyllacluster-7t8gg-mxsmm/nodeconfig-podinfo-70cb9e73-afd9-44ec-bbbf-cc8e41fdb038 created Sep 16 12:28:30.853: INFO: At 2022-09-16 12:22:59 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} FailedMount: MountVolume.SetUp failed for volume "scylla-client-config-volume" : failed to sync secret cache: timed out waiting for the condition Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:00 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} FailedMount: MountVolume.SetUp failed for volume "kube-api-access-bm7g6" : failed to sync configmap cache: timed out waiting for the condition Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:03 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Started: Started container sidecar-injection Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:03 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Pulled: Container image "docker.io/scylladb/scylla-operator:ci" already present on machine Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:03 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Created: Created container sidecar-injection Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:13 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Started: Started container scylla Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:13 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Pulled: Container image "docker.io/scylladb/scylla-manager-agent:2.6.3" already present on machine Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:13 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Created: Created container scylla-manager-agent Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:13 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Created: Created container scylla Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:13 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Pulled: Container image "docker.io/scylladb/scylla:4.6.3" already present on machine Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:14 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Started: Started container scylla-manager-agent Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:18 +0000 UTC - event for nodeconfig-podinfo-70cb9e73-afd9-44ec-bbbf-cc8e41fdb038: {NodeConfigCM-controller } ConfigMapUpdated: ConfigMap e2e-test-scyllacluster-7t8gg-mxsmm/nodeconfig-podinfo-70cb9e73-afd9-44ec-bbbf-cc8e41fdb038 updated Sep 16 12:28:30.853: INFO: At 2022-09-16 12:23:23 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Unhealthy: Startup probe failed: HTTP probe failed with statuscode: 503 Sep 16 12:28:30.854: INFO: At 2022-09-16 12:23:33 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Unhealthy: Readiness probe failed: HTTP probe failed with statuscode: 500 Sep 16 12:28:30.854: INFO: At 2022-09-16 12:23:43 +0000 UTC - event for basic-sxvd8-us-east-1-us-east-1a-0: {kubelet fv-az136-60} Unhealthy: Readiness probe failed: HTTP probe failed with statuscode: 503 STEP: Collecting dumps from namespace "e2e-test-scyllacluster-7t8gg-mxsmm". 09/16/22 12:28:30.854 STEP: Destroying namespace "e2e-test-scyllacluster-7t8gg-mxsmm". 09/16/22 12:28:31.094 STEP: Waiting for namespace "e2e-test-scyllacluster-7t8gg-mxsmm" to be removed. 09/16/22 12:28:31.106 << End Captured GinkgoWriter Output Unexpected error: <*errors.errorString | 0xc0002530b0>: { s: "timed out waiting for the condition", } timed out waiting for the condition occurred In [It] at: github.com/scylladb/scylla-operator/test/e2e/set/scyllacluster/scyllamanager.go:96 Full Stack Trace github.com/scylladb/scylla-operator/test/e2e/set/scyllacluster.glob..func15.1() github.com/scylladb/scylla-operator/test/e2e/set/scyllacluster/scyllamanager.go:96 +0x813 ```
non_test
flake scylla manager integration should discover cluster and sync tasks • scylla manager integration github com scylladb scylla operator test set scyllacluster scyllamanager go should discover cluster and sync tasks github com scylladb scylla operator test set scyllacluster scyllamanager go begin captured stdout stderr output framework framework go namespace removed namespace test scyllacluster mxsmm end captured stdout stderr output begin captured ginkgowriter output scylla manager integration github com scylladb scylla operator test framework framework go step creating a new namespace sep info created namespace test scyllacluster mxsmm step waiting for serviceaccount user in namespace test scyllacluster mxsmm step waiting for default serviceaccount in namespace test scyllacluster mxsmm should discover cluster and sync tasks github com scylladb scylla operator test set scyllacluster scyllamanager go step creating a scyllacluster step waiting for the scyllacluster to deploy sep info scyllacluster test scyllacluster mxsmm basic rv is rolled out step inserting data with rf step verifying the scyllacluster sep info found pvc s in namespace test scyllacluster mxsmm sep info found pvc s for scyllacluster test scyllacluster mxsmm basic step reading data with rf step verifying data consistency step waiting for the cluster sync with scylla manager scylla manager integration github com scylladb scylla operator test framework framework go step collecting events from namespace test scyllacluster mxsmm step found events sep info at utc event for basic auth token scyllacluster controller secretcreated secret test scyllacluster mxsmm basic auth token created sep info at utc event for basic member scyllacluster controller serviceaccountcreated serviceaccount test scyllacluster mxsmm basic member created sep info at utc event for basic member scyllacluster controller rolebindingcreated rolebinding test scyllacluster mxsmm basic member created sep info at utc event for basic us east us east scyllacluster controller statefulsetcreated statefulset test scyllacluster mxsmm basic us east us east created sep info at utc event for basic us east us east statefulset controller successfulcreate create claim data basic us east us east pod basic us east us east in statefulset basic us east us east success sep info at utc event for basic us east us east statefulset controller successfulcreate create pod basic us east us east in statefulset basic us east us east successful sep info at utc event for basic us east us east default scheduler failedscheduling nodes are available pod has unbound immediate persistentvolumeclaims sep info at utc event for data basic us east us east io minikube hostpath fv feac provisioningsucceeded successfully provisioned volume pvc sep info at utc event for data basic us east us east persistentvolume controller externalprovisioning waiting for a volume to be created either by external provisioner io minikube hostpath or manually created by system administrator sep info at utc event for data basic us east us east io minikube hostpath fv feac provisioning external provisioner is provisioning volume for claim test scyllacluster mxsmm data basic us east us east sep info at utc event for basic us east us east default scheduler scheduled successfully assigned test scyllacluster mxsmm basic us east us east to fv sep info at utc event for basic scyllacluster controller poddisruptionbudgetcreated poddisruptionbudget test scyllacluster mxsmm basic created sep info at utc event for basic client scyllacluster controller servicecreated service test scyllacluster mxsmm basic client created sep info at utc event for basic us east us east scyllacluster controller servicecreated service test scyllacluster mxsmm basic us east us east created sep info at utc event for nodeconfig podinfo bbbf nodeconfigcm controller configmapcreated configmap test scyllacluster mxsmm nodeconfig podinfo bbbf created sep info at utc event for basic us east us east kubelet fv failedmount mountvolume setup failed for volume scylla client config volume failed to sync secret cache timed out waiting for the condition sep info at utc event for basic us east us east kubelet fv failedmount mountvolume setup failed for volume kube api access failed to sync configmap cache timed out waiting for the condition sep info at utc event for basic us east us east kubelet fv started started container sidecar injection sep info at utc event for basic us east us east kubelet fv pulled container image docker io scylladb scylla operator ci already present on machine sep info at utc event for basic us east us east kubelet fv created created container sidecar injection sep info at utc event for basic us east us east kubelet fv started started container scylla sep info at utc event for basic us east us east kubelet fv pulled container image docker io scylladb scylla manager agent already present on machine sep info at utc event for basic us east us east kubelet fv created created container scylla manager agent sep info at utc event for basic us east us east kubelet fv created created container scylla sep info at utc event for basic us east us east kubelet fv pulled container image docker io scylladb scylla already present on machine sep info at utc event for basic us east us east kubelet fv started started container scylla manager agent sep info at utc event for nodeconfig podinfo bbbf nodeconfigcm controller configmapupdated configmap test scyllacluster mxsmm nodeconfig podinfo bbbf updated sep info at utc event for basic us east us east kubelet fv unhealthy startup probe failed http probe failed with statuscode sep info at utc event for basic us east us east kubelet fv unhealthy readiness probe failed http probe failed with statuscode sep info at utc event for basic us east us east kubelet fv unhealthy readiness probe failed http probe failed with statuscode step collecting dumps from namespace test scyllacluster mxsmm step destroying namespace test scyllacluster mxsmm step waiting for namespace test scyllacluster mxsmm to be removed end captured ginkgowriter output unexpected error s timed out waiting for the condition timed out waiting for the condition occurred in at github com scylladb scylla operator test set scyllacluster scyllamanager go full stack trace github com scylladb scylla operator test set scyllacluster glob github com scylladb scylla operator test set scyllacluster scyllamanager go
0