Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 4
112
| repo_url
stringlengths 33
141
| action
stringclasses 3
values | title
stringlengths 1
1.02k
| labels
stringlengths 4
1.54k
| body
stringlengths 1
262k
| index
stringclasses 17
values | text_combine
stringlengths 95
262k
| label
stringclasses 2
values | text
stringlengths 96
252k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
29,962
| 4,543,058,451
|
IssuesEvent
|
2016-09-10 00:06:17
|
briansmith/ring
|
https://api.github.com/repos/briansmith/ring
|
opened
|
Ensure there are no linking conflicts if OpenSSL/BoringSSL/LibreSSL are linked along with *ring*
|
static-analysis-and-type-safety test-coverage usability
|
If we didn't do anything, then linking OpenSSL and *ring* into the same executable would cause very bad things to happen, because each defines functions with the same names but with different signatures.
The main work to resolve this was done in https://github.com/briansmith/ring/pull/280. The remaining work is:
- [ ] Verify that there are no conflicting names.
- [ ] Write a tool that can verify that there are no conflicting names
- [ ] Run that tool as part of the build on CI.
|
1.0
|
Ensure there are no linking conflicts if OpenSSL/BoringSSL/LibreSSL are linked along with *ring* - If we didn't do anything, then linking OpenSSL and *ring* into the same executable would cause very bad things to happen, because each defines functions with the same names but with different signatures.
The main work to resolve this was done in https://github.com/briansmith/ring/pull/280. The remaining work is:
- [ ] Verify that there are no conflicting names.
- [ ] Write a tool that can verify that there are no conflicting names
- [ ] Run that tool as part of the build on CI.
|
test
|
ensure there are no linking conflicts if openssl boringssl libressl are linked along with ring if we didn t do anything then linking openssl and ring into the same executable would cause very bad things to happen because each defines functions with the same names but with different signatures the main work to resolve this was done in the remaining work is verify that there are no conflicting names write a tool that can verify that there are no conflicting names run that tool as part of the build on ci
| 1
|
74,067
| 7,374,164,667
|
IssuesEvent
|
2018-03-13 19:27:05
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
"Cluster level logging implemented by Stackdriver should ingest events" fails for GKE Regional Clusters
|
kind/bug priority/critical-urgent priority/failing-test sig/instrumentation status/approved-for-milestone status/in-progress
|
This test consistently fails for GKE Regional Clusters:
https://k8s-testgrid.appspot.com/google-gke#gke-regional
```
go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/instrumentation/logging/stackdrvier/basic.go:141
Expected error:
<*errors.errorString | 0xc420f6a3f0>: {
s: "timed out waiting for ingestion, still not ingested: ",
}
timed out waiting for ingestion, still not ingested:
not to have occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/instrumentation/logging/stackdrvier/basic.go:168
```
@piosz @crassirostris - can you please triage?
|
1.0
|
"Cluster level logging implemented by Stackdriver should ingest events" fails for GKE Regional Clusters - This test consistently fails for GKE Regional Clusters:
https://k8s-testgrid.appspot.com/google-gke#gke-regional
```
go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/instrumentation/logging/stackdrvier/basic.go:141
Expected error:
<*errors.errorString | 0xc420f6a3f0>: {
s: "timed out waiting for ingestion, still not ingested: ",
}
timed out waiting for ingestion, still not ingested:
not to have occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/instrumentation/logging/stackdrvier/basic.go:168
```
@piosz @crassirostris - can you please triage?
|
test
|
cluster level logging implemented by stackdriver should ingest events fails for gke regional clusters this test consistently fails for gke regional clusters go src io kubernetes output dockerized go src io kubernetes test instrumentation logging stackdrvier basic go expected error s timed out waiting for ingestion still not ingested timed out waiting for ingestion still not ingested not to have occurred go src io kubernetes output dockerized go src io kubernetes test instrumentation logging stackdrvier basic go piosz crassirostris can you please triage
| 1
|
95,478
| 16,096,585,078
|
IssuesEvent
|
2021-04-27 01:18:48
|
camcrosbie/Angular-GettingStarted
|
https://api.github.com/repos/camcrosbie/Angular-GettingStarted
|
closed
|
CVE-2021-23337 (High) detected in lodash-4.17.20.tgz - autoclosed
|
security vulnerability
|
## CVE-2021-23337 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.20.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.20.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.20.tgz</a></p>
<p>Path to dependency file: Angular-GettingStarted/APM-Final/package.json</p>
<p>Path to vulnerable library: Angular-GettingStarted/APM-Final/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- karma-5.0.9.tgz (Root Library)
- :x: **lodash-4.17.20.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/camcrosbie/Angular-GettingStarted/commit/b46d7af5159dce3099396a98e8b99a89e538e1bc">b46d7af5159dce3099396a98e8b99a89e538e1bc</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of package lodash; all versions of package org.fujion.webjars:lodash are vulnerable to Command Injection via template.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution: lodash - 4.17.21</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-23337 (High) detected in lodash-4.17.20.tgz - autoclosed - ## CVE-2021-23337 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.20.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.20.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.20.tgz</a></p>
<p>Path to dependency file: Angular-GettingStarted/APM-Final/package.json</p>
<p>Path to vulnerable library: Angular-GettingStarted/APM-Final/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- karma-5.0.9.tgz (Root Library)
- :x: **lodash-4.17.20.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/camcrosbie/Angular-GettingStarted/commit/b46d7af5159dce3099396a98e8b99a89e538e1bc">b46d7af5159dce3099396a98e8b99a89e538e1bc</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of package lodash; all versions of package org.fujion.webjars:lodash are vulnerable to Command Injection via template.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution: lodash - 4.17.21</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve high detected in lodash tgz autoclosed cve high severity vulnerability vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file angular gettingstarted apm final package json path to vulnerable library angular gettingstarted apm final node modules lodash package json dependency hierarchy karma tgz root library x lodash tgz vulnerable library found in head commit a href found in base branch main vulnerability details all versions of package lodash all versions of package org fujion webjars lodash are vulnerable to command injection via template publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash step up your open source security game with whitesource
| 0
|
95,128
| 3,934,468,350
|
IssuesEvent
|
2016-04-25 22:49:23
|
yola/drf-paginator
|
https://api.github.com/repos/yola/drf-paginator
|
closed
|
Initial Implementation
|
frontend priority
|
The goal is to make pagination easier to deal with for our clients. Clients have request functions in the form of `request(queryParams)` that return a promise for the response.
## Paginator Object
The paginator wraps a request and provides a clean API for returning promises for the results of each page.
- `new Paginator(request, queryParams)`: constructor that takes a request and query params for said request
- `paginator.next()`: returns the next page.
- Initial call should either be page 1 or the page given as a key on `queryParams` during initialization of the paginator.
- Caches responses to prevent duplicate requests
- `paginator.previous()`: returns the previous page
- Functions much like `next` in that it will return cached responses or make new requests
## Static methods:
- `Paginator.all(request, queryParams)`: returns a promise for all the pages of an endpoint concatenated together as a single result list
|
1.0
|
Initial Implementation - The goal is to make pagination easier to deal with for our clients. Clients have request functions in the form of `request(queryParams)` that return a promise for the response.
## Paginator Object
The paginator wraps a request and provides a clean API for returning promises for the results of each page.
- `new Paginator(request, queryParams)`: constructor that takes a request and query params for said request
- `paginator.next()`: returns the next page.
- Initial call should either be page 1 or the page given as a key on `queryParams` during initialization of the paginator.
- Caches responses to prevent duplicate requests
- `paginator.previous()`: returns the previous page
- Functions much like `next` in that it will return cached responses or make new requests
## Static methods:
- `Paginator.all(request, queryParams)`: returns a promise for all the pages of an endpoint concatenated together as a single result list
|
non_test
|
initial implementation the goal is to make pagination easier to deal with for our clients clients have request functions in the form of request queryparams that return a promise for the response paginator object the paginator wraps a request and provides a clean api for returning promises for the results of each page new paginator request queryparams constructor that takes a request and query params for said request paginator next returns the next page initial call should either be page or the page given as a key on queryparams during initialization of the paginator caches responses to prevent duplicate requests paginator previous returns the previous page functions much like next in that it will return cached responses or make new requests static methods paginator all request queryparams returns a promise for all the pages of an endpoint concatenated together as a single result list
| 0
|
57,217
| 6,540,781,619
|
IssuesEvent
|
2017-09-01 16:51:05
|
openbmc/openbmc-test-automation
|
https://api.github.com/repos/openbmc/openbmc-test-automation
|
opened
|
Power supply Sensors test
|
Test
|
- [x] Check inventory and see how many list and functional
- [x] Verify the ps0/ps1 sensors
|
1.0
|
Power supply Sensors test - - [x] Check inventory and see how many list and functional
- [x] Verify the ps0/ps1 sensors
|
test
|
power supply sensors test check inventory and see how many list and functional verify the sensors
| 1
|
193,272
| 14,645,522,561
|
IssuesEvent
|
2020-12-26 08:21:27
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
opened
|
Failing test: Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/dashboard/feature_controls/dashboard_security·ts - dashboard feature controls dashboard feature controls security global dashboard read-only privileges "before all" hook for "shows dashboard navlink"
|
failed-test
|
A test failed on a tracked branch
```
Error: retry.tryForTime timeout: Error: retry.try timeout: TimeoutError: timeout: Timed out receiving message from renderer: 0.000
(Session info: headless chrome=87.0.4280.88)
at Object.throwDecodedError (/dev/shm/workspace/kibana/node_modules/selenium-webdriver/lib/error.js:550:15)
at parseHttpResponse (/dev/shm/workspace/kibana/node_modules/selenium-webdriver/lib/http.js:565:13)
at Executor.execute (/dev/shm/workspace/kibana/node_modules/selenium-webdriver/lib/http.js:491:26)
at runMicrotasks (<anonymous>)
at processTicksAndRejections (internal/process/task_queues.js:93:5)
at Task.exec (/dev/shm/workspace/parallel/17/kibana/test/functional/services/remote/prevent_parallel_calls.ts:40:22)
at onFailure (/dev/shm/workspace/parallel/17/kibana/test/common/services/retry/retry_for_success.ts:28:9)
at retryForSuccess (/dev/shm/workspace/parallel/17/kibana/test/common/services/retry/retry_for_success.ts:68:13)
at Retry.try (/dev/shm/workspace/parallel/17/kibana/test/common/services/retry/retry.ts:43:14)
at /dev/shm/workspace/parallel/17/kibana/test/functional/page_objects/common_page.ts:238:23
at runAttempt (/dev/shm/workspace/parallel/17/kibana/test/common/services/retry/retry_for_success.ts:38:15)
at retryForSuccess (/dev/shm/workspace/parallel/17/kibana/test/common/services/retry/retry_for_success.ts:77:21)
at Retry.tryForTime (/dev/shm/workspace/parallel/17/kibana/test/common/services/retry/retry.ts:34:14)
at CommonPage.navigateToApp (/dev/shm/workspace/parallel/17/kibana/test/functional/page_objects/common_page.ts:237:7)
at Object.login (test/functional/page_objects/security_page.ts:117:9)
at SecurityPage.login (test/functional/page_objects/security_page.ts:208:7)
at Context.<anonymous> (test/functional/apps/dashboard/feature_controls/dashboard_security.ts:274:9)
at Object.apply (/dev/shm/workspace/parallel/17/kibana/packages/kbn-test/src/functional_test_runner/lib/mocha/wrap_function.js:84:16)
at onFailure (/dev/shm/workspace/parallel/17/kibana/test/common/services/retry/retry_for_success.ts:28:9)
at retryForSuccess (/dev/shm/workspace/parallel/17/kibana/test/common/services/retry/retry_for_success.ts:68:13)
at Retry.tryForTime (/dev/shm/workspace/parallel/17/kibana/test/common/services/retry/retry.ts:34:14)
at CommonPage.navigateToApp (/dev/shm/workspace/parallel/17/kibana/test/functional/page_objects/common_page.ts:237:7)
at Object.login (test/functional/page_objects/security_page.ts:117:9)
at SecurityPage.login (test/functional/page_objects/security_page.ts:208:7)
at Context.<anonymous> (test/functional/apps/dashboard/feature_controls/dashboard_security.ts:274:9)
at Object.apply (/dev/shm/workspace/parallel/17/kibana/packages/kbn-test/src/functional_test_runner/lib/mocha/wrap_function.js:84:16)
```
First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/10669/)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/dashboard/feature_controls/dashboard_security·ts","test.name":"dashboard feature controls dashboard feature controls security global dashboard read-only privileges \"before all\" hook for \"shows dashboard navlink\"","test.failCount":1}} -->
|
1.0
|
Failing test: Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/dashboard/feature_controls/dashboard_security·ts - dashboard feature controls dashboard feature controls security global dashboard read-only privileges "before all" hook for "shows dashboard navlink" - A test failed on a tracked branch
```
Error: retry.tryForTime timeout: Error: retry.try timeout: TimeoutError: timeout: Timed out receiving message from renderer: 0.000
(Session info: headless chrome=87.0.4280.88)
at Object.throwDecodedError (/dev/shm/workspace/kibana/node_modules/selenium-webdriver/lib/error.js:550:15)
at parseHttpResponse (/dev/shm/workspace/kibana/node_modules/selenium-webdriver/lib/http.js:565:13)
at Executor.execute (/dev/shm/workspace/kibana/node_modules/selenium-webdriver/lib/http.js:491:26)
at runMicrotasks (<anonymous>)
at processTicksAndRejections (internal/process/task_queues.js:93:5)
at Task.exec (/dev/shm/workspace/parallel/17/kibana/test/functional/services/remote/prevent_parallel_calls.ts:40:22)
at onFailure (/dev/shm/workspace/parallel/17/kibana/test/common/services/retry/retry_for_success.ts:28:9)
at retryForSuccess (/dev/shm/workspace/parallel/17/kibana/test/common/services/retry/retry_for_success.ts:68:13)
at Retry.try (/dev/shm/workspace/parallel/17/kibana/test/common/services/retry/retry.ts:43:14)
at /dev/shm/workspace/parallel/17/kibana/test/functional/page_objects/common_page.ts:238:23
at runAttempt (/dev/shm/workspace/parallel/17/kibana/test/common/services/retry/retry_for_success.ts:38:15)
at retryForSuccess (/dev/shm/workspace/parallel/17/kibana/test/common/services/retry/retry_for_success.ts:77:21)
at Retry.tryForTime (/dev/shm/workspace/parallel/17/kibana/test/common/services/retry/retry.ts:34:14)
at CommonPage.navigateToApp (/dev/shm/workspace/parallel/17/kibana/test/functional/page_objects/common_page.ts:237:7)
at Object.login (test/functional/page_objects/security_page.ts:117:9)
at SecurityPage.login (test/functional/page_objects/security_page.ts:208:7)
at Context.<anonymous> (test/functional/apps/dashboard/feature_controls/dashboard_security.ts:274:9)
at Object.apply (/dev/shm/workspace/parallel/17/kibana/packages/kbn-test/src/functional_test_runner/lib/mocha/wrap_function.js:84:16)
at onFailure (/dev/shm/workspace/parallel/17/kibana/test/common/services/retry/retry_for_success.ts:28:9)
at retryForSuccess (/dev/shm/workspace/parallel/17/kibana/test/common/services/retry/retry_for_success.ts:68:13)
at Retry.tryForTime (/dev/shm/workspace/parallel/17/kibana/test/common/services/retry/retry.ts:34:14)
at CommonPage.navigateToApp (/dev/shm/workspace/parallel/17/kibana/test/functional/page_objects/common_page.ts:237:7)
at Object.login (test/functional/page_objects/security_page.ts:117:9)
at SecurityPage.login (test/functional/page_objects/security_page.ts:208:7)
at Context.<anonymous> (test/functional/apps/dashboard/feature_controls/dashboard_security.ts:274:9)
at Object.apply (/dev/shm/workspace/parallel/17/kibana/packages/kbn-test/src/functional_test_runner/lib/mocha/wrap_function.js:84:16)
```
First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/10669/)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/dashboard/feature_controls/dashboard_security·ts","test.name":"dashboard feature controls dashboard feature controls security global dashboard read-only privileges \"before all\" hook for \"shows dashboard navlink\"","test.failCount":1}} -->
|
test
|
failing test chrome x pack ui functional tests x pack test functional apps dashboard feature controls dashboard security·ts dashboard feature controls dashboard feature controls security global dashboard read only privileges before all hook for shows dashboard navlink a test failed on a tracked branch error retry tryfortime timeout error retry try timeout timeouterror timeout timed out receiving message from renderer session info headless chrome at object throwdecodederror dev shm workspace kibana node modules selenium webdriver lib error js at parsehttpresponse dev shm workspace kibana node modules selenium webdriver lib http js at executor execute dev shm workspace kibana node modules selenium webdriver lib http js at runmicrotasks at processticksandrejections internal process task queues js at task exec dev shm workspace parallel kibana test functional services remote prevent parallel calls ts at onfailure dev shm workspace parallel kibana test common services retry retry for success ts at retryforsuccess dev shm workspace parallel kibana test common services retry retry for success ts at retry try dev shm workspace parallel kibana test common services retry retry ts at dev shm workspace parallel kibana test functional page objects common page ts at runattempt dev shm workspace parallel kibana test common services retry retry for success ts at retryforsuccess dev shm workspace parallel kibana test common services retry retry for success ts at retry tryfortime dev shm workspace parallel kibana test common services retry retry ts at commonpage navigatetoapp dev shm workspace parallel kibana test functional page objects common page ts at object login test functional page objects security page ts at securitypage login test functional page objects security page ts at context test functional apps dashboard feature controls dashboard security ts at object apply dev shm workspace parallel kibana packages kbn test src functional test runner lib mocha wrap function js at onfailure dev shm workspace parallel kibana test common services retry retry for success ts at retryforsuccess dev shm workspace parallel kibana test common services retry retry for success ts at retry tryfortime dev shm workspace parallel kibana test common services retry retry ts at commonpage navigatetoapp dev shm workspace parallel kibana test functional page objects common page ts at object login test functional page objects security page ts at securitypage login test functional page objects security page ts at context test functional apps dashboard feature controls dashboard security ts at object apply dev shm workspace parallel kibana packages kbn test src functional test runner lib mocha wrap function js first failure
| 1
|
9,170
| 3,025,066,613
|
IssuesEvent
|
2015-08-03 04:37:53
|
servo/servo
|
https://api.github.com/repos/servo/servo
|
closed
|
Intermittent timeout in /css21_dev/html4/absolute-replaced-height-012.htm
|
A-testing I-intermittent P-linux
|
```
0:31.85 TEST_START: Thread-TestrunnerManager-8 /css21_dev/html4/absolute-replaced-height-012.htm
0:47.89 TEST_END: Thread-TestrunnerManager-8 TIMEOUT, expected FAIL
```
|
1.0
|
Intermittent timeout in /css21_dev/html4/absolute-replaced-height-012.htm - ```
0:31.85 TEST_START: Thread-TestrunnerManager-8 /css21_dev/html4/absolute-replaced-height-012.htm
0:47.89 TEST_END: Thread-TestrunnerManager-8 TIMEOUT, expected FAIL
```
|
test
|
intermittent timeout in dev absolute replaced height htm test start thread testrunnermanager dev absolute replaced height htm test end thread testrunnermanager timeout expected fail
| 1
|
283,389
| 8,719,405,683
|
IssuesEvent
|
2018-12-08 00:38:19
|
aowen87/BAR
|
https://api.github.com/repos/aowen87/BAR
|
closed
|
VisIt hangs during re-execution prompted by pick.
|
bug crash likelihood medium priority reviewed severity high wrong results
|
This is a bug Bruce Hammel claims to have been experiencing with VisIt for years, so I've set the priority as high, now that I have reproducible steps.
This seems only to occur with 2 nodes. Multiple processors on single node does not replicate.
Also seems only to occur in conjunction with CoordSwap operator, and with Pick var being set to an expression (and not the pipeline var)
On surface: (ensure parallel engine with 2 nodes)
Open multi_curv2d.silo
Add PC Plot of d
Add CoordSwap operator, swap x and y coords
Draw.
Create a scalar expression d+p,
Open Pick window, set variable to d+p.
Apply
Do a Zone Pick
Using Navigation, change the view either by zooming or panning.
Do another Zone Pick.
Engine will hang, must cancel the engine_par job in order to interact with VisIt again.
Information window shows Pick wanting to re-execute, and a merge exception:
VisIt does not have all the information it needs to perform a pick. Please wait while the necessary information is calculated. All current pick selections have been cached and will be performed when calculations are complete. VisIt will notify you when it is fully ready for more picks.
Shortly thereafter, the following occured...
Pseudocolor: (InvalidMergeException)
viewer: Cannot merge datasets because of an incompatible field 1 and 2.
Pick mode now fully ready.
+
This shows error seeming to come from the viewer, but if you run with -debug 5, then 2 processors' log files will show this error:
+
This source should not load balance the data.
Exception: (InvalidMergeException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/avt/Pipeline/Data/avtDataAttributes.C, line 1360: Cannot merge datasets because of an incompatible field 1 and 2.
catch(VisItException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/engine/main/Executors.h:1027
This is consistent with what I saw with Bruce's real data, running on Muir in pdebug with 10 nodes and 120 processors, but also 2 nodes and 24 processors.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2169
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: VisIt hangs during re-execution prompted by pick.
Assigned to: Kathleen Biagas
Category:
Target version: 2.9.1
Author: Kathleen Biagas
Start: 03/03/2015
Due date:
% Done: 100
Estimated time:
Created: 03/03/2015 04:55 pm
Updated: 03/20/2015 05:44 pm
Likelihood: 3 - Occasional
Severity: 4 - Crash / Wrong Results
Found in version: 2.8.2
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
This is a bug Bruce Hammel claims to have been experiencing with VisIt for years, so I've set the priority as high, now that I have reproducible steps.
This seems only to occur with 2 nodes. Multiple processors on single node does not replicate.
Also seems only to occur in conjunction with CoordSwap operator, and with Pick var being set to an expression (and not the pipeline var)
On surface: (ensure parallel engine with 2 nodes)
Open multi_curv2d.silo
Add PC Plot of d
Add CoordSwap operator, swap x and y coords
Draw.
Create a scalar expression d+p,
Open Pick window, set variable to d+p.
Apply
Do a Zone Pick
Using Navigation, change the view either by zooming or panning.
Do another Zone Pick.
Engine will hang, must cancel the engine_par job in order to interact with VisIt again.
Information window shows Pick wanting to re-execute, and a merge exception:
++++++++++++++++
VisIt does not have all the information it needs to perform a pick. Please wait while the necessary information is calculated. All current pick selections have been cached and will be performed when calculations are complete. VisIt will notify you when it is fully ready for more picks.
Shortly thereafter, the following occured...
Pseudocolor: (InvalidMergeException)
viewer: Cannot merge datasets because of an incompatible field 1 and 2.
Pick mode now fully ready.
+++++++++++++++++++++
This shows error seeming to come from the viewer, but if you run with -debug 5, then 2 processors' log files will show this error:
+++++++++++++++++++++++++++++++++++++
This source should not load balance the data.
Exception: (InvalidMergeException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/avt/Pipeline/Data/avtDataAttributes.C, line 1360: Cannot merge datasets because of an incompatible field 1 and 2.
catch(VisItException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/engine/main/Executors.h:1027
++++++++++++++++++++++++++++++
This is consistent with what I saw with Bruce's real data, running on Muir in pdebug with 10 nodes and 120 processors, but also 2 nodes and 24 processors.
Comments:
Turns out that Pick was sending the secondary variable request to only 1 processor, which caused an 'Invalid Merge' exception during pipeline re-execution.I modified Pick to request SecondaryVars of all processors.SVN update 25986 (2.9RC), 25988 (trunk)M /src/avt/Queries/Pick/avtPickQuery.C
|
1.0
|
VisIt hangs during re-execution prompted by pick. - This is a bug Bruce Hammel claims to have been experiencing with VisIt for years, so I've set the priority as high, now that I have reproducible steps.
This seems only to occur with 2 nodes. Multiple processors on single node does not replicate.
Also seems only to occur in conjunction with CoordSwap operator, and with Pick var being set to an expression (and not the pipeline var)
On surface: (ensure parallel engine with 2 nodes)
Open multi_curv2d.silo
Add PC Plot of d
Add CoordSwap operator, swap x and y coords
Draw.
Create a scalar expression d+p,
Open Pick window, set variable to d+p.
Apply
Do a Zone Pick
Using Navigation, change the view either by zooming or panning.
Do another Zone Pick.
Engine will hang, must cancel the engine_par job in order to interact with VisIt again.
Information window shows Pick wanting to re-execute, and a merge exception:
VisIt does not have all the information it needs to perform a pick. Please wait while the necessary information is calculated. All current pick selections have been cached and will be performed when calculations are complete. VisIt will notify you when it is fully ready for more picks.
Shortly thereafter, the following occured...
Pseudocolor: (InvalidMergeException)
viewer: Cannot merge datasets because of an incompatible field 1 and 2.
Pick mode now fully ready.
+
This shows error seeming to come from the viewer, but if you run with -debug 5, then 2 processors' log files will show this error:
+
This source should not load balance the data.
Exception: (InvalidMergeException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/avt/Pipeline/Data/avtDataAttributes.C, line 1360: Cannot merge datasets because of an incompatible field 1 and 2.
catch(VisItException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/engine/main/Executors.h:1027
This is consistent with what I saw with Bruce's real data, running on Muir in pdebug with 10 nodes and 120 processors, but also 2 nodes and 24 processors.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2169
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: VisIt hangs during re-execution prompted by pick.
Assigned to: Kathleen Biagas
Category:
Target version: 2.9.1
Author: Kathleen Biagas
Start: 03/03/2015
Due date:
% Done: 100
Estimated time:
Created: 03/03/2015 04:55 pm
Updated: 03/20/2015 05:44 pm
Likelihood: 3 - Occasional
Severity: 4 - Crash / Wrong Results
Found in version: 2.8.2
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
This is a bug Bruce Hammel claims to have been experiencing with VisIt for years, so I've set the priority as high, now that I have reproducible steps.
This seems only to occur with 2 nodes. Multiple processors on single node does not replicate.
Also seems only to occur in conjunction with CoordSwap operator, and with Pick var being set to an expression (and not the pipeline var)
On surface: (ensure parallel engine with 2 nodes)
Open multi_curv2d.silo
Add PC Plot of d
Add CoordSwap operator, swap x and y coords
Draw.
Create a scalar expression d+p,
Open Pick window, set variable to d+p.
Apply
Do a Zone Pick
Using Navigation, change the view either by zooming or panning.
Do another Zone Pick.
Engine will hang, must cancel the engine_par job in order to interact with VisIt again.
Information window shows Pick wanting to re-execute, and a merge exception:
++++++++++++++++
VisIt does not have all the information it needs to perform a pick. Please wait while the necessary information is calculated. All current pick selections have been cached and will be performed when calculations are complete. VisIt will notify you when it is fully ready for more picks.
Shortly thereafter, the following occured...
Pseudocolor: (InvalidMergeException)
viewer: Cannot merge datasets because of an incompatible field 1 and 2.
Pick mode now fully ready.
+++++++++++++++++++++
This shows error seeming to come from the viewer, but if you run with -debug 5, then 2 processors' log files will show this error:
+++++++++++++++++++++++++++++++++++++
This source should not load balance the data.
Exception: (InvalidMergeException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/avt/Pipeline/Data/avtDataAttributes.C, line 1360: Cannot merge datasets because of an incompatible field 1 and 2.
catch(VisItException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/engine/main/Executors.h:1027
++++++++++++++++++++++++++++++
This is consistent with what I saw with Bruce's real data, running on Muir in pdebug with 10 nodes and 120 processors, but also 2 nodes and 24 processors.
Comments:
Turns out that Pick was sending the secondary variable request to only 1 processor, which caused an 'Invalid Merge' exception during pipeline re-execution.I modified Pick to request SecondaryVars of all processors.SVN update 25986 (2.9RC), 25988 (trunk)M /src/avt/Queries/Pick/avtPickQuery.C
|
non_test
|
visit hangs during re execution prompted by pick this is a bug bruce hammel claims to have been experiencing with visit for years so i ve set the priority as high now that i have reproducible steps this seems only to occur with nodes multiple processors on single node does not replicate also seems only to occur in conjunction with coordswap operator and with pick var being set to an expression and not the pipeline var on surface ensure parallel engine with nodes open multi silo add pc plot of d add coordswap operator swap x and y coords draw create a scalar expression d p open pick window set variable to d p apply do a zone pick using navigation change the view either by zooming or panning do another zone pick engine will hang must cancel the engine par job in order to interact with visit again information window shows pick wanting to re execute and a merge exception visit does not have all the information it needs to perform a pick please wait while the necessary information is calculated all current pick selections have been cached and will be performed when calculations are complete visit will notify you when it is fully ready for more picks shortly thereafter the following occured pseudocolor invalidmergeexception viewer cannot merge datasets because of an incompatible field and pick mode now fully ready this shows error seeming to come from the viewer but if you run with debug then processors log files will show this error this source should not load balance the data exception invalidmergeexception usr tmp brugger aztec visitbuild src avt pipeline data avtdataattributes c line cannot merge datasets because of an incompatible field and catch visitexception usr tmp brugger aztec visitbuild src engine main executors h this is consistent with what i saw with bruce s real data running on muir in pdebug with nodes and processors but also nodes and processors redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject visit hangs during re execution prompted by pick assigned to kathleen biagas category target version author kathleen biagas start due date done estimated time created pm updated pm likelihood occasional severity crash wrong results found in version impact expected use os all support group any description this is a bug bruce hammel claims to have been experiencing with visit for years so i ve set the priority as high now that i have reproducible steps this seems only to occur with nodes multiple processors on single node does not replicate also seems only to occur in conjunction with coordswap operator and with pick var being set to an expression and not the pipeline var on surface ensure parallel engine with nodes open multi silo add pc plot of d add coordswap operator swap x and y coords draw create a scalar expression d p open pick window set variable to d p apply do a zone pick using navigation change the view either by zooming or panning do another zone pick engine will hang must cancel the engine par job in order to interact with visit again information window shows pick wanting to re execute and a merge exception visit does not have all the information it needs to perform a pick please wait while the necessary information is calculated all current pick selections have been cached and will be performed when calculations are complete visit will notify you when it is fully ready for more picks shortly thereafter the following occured pseudocolor invalidmergeexception viewer cannot merge datasets because of an incompatible field and pick mode now fully ready this shows error seeming to come from the viewer but if you run with debug then processors log files will show this error this source should not load balance the data exception invalidmergeexception usr tmp brugger aztec visitbuild src avt pipeline data avtdataattributes c line cannot merge datasets because of an incompatible field and catch visitexception usr tmp brugger aztec visitbuild src engine main executors h this is consistent with what i saw with bruce s real data running on muir in pdebug with nodes and processors but also nodes and processors comments turns out that pick was sending the secondary variable request to only processor which caused an invalid merge exception during pipeline re execution i modified pick to request secondaryvars of all processors svn update trunk m src avt queries pick avtpickquery c
| 0
|
13,385
| 3,330,801,089
|
IssuesEvent
|
2015-11-11 12:46:11
|
krampstudio/future.js
|
https://api.github.com/repos/krampstudio/future.js
|
opened
|
Replace saue labs
|
Tests
|
As Sauce Labs closes it's free tier, we need to figure out how to replace it. Or at least run locally the tests against multiple browsers.
|
1.0
|
Replace saue labs - As Sauce Labs closes it's free tier, we need to figure out how to replace it. Or at least run locally the tests against multiple browsers.
|
test
|
replace saue labs as sauce labs closes it s free tier we need to figure out how to replace it or at least run locally the tests against multiple browsers
| 1
|
96,822
| 28,021,372,857
|
IssuesEvent
|
2023-03-28 05:48:26
|
pmem/pmdk
|
https://api.github.com/repos/pmem/pmdk
|
closed
|
RUNTESTS.py crashes when ndctl is not available.
|
Exposure: Medium Priority: 3 medium Type: Bug build system
|
# ISSUE: RUNTESTS.py crashes when ndctl is not available.
## Environment Information
- PMDK package version(s):
- OS(es) version(s): Ubuntu 22.04
- ndctl version(s): **None**
- kernel version(s): 5.15.90.1-microsoft-standard-WSL2
- compiler, libraries, packaging and other related tools version(s): Python 3.10.6
## Please provide a reproduction of the bug:
```
$sudo apt remove ndctl
$make
$cd src/test ; RUNTESTS.py
```
## How often bug is revealed: (always, often, rare): always
## Actual behavior:
```
$cd src/test ; RUNTESTS.py
obj_critnib/TEST0: SETUP (medium/static_release/non)
obj_critnib/TEST0: PASS [01.104 s]
obj_critnib/TEST0: SETUP (medium/debug/non)
obj_critnib/TEST0: PASS [01.134 s]
obj_critnib/TEST0: SETUP (medium/release/non)
obj_critnib/TEST0: PASS [01.389 s]
obj_critnib/TEST0: SETUP (medium/static_debug/non)
obj_critnib/TEST0: PASS [01.497 s]
pmem2_movnt_align/TEST5: SKIP: Valgrind tool 'pmemcheck' was not found
...
pmem2_vm_reservation/TEST32: SKIP: No dax devices defined in testconfig
Traceback (most recent call last):
File "/home/tgromadz/repos/pmdk/src/test/./RUNTESTS.py", line 199, in <module>
main()
File "/home/tgromadz/repos/pmdk/src/test/./RUNTESTS.py", line 195, in main
sys.exit(runner.run_tests())
File "/home/tgromadz/repos/pmdk/src/test/./RUNTESTS.py", line 107, in run_tests
for c in cf.get_contexts():
File "/home/tgromadz/repos/pmdk/src/test/unittest/ctx_filter.py", line 52, in get_contexts
if not reqs.check_if_all_requirements_are_met(self.tc):
File "/home/tgromadz/repos/pmdk/src/test/unittest/requirements.py", line 114, in check_if_all_requirements_are_met
if not self._check_ndctl_req_is_met(tc):
File "/home/tgromadz/repos/pmdk/src/test/unittest/requirements.py", line 89, in _check_ndctl_req_is_met
self.check_namespace()
File "/home/tgromadz/repos/pmdk/src/test/unittest/requirements.py", line 69, in check_namespace
proc = sp.run(cmd, stdout=sp.PIPE, stderr=sp.STDOUT,
File "/usr/lib/python3.10/subprocess.py", line 501, in run
with Popen(*popenargs, **kwargs) as process:
File "/usr/lib/python3.10/subprocess.py", line 969, in __init__
self._execute_child(args, executable, preexec_fn, close_fds,
File "/usr/lib/python3.10/subprocess.py", line 1845, in _execute_child
raise child_exception_type(errno_num, err_msg, err_filename)
FileNotFoundError: [Errno 2] No such file or directory: 'ndctl'
```
## Expected behavior:
PMem related tests shall be skiped:
```
pmem2_vm_reservation/TEST32: SKIP: No dax devices defined in testconfig
pmem2_badblock/TEST0: SKIP: no ndctl namespace set
```
## Details
## Additional information about Priority and Help Requested:
Are you willing to submit a pull request with a proposed change? (Yes, No)
Requested priority: (Showstopper, High, Medium, Low)
|
1.0
|
RUNTESTS.py crashes when ndctl is not available. - # ISSUE: RUNTESTS.py crashes when ndctl is not available.
## Environment Information
- PMDK package version(s):
- OS(es) version(s): Ubuntu 22.04
- ndctl version(s): **None**
- kernel version(s): 5.15.90.1-microsoft-standard-WSL2
- compiler, libraries, packaging and other related tools version(s): Python 3.10.6
## Please provide a reproduction of the bug:
```
$sudo apt remove ndctl
$make
$cd src/test ; RUNTESTS.py
```
## How often bug is revealed: (always, often, rare): always
## Actual behavior:
```
$cd src/test ; RUNTESTS.py
obj_critnib/TEST0: SETUP (medium/static_release/non)
obj_critnib/TEST0: PASS [01.104 s]
obj_critnib/TEST0: SETUP (medium/debug/non)
obj_critnib/TEST0: PASS [01.134 s]
obj_critnib/TEST0: SETUP (medium/release/non)
obj_critnib/TEST0: PASS [01.389 s]
obj_critnib/TEST0: SETUP (medium/static_debug/non)
obj_critnib/TEST0: PASS [01.497 s]
pmem2_movnt_align/TEST5: SKIP: Valgrind tool 'pmemcheck' was not found
...
pmem2_vm_reservation/TEST32: SKIP: No dax devices defined in testconfig
Traceback (most recent call last):
File "/home/tgromadz/repos/pmdk/src/test/./RUNTESTS.py", line 199, in <module>
main()
File "/home/tgromadz/repos/pmdk/src/test/./RUNTESTS.py", line 195, in main
sys.exit(runner.run_tests())
File "/home/tgromadz/repos/pmdk/src/test/./RUNTESTS.py", line 107, in run_tests
for c in cf.get_contexts():
File "/home/tgromadz/repos/pmdk/src/test/unittest/ctx_filter.py", line 52, in get_contexts
if not reqs.check_if_all_requirements_are_met(self.tc):
File "/home/tgromadz/repos/pmdk/src/test/unittest/requirements.py", line 114, in check_if_all_requirements_are_met
if not self._check_ndctl_req_is_met(tc):
File "/home/tgromadz/repos/pmdk/src/test/unittest/requirements.py", line 89, in _check_ndctl_req_is_met
self.check_namespace()
File "/home/tgromadz/repos/pmdk/src/test/unittest/requirements.py", line 69, in check_namespace
proc = sp.run(cmd, stdout=sp.PIPE, stderr=sp.STDOUT,
File "/usr/lib/python3.10/subprocess.py", line 501, in run
with Popen(*popenargs, **kwargs) as process:
File "/usr/lib/python3.10/subprocess.py", line 969, in __init__
self._execute_child(args, executable, preexec_fn, close_fds,
File "/usr/lib/python3.10/subprocess.py", line 1845, in _execute_child
raise child_exception_type(errno_num, err_msg, err_filename)
FileNotFoundError: [Errno 2] No such file or directory: 'ndctl'
```
## Expected behavior:
PMem related tests shall be skiped:
```
pmem2_vm_reservation/TEST32: SKIP: No dax devices defined in testconfig
pmem2_badblock/TEST0: SKIP: no ndctl namespace set
```
## Details
## Additional information about Priority and Help Requested:
Are you willing to submit a pull request with a proposed change? (Yes, No)
Requested priority: (Showstopper, High, Medium, Low)
|
non_test
|
runtests py crashes when ndctl is not available issue runtests py crashes when ndctl is not available environment information pmdk package version s os es version s ubuntu ndctl version s none kernel version s microsoft standard compiler libraries packaging and other related tools version s python please provide a reproduction of the bug sudo apt remove ndctl make cd src test runtests py how often bug is revealed always often rare always actual behavior cd src test runtests py obj critnib setup medium static release non obj critnib pass obj critnib setup medium debug non obj critnib pass obj critnib setup medium release non obj critnib pass obj critnib setup medium static debug non obj critnib pass movnt align skip valgrind tool pmemcheck was not found vm reservation skip no dax devices defined in testconfig traceback most recent call last file home tgromadz repos pmdk src test runtests py line in main file home tgromadz repos pmdk src test runtests py line in main sys exit runner run tests file home tgromadz repos pmdk src test runtests py line in run tests for c in cf get contexts file home tgromadz repos pmdk src test unittest ctx filter py line in get contexts if not reqs check if all requirements are met self tc file home tgromadz repos pmdk src test unittest requirements py line in check if all requirements are met if not self check ndctl req is met tc file home tgromadz repos pmdk src test unittest requirements py line in check ndctl req is met self check namespace file home tgromadz repos pmdk src test unittest requirements py line in check namespace proc sp run cmd stdout sp pipe stderr sp stdout file usr lib subprocess py line in run with popen popenargs kwargs as process file usr lib subprocess py line in init self execute child args executable preexec fn close fds file usr lib subprocess py line in execute child raise child exception type errno num err msg err filename filenotfounderror no such file or directory ndctl expected behavior pmem related tests shall be skiped vm reservation skip no dax devices defined in testconfig badblock skip no ndctl namespace set details additional information about priority and help requested are you willing to submit a pull request with a proposed change yes no requested priority showstopper high medium low
| 0
|
219,704
| 24,523,246,205
|
IssuesEvent
|
2022-10-11 11:11:11
|
QubesOS/qubes-issues
|
https://api.github.com/repos/QubesOS/qubes-issues
|
closed
|
Update zlib in dom0
|
T: bug C: other P: critical security
|
[How to file a helpful issue](https://www.qubes-os.org/doc/issue-tracking/)
### Qubes OS release
R4.1
### Brief summary
dom0 uses an old zlib which has at least two known vulnerabilities. zlib is used to e.g. compress and decompress PNG images and backups. Qubes OS assumes that compressing untrusted data is safe, and that decompressing the resulting data is also safe. One of the vulnerabilities breaks this assumpiton, provided that specific compression parameters (which are not the defaults) are used. It is not known if any code in Qubes OS uses parameters that expose the vulnerability, but zlib should still be updated out of caution.
### Steps to reproduce
Check version of zlib in dom0
### Expected behavior
dom0 version of zlib is secure
### Actual behavior
dom0 version of zlib is vulnerable
|
True
|
Update zlib in dom0 - [How to file a helpful issue](https://www.qubes-os.org/doc/issue-tracking/)
### Qubes OS release
R4.1
### Brief summary
dom0 uses an old zlib which has at least two known vulnerabilities. zlib is used to e.g. compress and decompress PNG images and backups. Qubes OS assumes that compressing untrusted data is safe, and that decompressing the resulting data is also safe. One of the vulnerabilities breaks this assumpiton, provided that specific compression parameters (which are not the defaults) are used. It is not known if any code in Qubes OS uses parameters that expose the vulnerability, but zlib should still be updated out of caution.
### Steps to reproduce
Check version of zlib in dom0
### Expected behavior
dom0 version of zlib is secure
### Actual behavior
dom0 version of zlib is vulnerable
|
non_test
|
update zlib in qubes os release brief summary uses an old zlib which has at least two known vulnerabilities zlib is used to e g compress and decompress png images and backups qubes os assumes that compressing untrusted data is safe and that decompressing the resulting data is also safe one of the vulnerabilities breaks this assumpiton provided that specific compression parameters which are not the defaults are used it is not known if any code in qubes os uses parameters that expose the vulnerability but zlib should still be updated out of caution steps to reproduce check version of zlib in expected behavior version of zlib is secure actual behavior version of zlib is vulnerable
| 0
|
202,593
| 15,287,043,125
|
IssuesEvent
|
2021-02-23 15:21:06
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachtest: ycsb/A/nodes=3/cpu=32 failed
|
C-test-failure O-roachtest O-robot branch-release-20.2 release-blocker
|
[(roachtest).ycsb/A/nodes=3/cpu=32 failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2657159&tab=buildLog) on [release-20.2@8c79e2bc4b35d36c8527f4c40c974f03d9034f46](https://github.com/cockroachdb/cockroach/commits/8c79e2bc4b35d36c8527f4c40c974f03d9034f46):
```
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1374
Wraps: (2) output in run_083043.166_n4_workload_run_ycsb
Wraps: (3) /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-2657159-1612856700-34-n4cpu32:4 -- ./workload run ycsb --init --insert-count=1000000 --workload=A --concurrency=144 --splits=3 --histograms=perf/stats.json --select-for-update=true --ramp=1m --duration=10m {pgurl:1-3} returned
| stderr:
| ./workload: /lib/x86_64-linux-gnu/libm.so.6: version `GLIBC_2.29' not found (required by ./workload)
| Error: COMMAND_PROBLEM: exit status 1
| (1) COMMAND_PROBLEM
| Wraps: (2) Node 4. Command with error:
| | ```
| | ./workload run ycsb --init --insert-count=1000000 --workload=A --concurrency=144 --splits=3 --histograms=perf/stats.json --select-for-update=true --ramp=1m --duration=10m {pgurl:1-3}
| | ```
| Wraps: (3) exit status 1
| Error types: (1) errors.Cmd (2) *hintdetail.withDetail (3) *exec.ExitError
|
| stdout:
Wraps: (4) exit status 20
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *main.withCommandDetails (4) *exec.ExitError
cluster.go:2654,ycsb.go:62,ycsb.go:79,test_runner.go:755: monitor failure: monitor task failed: t.Fatal() was called
(1) attached stack trace
-- stack trace:
| main.(*monitor).WaitE
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2642
| main.(*monitor).Wait
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2650
| main.registerYCSB.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/ycsb.go:62
| main.registerYCSB.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/ycsb.go:79
| main.(*testRunner).runTest.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:755
Wraps: (2) monitor failure
Wraps: (3) attached stack trace
-- stack trace:
| main.(*monitor).wait.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2698
Wraps: (4) monitor task failed
Wraps: (5) attached stack trace
-- stack trace:
| main.init
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2612
| runtime.doInit
| /usr/local/go/src/runtime/proc.go:5652
| runtime.main
| /usr/local/go/src/runtime/proc.go:191
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1374
Wraps: (6) t.Fatal() was called
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.withPrefix (5) *withstack.withStack (6) *errutil.leafError
```
<details><summary>More</summary><p>
Artifacts: [/ycsb/A/nodes=3/cpu=32](https://teamcity.cockroachdb.com/viewLog.html?buildId=2657159&tab=artifacts#/ycsb/A/nodes=3/cpu=32)
Related:
- #59933 roachtest: ycsb/A/nodes=3/cpu=32 failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Aycsb%2FA%2Fnodes%3D3%2Fcpu%3D32.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
2.0
|
roachtest: ycsb/A/nodes=3/cpu=32 failed - [(roachtest).ycsb/A/nodes=3/cpu=32 failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2657159&tab=buildLog) on [release-20.2@8c79e2bc4b35d36c8527f4c40c974f03d9034f46](https://github.com/cockroachdb/cockroach/commits/8c79e2bc4b35d36c8527f4c40c974f03d9034f46):
```
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1374
Wraps: (2) output in run_083043.166_n4_workload_run_ycsb
Wraps: (3) /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-2657159-1612856700-34-n4cpu32:4 -- ./workload run ycsb --init --insert-count=1000000 --workload=A --concurrency=144 --splits=3 --histograms=perf/stats.json --select-for-update=true --ramp=1m --duration=10m {pgurl:1-3} returned
| stderr:
| ./workload: /lib/x86_64-linux-gnu/libm.so.6: version `GLIBC_2.29' not found (required by ./workload)
| Error: COMMAND_PROBLEM: exit status 1
| (1) COMMAND_PROBLEM
| Wraps: (2) Node 4. Command with error:
| | ```
| | ./workload run ycsb --init --insert-count=1000000 --workload=A --concurrency=144 --splits=3 --histograms=perf/stats.json --select-for-update=true --ramp=1m --duration=10m {pgurl:1-3}
| | ```
| Wraps: (3) exit status 1
| Error types: (1) errors.Cmd (2) *hintdetail.withDetail (3) *exec.ExitError
|
| stdout:
Wraps: (4) exit status 20
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *main.withCommandDetails (4) *exec.ExitError
cluster.go:2654,ycsb.go:62,ycsb.go:79,test_runner.go:755: monitor failure: monitor task failed: t.Fatal() was called
(1) attached stack trace
-- stack trace:
| main.(*monitor).WaitE
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2642
| main.(*monitor).Wait
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2650
| main.registerYCSB.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/ycsb.go:62
| main.registerYCSB.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/ycsb.go:79
| main.(*testRunner).runTest.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:755
Wraps: (2) monitor failure
Wraps: (3) attached stack trace
-- stack trace:
| main.(*monitor).wait.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2698
Wraps: (4) monitor task failed
Wraps: (5) attached stack trace
-- stack trace:
| main.init
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2612
| runtime.doInit
| /usr/local/go/src/runtime/proc.go:5652
| runtime.main
| /usr/local/go/src/runtime/proc.go:191
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1374
Wraps: (6) t.Fatal() was called
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.withPrefix (5) *withstack.withStack (6) *errutil.leafError
```
<details><summary>More</summary><p>
Artifacts: [/ycsb/A/nodes=3/cpu=32](https://teamcity.cockroachdb.com/viewLog.html?buildId=2657159&tab=artifacts#/ycsb/A/nodes=3/cpu=32)
Related:
- #59933 roachtest: ycsb/A/nodes=3/cpu=32 failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Aycsb%2FA%2Fnodes%3D3%2Fcpu%3D32.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
test
|
roachtest ycsb a nodes cpu failed on runtime goexit usr local go src runtime asm s wraps output in run workload run ycsb wraps home agent work go src github com cockroachdb cockroach bin roachprod run teamcity workload run ycsb init insert count workload a concurrency splits histograms perf stats json select for update true ramp duration pgurl returned stderr workload lib linux gnu libm so version glibc not found required by workload error command problem exit status command problem wraps node command with error workload run ycsb init insert count workload a concurrency splits histograms perf stats json select for update true ramp duration pgurl wraps exit status error types errors cmd hintdetail withdetail exec exiterror stdout wraps exit status error types withstack withstack errutil withprefix main withcommanddetails exec exiterror cluster go ycsb go ycsb go test runner go monitor failure monitor task failed t fatal was called attached stack trace stack trace main monitor waite home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main monitor wait home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main registerycsb home agent work go src github com cockroachdb cockroach pkg cmd roachtest ycsb go main registerycsb home agent work go src github com cockroachdb cockroach pkg cmd roachtest ycsb go main testrunner runtest home agent work go src github com cockroachdb cockroach pkg cmd roachtest test runner go wraps monitor failure wraps attached stack trace stack trace main monitor wait home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go wraps monitor task failed wraps attached stack trace stack trace main init home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go runtime doinit usr local go src runtime proc go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps t fatal was called error types withstack withstack errutil withprefix withstack withstack errutil withprefix withstack withstack errutil leaferror more artifacts related roachtest ycsb a nodes cpu failed powered by
| 1
|
67,891
| 7,068,184,543
|
IssuesEvent
|
2018-01-08 06:44:47
|
AIE-Guild/GreenWall
|
https://api.github.com/repos/AIE-Guild/GreenWall
|
closed
|
Add update default option
|
feature low testing
|
It would be great if this addon had the ability to change the default options for all your characters. As someone who plays a lot of alts I often forget to set all the options I like for each toon. If I could set the options on one toon and have that be the new default for all my toons that would be awesome. Love the addon. It allowed my guild to grow in ways it couldn't before. Thank you.
|
1.0
|
Add update default option - It would be great if this addon had the ability to change the default options for all your characters. As someone who plays a lot of alts I often forget to set all the options I like for each toon. If I could set the options on one toon and have that be the new default for all my toons that would be awesome. Love the addon. It allowed my guild to grow in ways it couldn't before. Thank you.
|
test
|
add update default option it would be great if this addon had the ability to change the default options for all your characters as someone who plays a lot of alts i often forget to set all the options i like for each toon if i could set the options on one toon and have that be the new default for all my toons that would be awesome love the addon it allowed my guild to grow in ways it couldn t before thank you
| 1
|
328,466
| 28,122,368,795
|
IssuesEvent
|
2023-03-31 15:02:32
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
reopened
|
Fix layers.test_conv_general_dilated
|
Sub Task Ivy Functional API Failing Test
|
| | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4574198534/jobs/8075520768" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4574198534/jobs/8075520768" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4574198534/jobs/8075520768" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4574198534/jobs/8075520768" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_nn/test_layers.py::test_conv_general_dilated[gpu:0-ivy.functional.backends.tensorflow-False-False]</summary>
2023-03-31T10:42:36.5669448Z E RuntimeError: Unknown backend: 'gpu' requested, but no platforms that are instances of gpu are present. Platforms present are: interpreter,cpu
2023-03-31T10:42:36.5672967Z E ivy.utils.exceptions.IvyBackendException: jax: as_native_dev: Unknown backend: 'gpu' requested, but no platforms that are instances of gpu are present. Platforms present are: interpreter,cpu
2023-03-31T10:42:36.5682078Z E ivy.utils.exceptions.IvyBackendException: jax: default_device: jax: as_native_dev: Unknown backend: 'gpu' requested, but no platforms that are instances of gpu are present. Platforms present are: interpreter,cpu
2023-03-31T10:42:36.5682654Z E Falsifying example: test_conv_general_dilated(
2023-03-31T10:42:36.5682910Z E dims=1,
2023-03-31T10:42:36.5683190Z E x_f_d_df=(['float64'],
2023-03-31T10:42:36.5683409Z E array([[[0.5]]]),
2023-03-31T10:42:36.5683837Z E array([[[0.5]]]),
2023-03-31T10:42:36.5684142Z E (1, 1),
2023-03-31T10:42:36.5684471Z E 'channel_last',
2023-03-31T10:42:36.5684761Z E 1,
2023-03-31T10:42:36.5685115Z E 'SAME',
2023-03-31T10:42:36.5685359Z E 1,
2023-03-31T10:42:36.5685640Z E array([0.5])),
2023-03-31T10:42:36.5685971Z E test_flags=FunctionTestFlags(
2023-03-31T10:42:36.5686346Z E num_positional_args=4,
2023-03-31T10:42:36.5686688Z E with_out=False,
2023-03-31T10:42:36.5687014Z E instance_method=False,
2023-03-31T10:42:36.5687349Z E test_gradients=None,
2023-03-31T10:42:36.5687711Z E test_compile=None,
2023-03-31T10:42:36.5688054Z E as_variable=[False],
2023-03-31T10:42:36.5688389Z E native_arrays=[False],
2023-03-31T10:42:36.5688717Z E container=[False],
2023-03-31T10:42:36.5689062Z E ),
2023-03-31T10:42:36.5689442Z E ground_truth_backend='jax',
2023-03-31T10:42:36.5689853Z E fn_name='conv_general_dilated',
2023-03-31T10:42:36.5690504Z E backend_fw=<module 'ivy.functional.backends.tensorflow' from '/ivy/ivy/functional/backends/tensorflow/__init__.py'>,
2023-03-31T10:42:36.5691682Z E on_device='gpu:0',
2023-03-31T10:42:36.5692019Z E )
2023-03-31T10:42:36.5692347Z E
2023-03-31T10:42:36.5693020Z E You can reproduce this example by temporarily adding @reproduce_failure('6.68.2', b'AXicY2CAAUYweQDOx8GEAQBD6AJC') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_nn/test_layers.py::test_conv_general_dilated[gpu:0-ivy.functional.backends.tensorflow-False-False]</summary>
2023-03-31T10:42:36.5669448Z E RuntimeError: Unknown backend: 'gpu' requested, but no platforms that are instances of gpu are present. Platforms present are: interpreter,cpu
2023-03-31T10:42:36.5672967Z E ivy.utils.exceptions.IvyBackendException: jax: as_native_dev: Unknown backend: 'gpu' requested, but no platforms that are instances of gpu are present. Platforms present are: interpreter,cpu
2023-03-31T10:42:36.5682078Z E ivy.utils.exceptions.IvyBackendException: jax: default_device: jax: as_native_dev: Unknown backend: 'gpu' requested, but no platforms that are instances of gpu are present. Platforms present are: interpreter,cpu
2023-03-31T10:42:36.5682654Z E Falsifying example: test_conv_general_dilated(
2023-03-31T10:42:36.5682910Z E dims=1,
2023-03-31T10:42:36.5683190Z E x_f_d_df=(['float64'],
2023-03-31T10:42:36.5683409Z E array([[[0.5]]]),
2023-03-31T10:42:36.5683837Z E array([[[0.5]]]),
2023-03-31T10:42:36.5684142Z E (1, 1),
2023-03-31T10:42:36.5684471Z E 'channel_last',
2023-03-31T10:42:36.5684761Z E 1,
2023-03-31T10:42:36.5685115Z E 'SAME',
2023-03-31T10:42:36.5685359Z E 1,
2023-03-31T10:42:36.5685640Z E array([0.5])),
2023-03-31T10:42:36.5685971Z E test_flags=FunctionTestFlags(
2023-03-31T10:42:36.5686346Z E num_positional_args=4,
2023-03-31T10:42:36.5686688Z E with_out=False,
2023-03-31T10:42:36.5687014Z E instance_method=False,
2023-03-31T10:42:36.5687349Z E test_gradients=None,
2023-03-31T10:42:36.5687711Z E test_compile=None,
2023-03-31T10:42:36.5688054Z E as_variable=[False],
2023-03-31T10:42:36.5688389Z E native_arrays=[False],
2023-03-31T10:42:36.5688717Z E container=[False],
2023-03-31T10:42:36.5689062Z E ),
2023-03-31T10:42:36.5689442Z E ground_truth_backend='jax',
2023-03-31T10:42:36.5689853Z E fn_name='conv_general_dilated',
2023-03-31T10:42:36.5690504Z E backend_fw=<module 'ivy.functional.backends.tensorflow' from '/ivy/ivy/functional/backends/tensorflow/__init__.py'>,
2023-03-31T10:42:36.5691682Z E on_device='gpu:0',
2023-03-31T10:42:36.5692019Z E )
2023-03-31T10:42:36.5692347Z E
2023-03-31T10:42:36.5693020Z E You can reproduce this example by temporarily adding @reproduce_failure('6.68.2', b'AXicY2CAAUYweQDOx8GEAQBD6AJC') as a decorator on your test case
</details>
|
1.0
|
Fix layers.test_conv_general_dilated - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4574198534/jobs/8075520768" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4574198534/jobs/8075520768" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4574198534/jobs/8075520768" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4574198534/jobs/8075520768" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_nn/test_layers.py::test_conv_general_dilated[gpu:0-ivy.functional.backends.tensorflow-False-False]</summary>
2023-03-31T10:42:36.5669448Z E RuntimeError: Unknown backend: 'gpu' requested, but no platforms that are instances of gpu are present. Platforms present are: interpreter,cpu
2023-03-31T10:42:36.5672967Z E ivy.utils.exceptions.IvyBackendException: jax: as_native_dev: Unknown backend: 'gpu' requested, but no platforms that are instances of gpu are present. Platforms present are: interpreter,cpu
2023-03-31T10:42:36.5682078Z E ivy.utils.exceptions.IvyBackendException: jax: default_device: jax: as_native_dev: Unknown backend: 'gpu' requested, but no platforms that are instances of gpu are present. Platforms present are: interpreter,cpu
2023-03-31T10:42:36.5682654Z E Falsifying example: test_conv_general_dilated(
2023-03-31T10:42:36.5682910Z E dims=1,
2023-03-31T10:42:36.5683190Z E x_f_d_df=(['float64'],
2023-03-31T10:42:36.5683409Z E array([[[0.5]]]),
2023-03-31T10:42:36.5683837Z E array([[[0.5]]]),
2023-03-31T10:42:36.5684142Z E (1, 1),
2023-03-31T10:42:36.5684471Z E 'channel_last',
2023-03-31T10:42:36.5684761Z E 1,
2023-03-31T10:42:36.5685115Z E 'SAME',
2023-03-31T10:42:36.5685359Z E 1,
2023-03-31T10:42:36.5685640Z E array([0.5])),
2023-03-31T10:42:36.5685971Z E test_flags=FunctionTestFlags(
2023-03-31T10:42:36.5686346Z E num_positional_args=4,
2023-03-31T10:42:36.5686688Z E with_out=False,
2023-03-31T10:42:36.5687014Z E instance_method=False,
2023-03-31T10:42:36.5687349Z E test_gradients=None,
2023-03-31T10:42:36.5687711Z E test_compile=None,
2023-03-31T10:42:36.5688054Z E as_variable=[False],
2023-03-31T10:42:36.5688389Z E native_arrays=[False],
2023-03-31T10:42:36.5688717Z E container=[False],
2023-03-31T10:42:36.5689062Z E ),
2023-03-31T10:42:36.5689442Z E ground_truth_backend='jax',
2023-03-31T10:42:36.5689853Z E fn_name='conv_general_dilated',
2023-03-31T10:42:36.5690504Z E backend_fw=<module 'ivy.functional.backends.tensorflow' from '/ivy/ivy/functional/backends/tensorflow/__init__.py'>,
2023-03-31T10:42:36.5691682Z E on_device='gpu:0',
2023-03-31T10:42:36.5692019Z E )
2023-03-31T10:42:36.5692347Z E
2023-03-31T10:42:36.5693020Z E You can reproduce this example by temporarily adding @reproduce_failure('6.68.2', b'AXicY2CAAUYweQDOx8GEAQBD6AJC') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_nn/test_layers.py::test_conv_general_dilated[gpu:0-ivy.functional.backends.tensorflow-False-False]</summary>
2023-03-31T10:42:36.5669448Z E RuntimeError: Unknown backend: 'gpu' requested, but no platforms that are instances of gpu are present. Platforms present are: interpreter,cpu
2023-03-31T10:42:36.5672967Z E ivy.utils.exceptions.IvyBackendException: jax: as_native_dev: Unknown backend: 'gpu' requested, but no platforms that are instances of gpu are present. Platforms present are: interpreter,cpu
2023-03-31T10:42:36.5682078Z E ivy.utils.exceptions.IvyBackendException: jax: default_device: jax: as_native_dev: Unknown backend: 'gpu' requested, but no platforms that are instances of gpu are present. Platforms present are: interpreter,cpu
2023-03-31T10:42:36.5682654Z E Falsifying example: test_conv_general_dilated(
2023-03-31T10:42:36.5682910Z E dims=1,
2023-03-31T10:42:36.5683190Z E x_f_d_df=(['float64'],
2023-03-31T10:42:36.5683409Z E array([[[0.5]]]),
2023-03-31T10:42:36.5683837Z E array([[[0.5]]]),
2023-03-31T10:42:36.5684142Z E (1, 1),
2023-03-31T10:42:36.5684471Z E 'channel_last',
2023-03-31T10:42:36.5684761Z E 1,
2023-03-31T10:42:36.5685115Z E 'SAME',
2023-03-31T10:42:36.5685359Z E 1,
2023-03-31T10:42:36.5685640Z E array([0.5])),
2023-03-31T10:42:36.5685971Z E test_flags=FunctionTestFlags(
2023-03-31T10:42:36.5686346Z E num_positional_args=4,
2023-03-31T10:42:36.5686688Z E with_out=False,
2023-03-31T10:42:36.5687014Z E instance_method=False,
2023-03-31T10:42:36.5687349Z E test_gradients=None,
2023-03-31T10:42:36.5687711Z E test_compile=None,
2023-03-31T10:42:36.5688054Z E as_variable=[False],
2023-03-31T10:42:36.5688389Z E native_arrays=[False],
2023-03-31T10:42:36.5688717Z E container=[False],
2023-03-31T10:42:36.5689062Z E ),
2023-03-31T10:42:36.5689442Z E ground_truth_backend='jax',
2023-03-31T10:42:36.5689853Z E fn_name='conv_general_dilated',
2023-03-31T10:42:36.5690504Z E backend_fw=<module 'ivy.functional.backends.tensorflow' from '/ivy/ivy/functional/backends/tensorflow/__init__.py'>,
2023-03-31T10:42:36.5691682Z E on_device='gpu:0',
2023-03-31T10:42:36.5692019Z E )
2023-03-31T10:42:36.5692347Z E
2023-03-31T10:42:36.5693020Z E You can reproduce this example by temporarily adding @reproduce_failure('6.68.2', b'AXicY2CAAUYweQDOx8GEAQBD6AJC') as a decorator on your test case
</details>
|
test
|
fix layers test conv general dilated tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test functional test nn test layers py test conv general dilated e runtimeerror unknown backend gpu requested but no platforms that are instances of gpu are present platforms present are interpreter cpu e ivy utils exceptions ivybackendexception jax as native dev unknown backend gpu requested but no platforms that are instances of gpu are present platforms present are interpreter cpu e ivy utils exceptions ivybackendexception jax default device jax as native dev unknown backend gpu requested but no platforms that are instances of gpu are present platforms present are interpreter cpu e falsifying example test conv general dilated e dims e x f d df e array e array e e channel last e e same e e array e test flags functiontestflags e num positional args e with out false e instance method false e test gradients none e test compile none e as variable e native arrays e container e e ground truth backend jax e fn name conv general dilated e backend fw e on device gpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test functional test nn test layers py test conv general dilated e runtimeerror unknown backend gpu requested but no platforms that are instances of gpu are present platforms present are interpreter cpu e ivy utils exceptions ivybackendexception jax as native dev unknown backend gpu requested but no platforms that are instances of gpu are present platforms present are interpreter cpu e ivy utils exceptions ivybackendexception jax default device jax as native dev unknown backend gpu requested but no platforms that are instances of gpu are present platforms present are interpreter cpu e falsifying example test conv general dilated e dims e x f d df e array e array e e channel last e e same e e array e test flags functiontestflags e num positional args e with out false e instance method false e test gradients none e test compile none e as variable e native arrays e container e e ground truth backend jax e fn name conv general dilated e backend fw e on device gpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case
| 1
|
470,560
| 13,540,462,032
|
IssuesEvent
|
2020-09-16 14:41:38
|
frontity/docs
|
https://api.github.com/repos/frontity/docs
|
opened
|
Document the process of updating links in the WP content
|
effort: med getting started good first issue priority: low source: community
|
As the main domain needs to point to the Frontity installation in decoupled mode, the WP installation needs to be on a different domain/sub-domain.
We should document how to update all the links in the WP content when the site URL is changed. As per: https://wordpress.org/support/article/moving-wordpress/
Issue raised here:
https://community.frontity.org/t/my-experience-developing-a-new-frontity-theme-and-remaking-my-blog-with-frontity/2902
|
1.0
|
Document the process of updating links in the WP content - As the main domain needs to point to the Frontity installation in decoupled mode, the WP installation needs to be on a different domain/sub-domain.
We should document how to update all the links in the WP content when the site URL is changed. As per: https://wordpress.org/support/article/moving-wordpress/
Issue raised here:
https://community.frontity.org/t/my-experience-developing-a-new-frontity-theme-and-remaking-my-blog-with-frontity/2902
|
non_test
|
document the process of updating links in the wp content as the main domain needs to point to the frontity installation in decoupled mode the wp installation needs to be on a different domain sub domain we should document how to update all the links in the wp content when the site url is changed as per issue raised here
| 0
|
213,196
| 23,966,148,776
|
IssuesEvent
|
2022-09-13 01:16:38
|
nanopathi/linux-4.19.72_CVE-2020-25705
|
https://api.github.com/repos/nanopathi/linux-4.19.72_CVE-2020-25705
|
closed
|
WS-2021-0570 (Medium) detected in linuxlinux-4.19.236 - autoclosed
|
security vulnerability
|
## WS-2021-0570 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.236</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nanopathi/linux-4.19.72_CVE-2020-25705/commit/bb043298c885ef0a0fbb57f40b74a3b5d8848f4c">bb043298c885ef0a0fbb57f40b74a3b5d8848f4c</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/mtd/nand/raw/fsmc_nand.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/mtd/nand/raw/fsmc_nand.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
mtd: rawnand: fsmc: Fix timing computation
This is an automated ID intended to aid in discovery of potential security vulnerabilities. The actual impact and attack plausibility have not yet been proven.
This ID is fixed in Linux Kernel version v5.15.8 by commit a4953e7b6f6ab5e1faf3267b5728e40eac1b6f8d, it was introduced in version v4.13 by commit d9fb0795718333e36f7e472d7d81b7b8efe347c8. For more details please see the references link.
<p>Publish Date: 2021-12-19
<p>URL: <a href=https://github.com/gregkh/linux/commit/a4953e7b6f6ab5e1faf3267b5728e40eac1b6f8d>WS-2021-0570</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://osv.dev/vulnerability/GSD-2021-1002605">https://osv.dev/vulnerability/GSD-2021-1002605</a></p>
<p>Release Date: 2021-12-19</p>
<p>Fix Resolution: v5.15.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2021-0570 (Medium) detected in linuxlinux-4.19.236 - autoclosed - ## WS-2021-0570 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.236</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nanopathi/linux-4.19.72_CVE-2020-25705/commit/bb043298c885ef0a0fbb57f40b74a3b5d8848f4c">bb043298c885ef0a0fbb57f40b74a3b5d8848f4c</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/mtd/nand/raw/fsmc_nand.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/mtd/nand/raw/fsmc_nand.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
mtd: rawnand: fsmc: Fix timing computation
This is an automated ID intended to aid in discovery of potential security vulnerabilities. The actual impact and attack plausibility have not yet been proven.
This ID is fixed in Linux Kernel version v5.15.8 by commit a4953e7b6f6ab5e1faf3267b5728e40eac1b6f8d, it was introduced in version v4.13 by commit d9fb0795718333e36f7e472d7d81b7b8efe347c8. For more details please see the references link.
<p>Publish Date: 2021-12-19
<p>URL: <a href=https://github.com/gregkh/linux/commit/a4953e7b6f6ab5e1faf3267b5728e40eac1b6f8d>WS-2021-0570</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://osv.dev/vulnerability/GSD-2021-1002605">https://osv.dev/vulnerability/GSD-2021-1002605</a></p>
<p>Release Date: 2021-12-19</p>
<p>Fix Resolution: v5.15.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
ws medium detected in linuxlinux autoclosed ws medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files drivers mtd nand raw fsmc nand c drivers mtd nand raw fsmc nand c vulnerability details mtd rawnand fsmc fix timing computation this is an automated id intended to aid in discovery of potential security vulnerabilities the actual impact and attack plausibility have not yet been proven this id is fixed in linux kernel version by commit it was introduced in version by commit for more details please see the references link publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
269,674
| 28,960,245,962
|
IssuesEvent
|
2023-05-10 01:26:27
|
dpteam/RK3188_TABLET
|
https://api.github.com/repos/dpteam/RK3188_TABLET
|
reopened
|
CVE-2013-1773 (High) detected in randomv3.0.66
|
Mend: dependency security vulnerability
|
## CVE-2013-1773 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>randomv3.0.66</b></p></summary>
<p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/tytso/random.git>https://git.kernel.org/pub/scm/linux/kernel/git/tytso/random.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/include/linux/nls.h</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/include/linux/nls.h</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/include/linux/nls.h</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Buffer overflow in the VFAT filesystem implementation in the Linux kernel before 3.3 allows local users to gain privileges or cause a denial of service (system crash) via a VFAT write operation on a filesystem with the utf8 mount option, which is not properly handled during UTF-8 to UTF-16 conversion.
<p>Publish Date: 2013-02-28
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2013-1773>CVE-2013-1773</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2013-1773">https://nvd.nist.gov/vuln/detail/CVE-2013-1773</a></p>
<p>Release Date: 2013-02-28</p>
<p>Fix Resolution: 3.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2013-1773 (High) detected in randomv3.0.66 - ## CVE-2013-1773 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>randomv3.0.66</b></p></summary>
<p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/tytso/random.git>https://git.kernel.org/pub/scm/linux/kernel/git/tytso/random.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/include/linux/nls.h</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/include/linux/nls.h</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/include/linux/nls.h</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Buffer overflow in the VFAT filesystem implementation in the Linux kernel before 3.3 allows local users to gain privileges or cause a denial of service (system crash) via a VFAT write operation on a filesystem with the utf8 mount option, which is not properly handled during UTF-8 to UTF-16 conversion.
<p>Publish Date: 2013-02-28
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2013-1773>CVE-2013-1773</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2013-1773">https://nvd.nist.gov/vuln/detail/CVE-2013-1773</a></p>
<p>Release Date: 2013-02-28</p>
<p>Fix Resolution: 3.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve high detected in cve high severity vulnerability vulnerable library library home page a href found in head commit a href found in base branch master vulnerable source files include linux nls h include linux nls h include linux nls h vulnerability details buffer overflow in the vfat filesystem implementation in the linux kernel before allows local users to gain privileges or cause a denial of service system crash via a vfat write operation on a filesystem with the mount option which is not properly handled during utf to utf conversion publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
325,328
| 24,043,944,249
|
IssuesEvent
|
2022-09-16 06:23:22
|
ffoodd/a11y.css
|
https://api.github.com/repos/ffoodd/a11y.css
|
closed
|
Mention languages on home page
|
website documentation
|
Since there's no bookmarklet form anymore, there's no indication of available languages in the site itself. This should be improved, obviously!
|
1.0
|
Mention languages on home page - Since there's no bookmarklet form anymore, there's no indication of available languages in the site itself. This should be improved, obviously!
|
non_test
|
mention languages on home page since there s no bookmarklet form anymore there s no indication of available languages in the site itself this should be improved obviously
| 0
|
42,004
| 9,114,674,866
|
IssuesEvent
|
2019-02-22 01:19:06
|
builderbook/builderbook
|
https://api.github.com/repos/builderbook/builderbook
|
closed
|
old import/export syntax for server code
|
code inside book
|
@whitefire0 Thanks for reporting issues. As we discussed in #108 - there is a confusion about ES6 import/export syntax. You can see how it's done in main app, for example: https://github.com/builderbook/builderbook/blob/master/server/app.js
Would you like update import/export syntax in all 8 Chapters so we can drop babel-node? Could be a good exercise!
|
1.0
|
old import/export syntax for server code - @whitefire0 Thanks for reporting issues. As we discussed in #108 - there is a confusion about ES6 import/export syntax. You can see how it's done in main app, for example: https://github.com/builderbook/builderbook/blob/master/server/app.js
Would you like update import/export syntax in all 8 Chapters so we can drop babel-node? Could be a good exercise!
|
non_test
|
old import export syntax for server code thanks for reporting issues as we discussed in there is a confusion about import export syntax you can see how it s done in main app for example would you like update import export syntax in all chapters so we can drop babel node could be a good exercise
| 0
|
9,672
| 3,066,936,162
|
IssuesEvent
|
2015-08-18 07:04:47
|
CasparCG/Server
|
https://api.github.com/repos/CasparCG/Server
|
closed
|
LOAD does not allways execute
|
bug testable
|
When three commands
```
LOAD 2-0 S0
LOAD 2-1 S16
CLEAR 2-2
```
are send within single frame period (to test the case, I use Putty), the second command is not executed - server response is:
```
202 LOAD OK
202 CLEAR OK
```
Of course, there is a slight chance to execute all three commands (I suppose it happens when not all three command a executed i the same window frame).
The S0, S16 files are simple tif images (so image_producer is executed)
More detailed console output:
```[2014-03-18 08:31:43.877] [03B24328] [info] Received message from 10.191.100.16: LOAD 2-0 S0\r\n
[2014-03-18 08:31:43.886] [03B24328] [info] Received message from 10.191.100.16: LOAD 2-1 S16\r\n
[2014-03-18 08:31:43.898] [03B24328] [info] Received message from 10.191.100.16: CLEAR 2-2\r\n
[2014-03-18 08:31:43.910] [03B55628] [info] image_producer[media\S0.tif] Initialized.
[2014-03-18 08:31:43.921] [03B55598] [info] image_producer[media\S0.tif] Uninitialized.
[2014-03-18 08:31:43.930] [03B55628] [debug] Executed command: LoadCommand
[2014-03-18 08:31:43.938] [03B55628] [info] Sent message to 10.191.100.16: 202 LOAD OK\r\n
[2014-03-18 08:31:43.948] [03B55628] [debug] Executed command: ClearCommand
[2014-03-18 08:31:43.957] [03B55628] [info] Sent message to 10.191.100.16: 202 CLEAR OK\r\n```
This behavior was observed with 2.06 stable and 2.042 community builds (I did not tested the others).
|
1.0
|
LOAD does not allways execute - When three commands
```
LOAD 2-0 S0
LOAD 2-1 S16
CLEAR 2-2
```
are send within single frame period (to test the case, I use Putty), the second command is not executed - server response is:
```
202 LOAD OK
202 CLEAR OK
```
Of course, there is a slight chance to execute all three commands (I suppose it happens when not all three command a executed i the same window frame).
The S0, S16 files are simple tif images (so image_producer is executed)
More detailed console output:
```[2014-03-18 08:31:43.877] [03B24328] [info] Received message from 10.191.100.16: LOAD 2-0 S0\r\n
[2014-03-18 08:31:43.886] [03B24328] [info] Received message from 10.191.100.16: LOAD 2-1 S16\r\n
[2014-03-18 08:31:43.898] [03B24328] [info] Received message from 10.191.100.16: CLEAR 2-2\r\n
[2014-03-18 08:31:43.910] [03B55628] [info] image_producer[media\S0.tif] Initialized.
[2014-03-18 08:31:43.921] [03B55598] [info] image_producer[media\S0.tif] Uninitialized.
[2014-03-18 08:31:43.930] [03B55628] [debug] Executed command: LoadCommand
[2014-03-18 08:31:43.938] [03B55628] [info] Sent message to 10.191.100.16: 202 LOAD OK\r\n
[2014-03-18 08:31:43.948] [03B55628] [debug] Executed command: ClearCommand
[2014-03-18 08:31:43.957] [03B55628] [info] Sent message to 10.191.100.16: 202 CLEAR OK\r\n```
This behavior was observed with 2.06 stable and 2.042 community builds (I did not tested the others).
|
test
|
load does not allways execute when three commands load load clear are send within single frame period to test the case i use putty the second command is not executed server response is load ok clear ok of course there is a slight chance to execute all three commands i suppose it happens when not all three command a executed i the same window frame the files are simple tif images so image producer is executed more detailed console output received message from load r n received message from load r n received message from clear r n image producer initialized image producer uninitialized executed command loadcommand sent message to load ok r n executed command clearcommand sent message to clear ok r n this behavior was observed with stable and community builds i did not tested the others
| 1
|
151,386
| 12,034,951,665
|
IssuesEvent
|
2020-04-13 16:59:59
|
NuGet/Home
|
https://api.github.com/repos/NuGet/Home
|
closed
|
[CPVM] Add E2E tests for cpvm enabled and lock enforced projects.
|
Area:CPVM Area:Test Category:Quality Week Pipeline: Backlog Priority:2
|
## Details about Problem
Add E2E tests for cpvm enabled and lock enforced projects.
|
1.0
|
[CPVM] Add E2E tests for cpvm enabled and lock enforced projects. -
## Details about Problem
Add E2E tests for cpvm enabled and lock enforced projects.
|
test
|
add tests for cpvm enabled and lock enforced projects details about problem add tests for cpvm enabled and lock enforced projects
| 1
|
248,859
| 26,848,593,993
|
IssuesEvent
|
2023-02-03 09:17:44
|
ArtyomZemlyak/expert-system
|
https://api.github.com/repos/ArtyomZemlyak/expert-system
|
opened
|
websockets-8.1-cp37-cp37m-manylinux2010_x86_64.whl: 1 vulnerabilities (highest severity is: 5.9)
|
security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>websockets-8.1-cp37-cp37m-manylinux2010_x86_64.whl</b></p></summary>
<p>An implementation of the WebSocket Protocol (RFC 6455 & 7692)</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/5a/0b/3ebc752392a368af14dd24ee041683416ac6d2463eead94b311b11e41c82/websockets-8.1-cp37-cp37m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/5a/0b/3ebc752392a368af14dd24ee041683416ac6d2463eead94b311b11e41c82/websockets-8.1-cp37-cp37m-manylinux2010_x86_64.whl</a></p>
<p>Path to dependency file: /scripts/testing/cython/testing</p>
<p>Path to vulnerable library: /scripts/testing/cython/testing,/DeepPavlov/requirements.txt,/DeepPavlov</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/ArtyomZemlyak/expert-system/commit/3c75cc05a886394a1c59833f685a90669bc1384b">3c75cc05a886394a1c59833f685a90669bc1384b</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (websockets version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2021-33880](https://www.mend.io/vulnerability-database/CVE-2021-33880) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.9 | websockets-8.1-cp37-cp37m-manylinux2010_x86_64.whl | Direct | 9.1 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2021-33880</summary>
### Vulnerable Library - <b>websockets-8.1-cp37-cp37m-manylinux2010_x86_64.whl</b></p>
<p>An implementation of the WebSocket Protocol (RFC 6455 & 7692)</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/5a/0b/3ebc752392a368af14dd24ee041683416ac6d2463eead94b311b11e41c82/websockets-8.1-cp37-cp37m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/5a/0b/3ebc752392a368af14dd24ee041683416ac6d2463eead94b311b11e41c82/websockets-8.1-cp37-cp37m-manylinux2010_x86_64.whl</a></p>
<p>Path to dependency file: /scripts/testing/cython/testing</p>
<p>Path to vulnerable library: /scripts/testing/cython/testing,/DeepPavlov/requirements.txt,/DeepPavlov</p>
<p>
Dependency Hierarchy:
- :x: **websockets-8.1-cp37-cp37m-manylinux2010_x86_64.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ArtyomZemlyak/expert-system/commit/3c75cc05a886394a1c59833f685a90669bc1384b">3c75cc05a886394a1c59833f685a90669bc1384b</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The aaugustin websockets library before 9.1 for Python has an Observable Timing Discrepancy on servers when HTTP Basic Authentication is enabled with basic_auth_protocol_factory(credentials=...). An attacker may be able to guess a password via a timing attack.
<p>Publish Date: 2021-06-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-33880>CVE-2021-33880</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.9</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33880">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33880</a></p>
<p>Release Date: 2021-06-06</p>
<p>Fix Resolution: 9.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
True
|
websockets-8.1-cp37-cp37m-manylinux2010_x86_64.whl: 1 vulnerabilities (highest severity is: 5.9) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>websockets-8.1-cp37-cp37m-manylinux2010_x86_64.whl</b></p></summary>
<p>An implementation of the WebSocket Protocol (RFC 6455 & 7692)</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/5a/0b/3ebc752392a368af14dd24ee041683416ac6d2463eead94b311b11e41c82/websockets-8.1-cp37-cp37m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/5a/0b/3ebc752392a368af14dd24ee041683416ac6d2463eead94b311b11e41c82/websockets-8.1-cp37-cp37m-manylinux2010_x86_64.whl</a></p>
<p>Path to dependency file: /scripts/testing/cython/testing</p>
<p>Path to vulnerable library: /scripts/testing/cython/testing,/DeepPavlov/requirements.txt,/DeepPavlov</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/ArtyomZemlyak/expert-system/commit/3c75cc05a886394a1c59833f685a90669bc1384b">3c75cc05a886394a1c59833f685a90669bc1384b</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (websockets version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2021-33880](https://www.mend.io/vulnerability-database/CVE-2021-33880) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.9 | websockets-8.1-cp37-cp37m-manylinux2010_x86_64.whl | Direct | 9.1 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2021-33880</summary>
### Vulnerable Library - <b>websockets-8.1-cp37-cp37m-manylinux2010_x86_64.whl</b></p>
<p>An implementation of the WebSocket Protocol (RFC 6455 & 7692)</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/5a/0b/3ebc752392a368af14dd24ee041683416ac6d2463eead94b311b11e41c82/websockets-8.1-cp37-cp37m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/5a/0b/3ebc752392a368af14dd24ee041683416ac6d2463eead94b311b11e41c82/websockets-8.1-cp37-cp37m-manylinux2010_x86_64.whl</a></p>
<p>Path to dependency file: /scripts/testing/cython/testing</p>
<p>Path to vulnerable library: /scripts/testing/cython/testing,/DeepPavlov/requirements.txt,/DeepPavlov</p>
<p>
Dependency Hierarchy:
- :x: **websockets-8.1-cp37-cp37m-manylinux2010_x86_64.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ArtyomZemlyak/expert-system/commit/3c75cc05a886394a1c59833f685a90669bc1384b">3c75cc05a886394a1c59833f685a90669bc1384b</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The aaugustin websockets library before 9.1 for Python has an Observable Timing Discrepancy on servers when HTTP Basic Authentication is enabled with basic_auth_protocol_factory(credentials=...). An attacker may be able to guess a password via a timing attack.
<p>Publish Date: 2021-06-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-33880>CVE-2021-33880</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.9</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33880">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33880</a></p>
<p>Release Date: 2021-06-06</p>
<p>Fix Resolution: 9.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
non_test
|
websockets whl vulnerabilities highest severity is vulnerable library websockets whl an implementation of the websocket protocol rfc library home page a href path to dependency file scripts testing cython testing path to vulnerable library scripts testing cython testing deeppavlov requirements txt deeppavlov found in head commit a href vulnerabilities cve severity cvss dependency type fixed in websockets version remediation available medium websockets whl direct details cve vulnerable library websockets whl an implementation of the websocket protocol rfc library home page a href path to dependency file scripts testing cython testing path to vulnerable library scripts testing cython testing deeppavlov requirements txt deeppavlov dependency hierarchy x websockets whl vulnerable library found in head commit a href found in base branch main vulnerability details the aaugustin websockets library before for python has an observable timing discrepancy on servers when http basic authentication is enabled with basic auth protocol factory credentials an attacker may be able to guess a password via a timing attack publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
346,305
| 10,410,513,770
|
IssuesEvent
|
2019-09-13 11:37:02
|
gitextensions/gitextensions
|
https://api.github.com/repos/gitextensions/gitextensions
|
closed
|
Latest Version of Git Extensions Installer Always Reverts For Global user Install
|
area: installer priority: high type: bug
|
## Current behavior
Installer should always work, prompt for install happened today, and had a new user install it.
## Expected behavior
Installer should work.
## Steps to reproduce
<!-- Take some time to try and reproduce the issue, then explain how to do so here. -->
## Screenshots
Install latest MSI for all users.
## Did this work in previous version of GitExtensions
3.2.0.6586.msi, I'm using 3.1, do not know which MSI
## Environment
- GitExtensions version:
- GIT version: unrelated.
- OS version: Windows 10
- .NET version: both standard user machine, and dev machine with many versions installed.
## Diagnostics
Telemetry was not enabled during the install .
|
1.0
|
Latest Version of Git Extensions Installer Always Reverts For Global user Install - ## Current behavior
Installer should always work, prompt for install happened today, and had a new user install it.
## Expected behavior
Installer should work.
## Steps to reproduce
<!-- Take some time to try and reproduce the issue, then explain how to do so here. -->
## Screenshots
Install latest MSI for all users.
## Did this work in previous version of GitExtensions
3.2.0.6586.msi, I'm using 3.1, do not know which MSI
## Environment
- GitExtensions version:
- GIT version: unrelated.
- OS version: Windows 10
- .NET version: both standard user machine, and dev machine with many versions installed.
## Diagnostics
Telemetry was not enabled during the install .
|
non_test
|
latest version of git extensions installer always reverts for global user install current behavior installer should always work prompt for install happened today and had a new user install it expected behavior installer should work steps to reproduce screenshots install latest msi for all users did this work in previous version of gitextensions msi i m using do not know which msi environment gitextensions version git version unrelated os version windows net version both standard user machine and dev machine with many versions installed diagnostics telemetry was not enabled during the install
| 0
|
429,900
| 30,108,336,508
|
IssuesEvent
|
2023-06-30 04:53:57
|
winglang/wing
|
https://api.github.com/repos/winglang/wing
|
closed
|
Links in the docs lead to 404
|
🐛 bug 📚 documentation 1️⃣ good first issue
|
### I tried this:
https://www.winglang.io/docs/concepts/compiler-targets
Portability section, found [here](https://www.winglang.io/docs/concepts/docs/docs/04-standard-library/03-winglang-support-matrix) link leads to 404.
Same for https://www.winglang.io/docs/language-guide/equality - Equality section, ([preflight or inflight)](https://www.winglang.io/docs/02-core-concepts/01-preflight-and-inflight.md) link.
### This happened:
404 page
### I expected this:
Lead to the right doc.
### Is there a workaround?
_No response_
### Component
Documentation
### Wing Version
_No response_
### Wing Console Version
_No response_
### Node.js Version
_No response_
### Platform(s)
_No response_
### Anything else?
_No response_
### Community Notes
<!-- Please keep this note for the community -->
* Please vote by adding a 👍 reaction to the issue to help us prioritize.
* If you are interested to work on this issue, please leave a comment.
|
1.0
|
Links in the docs lead to 404 - ### I tried this:
https://www.winglang.io/docs/concepts/compiler-targets
Portability section, found [here](https://www.winglang.io/docs/concepts/docs/docs/04-standard-library/03-winglang-support-matrix) link leads to 404.
Same for https://www.winglang.io/docs/language-guide/equality - Equality section, ([preflight or inflight)](https://www.winglang.io/docs/02-core-concepts/01-preflight-and-inflight.md) link.
### This happened:
404 page
### I expected this:
Lead to the right doc.
### Is there a workaround?
_No response_
### Component
Documentation
### Wing Version
_No response_
### Wing Console Version
_No response_
### Node.js Version
_No response_
### Platform(s)
_No response_
### Anything else?
_No response_
### Community Notes
<!-- Please keep this note for the community -->
* Please vote by adding a 👍 reaction to the issue to help us prioritize.
* If you are interested to work on this issue, please leave a comment.
|
non_test
|
links in the docs lead to i tried this portability section found link leads to same for equality section link this happened page i expected this lead to the right doc is there a workaround no response component documentation wing version no response wing console version no response node js version no response platform s no response anything else no response community notes please vote by adding a 👍 reaction to the issue to help us prioritize if you are interested to work on this issue please leave a comment
| 0
|
44,720
| 11,494,923,639
|
IssuesEvent
|
2020-02-12 03:07:26
|
GoogleCloudPlatform/python-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
|
closed
|
dataproc.quickstart.quickstart_test: test_quickstart failed
|
buildcop:issue priority: p1 type: bug
|
dataproc.quickstart.quickstart_test: test_quickstart failed
buildID: fbcdef163f965e363c6b1abf9d41b14fde28271e
buildURL: https://source.cloud.google.com/results/invocations/9b3b09c2-b687-42e2-9066-613031c2b1ec
status: failed
|
1.0
|
dataproc.quickstart.quickstart_test: test_quickstart failed - dataproc.quickstart.quickstart_test: test_quickstart failed
buildID: fbcdef163f965e363c6b1abf9d41b14fde28271e
buildURL: https://source.cloud.google.com/results/invocations/9b3b09c2-b687-42e2-9066-613031c2b1ec
status: failed
|
non_test
|
dataproc quickstart quickstart test test quickstart failed dataproc quickstart quickstart test test quickstart failed buildid buildurl status failed
| 0
|
83,730
| 7,880,816,785
|
IssuesEvent
|
2018-06-26 17:01:25
|
hazelcast/hazelcast
|
https://api.github.com/repos/hazelcast/hazelcast
|
closed
|
ClientCheckDependenciesIT.testNoMandatoryDependencyDeclared failure
|
Team: Core Type: Test-Failure
|
*master* (commit 60e11525cd85977b9ee6bd51611f463cc2452994)
*maintenance* (commit 3be1639dd4b5e497950aede41a41fbadecbafc34 )
Failed on various JDKs 8 (maintenance): https://hazelcast-l337.ci.cloudbees.com/view/Official%20Builds/job/Hazelcast-3.maintenance-OracleJDK8/com.hazelcast$hazelcast-client/642/testReport/junit/com.hazelcast.client.osgi/ClientCheckDependenciesIT/testNoMandatoryDependencyDeclared/
Failed on various JDKs 8 (master): https://hazelcast-l337.ci.cloudbees.com/view/Official%20Builds/job/Hazelcast-3.x-OracleJDK8/com.hazelcast$hazelcast-client/1537/testReport/junit/com.hazelcast.client.osgi/ClientCheckDependenciesIT/testNoMandatoryDependencyDeclared/
Stacktrace:
```
java.lang.AssertionError: Import com.eclipsesource.json is not declared as optional
```
|
1.0
|
ClientCheckDependenciesIT.testNoMandatoryDependencyDeclared failure - *master* (commit 60e11525cd85977b9ee6bd51611f463cc2452994)
*maintenance* (commit 3be1639dd4b5e497950aede41a41fbadecbafc34 )
Failed on various JDKs 8 (maintenance): https://hazelcast-l337.ci.cloudbees.com/view/Official%20Builds/job/Hazelcast-3.maintenance-OracleJDK8/com.hazelcast$hazelcast-client/642/testReport/junit/com.hazelcast.client.osgi/ClientCheckDependenciesIT/testNoMandatoryDependencyDeclared/
Failed on various JDKs 8 (master): https://hazelcast-l337.ci.cloudbees.com/view/Official%20Builds/job/Hazelcast-3.x-OracleJDK8/com.hazelcast$hazelcast-client/1537/testReport/junit/com.hazelcast.client.osgi/ClientCheckDependenciesIT/testNoMandatoryDependencyDeclared/
Stacktrace:
```
java.lang.AssertionError: Import com.eclipsesource.json is not declared as optional
```
|
test
|
clientcheckdependenciesit testnomandatorydependencydeclared failure master commit maintenance commit failed on various jdks maintenance failed on various jdks master stacktrace java lang assertionerror import com eclipsesource json is not declared as optional
| 1
|
208,452
| 15,890,601,106
|
IssuesEvent
|
2021-04-10 16:01:25
|
gii-is-psg2/PSG2-2021-G2-22
|
https://api.github.com/repos/gii-is-psg2/PSG2-2021-G2-22
|
closed
|
Tarea 12 : Crear una funcionalidad de hotel para mascotas
|
functionality tests
|
Funcionalidad que permita a los clientes enviar detalles sobre la mascota para la que quieren reservar una habitación, así como las fechas de inicio y finalización de la reserva.
- [x] Carmen Mª Muñoz Pérez (revisora)
|
1.0
|
Tarea 12 : Crear una funcionalidad de hotel para mascotas - Funcionalidad que permita a los clientes enviar detalles sobre la mascota para la que quieren reservar una habitación, así como las fechas de inicio y finalización de la reserva.
- [x] Carmen Mª Muñoz Pérez (revisora)
|
test
|
tarea crear una funcionalidad de hotel para mascotas funcionalidad que permita a los clientes enviar detalles sobre la mascota para la que quieren reservar una habitación así como las fechas de inicio y finalización de la reserva carmen mª muñoz pérez revisora
| 1
|
823,526
| 31,022,700,913
|
IssuesEvent
|
2023-08-10 06:57:11
|
Betarena/scores
|
https://api.github.com/repos/Betarena/scores
|
closed
|
Profile - Withdraw option
|
🌟 [A] enhancement 📣 [A] high priority
|
#### 📝 DESCRIPTION
We are adding a Withdraw option on the profile user section and a few methods that will allow users to withdraw funds from the Betarena platform.

---
#### 📱 TARGET PLATFORMS
This option will be available on all platform versions.
---
#### 🎬 LINK
https://scores.betarena.com/u/withdraw/pt
---
#### ℹ️ OTHER CONDITIONS (& RELEVANT INFO)
All withdrawal options have a form associated so that we can collect user information to be able to associate the request of funds with the user account and to be able to have access to the information needed to send user funds.
We need a preloader for this and other sections; we can use the following using Black instead of blue:
https://lottiefiles.com/73801-preloader-animation
**Initial withdrawal options:**
- Paypal
- Skrill
- Remessa Online
- Revolut
- Wired Transfer
- Wise
- Cryptocurrency
Since all information is coming from the DB, this means that we can add new withdrawal options without the need to change the front end.
---
### Versions:
All theme variations are supported.
---
#### 💠 DATA
➤ **Hasura**
| table | columns |
| ----- | --------- |
| withdraw_options | id + type + form_id + provider_logo + blacklist |
| forms | id + type + type_id + form_fields_step_1 + form_fields_step_2 + gateway + form_fields_success |
### Fields description:
**withdraw_options**
```
id = The ID of each withdrawal option
type = Gateway
form_id = The Form ID associated with this Gateway, available on forms
provider_logo = Gateway logo
blacklist = GEO countries where it's not available
```
===
```
**forms**
id = Form ID that allows us to associate the withdraw option with the correct Form
type = Description of what kind of form it is
type_id = Type of the form, withdraw, deposit, or others
form_fields_step_1 = Content and Fields of the form, including the type of field
form_fields_step_2 = In the case of multi-step form, here is where the second step data is available
gateway = What type of Gateway is associated with this form
form_fields_success = Content of the success message associated with the form
```
===
### Sending the Data:
After filling up the form, the data should be sent to the DB:
➤ **Hasura**
| table | columns |
| ----- | --------- |
| transaction_history | id + date + asset + quantity + amount + description + type + status + Gateway + first_name + last_name + payment_email + iban + bic_swift + wallet_address_erc20 + uid + payment_processor_fee|
```
id = unique identifier for each transaction.
date = Date when the transaction occurred.
asset = Specifies whether it involves fiat currency or cryptocurrency.
quantity = The quantity or amount involved in the transaction (e.g., fiat currency amount, crypto amount).
amount = The corresponding amount in BTA tokens.
description = Brief description of the transaction.
type = Indicates whether it is a deposit or withdrawal request.
status = Current status of the transaction (Completed, Pending, Processing, Canceled, [DEFAULT STATUS PENDING]))
Gateway = Payment processor or platform used for the transaction (e.g., Stripe, Coinbase, Wise).
first_name = User First Name if available
last_name = User Last Name if available
payment_email = Email associated with the Gateway
iban = IBAN associated with the Gateway
bic_swift = BIC/SWIFT CODE associated with the payment Gateway
wallet_address_erc20 = For withdrawals, the wallet address where the funds are being or has been sent.
uid = User Unique UID
payment_processor_fee = Fee charged by the payment processor for the transaction.
```
---
#### 🎏 TRANSLATIONS
**Translations endpoint**
Within the specific Schema:
| table | columns |
| ----- | --------- |
| forms |
The translations are included in the form json.
---
#### ⚡️ CACHE
Unavailable.
---
#### 🔎 SEO
Unavailable.
---
### Placeholder when there is no content:
Defined preloader on top.
---
### Resources
Figma Design:
https://www.figma.com/file/4SlIPyxRBX2qg1eLaPUk9U/BetArena-%2F-Desktop?type=design&node-id=32%3A1635&t=wLX73xuz9lYHGwZ7-1
|
1.0
|
Profile - Withdraw option - #### 📝 DESCRIPTION
We are adding a Withdraw option on the profile user section and a few methods that will allow users to withdraw funds from the Betarena platform.

---
#### 📱 TARGET PLATFORMS
This option will be available on all platform versions.
---
#### 🎬 LINK
https://scores.betarena.com/u/withdraw/pt
---
#### ℹ️ OTHER CONDITIONS (& RELEVANT INFO)
All withdrawal options have a form associated so that we can collect user information to be able to associate the request of funds with the user account and to be able to have access to the information needed to send user funds.
We need a preloader for this and other sections; we can use the following using Black instead of blue:
https://lottiefiles.com/73801-preloader-animation
**Initial withdrawal options:**
- Paypal
- Skrill
- Remessa Online
- Revolut
- Wired Transfer
- Wise
- Cryptocurrency
Since all information is coming from the DB, this means that we can add new withdrawal options without the need to change the front end.
---
### Versions:
All theme variations are supported.
---
#### 💠 DATA
➤ **Hasura**
| table | columns |
| ----- | --------- |
| withdraw_options | id + type + form_id + provider_logo + blacklist |
| forms | id + type + type_id + form_fields_step_1 + form_fields_step_2 + gateway + form_fields_success |
### Fields description:
**withdraw_options**
```
id = The ID of each withdrawal option
type = Gateway
form_id = The Form ID associated with this Gateway, available on forms
provider_logo = Gateway logo
blacklist = GEO countries where it's not available
```
===
```
**forms**
id = Form ID that allows us to associate the withdraw option with the correct Form
type = Description of what kind of form it is
type_id = Type of the form, withdraw, deposit, or others
form_fields_step_1 = Content and Fields of the form, including the type of field
form_fields_step_2 = In the case of multi-step form, here is where the second step data is available
gateway = What type of Gateway is associated with this form
form_fields_success = Content of the success message associated with the form
```
===
### Sending the Data:
After filling up the form, the data should be sent to the DB:
➤ **Hasura**
| table | columns |
| ----- | --------- |
| transaction_history | id + date + asset + quantity + amount + description + type + status + Gateway + first_name + last_name + payment_email + iban + bic_swift + wallet_address_erc20 + uid + payment_processor_fee|
```
id = unique identifier for each transaction.
date = Date when the transaction occurred.
asset = Specifies whether it involves fiat currency or cryptocurrency.
quantity = The quantity or amount involved in the transaction (e.g., fiat currency amount, crypto amount).
amount = The corresponding amount in BTA tokens.
description = Brief description of the transaction.
type = Indicates whether it is a deposit or withdrawal request.
status = Current status of the transaction (Completed, Pending, Processing, Canceled, [DEFAULT STATUS PENDING]))
Gateway = Payment processor or platform used for the transaction (e.g., Stripe, Coinbase, Wise).
first_name = User First Name if available
last_name = User Last Name if available
payment_email = Email associated with the Gateway
iban = IBAN associated with the Gateway
bic_swift = BIC/SWIFT CODE associated with the payment Gateway
wallet_address_erc20 = For withdrawals, the wallet address where the funds are being or has been sent.
uid = User Unique UID
payment_processor_fee = Fee charged by the payment processor for the transaction.
```
---
#### 🎏 TRANSLATIONS
**Translations endpoint**
Within the specific Schema:
| table | columns |
| ----- | --------- |
| forms |
The translations are included in the form json.
---
#### ⚡️ CACHE
Unavailable.
---
#### 🔎 SEO
Unavailable.
---
### Placeholder when there is no content:
Defined preloader on top.
---
### Resources
Figma Design:
https://www.figma.com/file/4SlIPyxRBX2qg1eLaPUk9U/BetArena-%2F-Desktop?type=design&node-id=32%3A1635&t=wLX73xuz9lYHGwZ7-1
|
non_test
|
profile withdraw option 📝 description we are adding a withdraw option on the profile user section and a few methods that will allow users to withdraw funds from the betarena platform 📱 target platforms this option will be available on all platform versions 🎬 link ℹ️ other conditions relevant info all withdrawal options have a form associated so that we can collect user information to be able to associate the request of funds with the user account and to be able to have access to the information needed to send user funds we need a preloader for this and other sections we can use the following using black instead of blue initial withdrawal options paypal skrill remessa online revolut wired transfer wise cryptocurrency since all information is coming from the db this means that we can add new withdrawal options without the need to change the front end versions all theme variations are supported 💠 data ➤ hasura table columns withdraw options id type form id provider logo blacklist forms id type type id form fields step form fields step gateway form fields success fields description withdraw options id the id of each withdrawal option type gateway form id the form id associated with this gateway available on forms provider logo gateway logo blacklist geo countries where it s not available forms id form id that allows us to associate the withdraw option with the correct form type description of what kind of form it is type id type of the form withdraw deposit or others form fields step content and fields of the form including the type of field form fields step in the case of multi step form here is where the second step data is available gateway what type of gateway is associated with this form form fields success content of the success message associated with the form sending the data after filling up the form the data should be sent to the db ➤ hasura table columns transaction history id date asset quantity amount description type status gateway first name last name payment email iban bic swift wallet address uid payment processor fee id unique identifier for each transaction date date when the transaction occurred asset specifies whether it involves fiat currency or cryptocurrency quantity the quantity or amount involved in the transaction e g fiat currency amount crypto amount amount the corresponding amount in bta tokens description brief description of the transaction type indicates whether it is a deposit or withdrawal request status current status of the transaction completed pending processing canceled gateway payment processor or platform used for the transaction e g stripe coinbase wise first name user first name if available last name user last name if available payment email email associated with the gateway iban iban associated with the gateway bic swift bic swift code associated with the payment gateway wallet address for withdrawals the wallet address where the funds are being or has been sent uid user unique uid payment processor fee fee charged by the payment processor for the transaction 🎏 translations translations endpoint within the specific schema table columns forms the translations are included in the form json ⚡️ cache unavailable 🔎 seo unavailable placeholder when there is no content defined preloader on top resources figma design
| 0
|
280,298
| 24,292,120,202
|
IssuesEvent
|
2022-09-29 07:07:26
|
bankidz/bankidz-client
|
https://api.github.com/repos/bankidz/bankidz-client
|
closed
|
[FIX] v2.0 QA 이슈 대응 등
|
For: API For: Markup/UX For: Script Type: Errors/Bugs Type: Feature Type: Refactor Type: Test
|
# 🤖 기능 개요
<!-- 이슈에 할당된 기능이 무엇인지 간략하게 한 줄로 적습니다 -->
### ✅ Implement TODO
<!-- 이슈에 할당된 TODO를 나름대로 항목화하여 적습니다 (PR할 때에는 모두 체크되어야함) -->
- [x] 미지급 이자 뷰 스켈레톤 UI
- [x] 모든 API 테스트
- [x] 탈퇴사유 textarea 크기 조절 제외
- [x] v2.0 QA 이슈 전부 대응
- [x] 당겨서 새로고침
- [x] 로딩 스피너 교체
### 📚 Remarks
<!-- 기능 개발에 있어 비고사항이 있었다면 적기 -->
|
1.0
|
[FIX] v2.0 QA 이슈 대응 등 - # 🤖 기능 개요
<!-- 이슈에 할당된 기능이 무엇인지 간략하게 한 줄로 적습니다 -->
### ✅ Implement TODO
<!-- 이슈에 할당된 TODO를 나름대로 항목화하여 적습니다 (PR할 때에는 모두 체크되어야함) -->
- [x] 미지급 이자 뷰 스켈레톤 UI
- [x] 모든 API 테스트
- [x] 탈퇴사유 textarea 크기 조절 제외
- [x] v2.0 QA 이슈 전부 대응
- [x] 당겨서 새로고침
- [x] 로딩 스피너 교체
### 📚 Remarks
<!-- 기능 개발에 있어 비고사항이 있었다면 적기 -->
|
test
|
qa 이슈 대응 등 🤖 기능 개요 ✅ implement todo 미지급 이자 뷰 스켈레톤 ui 모든 api 테스트 탈퇴사유 textarea 크기 조절 제외 qa 이슈 전부 대응 당겨서 새로고침 로딩 스피너 교체 📚 remarks
| 1
|
439,887
| 12,689,578,352
|
IssuesEvent
|
2020-06-21 07:03:53
|
rich-iannone/pointblank
|
https://api.github.com/repos/rich-iannone/pointblank
|
closed
|
Make `scan_data()` work with database tables
|
Difficulty: [3] Advanced Effort: [3] High Priority: [3] High Type: ★ Enhancement
|
Right now, `tbl_dbi` table objects are disallowed from use in `scan_data()`. Reason for this is that I couldn’t verify that the dplyr statements would generate valid SQL for the summary stats. The next phase with `scan_data()` is to make it work with any `tbl_dbi` object, aiming for high compatibility with different DB types.
|
1.0
|
Make `scan_data()` work with database tables - Right now, `tbl_dbi` table objects are disallowed from use in `scan_data()`. Reason for this is that I couldn’t verify that the dplyr statements would generate valid SQL for the summary stats. The next phase with `scan_data()` is to make it work with any `tbl_dbi` object, aiming for high compatibility with different DB types.
|
non_test
|
make scan data work with database tables right now tbl dbi table objects are disallowed from use in scan data reason for this is that i couldn’t verify that the dplyr statements would generate valid sql for the summary stats the next phase with scan data is to make it work with any tbl dbi object aiming for high compatibility with different db types
| 0
|
62,813
| 17,202,115,885
|
IssuesEvent
|
2021-07-17 13:11:44
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Papercut: alt-text on fullscreen images are annoying
|
A-Media T-Defect
|
<!-- A picture's worth a thousand words: PLEASE INCLUDE A SCREENSHOT :P -->
<!-- Please report security issues by email to security@matrix.org -->
<!-- This is a bug report template. By following the instructions below and
filling out the sections with your information, you will help the us to get all
the necessary data to fix your issue.
You can also preview your report before submitting it. You may remove sections
that aren't relevant to your particular case.
Text between <!-- and --> marks will be invisible in the report.
-->
So there I was, sitting in the TWiM room enjoying some late-night chatter about hacking doorbell systems - when I came across an image of a pcb. I squinted, attempting to stare at the tiny inscriptions upon the green metal, when I realised. "how foolish of me, Element has image zoom. We live in fruitful times."
I clicked the image to make it fullscreen and zoomed in, confident in my complete lack of understanding of what the circuit was actually doing. But as my eyes scanned across, trying to make sense of the hieroglyphs, a box appeared on top of the image out of nowhere, displaying: **image.png**.

My attention was lost, and instead I suddenly had the urge to make a bug report rather than pretending to understand circuit boards :(
TL;DR it's annoying that the image filename appears as alt text when you hover over it with your mouse, especially when this information is already available elsewhere.
#### Steps to reproduce
- Open an image from a room's timeline.
- Hover over it with your mouse.
- Alt-text appears.
<!-- Include screenshots if possible: you can drag and drop images below. -->
#### Version information
<!-- IMPORTANT: please answer the following questions, to help us narrow down the problem -->
- **Platform**: Desktop
- **OS**: Linux
- **Version**: Nightly 2021071601
|
1.0
|
Papercut: alt-text on fullscreen images are annoying - <!-- A picture's worth a thousand words: PLEASE INCLUDE A SCREENSHOT :P -->
<!-- Please report security issues by email to security@matrix.org -->
<!-- This is a bug report template. By following the instructions below and
filling out the sections with your information, you will help the us to get all
the necessary data to fix your issue.
You can also preview your report before submitting it. You may remove sections
that aren't relevant to your particular case.
Text between <!-- and --> marks will be invisible in the report.
-->
So there I was, sitting in the TWiM room enjoying some late-night chatter about hacking doorbell systems - when I came across an image of a pcb. I squinted, attempting to stare at the tiny inscriptions upon the green metal, when I realised. "how foolish of me, Element has image zoom. We live in fruitful times."
I clicked the image to make it fullscreen and zoomed in, confident in my complete lack of understanding of what the circuit was actually doing. But as my eyes scanned across, trying to make sense of the hieroglyphs, a box appeared on top of the image out of nowhere, displaying: **image.png**.

My attention was lost, and instead I suddenly had the urge to make a bug report rather than pretending to understand circuit boards :(
TL;DR it's annoying that the image filename appears as alt text when you hover over it with your mouse, especially when this information is already available elsewhere.
#### Steps to reproduce
- Open an image from a room's timeline.
- Hover over it with your mouse.
- Alt-text appears.
<!-- Include screenshots if possible: you can drag and drop images below. -->
#### Version information
<!-- IMPORTANT: please answer the following questions, to help us narrow down the problem -->
- **Platform**: Desktop
- **OS**: Linux
- **Version**: Nightly 2021071601
|
non_test
|
papercut alt text on fullscreen images are annoying this is a bug report template by following the instructions below and filling out the sections with your information you will help the us to get all the necessary data to fix your issue you can also preview your report before submitting it you may remove sections that aren t relevant to your particular case text between marks will be invisible in the report so there i was sitting in the twim room enjoying some late night chatter about hacking doorbell systems when i came across an image of a pcb i squinted attempting to stare at the tiny inscriptions upon the green metal when i realised how foolish of me element has image zoom we live in fruitful times i clicked the image to make it fullscreen and zoomed in confident in my complete lack of understanding of what the circuit was actually doing but as my eyes scanned across trying to make sense of the hieroglyphs a box appeared on top of the image out of nowhere displaying image png my attention was lost and instead i suddenly had the urge to make a bug report rather than pretending to understand circuit boards tl dr it s annoying that the image filename appears as alt text when you hover over it with your mouse especially when this information is already available elsewhere steps to reproduce open an image from a room s timeline hover over it with your mouse alt text appears version information platform desktop os linux version nightly
| 0
|
291,601
| 25,159,031,610
|
IssuesEvent
|
2022-11-10 15:30:27
|
apache/skywalking
|
https://api.github.com/repos/apache/skywalking
|
closed
|
[Bug] [Test] K8sTagTest can't pass on M1
|
bug backend test
|
### Search before asking
- [X] I had searched in the [issues](https://github.com/apache/skywalking/issues?q=is%3Aissue) and found no similar issues.
### Apache SkyWalking Component
OAP server (apache/skywalking)
### What happened
Zulu 1.8 on aarm
```
ScriptEngineManager providers.next(): javax.script.ScriptEngineFactory: Provider jdk.nashorn.api.scripting.NashornScriptEngineFactory not a subtype
ScriptEngineManager providers.next(): javax.script.ScriptEngineFactory: Provider jdk.nashorn.api.scripting.NashornScriptEngineFactory not a subtype
```
Zulu 11 on aarm
```
java.lang.NoClassDefFoundError: Could not initialize class org.apache.skywalking.library.kubernetes.KubernetesServices$MockitoMock$54094179
```
### What you expected to happen
The test could pass on M1
### How to reproduce
Run it on M1. We may consider don't mock `enum` through powermock. As this is the only one test using it, and only it fails.
### Anything else
_No response_
### Are you willing to submit PR?
- [ ] Yes I am willing to submit a PR!
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
|
1.0
|
[Bug] [Test] K8sTagTest can't pass on M1 - ### Search before asking
- [X] I had searched in the [issues](https://github.com/apache/skywalking/issues?q=is%3Aissue) and found no similar issues.
### Apache SkyWalking Component
OAP server (apache/skywalking)
### What happened
Zulu 1.8 on aarm
```
ScriptEngineManager providers.next(): javax.script.ScriptEngineFactory: Provider jdk.nashorn.api.scripting.NashornScriptEngineFactory not a subtype
ScriptEngineManager providers.next(): javax.script.ScriptEngineFactory: Provider jdk.nashorn.api.scripting.NashornScriptEngineFactory not a subtype
```
Zulu 11 on aarm
```
java.lang.NoClassDefFoundError: Could not initialize class org.apache.skywalking.library.kubernetes.KubernetesServices$MockitoMock$54094179
```
### What you expected to happen
The test could pass on M1
### How to reproduce
Run it on M1. We may consider don't mock `enum` through powermock. As this is the only one test using it, and only it fails.
### Anything else
_No response_
### Are you willing to submit PR?
- [ ] Yes I am willing to submit a PR!
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
|
test
|
can t pass on search before asking i had searched in the and found no similar issues apache skywalking component oap server apache skywalking what happened zulu on aarm scriptenginemanager providers next javax script scriptenginefactory provider jdk nashorn api scripting nashornscriptenginefactory not a subtype scriptenginemanager providers next javax script scriptenginefactory provider jdk nashorn api scripting nashornscriptenginefactory not a subtype zulu on aarm java lang noclassdeffounderror could not initialize class org apache skywalking library kubernetes kubernetesservices mockitomock what you expected to happen the test could pass on how to reproduce run it on we may consider don t mock enum through powermock as this is the only one test using it and only it fails anything else no response are you willing to submit pr yes i am willing to submit a pr code of conduct i agree to follow this project s
| 1
|
675,999
| 23,113,064,080
|
IssuesEvent
|
2022-07-27 14:29:34
|
FEeasy404/GameUs
|
https://api.github.com/repos/FEeasy404/GameUs
|
closed
|
하단 탭 메뉴 구현
|
✨Feat 🖐Priority: Medium
|
## 추가 기능 설명
3.1.14. 하단 탭 메뉴
- 하단 탭 메뉴는 홈, 채팅, 게시물 작성, 프로필 4개의 메뉴로 구성되어 있습니다.
- 모든 페이지는 페이지 경로에 해당하는 탭 메뉴가 활성화됩니다.
## 할 일
- [x] 홈, 채팅, 게시물 작성, 프로필 4개의 메뉴 구현
- [ ] 페이지 경로에 해당하는 탭 메뉴 활성화
## ETC
|
1.0
|
하단 탭 메뉴 구현 - ## 추가 기능 설명
3.1.14. 하단 탭 메뉴
- 하단 탭 메뉴는 홈, 채팅, 게시물 작성, 프로필 4개의 메뉴로 구성되어 있습니다.
- 모든 페이지는 페이지 경로에 해당하는 탭 메뉴가 활성화됩니다.
## 할 일
- [x] 홈, 채팅, 게시물 작성, 프로필 4개의 메뉴 구현
- [ ] 페이지 경로에 해당하는 탭 메뉴 활성화
## ETC
|
non_test
|
하단 탭 메뉴 구현 추가 기능 설명 하단 탭 메뉴 하단 탭 메뉴는 홈 채팅 게시물 작성 프로필 메뉴로 구성되어 있습니다 모든 페이지는 페이지 경로에 해당하는 탭 메뉴가 활성화됩니다 할 일 홈 채팅 게시물 작성 프로필 메뉴 구현 페이지 경로에 해당하는 탭 메뉴 활성화 etc
| 0
|
220,963
| 17,272,035,051
|
IssuesEvent
|
2021-07-22 21:17:00
|
urapadmin/kiosk
|
https://api.github.com/repos/urapadmin/kiosk
|
closed
|
"modified on" for image changes when image added to a new context in file repository
|
bug kiosk minor test-stage
|
In testing #441 and seeing that indeed, a pre-existing image that is put into an additional context in the recording system retains its original "created on" date, as we expected, a small other issue was noted. The "modified on" field was the same as the original "created on" date after synchronization, which Lutz thinks is proper - the image wasn't modified, it was simply attached to a new context. But when I then added yet another new context using the file repository, not the recording system, the "modified on" was updated to today. (When I clicked "ok".)

|
1.0
|
"modified on" for image changes when image added to a new context in file repository - In testing #441 and seeing that indeed, a pre-existing image that is put into an additional context in the recording system retains its original "created on" date, as we expected, a small other issue was noted. The "modified on" field was the same as the original "created on" date after synchronization, which Lutz thinks is proper - the image wasn't modified, it was simply attached to a new context. But when I then added yet another new context using the file repository, not the recording system, the "modified on" was updated to today. (When I clicked "ok".)

|
test
|
modified on for image changes when image added to a new context in file repository in testing and seeing that indeed a pre existing image that is put into an additional context in the recording system retains its original created on date as we expected a small other issue was noted the modified on field was the same as the original created on date after synchronization which lutz thinks is proper the image wasn t modified it was simply attached to a new context but when i then added yet another new context using the file repository not the recording system the modified on was updated to today when i clicked ok
| 1
|
49,897
| 6,045,007,906
|
IssuesEvent
|
2017-06-12 08:04:24
|
telerik/UI-For-UWP
|
https://api.github.com/repos/telerik/UI-For-UWP
|
opened
|
Chart: Updating properties of PropertyNameDataPointBinding is not reflected in the chart.
|
status: ReadyForTest
|
If a property of a PropertyNameDataPointBinding is updated after the series is added to the chart, the change is not reflected.
|
1.0
|
Chart: Updating properties of PropertyNameDataPointBinding is not reflected in the chart. - If a property of a PropertyNameDataPointBinding is updated after the series is added to the chart, the change is not reflected.
|
test
|
chart updating properties of propertynamedatapointbinding is not reflected in the chart if a property of a propertynamedatapointbinding is updated after the series is added to the chart the change is not reflected
| 1
|
299,870
| 9,205,931,255
|
IssuesEvent
|
2019-03-08 12:08:43
|
wso2/product-is
|
https://api.github.com/repos/wso2/product-is
|
closed
|
Disabling the feature SP template
|
Affected/5.8.0-M24 Component/Identity Mgt Priority/High Severity/Major
|
Disabling the feature service provider template creation since the feature is not in production ready state.
|
1.0
|
Disabling the feature SP template - Disabling the feature service provider template creation since the feature is not in production ready state.
|
non_test
|
disabling the feature sp template disabling the feature service provider template creation since the feature is not in production ready state
| 0
|
16,165
| 20,602,524,769
|
IssuesEvent
|
2022-03-06 13:43:03
|
NationalSecurityAgency/ghidra
|
https://api.github.com/repos/NationalSecurityAgency/ghidra
|
closed
|
XCHG RSP,RBP
|
Feature: Decompiler Feature: Processor/x86
|
**Describe the bug**
```
MOV RBP,LAB_7ff7f9a01b44
XCHG qword ptr [RSP],RBP=>LAB_7ff7f9a01b44
```
If you use this assembler code (anti reversing trick) it will not show it as call!
Used by Arxan Anti Debugging and Anti reversing (Arxan Technologies or Digital.Ai)
**To Reproduce**
Steps to reproduce the behavior:
Compile the assembler code
Decompile it it will not show it as code!
**Expected behavior**
Should show as a call
**Screenshots**
https://imgur.com/zUdJMrh
**Attachments**
If applicable, please attach any files that caused problems or log files generated by the software.
**Environment (please complete the following information):**
- OS: Win 10
- Java Version: 11.X
- Ghidra Version: 10.1.2
- Ghidra Origin: ghidra-sre.org
|
1.0
|
XCHG RSP,RBP - **Describe the bug**
```
MOV RBP,LAB_7ff7f9a01b44
XCHG qword ptr [RSP],RBP=>LAB_7ff7f9a01b44
```
If you use this assembler code (anti reversing trick) it will not show it as call!
Used by Arxan Anti Debugging and Anti reversing (Arxan Technologies or Digital.Ai)
**To Reproduce**
Steps to reproduce the behavior:
Compile the assembler code
Decompile it it will not show it as code!
**Expected behavior**
Should show as a call
**Screenshots**
https://imgur.com/zUdJMrh
**Attachments**
If applicable, please attach any files that caused problems or log files generated by the software.
**Environment (please complete the following information):**
- OS: Win 10
- Java Version: 11.X
- Ghidra Version: 10.1.2
- Ghidra Origin: ghidra-sre.org
|
non_test
|
xchg rsp rbp describe the bug mov rbp lab xchg qword ptr rbp lab if you use this assembler code anti reversing trick it will not show it as call used by arxan anti debugging and anti reversing arxan technologies or digital ai to reproduce steps to reproduce the behavior compile the assembler code decompile it it will not show it as code expected behavior should show as a call screenshots attachments if applicable please attach any files that caused problems or log files generated by the software environment please complete the following information os win java version x ghidra version ghidra origin ghidra sre org
| 0
|
30,736
| 7,247,923,024
|
IssuesEvent
|
2018-02-15 06:42:35
|
mccabeshell/PSComputerAuditing
|
https://api.github.com/repos/mccabeshell/PSComputerAuditing
|
opened
|
Get-ComputerWsusAudit - ComputerScope
|
code improvement enhancement
|
## Handle WSUS Name Searching
Create two parameter sets:
1. FullDomainName i.e. must be an exact match and return one computer object.
2. NameIncludes i.e. can return multiple results
The following needs to be done, it will also greatly improve efficiency:
- Move ComputerScope and GetSummariesPerComputerTarget into PROCESS block
- Set ComputerScope NameIncludes parameter `$ComputerScope.NameIncludes = 'Computer1'`
- For FullDomainName continue to use GetComputerTargetByName & Where-Object IDs match, skip it for NameIncludes and set $ComputerTargetSummary = $AllComputerTargetSummaries
- ComputerTargetSummary can now contain multiple results, so add a loop around creation and output of object
|
1.0
|
Get-ComputerWsusAudit - ComputerScope - ## Handle WSUS Name Searching
Create two parameter sets:
1. FullDomainName i.e. must be an exact match and return one computer object.
2. NameIncludes i.e. can return multiple results
The following needs to be done, it will also greatly improve efficiency:
- Move ComputerScope and GetSummariesPerComputerTarget into PROCESS block
- Set ComputerScope NameIncludes parameter `$ComputerScope.NameIncludes = 'Computer1'`
- For FullDomainName continue to use GetComputerTargetByName & Where-Object IDs match, skip it for NameIncludes and set $ComputerTargetSummary = $AllComputerTargetSummaries
- ComputerTargetSummary can now contain multiple results, so add a loop around creation and output of object
|
non_test
|
get computerwsusaudit computerscope handle wsus name searching create two parameter sets fulldomainname i e must be an exact match and return one computer object nameincludes i e can return multiple results the following needs to be done it will also greatly improve efficiency move computerscope and getsummariespercomputertarget into process block set computerscope nameincludes parameter computerscope nameincludes for fulldomainname continue to use getcomputertargetbyname where object ids match skip it for nameincludes and set computertargetsummary allcomputertargetsummaries computertargetsummary can now contain multiple results so add a loop around creation and output of object
| 0
|
755,568
| 26,433,152,113
|
IssuesEvent
|
2023-01-15 03:07:43
|
Tedeapolis/development
|
https://api.github.com/repos/Tedeapolis/development
|
closed
|
[EUP]: Steekvest AMBU
|
EUP - interiors - retextures accepted low priority
|
### Contact Details
Wilma Wonka#4312
### Waar bevind de issue zich?
EUP/Kleding
### Wat klopt er niet
Suggestie: Scherf/steek/kogelvrij vesten / helmen
Aangezien we merken dat we als ambulanciers in steeds minder veilige situaties komen in TDA-stad denk ik dat het handig is om de ambulanciers uit te rusten met een scherf/steek/kogelvrij vest dit omdat ambulanciers steeds vaker moeten werken bij bvb schiet/steekmeldingen van de politie/kmar waarbij ik altijd denk is het niet handig dat we de mogelijkheid hebben om een vest aan te trekken opdat onze vitale organen beter beschermt zijn. In het echt hebben veel regionale ambulance diensten ook een vest ter bescherming van de medewerks in ieder geval in Almelo wel. Ik zou het fijn vinden of er hier mogelijk naar gekeken kan worden.
Ik weet niet of deze hier hoort. Sorry als ie verkeerd staat. Maar volgens mij is deze ooit aangemaakt maar kan het nergens meer terug vinden wat hier op was gezegt!
|
1.0
|
[EUP]: Steekvest AMBU - ### Contact Details
Wilma Wonka#4312
### Waar bevind de issue zich?
EUP/Kleding
### Wat klopt er niet
Suggestie: Scherf/steek/kogelvrij vesten / helmen
Aangezien we merken dat we als ambulanciers in steeds minder veilige situaties komen in TDA-stad denk ik dat het handig is om de ambulanciers uit te rusten met een scherf/steek/kogelvrij vest dit omdat ambulanciers steeds vaker moeten werken bij bvb schiet/steekmeldingen van de politie/kmar waarbij ik altijd denk is het niet handig dat we de mogelijkheid hebben om een vest aan te trekken opdat onze vitale organen beter beschermt zijn. In het echt hebben veel regionale ambulance diensten ook een vest ter bescherming van de medewerks in ieder geval in Almelo wel. Ik zou het fijn vinden of er hier mogelijk naar gekeken kan worden.
Ik weet niet of deze hier hoort. Sorry als ie verkeerd staat. Maar volgens mij is deze ooit aangemaakt maar kan het nergens meer terug vinden wat hier op was gezegt!
|
non_test
|
steekvest ambu contact details wilma wonka waar bevind de issue zich eup kleding wat klopt er niet suggestie scherf steek kogelvrij vesten helmen aangezien we merken dat we als ambulanciers in steeds minder veilige situaties komen in tda stad denk ik dat het handig is om de ambulanciers uit te rusten met een scherf steek kogelvrij vest dit omdat ambulanciers steeds vaker moeten werken bij bvb schiet steekmeldingen van de politie kmar waarbij ik altijd denk is het niet handig dat we de mogelijkheid hebben om een vest aan te trekken opdat onze vitale organen beter beschermt zijn in het echt hebben veel regionale ambulance diensten ook een vest ter bescherming van de medewerks in ieder geval in almelo wel ik zou het fijn vinden of er hier mogelijk naar gekeken kan worden ik weet niet of deze hier hoort sorry als ie verkeerd staat maar volgens mij is deze ooit aangemaakt maar kan het nergens meer terug vinden wat hier op was gezegt
| 0
|
260,568
| 8,211,795,056
|
IssuesEvent
|
2018-09-04 14:41:19
|
poanetwork/token-wizard
|
https://api.github.com/repos/poanetwork/token-wizard
|
reopened
|
(Feature) metadata should be added to zip archive for download
|
enhancement medium priority migration to auth-os
|
**Problem**: the last part of #1035
**Solution**: file with metadata (smart-contract name, ABI encoded parameters, version of compiler, optimization flag value) should be added to zip archive for download. Example of such file could be gotten from the live version of Token Wizard (https://wizard.poa.network/)
|
1.0
|
(Feature) metadata should be added to zip archive for download - **Problem**: the last part of #1035
**Solution**: file with metadata (smart-contract name, ABI encoded parameters, version of compiler, optimization flag value) should be added to zip archive for download. Example of such file could be gotten from the live version of Token Wizard (https://wizard.poa.network/)
|
non_test
|
feature metadata should be added to zip archive for download problem the last part of solution file with metadata smart contract name abi encoded parameters version of compiler optimization flag value should be added to zip archive for download example of such file could be gotten from the live version of token wizard
| 0
|
189,228
| 14,494,902,645
|
IssuesEvent
|
2020-12-11 10:24:34
|
zephyrproject-rtos/zephyr
|
https://api.github.com/repos/zephyrproject-rtos/zephyr
|
closed
|
tests: kernel: Test kernel.memory_protection.gap_filling fails on nrf5340dk_nrf5340_cpuapp
|
area: Kernel area: Tests bug platform: nRF priority: medium
|
**Describe the bug**
Test from zephyr/tests/kernel/mem_protect/mem_protect fails on nrf5340dk_nrf5340_cpuapp.
**To Reproduce**
Steps to reproduce the behavior:
1. have nrf5340dk_nrf5340_cpuapp
2. go to zephyr dir
3. run `./scripts/sanitycheck --device-testing -p nrf5340dk_nrf5340_cpuapp -T tests/kernel/mem_protect/mem_protect --device-serial /dev/ttyACM2 -v -v
`
4. See error
**Expected behavior**
Test passes
**Impact**
Not clear
**Logs and console output**
```
*** Booting Zephyr OS build zephyr-v2.4.0-1784-g2ce570b03f97 ***
Running test suite memory_protection_test_suite
===================================================================
START - test_permission_inheritance
E: ***** MPU FAULT *****
E: Data Access Violation
E: MMFAR Address: 0x20001030
E: r0/a1: 0x00000000 r1/a2: 0x00000000 r2/a3: 0x00000001
E: r3/a4: 0x20001030 r12/ip: 0x00000765 r14/lr: 0x0000075d
E: xpsr: 0x21000000
E: Faulting instruction address (r15/pc): 0x000007e4
E: >>> ZEPHYR FATAL ERROR 0: CPU exception on CPU 0
E: Current thread: 0x20000750 (unknown)
Caught system error -- reason 0 1
fatal error expected as part of test case
PASS - test_permission_inheritance
===================================================================
START - test_inherit_resource_pool
PASS - test_inherit_resource_pool
===================================================================
START - test_mem_domain_setup
Assertion failed at WEST_TOPDIR/zephyr/tests/kernel/mem_protect/mem_protect/src/mem_domain.c:51: test_mem_domain_setup: (num_rw_parts > 0 is false)
no free memory partitions
FAIL - test_mem_domain_setup
===================================================================
START - test_mem_domain_valid_access
E: ***** MPU FAULT *****
E: Data Access Violation
E: MMFAR Address: 0x0
E: r0/a1: 0x200011dc r1/a2: 0x20000b38 r2/a3: 0x00000000
E: r3/a4: 0x20000bd4 r12/ip: 0x00000000 r14/lr: 0x0000be77
E: xpsr: 0x61000000
E: Faulting instruction address (r15/pc): 0x0000b876
E: >>> ZEPHYR FATAL ERROR 0: CPU exception on CPU 0
E: Current thread: 0x20000c00 (ztest_thread)
Caught system error -- reason 0 0
fatal error was unexpected, aborting
```
**Environment (please complete the following information):**
- OS: Linux 18.04
- Toolchain Zephyr SDK
- Commit SHA or Version used: zephyr-v2.4.0-1784-g2ce570b03f97
**Additional context**
The test passes on pdk version of nrf53
|
1.0
|
tests: kernel: Test kernel.memory_protection.gap_filling fails on nrf5340dk_nrf5340_cpuapp - **Describe the bug**
Test from zephyr/tests/kernel/mem_protect/mem_protect fails on nrf5340dk_nrf5340_cpuapp.
**To Reproduce**
Steps to reproduce the behavior:
1. have nrf5340dk_nrf5340_cpuapp
2. go to zephyr dir
3. run `./scripts/sanitycheck --device-testing -p nrf5340dk_nrf5340_cpuapp -T tests/kernel/mem_protect/mem_protect --device-serial /dev/ttyACM2 -v -v
`
4. See error
**Expected behavior**
Test passes
**Impact**
Not clear
**Logs and console output**
```
*** Booting Zephyr OS build zephyr-v2.4.0-1784-g2ce570b03f97 ***
Running test suite memory_protection_test_suite
===================================================================
START - test_permission_inheritance
E: ***** MPU FAULT *****
E: Data Access Violation
E: MMFAR Address: 0x20001030
E: r0/a1: 0x00000000 r1/a2: 0x00000000 r2/a3: 0x00000001
E: r3/a4: 0x20001030 r12/ip: 0x00000765 r14/lr: 0x0000075d
E: xpsr: 0x21000000
E: Faulting instruction address (r15/pc): 0x000007e4
E: >>> ZEPHYR FATAL ERROR 0: CPU exception on CPU 0
E: Current thread: 0x20000750 (unknown)
Caught system error -- reason 0 1
fatal error expected as part of test case
PASS - test_permission_inheritance
===================================================================
START - test_inherit_resource_pool
PASS - test_inherit_resource_pool
===================================================================
START - test_mem_domain_setup
Assertion failed at WEST_TOPDIR/zephyr/tests/kernel/mem_protect/mem_protect/src/mem_domain.c:51: test_mem_domain_setup: (num_rw_parts > 0 is false)
no free memory partitions
FAIL - test_mem_domain_setup
===================================================================
START - test_mem_domain_valid_access
E: ***** MPU FAULT *****
E: Data Access Violation
E: MMFAR Address: 0x0
E: r0/a1: 0x200011dc r1/a2: 0x20000b38 r2/a3: 0x00000000
E: r3/a4: 0x20000bd4 r12/ip: 0x00000000 r14/lr: 0x0000be77
E: xpsr: 0x61000000
E: Faulting instruction address (r15/pc): 0x0000b876
E: >>> ZEPHYR FATAL ERROR 0: CPU exception on CPU 0
E: Current thread: 0x20000c00 (ztest_thread)
Caught system error -- reason 0 0
fatal error was unexpected, aborting
```
**Environment (please complete the following information):**
- OS: Linux 18.04
- Toolchain Zephyr SDK
- Commit SHA or Version used: zephyr-v2.4.0-1784-g2ce570b03f97
**Additional context**
The test passes on pdk version of nrf53
|
test
|
tests kernel test kernel memory protection gap filling fails on cpuapp describe the bug test from zephyr tests kernel mem protect mem protect fails on cpuapp to reproduce steps to reproduce the behavior have cpuapp go to zephyr dir run scripts sanitycheck device testing p cpuapp t tests kernel mem protect mem protect device serial dev v v see error expected behavior test passes impact not clear logs and console output booting zephyr os build zephyr running test suite memory protection test suite start test permission inheritance e mpu fault e data access violation e mmfar address e e ip lr e xpsr e faulting instruction address pc e zephyr fatal error cpu exception on cpu e current thread unknown caught system error reason fatal error expected as part of test case pass test permission inheritance start test inherit resource pool pass test inherit resource pool start test mem domain setup assertion failed at west topdir zephyr tests kernel mem protect mem protect src mem domain c test mem domain setup num rw parts is false no free memory partitions fail test mem domain setup start test mem domain valid access e mpu fault e data access violation e mmfar address e e ip lr e xpsr e faulting instruction address pc e zephyr fatal error cpu exception on cpu e current thread ztest thread caught system error reason fatal error was unexpected aborting environment please complete the following information os linux toolchain zephyr sdk commit sha or version used zephyr additional context the test passes on pdk version of
| 1
|
349,531
| 31,810,776,074
|
IssuesEvent
|
2023-09-13 16:40:16
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
opened
|
Failing test: Chrome X-Pack UI Functional Tests with ES SSL - Discover, Uptime, ML.x-pack/test/functional_with_es_ssl/apps/discover_ml_uptime/ml/alert_flyout·ts - ML app anomaly detection alert "before all" hook in "anomaly detection alert"
|
failed-test
|
A test failed on a tracked branch
```
WebDriverError: unknown error: session deleted because of page crash
from unknown error: cannot determine loading status
from tab crashed
(Session info: chrome=116.0.5845.187)
at Object.throwDecodedError (node_modules/selenium-webdriver/lib/error.js:524:15)
at parseHttpResponse (node_modules/selenium-webdriver/lib/http.js:601:13)
at Executor.execute (node_modules/selenium-webdriver/lib/http.js:529:28)
at processTicksAndRejections (node:internal/process/task_queues:95:5)
at Task.exec (prevent_parallel_calls.ts:28:20) {
remoteStacktrace: '#0 0x55c57db46e23 <unknown>\n' +
'#1 0x55c57d86f5f6 <unknown>\n' +
'#2 0x55c57d856567 <unknown>\n' +
'#3 0x55c57d85528a <unknown>\n' +
'#4 0x55c57d85584f <unknown>\n' +
'#5 0x55c57d865b47 <unknown>\n' +
'#6 0x55c57d866ca2 <unknown>\n' +
'#7 0x55c57d879a31 <unknown>\n' +
'#8 0x55c57d8e74b7 <unknown>\n' +
'#9 0x55c57d8cede3 <unknown>\n' +
'#10 0x55c57d8a2a7b <unknown>\n' +
'#11 0x55c57d8a381e <unknown>\n' +
'#12 0x55c57db08638 <unknown>\n' +
'#13 0x55c57db0c507 <unknown>\n' +
'#14 0x55c57db16c4c <unknown>\n' +
'#15 0x55c57db0d136 <unknown>\n' +
'#16 0x55c57dadb9cf <unknown>\n' +
'#17 0x55c57db30b98 <unknown>\n' +
'#18 0x55c57db30d68 <unknown>\n' +
'#19 0x55c57db3fcb3 <unknown>\n' +
'#20 0x7fdd04ec8609 start_thread\n'
}
```
First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/35422#018a8f37-87fa-48d5-9fd8-a6a912c670d5)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests with ES SSL - Discover, Uptime, ML.x-pack/test/functional_with_es_ssl/apps/discover_ml_uptime/ml/alert_flyout·ts","test.name":"ML app anomaly detection alert \"before all\" hook in \"anomaly detection alert\"","test.failCount":1}} -->
|
1.0
|
Failing test: Chrome X-Pack UI Functional Tests with ES SSL - Discover, Uptime, ML.x-pack/test/functional_with_es_ssl/apps/discover_ml_uptime/ml/alert_flyout·ts - ML app anomaly detection alert "before all" hook in "anomaly detection alert" - A test failed on a tracked branch
```
WebDriverError: unknown error: session deleted because of page crash
from unknown error: cannot determine loading status
from tab crashed
(Session info: chrome=116.0.5845.187)
at Object.throwDecodedError (node_modules/selenium-webdriver/lib/error.js:524:15)
at parseHttpResponse (node_modules/selenium-webdriver/lib/http.js:601:13)
at Executor.execute (node_modules/selenium-webdriver/lib/http.js:529:28)
at processTicksAndRejections (node:internal/process/task_queues:95:5)
at Task.exec (prevent_parallel_calls.ts:28:20) {
remoteStacktrace: '#0 0x55c57db46e23 <unknown>\n' +
'#1 0x55c57d86f5f6 <unknown>\n' +
'#2 0x55c57d856567 <unknown>\n' +
'#3 0x55c57d85528a <unknown>\n' +
'#4 0x55c57d85584f <unknown>\n' +
'#5 0x55c57d865b47 <unknown>\n' +
'#6 0x55c57d866ca2 <unknown>\n' +
'#7 0x55c57d879a31 <unknown>\n' +
'#8 0x55c57d8e74b7 <unknown>\n' +
'#9 0x55c57d8cede3 <unknown>\n' +
'#10 0x55c57d8a2a7b <unknown>\n' +
'#11 0x55c57d8a381e <unknown>\n' +
'#12 0x55c57db08638 <unknown>\n' +
'#13 0x55c57db0c507 <unknown>\n' +
'#14 0x55c57db16c4c <unknown>\n' +
'#15 0x55c57db0d136 <unknown>\n' +
'#16 0x55c57dadb9cf <unknown>\n' +
'#17 0x55c57db30b98 <unknown>\n' +
'#18 0x55c57db30d68 <unknown>\n' +
'#19 0x55c57db3fcb3 <unknown>\n' +
'#20 0x7fdd04ec8609 start_thread\n'
}
```
First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/35422#018a8f37-87fa-48d5-9fd8-a6a912c670d5)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests with ES SSL - Discover, Uptime, ML.x-pack/test/functional_with_es_ssl/apps/discover_ml_uptime/ml/alert_flyout·ts","test.name":"ML app anomaly detection alert \"before all\" hook in \"anomaly detection alert\"","test.failCount":1}} -->
|
test
|
failing test chrome x pack ui functional tests with es ssl discover uptime ml x pack test functional with es ssl apps discover ml uptime ml alert flyout·ts ml app anomaly detection alert before all hook in anomaly detection alert a test failed on a tracked branch webdrivererror unknown error session deleted because of page crash from unknown error cannot determine loading status from tab crashed session info chrome at object throwdecodederror node modules selenium webdriver lib error js at parsehttpresponse node modules selenium webdriver lib http js at executor execute node modules selenium webdriver lib http js at processticksandrejections node internal process task queues at task exec prevent parallel calls ts remotestacktrace n n n n n n n n n n n n n n n n n n n n start thread n first failure
| 1
|
111,070
| 9,489,685,115
|
IssuesEvent
|
2019-04-22 23:38:06
|
ihhub/penguinV
|
https://api.github.com/repos/ihhub/penguinV
|
closed
|
Add unit tests for Merge() function
|
good first issue unit tests
|
We have an implementation of **Merge**() function in `src/image_function.h` and `src/image_function.cpp`. Our unit tests locate in `test/unit_tests/unit_test_image_function.cpp`. We do not have unit tests for this function. What we need to do is to add 4 unit test functions to cover 4 different types of Merge() function. Please refer to Invert() function as an example.
|
1.0
|
Add unit tests for Merge() function - We have an implementation of **Merge**() function in `src/image_function.h` and `src/image_function.cpp`. Our unit tests locate in `test/unit_tests/unit_test_image_function.cpp`. We do not have unit tests for this function. What we need to do is to add 4 unit test functions to cover 4 different types of Merge() function. Please refer to Invert() function as an example.
|
test
|
add unit tests for merge function we have an implementation of merge function in src image function h and src image function cpp our unit tests locate in test unit tests unit test image function cpp we do not have unit tests for this function what we need to do is to add unit test functions to cover different types of merge function please refer to invert function as an example
| 1
|
331,641
| 29,044,890,735
|
IssuesEvent
|
2023-05-13 12:33:14
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
Failing test: X-Pack Case API Integration Tests.x-pack/test/cases_api_integration/security_and_spaces/tests/trial/cases/push_case·ts - cases security and spaces enabled: trial push_case incident recorder server should map the fields and add the backlink to Kibana correctly
|
failed-test Team:ResponseOps
|
A test failed on a tracked branch
```
Error: expected 200 "OK", got 403 "Forbidden"
at Test._assertStatus (node_modules/supertest/lib/test.js:268:12)
at Test._assertFunction (node_modules/supertest/lib/test.js:283:11)
at Test.assert (node_modules/supertest/lib/test.js:173:18)
at localAssert (node_modules/supertest/lib/test.js:131:12)
at /home/buildkite-agent/builds/kb-n2-4-spot-0431e15284df11c1/elastic/kibana-on-merge/kibana/node_modules/supertest/lib/test.js:128:5
at Test.Request.callback (node_modules/superagent/lib/node/index.js:728:3)
at /home/buildkite-agent/builds/kb-n2-4-spot-0431e15284df11c1/elastic/kibana-on-merge/kibana/node_modules/superagent/lib/node/index.js:916:18
at IncomingMessage.<anonymous> (node_modules/superagent/lib/node/parsers/json.js:19:7)
at IncomingMessage.emit (node:events:525:35)
at endReadableNT (node:internal/streams/readable:1358:12)
at processTicksAndRejections (node:internal/process/task_queues:83:21)
```
First failure: [CI Build - 8.8](https://buildkite.com/elastic/kibana-on-merge/builds/30378#01881150-c3e6-4d3d-8eb0-32da324fbeb3)
<!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Case API Integration Tests.x-pack/test/cases_api_integration/security_and_spaces/tests/trial/cases/push_case·ts","test.name":"cases security and spaces enabled: trial push_case incident recorder server should map the fields and add the backlink to Kibana correctly","test.failCount":2}} -->
|
1.0
|
Failing test: X-Pack Case API Integration Tests.x-pack/test/cases_api_integration/security_and_spaces/tests/trial/cases/push_case·ts - cases security and spaces enabled: trial push_case incident recorder server should map the fields and add the backlink to Kibana correctly - A test failed on a tracked branch
```
Error: expected 200 "OK", got 403 "Forbidden"
at Test._assertStatus (node_modules/supertest/lib/test.js:268:12)
at Test._assertFunction (node_modules/supertest/lib/test.js:283:11)
at Test.assert (node_modules/supertest/lib/test.js:173:18)
at localAssert (node_modules/supertest/lib/test.js:131:12)
at /home/buildkite-agent/builds/kb-n2-4-spot-0431e15284df11c1/elastic/kibana-on-merge/kibana/node_modules/supertest/lib/test.js:128:5
at Test.Request.callback (node_modules/superagent/lib/node/index.js:728:3)
at /home/buildkite-agent/builds/kb-n2-4-spot-0431e15284df11c1/elastic/kibana-on-merge/kibana/node_modules/superagent/lib/node/index.js:916:18
at IncomingMessage.<anonymous> (node_modules/superagent/lib/node/parsers/json.js:19:7)
at IncomingMessage.emit (node:events:525:35)
at endReadableNT (node:internal/streams/readable:1358:12)
at processTicksAndRejections (node:internal/process/task_queues:83:21)
```
First failure: [CI Build - 8.8](https://buildkite.com/elastic/kibana-on-merge/builds/30378#01881150-c3e6-4d3d-8eb0-32da324fbeb3)
<!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Case API Integration Tests.x-pack/test/cases_api_integration/security_and_spaces/tests/trial/cases/push_case·ts","test.name":"cases security and spaces enabled: trial push_case incident recorder server should map the fields and add the backlink to Kibana correctly","test.failCount":2}} -->
|
test
|
failing test x pack case api integration tests x pack test cases api integration security and spaces tests trial cases push case·ts cases security and spaces enabled trial push case incident recorder server should map the fields and add the backlink to kibana correctly a test failed on a tracked branch error expected ok got forbidden at test assertstatus node modules supertest lib test js at test assertfunction node modules supertest lib test js at test assert node modules supertest lib test js at localassert node modules supertest lib test js at home buildkite agent builds kb spot elastic kibana on merge kibana node modules supertest lib test js at test request callback node modules superagent lib node index js at home buildkite agent builds kb spot elastic kibana on merge kibana node modules superagent lib node index js at incomingmessage node modules superagent lib node parsers json js at incomingmessage emit node events at endreadablent node internal streams readable at processticksandrejections node internal process task queues first failure
| 1
|
22,668
| 3,968,209,404
|
IssuesEvent
|
2016-05-03 18:51:10
|
angular/angular
|
https://api.github.com/repos/angular/angular
|
closed
|
ComponentFixture.onStable does not wait for macrotasks
|
comp: core/testbed
|
Currently, `ComponentFixture.onStable` waits for microtasks (via `NgZone.onStable`), but it does not check if there are pending macrotasks. It should do this, via `NgZone.hasPendingMacrotasks`.
cc@vikerman
|
1.0
|
ComponentFixture.onStable does not wait for macrotasks - Currently, `ComponentFixture.onStable` waits for microtasks (via `NgZone.onStable`), but it does not check if there are pending macrotasks. It should do this, via `NgZone.hasPendingMacrotasks`.
cc@vikerman
|
test
|
componentfixture onstable does not wait for macrotasks currently componentfixture onstable waits for microtasks via ngzone onstable but it does not check if there are pending macrotasks it should do this via ngzone haspendingmacrotasks cc vikerman
| 1
|
716,178
| 24,624,715,037
|
IssuesEvent
|
2022-10-16 11:15:56
|
Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2
|
https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2
|
opened
|
Patreon-characters branch crashes
|
priority high :exclamation: bug :bug:
|
<!--
**DO NOT REMOVE PRE-EXISTING LINES**
------------------------------------------------------------------------------------------------------------
-->
**Your mod version is:**
d70d08b6b5fbc90796f3030b2213daa9eed93b2b (CK3 `1.5.0.1`)
**What expansions do you have installed?**
- Garments of the Holy Roman Empire (Sep 2020)
- Northern Lords (Mar 2021)
**Are you using any submods/mods? If so, which?**
None.
**Please explain your issue in as much detail as possible:**
Game crashes repeatedly in loading screen. The following reports in `exception.txt` and `error.log` stood out to me. Full `error.log` download available at the bottom.
<details><summary>Click to expand</summary>
**exception.txt**
```
Unhandled Exception C0000005 (EXCEPTION_ACCESS_VIOLATION) at address 0x00007FF768E88566
Stack Trace:
1 ck3.exe (function-name not available) (+ 0)
2 ck3.exe (function-name not available) (+ 0)
3 ck3.exe (function-name not available) (+ 0)
4 ck3.exe (function-name not available) (+ 0)
5 ck3.exe (function-name not available) (+ 0)
6 ck3.exe (function-name not available) (+ 0)
7 ck3.exe (function-name not available) (+ 0)
8 ck3.exe SDL_StopTextInput (+ 2446547)
9 ck3.exe (function-name not available) (+ 0)
10 ck3.exe (function-name not available) (+ 0)
11 ck3.exe (function-name not available) (+ 0)
12 ck3.exe (function-name not available) (+ 0)
13 ck3.exe SDL_StopTextInput (+ 2453446)
14 ck3.exe SDL_StopTextInput (+ 2453579)
15 ck3.exe SDL_StopTextInput (+ 2470001)
16 ck3.exe SDL_StopTextInput (+ 2464510)
17 ck3.exe SDL_StopTextInput (+ 2617513)
18 ck3.exe SDL_StopTextInput (+ 2617927)
19 ck3.exe SDL_DYNAPI_entry (+ 1516804)
20 KERNEL32.DLL BaseThreadInitThunk (+ 20)
21 ntdll.dll RtlUserThreadStart (+ 33)
22 ntdll.dll RtlUserThreadStart (+ 33)
```
**error.log**
```
[13:43:47][log.cpp:466]: Instance has not been created!
00007FF7D4E86D00 (ck3): (filename not available): SDL_StopTextInput
00007FF7D34D99C5 (ck3): (filename not available): (function-name not available)
00007FF7D412D039 (ck3): (filename not available): (function-name not available)
00007FF7D412D44F (ck3): (filename not available): (function-name not available)
00007FF7D412D772 (ck3): (filename not available): (function-name not available)
00007FF7D412B7DD (ck3): (filename not available): (function-name not available)
00007FF7D3A6F7C1 (ck3): (filename not available): (function-name not available)
00007FF7D40DE142 (ck3): (filename not available): (function-name not available)
00007FF7D4E9F473 (ck3): (filename not available): SDL_StopTextInput
00007FF7D3DAF34B (ck3): (filename not available): (function-name not available)
00007FF7D3D6A18E (ck3): (filename not available): (function-name not available)
00007FF7D3DAF504 (ck3): (filename not available): (function-name not available)
00007FF7D3D69B1E (ck3): (filename not available): (function-name not available)
00007FF7D4EA0F66 (ck3): (filename not available): SDL_StopTextInput
00007FF7D4EA0FEB (ck3): (filename not available): SDL_StopTextInput
00007FF7D4EA5011 (ck3): (filename not available): SDL_StopTextInput
00007FF7D4EA3A9E (ck3): (filename not available): SDL_StopTextInput
00007FF7D4EC9049 (ck3): (filename not available): SDL_StopTextInput
00007FF7D4EC91E7 (ck3): (filename not available): SDL_StopTextInput
00007FF7D507D9A4 (ck3): (filename not available): SDL_DYNAPI_entry
00007FFF584B7034 (KERNEL32): (filename not available): BaseThreadInitThunk
00007FFF590A26A1 (ntdll): (filename not available): RtlUserThreadStart
```
</details>
**Steps to reproduce the issue:**
Downgrade the game to `1.5.0.1` and try starting up `Patreon-characters` branch.
**Upload an attachment below: .zip of your save, or screenshots:**
Regarding `EXCEPTION_ACCESS_VIOLATION`, I was able to find the following info from the **Mod Co-Op**:
> "That says the game engine is trying to write to memory it is forbidden to write to (i.e., stuff the operating system can only access)"
General advice seemed to be to remove folders from the mod one by one until the game no longer crashes.
[error.log](https://github.com/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2/files/9794294/error.log)
|
1.0
|
Patreon-characters branch crashes - <!--
**DO NOT REMOVE PRE-EXISTING LINES**
------------------------------------------------------------------------------------------------------------
-->
**Your mod version is:**
d70d08b6b5fbc90796f3030b2213daa9eed93b2b (CK3 `1.5.0.1`)
**What expansions do you have installed?**
- Garments of the Holy Roman Empire (Sep 2020)
- Northern Lords (Mar 2021)
**Are you using any submods/mods? If so, which?**
None.
**Please explain your issue in as much detail as possible:**
Game crashes repeatedly in loading screen. The following reports in `exception.txt` and `error.log` stood out to me. Full `error.log` download available at the bottom.
<details><summary>Click to expand</summary>
**exception.txt**
```
Unhandled Exception C0000005 (EXCEPTION_ACCESS_VIOLATION) at address 0x00007FF768E88566
Stack Trace:
1 ck3.exe (function-name not available) (+ 0)
2 ck3.exe (function-name not available) (+ 0)
3 ck3.exe (function-name not available) (+ 0)
4 ck3.exe (function-name not available) (+ 0)
5 ck3.exe (function-name not available) (+ 0)
6 ck3.exe (function-name not available) (+ 0)
7 ck3.exe (function-name not available) (+ 0)
8 ck3.exe SDL_StopTextInput (+ 2446547)
9 ck3.exe (function-name not available) (+ 0)
10 ck3.exe (function-name not available) (+ 0)
11 ck3.exe (function-name not available) (+ 0)
12 ck3.exe (function-name not available) (+ 0)
13 ck3.exe SDL_StopTextInput (+ 2453446)
14 ck3.exe SDL_StopTextInput (+ 2453579)
15 ck3.exe SDL_StopTextInput (+ 2470001)
16 ck3.exe SDL_StopTextInput (+ 2464510)
17 ck3.exe SDL_StopTextInput (+ 2617513)
18 ck3.exe SDL_StopTextInput (+ 2617927)
19 ck3.exe SDL_DYNAPI_entry (+ 1516804)
20 KERNEL32.DLL BaseThreadInitThunk (+ 20)
21 ntdll.dll RtlUserThreadStart (+ 33)
22 ntdll.dll RtlUserThreadStart (+ 33)
```
**error.log**
```
[13:43:47][log.cpp:466]: Instance has not been created!
00007FF7D4E86D00 (ck3): (filename not available): SDL_StopTextInput
00007FF7D34D99C5 (ck3): (filename not available): (function-name not available)
00007FF7D412D039 (ck3): (filename not available): (function-name not available)
00007FF7D412D44F (ck3): (filename not available): (function-name not available)
00007FF7D412D772 (ck3): (filename not available): (function-name not available)
00007FF7D412B7DD (ck3): (filename not available): (function-name not available)
00007FF7D3A6F7C1 (ck3): (filename not available): (function-name not available)
00007FF7D40DE142 (ck3): (filename not available): (function-name not available)
00007FF7D4E9F473 (ck3): (filename not available): SDL_StopTextInput
00007FF7D3DAF34B (ck3): (filename not available): (function-name not available)
00007FF7D3D6A18E (ck3): (filename not available): (function-name not available)
00007FF7D3DAF504 (ck3): (filename not available): (function-name not available)
00007FF7D3D69B1E (ck3): (filename not available): (function-name not available)
00007FF7D4EA0F66 (ck3): (filename not available): SDL_StopTextInput
00007FF7D4EA0FEB (ck3): (filename not available): SDL_StopTextInput
00007FF7D4EA5011 (ck3): (filename not available): SDL_StopTextInput
00007FF7D4EA3A9E (ck3): (filename not available): SDL_StopTextInput
00007FF7D4EC9049 (ck3): (filename not available): SDL_StopTextInput
00007FF7D4EC91E7 (ck3): (filename not available): SDL_StopTextInput
00007FF7D507D9A4 (ck3): (filename not available): SDL_DYNAPI_entry
00007FFF584B7034 (KERNEL32): (filename not available): BaseThreadInitThunk
00007FFF590A26A1 (ntdll): (filename not available): RtlUserThreadStart
```
</details>
**Steps to reproduce the issue:**
Downgrade the game to `1.5.0.1` and try starting up `Patreon-characters` branch.
**Upload an attachment below: .zip of your save, or screenshots:**
Regarding `EXCEPTION_ACCESS_VIOLATION`, I was able to find the following info from the **Mod Co-Op**:
> "That says the game engine is trying to write to memory it is forbidden to write to (i.e., stuff the operating system can only access)"
General advice seemed to be to remove folders from the mod one by one until the game no longer crashes.
[error.log](https://github.com/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2/files/9794294/error.log)
|
non_test
|
patreon characters branch crashes do not remove pre existing lines your mod version is what expansions do you have installed garments of the holy roman empire sep northern lords mar are you using any submods mods if so which none please explain your issue in as much detail as possible game crashes repeatedly in loading screen the following reports in exception txt and error log stood out to me full error log download available at the bottom click to expand exception txt unhandled exception exception access violation at address stack trace exe function name not available exe function name not available exe function name not available exe function name not available exe function name not available exe function name not available exe function name not available exe sdl stoptextinput exe function name not available exe function name not available exe function name not available exe function name not available exe sdl stoptextinput exe sdl stoptextinput exe sdl stoptextinput exe sdl stoptextinput exe sdl stoptextinput exe sdl stoptextinput exe sdl dynapi entry dll basethreadinitthunk ntdll dll rtluserthreadstart ntdll dll rtluserthreadstart error log instance has not been created filename not available sdl stoptextinput filename not available function name not available filename not available function name not available filename not available function name not available filename not available function name not available filename not available function name not available filename not available function name not available filename not available function name not available filename not available sdl stoptextinput filename not available function name not available filename not available function name not available filename not available function name not available filename not available function name not available filename not available sdl stoptextinput filename not available sdl stoptextinput filename not available sdl stoptextinput filename not available sdl stoptextinput filename not available sdl stoptextinput filename not available sdl stoptextinput filename not available sdl dynapi entry filename not available basethreadinitthunk ntdll filename not available rtluserthreadstart steps to reproduce the issue downgrade the game to and try starting up patreon characters branch upload an attachment below zip of your save or screenshots regarding exception access violation i was able to find the following info from the mod co op that says the game engine is trying to write to memory it is forbidden to write to i e stuff the operating system can only access general advice seemed to be to remove folders from the mod one by one until the game no longer crashes
| 0
|
56,245
| 13,780,008,111
|
IssuesEvent
|
2020-10-08 14:25:45
|
odpi/egeria
|
https://api.github.com/repos/odpi/egeria
|
closed
|
Partial build of any module depending on assembly not possible - requires full build
|
build-improvement
|
In our maven build, a number of modules including fvt & containers are dependent on having a valid assembly.
A full maven build works fine.
However if I for example try to build the data manager fvt only - having previously done a full build - I see:
```
$ mvn install [17:08:06]
[INFO] Scanning for projects...
[INFO]
[INFO] ------------------< org.odpi.egeria:data-manager-fvt >------------------
[INFO] Building Data Manager OMAS FVT 2.4-SNAPSHOT
[INFO] --------------------------------[ jar ]---------------------------------
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2.871 s
[INFO] Finished at: 2020-10-05T17:08:12+01:00
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal on project data-manager-fvt: Could not resolve dependencies for project org.odpi.egeria:data-manager-fvt:jar:2.4-SNAPSHOT: Could not find artifact org.odpi.egeria:open-metadata-assemblies:pom:distribution:2.4-SNAPSHOT -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/DependencyResolutionException
```
I would have expected this dependency to be available in my local ~/.m2 - but it's possible maven insists it is available in the reactor.
This makes quicker builds increasingly difficult. Gradle addresses this issue but currently does not handle execution of FVT., packaging, spring apps, or assembly.
|
1.0
|
Partial build of any module depending on assembly not possible - requires full build - In our maven build, a number of modules including fvt & containers are dependent on having a valid assembly.
A full maven build works fine.
However if I for example try to build the data manager fvt only - having previously done a full build - I see:
```
$ mvn install [17:08:06]
[INFO] Scanning for projects...
[INFO]
[INFO] ------------------< org.odpi.egeria:data-manager-fvt >------------------
[INFO] Building Data Manager OMAS FVT 2.4-SNAPSHOT
[INFO] --------------------------------[ jar ]---------------------------------
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2.871 s
[INFO] Finished at: 2020-10-05T17:08:12+01:00
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal on project data-manager-fvt: Could not resolve dependencies for project org.odpi.egeria:data-manager-fvt:jar:2.4-SNAPSHOT: Could not find artifact org.odpi.egeria:open-metadata-assemblies:pom:distribution:2.4-SNAPSHOT -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/DependencyResolutionException
```
I would have expected this dependency to be available in my local ~/.m2 - but it's possible maven insists it is available in the reactor.
This makes quicker builds increasingly difficult. Gradle addresses this issue but currently does not handle execution of FVT., packaging, spring apps, or assembly.
|
non_test
|
partial build of any module depending on assembly not possible requires full build in our maven build a number of modules including fvt containers are dependent on having a valid assembly a full maven build works fine however if i for example try to build the data manager fvt only having previously done a full build i see mvn install scanning for projects building data manager omas fvt snapshot build failure total time s finished at failed to execute goal on project data manager fvt could not resolve dependencies for project org odpi egeria data manager fvt jar snapshot could not find artifact org odpi egeria open metadata assemblies pom distribution snapshot to see the full stack trace of the errors re run maven with the e switch re run maven using the x switch to enable full debug logging for more information about the errors and possible solutions please read the following articles i would have expected this dependency to be available in my local but it s possible maven insists it is available in the reactor this makes quicker builds increasingly difficult gradle addresses this issue but currently does not handle execution of fvt packaging spring apps or assembly
| 0
|
334,508
| 29,871,497,678
|
IssuesEvent
|
2023-06-20 08:48:06
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Unlimited amount of visual regression tests with `cypress-visual-regression`
|
T-Enhancement A-Testing
|
### Your use case
#### What would you like to do?
Introduce `cypress-visual-regression` https://github.com/cypress-visual-regression/cypress-visual-regression to add unlimited amount of visual regression tests, regardless of the budget for Percy.
#### Why would you like to do it?
Because it is not optimal to limit the amount of tests due to budget limit. Add tests as many as you need.
#### How would you like to achieve it?
### Have you considered any alternatives?
_No response_
### Additional context
I do not know whether `cypress-visual-regression` had been tested by the Element company or not.
|
1.0
|
Unlimited amount of visual regression tests with `cypress-visual-regression` - ### Your use case
#### What would you like to do?
Introduce `cypress-visual-regression` https://github.com/cypress-visual-regression/cypress-visual-regression to add unlimited amount of visual regression tests, regardless of the budget for Percy.
#### Why would you like to do it?
Because it is not optimal to limit the amount of tests due to budget limit. Add tests as many as you need.
#### How would you like to achieve it?
### Have you considered any alternatives?
_No response_
### Additional context
I do not know whether `cypress-visual-regression` had been tested by the Element company or not.
|
test
|
unlimited amount of visual regression tests with cypress visual regression your use case what would you like to do introduce cypress visual regression to add unlimited amount of visual regression tests regardless of the budget for percy why would you like to do it because it is not optimal to limit the amount of tests due to budget limit add tests as many as you need how would you like to achieve it have you considered any alternatives no response additional context i do not know whether cypress visual regression had been tested by the element company or not
| 1
|
58,716
| 11,905,123,663
|
IssuesEvent
|
2020-03-30 18:00:59
|
home-assistant/brands
|
https://api.github.com/repos/home-assistant/brands
|
opened
|
Ecovacs is missing brand images
|
domain-missing has-codeowner
|
## The problem
The Ecovacs integration does not have brand images in
this repository.
We recently started this Brands repository, to create a centralized storage of all brand-related images. These images are used on our website and the Home Assistant frontend.
The following images are missing and would ideally be added:
- `src/ecovacs/icon.png`
- `src/ecovacs/logo.png`
- `src/ecovacs/icon@2x.png`
- `src/ecovacs/logo@2x.png`
For image specifications and requirements, please see [README.md](https://github.com/home-assistant/brands/blob/master/README.md).
## Updating the documentation repository
Our documentation repository already has a logo for this integration, however, it does not meet the image requirements of this new Brands repository.
If adding images to this repository, please open up a PR to the documentation repository as well, removing the `logo: ecovacs.png` line from this file:
<https://github.com/home-assistant/home-assistant.io/blob/current/source/_integrations/ecovacs.markdown>
**Note**: The documentation PR needs to be opened against the `current` branch.
**Note2**: Please leave the actual logo file in the documentation repository. It will be cleaned up differently.
## Additional information
For more information about this repository, read the [README.md](https://github.com/home-assistant/brands/blob/master/README.md) file of this repository. It contains information on how this repository works, and image specification and requirements.
## Codeowner mention
Hi there, @OverloadUT! Mind taking a look at this issue as it is with an integration (ecovacs) you are listed as a [codeowner](https://github.com/home-assistant/core/blob/dev/homeassistant/components/ecovacs/manifest.json) for? Thanks!
Resolving this issue is not limited to codeowners! If you want to help us out, feel free to resolve this issue! Thanks already!
|
1.0
|
Ecovacs is missing brand images -
## The problem
The Ecovacs integration does not have brand images in
this repository.
We recently started this Brands repository, to create a centralized storage of all brand-related images. These images are used on our website and the Home Assistant frontend.
The following images are missing and would ideally be added:
- `src/ecovacs/icon.png`
- `src/ecovacs/logo.png`
- `src/ecovacs/icon@2x.png`
- `src/ecovacs/logo@2x.png`
For image specifications and requirements, please see [README.md](https://github.com/home-assistant/brands/blob/master/README.md).
## Updating the documentation repository
Our documentation repository already has a logo for this integration, however, it does not meet the image requirements of this new Brands repository.
If adding images to this repository, please open up a PR to the documentation repository as well, removing the `logo: ecovacs.png` line from this file:
<https://github.com/home-assistant/home-assistant.io/blob/current/source/_integrations/ecovacs.markdown>
**Note**: The documentation PR needs to be opened against the `current` branch.
**Note2**: Please leave the actual logo file in the documentation repository. It will be cleaned up differently.
## Additional information
For more information about this repository, read the [README.md](https://github.com/home-assistant/brands/blob/master/README.md) file of this repository. It contains information on how this repository works, and image specification and requirements.
## Codeowner mention
Hi there, @OverloadUT! Mind taking a look at this issue as it is with an integration (ecovacs) you are listed as a [codeowner](https://github.com/home-assistant/core/blob/dev/homeassistant/components/ecovacs/manifest.json) for? Thanks!
Resolving this issue is not limited to codeowners! If you want to help us out, feel free to resolve this issue! Thanks already!
|
non_test
|
ecovacs is missing brand images the problem the ecovacs integration does not have brand images in this repository we recently started this brands repository to create a centralized storage of all brand related images these images are used on our website and the home assistant frontend the following images are missing and would ideally be added src ecovacs icon png src ecovacs logo png src ecovacs icon png src ecovacs logo png for image specifications and requirements please see updating the documentation repository our documentation repository already has a logo for this integration however it does not meet the image requirements of this new brands repository if adding images to this repository please open up a pr to the documentation repository as well removing the logo ecovacs png line from this file note the documentation pr needs to be opened against the current branch please leave the actual logo file in the documentation repository it will be cleaned up differently additional information for more information about this repository read the file of this repository it contains information on how this repository works and image specification and requirements codeowner mention hi there overloadut mind taking a look at this issue as it is with an integration ecovacs you are listed as a for thanks resolving this issue is not limited to codeowners if you want to help us out feel free to resolve this issue thanks already
| 0
|
135,305
| 30,281,989,512
|
IssuesEvent
|
2023-07-08 07:08:28
|
EddieHubCommunity/LinkFree
|
https://api.github.com/repos/EddieHubCommunity/LinkFree
|
closed
|
[FEATURE] News letter form should below the text in mobile view
|
⭐ goal: addition 💻 aspect: code 🏁 status: ready for dev good first issue 🔢 points: 1
|
### Description
As the suggestion given by @amandamartin-dev in PR number #7681. The form should be below the text for better context.
I am giving screenshot below
### Screenshots

### Additional information
_No response_
|
1.0
|
[FEATURE] News letter form should below the text in mobile view - ### Description
As the suggestion given by @amandamartin-dev in PR number #7681. The form should be below the text for better context.
I am giving screenshot below
### Screenshots

### Additional information
_No response_
|
non_test
|
news letter form should below the text in mobile view description as the suggestion given by amandamartin dev in pr number the form should be below the text for better context i am giving screenshot below screenshots additional information no response
| 0
|
139,584
| 31,710,682,104
|
IssuesEvent
|
2023-09-09 08:16:32
|
SFDO-Community/Salesforce-Indicators
|
https://api.github.com/repos/SFDO-Community/Salesforce-Indicators
|
closed
|
Packaging
|
code future c-indicators
|
The Component needs to be packaged so it can be installed in orgs easily.
Preferably using the same tools as other Open Source Commons uses - MetaDeploy (similar to [DLRS](https://install.salesforce.org/products/dlrs/latest)), however that is now a Discussion item because it is bigger than getting it done in this sprint. See #55
|
1.0
|
Packaging - The Component needs to be packaged so it can be installed in orgs easily.
Preferably using the same tools as other Open Source Commons uses - MetaDeploy (similar to [DLRS](https://install.salesforce.org/products/dlrs/latest)), however that is now a Discussion item because it is bigger than getting it done in this sprint. See #55
|
non_test
|
packaging the component needs to be packaged so it can be installed in orgs easily preferably using the same tools as other open source commons uses metadeploy similar to however that is now a discussion item because it is bigger than getting it done in this sprint see
| 0
|
63,363
| 6,844,454,309
|
IssuesEvent
|
2017-11-13 01:31:48
|
wesnoth/wesnoth
|
https://api.github.com/repos/wesnoth/wesnoth
|
closed
|
Small wall-castle transition issue
|
bug Graphics ready for testing
|
See below:

@doofus-01 @ln-zookeeper perhaps either of you know how this could be fixed? The brazier makes no difference, BTW.
|
1.0
|
Small wall-castle transition issue - See below:

@doofus-01 @ln-zookeeper perhaps either of you know how this could be fixed? The brazier makes no difference, BTW.
|
test
|
small wall castle transition issue see below doofus ln zookeeper perhaps either of you know how this could be fixed the brazier makes no difference btw
| 1
|
85,140
| 3,687,019,203
|
IssuesEvent
|
2016-02-25 05:26:37
|
samolds/principia
|
https://api.github.com/repos/samolds/principia
|
closed
|
Add list of all existing bodies to the property window
|
enhancement priority-medium simulator
|
Add a new tab with a list of all of the bodies in the simulator.
|
1.0
|
Add list of all existing bodies to the property window - Add a new tab with a list of all of the bodies in the simulator.
|
non_test
|
add list of all existing bodies to the property window add a new tab with a list of all of the bodies in the simulator
| 0
|
88,170
| 17,481,542,624
|
IssuesEvent
|
2021-08-09 03:50:24
|
fastjengine/FastJ
|
https://api.github.com/repos/fastjengine/FastJ
|
closed
|
Add comment explaining use of cascading switch
|
documentation source code
|
> `case Paused: { audioEventListener.audioResumeAction.accept(audioEvent); }`
>
> I was wondering if you might want to add a `break` statement here?
> Actually, I've been meaning to include a comment explaining the lack of a break! Good catch.
> The audio event system includes audio events for when audio is paused, and when the audio's playing stream is stopped (either temporarily or permanently). As of right now, I've intentionally had both trigger because it made sense at the time -- when an audio stream stops, ot could be either paused or completely stopped. If it is paused, then an extra event should be created for that. Feel free to dispute this though -- I've been considering adding those break statements for a while.
_Originally posted by @lucasstarsz in https://github.com/fastjengine/FastJ/issues/71#issuecomment-893900855_
At some point I need to add a comment explaining the usage that I talk about here.
|
1.0
|
Add comment explaining use of cascading switch - > `case Paused: { audioEventListener.audioResumeAction.accept(audioEvent); }`
>
> I was wondering if you might want to add a `break` statement here?
> Actually, I've been meaning to include a comment explaining the lack of a break! Good catch.
> The audio event system includes audio events for when audio is paused, and when the audio's playing stream is stopped (either temporarily or permanently). As of right now, I've intentionally had both trigger because it made sense at the time -- when an audio stream stops, ot could be either paused or completely stopped. If it is paused, then an extra event should be created for that. Feel free to dispute this though -- I've been considering adding those break statements for a while.
_Originally posted by @lucasstarsz in https://github.com/fastjengine/FastJ/issues/71#issuecomment-893900855_
At some point I need to add a comment explaining the usage that I talk about here.
|
non_test
|
add comment explaining use of cascading switch case paused audioeventlistener audioresumeaction accept audioevent i was wondering if you might want to add a break statement here actually i ve been meaning to include a comment explaining the lack of a break good catch the audio event system includes audio events for when audio is paused and when the audio s playing stream is stopped either temporarily or permanently as of right now i ve intentionally had both trigger because it made sense at the time when an audio stream stops ot could be either paused or completely stopped if it is paused then an extra event should be created for that feel free to dispute this though i ve been considering adding those break statements for a while originally posted by lucasstarsz in at some point i need to add a comment explaining the usage that i talk about here
| 0
|
204,658
| 15,946,787,890
|
IssuesEvent
|
2021-04-15 01:56:03
|
Covidioten/UI
|
https://api.github.com/repos/Covidioten/UI
|
closed
|
[3h] Deployment der Analyseergebnisdatei
|
documentation question
|
Wie soll die Datei ausgeliefert werden? Weiterhin statisches deployment via gh-pages, containerized, tomcat?
|
1.0
|
[3h] Deployment der Analyseergebnisdatei - Wie soll die Datei ausgeliefert werden? Weiterhin statisches deployment via gh-pages, containerized, tomcat?
|
non_test
|
deployment der analyseergebnisdatei wie soll die datei ausgeliefert werden weiterhin statisches deployment via gh pages containerized tomcat
| 0
|
54,077
| 6,363,102,891
|
IssuesEvent
|
2017-07-31 16:16:54
|
intel-hpdd/intel-manager-for-lustre
|
https://api.github.com/repos/intel-hpdd/intel-manager-for-lustre
|
opened
|
timeout in set_value
|
failing tests
|
In a [test run](http://jenkins.lotus.hpdd.lab.intel.com/job/integration-tests-shared-storage-configuration/810/arch=x86_64,distro=el7):
```
test_hyphenated_named_filesystem (tests.integration.shared_storage_configuration.test_hyphenated_filesystem_name.TestHyphenatedFilesystemName) ... FAIL
FAIL
Traceback (most recent call last):
File "/usr/share/chroma-manager/tests/integration/shared_storage_configuration/test_hyphenated_filesystem_name.py", line 15, in test_hyphenated_named_filesystem
self.set_state(filesystem['resource_uri'], 'available')
File "/usr/share/chroma-manager/tests/integration/core/api_testcase_with_test_reset.py", line 495, in set_state
msg)
File "/usr/share/chroma-manager/tests/integration/core/api_testcase_with_test_reset.py", line 479, in set_value
command = self.wait_for_command(self.chroma_manager, command_id, verify_successful=(verify_successful != self.VERIFY_SUCCESS_NO), msg=msg, timeout=LONG_TEST_TIMEOUT)
File "/usr/share/chroma-manager/tests/integration/core/api_testcase_with_test_reset.py", line 336, in wait_for_command
self.assertFalse(command['errored'] or command['cancelled'], command)
AssertionError: {u'jobs': [u'/api/job/84/', u'/api/job/85/', u'/api/job/86/', u'/api/job/87/'], u'complete': True, u'created_at': u'2017-07-30T03:45:55.758492', u'message': u'Start file system test-fs', u'cancelled': True, u'errored': False, u'resource_uri': u'/api/command/24/', u'id': 24, u'logs': u'\n'}
-------------------- >> begin captured stdout << ---------------------
{u'status': u'configured-ha', u'kind': u'ZfsPool', u'storage_resource': u'/api/storage_resource/47/', u'volume_nodes': [{u'use': False, u'volume_id': 1, u'primary': False, u'host': u'/api/host/2/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk1', u'host_id': 2, u'resource_uri': u'/api/volume_node/3/', u'id': 3, u'host_label': u'lotus-53vm6.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 1, u'primary': False, u'host': u'/api/host/4/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk1', u'host_id': 4, u'resource_uri': u'/api/volume_node/8/', u'id': 8, u'host_label': u'lotus-53vm8.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 1, u'primary': False, u'host': u'/api/host/1/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk1', u'host_id': 1, u'resource_uri': u'/api/volume_node/13/', u'id': 13, u'host_label': u'lotus-53vm5.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 1, u'primary': True, u'host': u'/api/host/3/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk1', u'host_id': 3, u'resource_uri': u'/api/volume_node/18/', u'id': 18, u'host_label': u'lotus-53vm7.lotus.hpdd.lab.intel.com'}], u'filesystem_type': None, u'label': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk1', u'usable_for_lustre': True, u'resource_uri': u'/api/volume/1/', u'id': 1, u'size': u'10672993730'}
{u'status': u'configured-ha', u'kind': u'SCSI device', u'storage_resource': u'/api/storage_resource/26/', u'volume_nodes': [{u'use': True, u'volume_id': 2, u'primary': True, u'host': u'/api/host/2/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk4', u'host_id': 2, u'resource_uri': u'/api/volume_node/2/', u'id': 2, u'host_label': u'lotus-53vm6.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 2, u'primary': False, u'host': u'/api/host/4/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk4', u'host_id': 4, u'resource_uri': u'/api/volume_node/7/', u'id': 7, u'host_label': u'lotus-53vm8.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 2, u'primary': False, u'host': u'/api/host/1/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk4', u'host_id': 1, u'resource_uri': u'/api/volume_node/12/', u'id': 12, u'host_label': u'lotus-53vm5.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 2, u'primary': False, u'host': u'/api/host/3/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk4', u'host_id': 3, u'resource_uri': u'/api/volume_node/17/', u'id': 17, u'host_label': u'lotus-53vm7.lotus.hpdd.lab.intel.com'}], u'filesystem_type': None, u'label': u'disk4', u'usable_for_lustre': True, u'resource_uri': u'/api/volume/2/', u'id': 2, u'size': u'10737418240'}
{u'status': u'configured-ha', u'kind': u'ZfsPool', u'storage_resource': u'/api/storage_resource/49/', u'volume_nodes': [{u'use': True, u'volume_id': 3, u'primary': False, u'host': u'/api/host/2/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk5', u'host_id': 2, u'resource_uri': u'/api/volume_node/4/', u'id': 4, u'host_label': u'lotus-53vm6.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 3, u'primary': False, u'host': u'/api/host/4/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk5', u'host_id': 4, u'resource_uri': u'/api/volume_node/9/', u'id': 9, u'host_label': u'lotus-53vm8.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 3, u'primary': True, u'host': u'/api/host/1/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk5', u'host_id': 1, u'resource_uri': u'/api/volume_node/14/', u'id': 14, u'host_label': u'lotus-53vm5.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 3, u'primary': False, u'host': u'/api/host/3/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk5', u'host_id': 3, u'resource_uri': u'/api/volume_node/19/', u'id': 19, u'host_label': u'lotus-53vm7.lotus.hpdd.lab.intel.com'}], u'filesystem_type': None, u'label': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk5', u'usable_for_lustre': True, u'resource_uri': u'/api/volume/3/', u'id': 3, u'size': u'10672993730'}
{u'status': u'configured-ha', u'kind': u'ZfsPool', u'storage_resource': u'/api/storage_resource/51/', u'volume_nodes': [{u'use': False, u'volume_id': 4, u'primary': False, u'host': u'/api/host/2/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk3', u'host_id': 2, u'resource_uri': u'/api/volume_node/5/', u'id': 5, u'host_label': u'lotus-53vm6.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 4, u'primary': True, u'host': u'/api/host/4/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk3', u'host_id': 4, u'resource_uri': u'/api/volume_node/10/', u'id': 10, u'host_label': u'lotus-53vm8.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 4, u'primary': False, u'host': u'/api/host/1/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk3', u'host_id': 1, u'resource_uri': u'/api/volume_node/15/', u'id': 15, u'host_label': u'lotus-53vm5.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 4, u'primary': False, u'host': u'/api/host/3/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk3', u'host_id': 3, u'resource_uri': u'/api/volume_node/20/', u'id': 20, u'host_label': u'lotus-53vm7.lotus.hpdd.lab.intel.com'}], u'filesystem_type': None, u'label': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk3', u'usable_for_lustre': True, u'resource_uri': u'/api/volume/4/', u'id': 4, u'size': u'10672993730'}
{u'status': u'configured-ha', u'kind': u'SCSI device', u'storage_resource': u'/api/storage_resource/24/', u'volume_nodes': [{u'use': True, u'volume_id': 5, u'primary': False, u'host': u'/api/host/2/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk2', u'host_id': 2, u'resource_uri': u'/api/volume_node/1/', u'id': 1, u'host_label': u'lotus-53vm6.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 5, u'primary': False, u'host': u'/api/host/4/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk2', u'host_id': 4, u'resource_uri': u'/api/volume_node/6/', u'id': 6, u'host_label': u'lotus-53vm8.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 5, u'primary': True, u'host': u'/api/host/1/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk2', u'host_id': 1, u'resource_uri': u'/api/volume_node/11/', u'id': 11, u'host_label': u'lotus-53vm5.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 5, u'primary': False, u'host': u'/api/host/3/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk2', u'host_id': 3, u'resource_uri': u'/api/volume_node/16/', u'id': 16, u'host_label': u'lotus-53vm7.lotus.hpdd.lab.intel.com'}], u'filesystem_type': None, u'label': u'disk2', u'usable_for_lustre': True, u'resource_uri': u'/api/volume/5/', u'id': 5, u'size': u'10737418240'}
{u'status': u'configured-ha', u'kind': u'ZfsPool', u'storage_resource': u'/api/storage_resource/47/', u'volume_nodes': [{u'use': False, u'volume_id': 1, u'primary': False, u'host': u'/api/host/2/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk1', u'host_id': 2, u'resource_uri': u'/api/volume_node/3/', u'id': 3, u'host_label': u'lotus-53vm6.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 1, u'primary': False, u'host': u'/api/host/4/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk1', u'host_id': 4, u'resource_uri': u'/api/volume_node/8/', u'id': 8, u'host_label': u'lotus-53vm8.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 1, u'primary': False, u'host': u'/api/host/1/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk1', u'host_id': 1, u'resource_uri': u'/api/volume_node/13/', u'id': 13, u'host_label': u'lotus-53vm5.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 1, u'primary': True, u'host': u'/api/host/3/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk1', u'host_id': 3, u'resource_uri': u'/api/volume_node/18/', u'id': 18, u'host_label': u'lotus-53vm7.lotus.hpdd.lab.intel.com'}], u'filesystem_type': None, u'label': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk1', u'usable_for_lustre': True, u'resource_uri': u'/api/volume/1/', u'id': 1, u'size': u'10672993730'}
{u'status': u'configured-ha', u'kind': u'SCSI device', u'storage_resource': u'/api/storage_resource/26/', u'volume_nodes': [{u'use': True, u'volume_id': 2, u'primary': True, u'host': u'/api/host/2/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk4', u'host_id': 2, u'resource_uri': u'/api/volume_node/2/', u'id': 2, u'host_label': u'lotus-53vm6.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 2, u'primary': False, u'host': u'/api/host/4/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk4', u'host_id': 4, u'resource_uri': u'/api/volume_node/7/', u'id': 7, u'host_label': u'lotus-53vm8.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 2, u'primary': False, u'host': u'/api/host/1/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk4', u'host_id': 1, u'resource_uri': u'/api/volume_node/12/', u'id': 12, u'host_label': u'lotus-53vm5.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 2, u'primary': False, u'host': u'/api/host/3/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk4', u'host_id': 3, u'resource_uri': u'/api/volume_node/17/', u'id': 17, u'host_label': u'lotus-53vm7.lotus.hpdd.lab.intel.com'}], u'filesystem_type': None, u'label': u'disk4', u'usable_for_lustre': True, u'resource_uri': u'/api/volume/2/', u'id': 2, u'size': u'10737418240'}
{u'status': u'configured-ha', u'kind': u'ZfsPool', u'storage_resource': u'/api/storage_resource/49/', u'volume_nodes': [{u'use': True, u'volume_id': 3, u'primary': False, u'host': u'/api/host/2/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk5', u'host_id': 2, u'resource_uri': u'/api/volume_node/4/', u'id': 4, u'host_label': u'lotus-53vm6.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 3, u'primary': False, u'host': u'/api/host/4/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk5', u'host_id': 4, u'resource_uri': u'/api/volume_node/9/', u'id': 9, u'host_label': u'lotus-53vm8.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 3, u'primary': True, u'host': u'/api/host/1/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk5', u'host_id': 1, u'resource_uri': u'/api/volume_node/14/', u'id': 14, u'host_label': u'lotus-53vm5.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 3, u'primary': False, u'host': u'/api/host/3/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk5', u'host_id': 3, u'resource_uri': u'/api/volume_node/19/', u'id': 19, u'host_label': u'lotus-53vm7.lotus.hpdd.lab.intel.com'}], u'filesystem_type': None, u'label': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk5', u'usable_for_lustre': True, u'resource_uri': u'/api/volume/3/', u'id': 3, u'size': u'10672993730'}
{u'status': u'configured-ha', u'kind': u'ZfsPool', u'storage_resource': u'/api/storage_resource/51/', u'volume_nodes': [{u'use': False, u'volume_id': 4, u'primary': False, u'host': u'/api/host/2/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk3', u'host_id': 2, u'resource_uri': u'/api/volume_node/5/', u'id': 5, u'host_label': u'lotus-53vm6.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 4, u'primary': True, u'host': u'/api/host/4/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk3', u'host_id': 4, u'resource_uri': u'/api/volume_node/10/', u'id': 10, u'host_label': u'lotus-53vm8.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 4, u'primary': False, u'host': u'/api/host/1/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk3', u'host_id': 1, u'resource_uri': u'/api/volume_node/15/', u'id': 15, u'host_label': u'lotus-53vm5.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 4, u'primary': False, u'host': u'/api/host/3/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk3', u'host_id': 3, u'resource_uri': u'/api/volume_node/20/', u'id': 20, u'host_label': u'lotus-53vm7.lotus.hpdd.lab.intel.com'}], u'filesystem_type': None, u'label': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk3', u'usable_for_lustre': True, u'resource_uri': u'/api/volume/4/', u'id': 4, u'size': u'10672993730'}
{u'status': u'configured-ha', u'kind': u'SCSI device', u'storage_resource': u'/api/storage_resource/24/', u'volume_nodes': [{u'use': True, u'volume_id': 5, u'primary': False, u'host': u'/api/host/2/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk2', u'host_id': 2, u'resource_uri': u'/api/volume_node/1/', u'id': 1, u'host_label': u'lotus-53vm6.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 5, u'primary': False, u'host': u'/api/host/4/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk2', u'host_id': 4, u'resource_uri': u'/api/volume_node/6/', u'id': 6, u'host_label': u'lotus-53vm8.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 5, u'primary': True, u'host': u'/api/host/1/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk2', u'host_id': 1, u'resource_uri': u'/api/volume_node/11/', u'id': 11, u'host_label': u'lotus-53vm5.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 5, u'primary': False, u'host': u'/api/host/3/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk2', u'host_id': 3, u'resource_uri': u'/api/volume_node/16/', u'id': 16, u'host_label': u'lotus-53vm7.lotus.hpdd.lab.intel.com'}], u'filesystem_type': None, u'label': u'disk2', u'usable_for_lustre': True, u'resource_uri': u'/api/volume/5/', u'id': 5, u'size': u'10737418240'}
COMMAND 24: FAILED
-----------------------------------------------------------
{u'jobs': [u'/api/job/84/', u'/api/job/85/', u'/api/job/86/', u'/api/job/87/'], u'complete': True, u'created_at': u'2017-07-30T03:45:55.758492', u'message': u'Start file system test-fs', u'cancelled': True, u'errored': False, u'resource_uri': u'/api/command/24/', u'id': 24, u'logs': u'\n'}
--------------------- >> end captured stdout << ----------------------
```
|
1.0
|
timeout in set_value - In a [test run](http://jenkins.lotus.hpdd.lab.intel.com/job/integration-tests-shared-storage-configuration/810/arch=x86_64,distro=el7):
```
test_hyphenated_named_filesystem (tests.integration.shared_storage_configuration.test_hyphenated_filesystem_name.TestHyphenatedFilesystemName) ... FAIL
FAIL
Traceback (most recent call last):
File "/usr/share/chroma-manager/tests/integration/shared_storage_configuration/test_hyphenated_filesystem_name.py", line 15, in test_hyphenated_named_filesystem
self.set_state(filesystem['resource_uri'], 'available')
File "/usr/share/chroma-manager/tests/integration/core/api_testcase_with_test_reset.py", line 495, in set_state
msg)
File "/usr/share/chroma-manager/tests/integration/core/api_testcase_with_test_reset.py", line 479, in set_value
command = self.wait_for_command(self.chroma_manager, command_id, verify_successful=(verify_successful != self.VERIFY_SUCCESS_NO), msg=msg, timeout=LONG_TEST_TIMEOUT)
File "/usr/share/chroma-manager/tests/integration/core/api_testcase_with_test_reset.py", line 336, in wait_for_command
self.assertFalse(command['errored'] or command['cancelled'], command)
AssertionError: {u'jobs': [u'/api/job/84/', u'/api/job/85/', u'/api/job/86/', u'/api/job/87/'], u'complete': True, u'created_at': u'2017-07-30T03:45:55.758492', u'message': u'Start file system test-fs', u'cancelled': True, u'errored': False, u'resource_uri': u'/api/command/24/', u'id': 24, u'logs': u'\n'}
-------------------- >> begin captured stdout << ---------------------
{u'status': u'configured-ha', u'kind': u'ZfsPool', u'storage_resource': u'/api/storage_resource/47/', u'volume_nodes': [{u'use': False, u'volume_id': 1, u'primary': False, u'host': u'/api/host/2/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk1', u'host_id': 2, u'resource_uri': u'/api/volume_node/3/', u'id': 3, u'host_label': u'lotus-53vm6.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 1, u'primary': False, u'host': u'/api/host/4/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk1', u'host_id': 4, u'resource_uri': u'/api/volume_node/8/', u'id': 8, u'host_label': u'lotus-53vm8.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 1, u'primary': False, u'host': u'/api/host/1/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk1', u'host_id': 1, u'resource_uri': u'/api/volume_node/13/', u'id': 13, u'host_label': u'lotus-53vm5.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 1, u'primary': True, u'host': u'/api/host/3/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk1', u'host_id': 3, u'resource_uri': u'/api/volume_node/18/', u'id': 18, u'host_label': u'lotus-53vm7.lotus.hpdd.lab.intel.com'}], u'filesystem_type': None, u'label': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk1', u'usable_for_lustre': True, u'resource_uri': u'/api/volume/1/', u'id': 1, u'size': u'10672993730'}
{u'status': u'configured-ha', u'kind': u'SCSI device', u'storage_resource': u'/api/storage_resource/26/', u'volume_nodes': [{u'use': True, u'volume_id': 2, u'primary': True, u'host': u'/api/host/2/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk4', u'host_id': 2, u'resource_uri': u'/api/volume_node/2/', u'id': 2, u'host_label': u'lotus-53vm6.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 2, u'primary': False, u'host': u'/api/host/4/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk4', u'host_id': 4, u'resource_uri': u'/api/volume_node/7/', u'id': 7, u'host_label': u'lotus-53vm8.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 2, u'primary': False, u'host': u'/api/host/1/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk4', u'host_id': 1, u'resource_uri': u'/api/volume_node/12/', u'id': 12, u'host_label': u'lotus-53vm5.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 2, u'primary': False, u'host': u'/api/host/3/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk4', u'host_id': 3, u'resource_uri': u'/api/volume_node/17/', u'id': 17, u'host_label': u'lotus-53vm7.lotus.hpdd.lab.intel.com'}], u'filesystem_type': None, u'label': u'disk4', u'usable_for_lustre': True, u'resource_uri': u'/api/volume/2/', u'id': 2, u'size': u'10737418240'}
{u'status': u'configured-ha', u'kind': u'ZfsPool', u'storage_resource': u'/api/storage_resource/49/', u'volume_nodes': [{u'use': True, u'volume_id': 3, u'primary': False, u'host': u'/api/host/2/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk5', u'host_id': 2, u'resource_uri': u'/api/volume_node/4/', u'id': 4, u'host_label': u'lotus-53vm6.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 3, u'primary': False, u'host': u'/api/host/4/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk5', u'host_id': 4, u'resource_uri': u'/api/volume_node/9/', u'id': 9, u'host_label': u'lotus-53vm8.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 3, u'primary': True, u'host': u'/api/host/1/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk5', u'host_id': 1, u'resource_uri': u'/api/volume_node/14/', u'id': 14, u'host_label': u'lotus-53vm5.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 3, u'primary': False, u'host': u'/api/host/3/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk5', u'host_id': 3, u'resource_uri': u'/api/volume_node/19/', u'id': 19, u'host_label': u'lotus-53vm7.lotus.hpdd.lab.intel.com'}], u'filesystem_type': None, u'label': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk5', u'usable_for_lustre': True, u'resource_uri': u'/api/volume/3/', u'id': 3, u'size': u'10672993730'}
{u'status': u'configured-ha', u'kind': u'ZfsPool', u'storage_resource': u'/api/storage_resource/51/', u'volume_nodes': [{u'use': False, u'volume_id': 4, u'primary': False, u'host': u'/api/host/2/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk3', u'host_id': 2, u'resource_uri': u'/api/volume_node/5/', u'id': 5, u'host_label': u'lotus-53vm6.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 4, u'primary': True, u'host': u'/api/host/4/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk3', u'host_id': 4, u'resource_uri': u'/api/volume_node/10/', u'id': 10, u'host_label': u'lotus-53vm8.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 4, u'primary': False, u'host': u'/api/host/1/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk3', u'host_id': 1, u'resource_uri': u'/api/volume_node/15/', u'id': 15, u'host_label': u'lotus-53vm5.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 4, u'primary': False, u'host': u'/api/host/3/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk3', u'host_id': 3, u'resource_uri': u'/api/volume_node/20/', u'id': 20, u'host_label': u'lotus-53vm7.lotus.hpdd.lab.intel.com'}], u'filesystem_type': None, u'label': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk3', u'usable_for_lustre': True, u'resource_uri': u'/api/volume/4/', u'id': 4, u'size': u'10672993730'}
{u'status': u'configured-ha', u'kind': u'SCSI device', u'storage_resource': u'/api/storage_resource/24/', u'volume_nodes': [{u'use': True, u'volume_id': 5, u'primary': False, u'host': u'/api/host/2/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk2', u'host_id': 2, u'resource_uri': u'/api/volume_node/1/', u'id': 1, u'host_label': u'lotus-53vm6.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 5, u'primary': False, u'host': u'/api/host/4/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk2', u'host_id': 4, u'resource_uri': u'/api/volume_node/6/', u'id': 6, u'host_label': u'lotus-53vm8.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 5, u'primary': True, u'host': u'/api/host/1/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk2', u'host_id': 1, u'resource_uri': u'/api/volume_node/11/', u'id': 11, u'host_label': u'lotus-53vm5.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 5, u'primary': False, u'host': u'/api/host/3/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk2', u'host_id': 3, u'resource_uri': u'/api/volume_node/16/', u'id': 16, u'host_label': u'lotus-53vm7.lotus.hpdd.lab.intel.com'}], u'filesystem_type': None, u'label': u'disk2', u'usable_for_lustre': True, u'resource_uri': u'/api/volume/5/', u'id': 5, u'size': u'10737418240'}
{u'status': u'configured-ha', u'kind': u'ZfsPool', u'storage_resource': u'/api/storage_resource/47/', u'volume_nodes': [{u'use': False, u'volume_id': 1, u'primary': False, u'host': u'/api/host/2/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk1', u'host_id': 2, u'resource_uri': u'/api/volume_node/3/', u'id': 3, u'host_label': u'lotus-53vm6.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 1, u'primary': False, u'host': u'/api/host/4/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk1', u'host_id': 4, u'resource_uri': u'/api/volume_node/8/', u'id': 8, u'host_label': u'lotus-53vm8.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 1, u'primary': False, u'host': u'/api/host/1/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk1', u'host_id': 1, u'resource_uri': u'/api/volume_node/13/', u'id': 13, u'host_label': u'lotus-53vm5.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 1, u'primary': True, u'host': u'/api/host/3/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk1', u'host_id': 3, u'resource_uri': u'/api/volume_node/18/', u'id': 18, u'host_label': u'lotus-53vm7.lotus.hpdd.lab.intel.com'}], u'filesystem_type': None, u'label': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk1', u'usable_for_lustre': True, u'resource_uri': u'/api/volume/1/', u'id': 1, u'size': u'10672993730'}
{u'status': u'configured-ha', u'kind': u'SCSI device', u'storage_resource': u'/api/storage_resource/26/', u'volume_nodes': [{u'use': True, u'volume_id': 2, u'primary': True, u'host': u'/api/host/2/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk4', u'host_id': 2, u'resource_uri': u'/api/volume_node/2/', u'id': 2, u'host_label': u'lotus-53vm6.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 2, u'primary': False, u'host': u'/api/host/4/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk4', u'host_id': 4, u'resource_uri': u'/api/volume_node/7/', u'id': 7, u'host_label': u'lotus-53vm8.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 2, u'primary': False, u'host': u'/api/host/1/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk4', u'host_id': 1, u'resource_uri': u'/api/volume_node/12/', u'id': 12, u'host_label': u'lotus-53vm5.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 2, u'primary': False, u'host': u'/api/host/3/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk4', u'host_id': 3, u'resource_uri': u'/api/volume_node/17/', u'id': 17, u'host_label': u'lotus-53vm7.lotus.hpdd.lab.intel.com'}], u'filesystem_type': None, u'label': u'disk4', u'usable_for_lustre': True, u'resource_uri': u'/api/volume/2/', u'id': 2, u'size': u'10737418240'}
{u'status': u'configured-ha', u'kind': u'ZfsPool', u'storage_resource': u'/api/storage_resource/49/', u'volume_nodes': [{u'use': True, u'volume_id': 3, u'primary': False, u'host': u'/api/host/2/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk5', u'host_id': 2, u'resource_uri': u'/api/volume_node/4/', u'id': 4, u'host_label': u'lotus-53vm6.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 3, u'primary': False, u'host': u'/api/host/4/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk5', u'host_id': 4, u'resource_uri': u'/api/volume_node/9/', u'id': 9, u'host_label': u'lotus-53vm8.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 3, u'primary': True, u'host': u'/api/host/1/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk5', u'host_id': 1, u'resource_uri': u'/api/volume_node/14/', u'id': 14, u'host_label': u'lotus-53vm5.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 3, u'primary': False, u'host': u'/api/host/3/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk5', u'host_id': 3, u'resource_uri': u'/api/volume_node/19/', u'id': 19, u'host_label': u'lotus-53vm7.lotus.hpdd.lab.intel.com'}], u'filesystem_type': None, u'label': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk5', u'usable_for_lustre': True, u'resource_uri': u'/api/volume/3/', u'id': 3, u'size': u'10672993730'}
{u'status': u'configured-ha', u'kind': u'ZfsPool', u'storage_resource': u'/api/storage_resource/51/', u'volume_nodes': [{u'use': False, u'volume_id': 4, u'primary': False, u'host': u'/api/host/2/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk3', u'host_id': 2, u'resource_uri': u'/api/volume_node/5/', u'id': 5, u'host_label': u'lotus-53vm6.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 4, u'primary': True, u'host': u'/api/host/4/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk3', u'host_id': 4, u'resource_uri': u'/api/volume_node/10/', u'id': 10, u'host_label': u'lotus-53vm8.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 4, u'primary': False, u'host': u'/api/host/1/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk3', u'host_id': 1, u'resource_uri': u'/api/volume_node/15/', u'id': 15, u'host_label': u'lotus-53vm5.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 4, u'primary': False, u'host': u'/api/host/3/', u'path': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk3', u'host_id': 3, u'resource_uri': u'/api/volume_node/20/', u'id': 20, u'host_label': u'lotus-53vm7.lotus.hpdd.lab.intel.com'}], u'filesystem_type': None, u'label': u'zfs_pool_scsi0QEMU_QEMU_HARDDISK_disk3', u'usable_for_lustre': True, u'resource_uri': u'/api/volume/4/', u'id': 4, u'size': u'10672993730'}
{u'status': u'configured-ha', u'kind': u'SCSI device', u'storage_resource': u'/api/storage_resource/24/', u'volume_nodes': [{u'use': True, u'volume_id': 5, u'primary': False, u'host': u'/api/host/2/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk2', u'host_id': 2, u'resource_uri': u'/api/volume_node/1/', u'id': 1, u'host_label': u'lotus-53vm6.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 5, u'primary': False, u'host': u'/api/host/4/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk2', u'host_id': 4, u'resource_uri': u'/api/volume_node/6/', u'id': 6, u'host_label': u'lotus-53vm8.lotus.hpdd.lab.intel.com'}, {u'use': True, u'volume_id': 5, u'primary': True, u'host': u'/api/host/1/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk2', u'host_id': 1, u'resource_uri': u'/api/volume_node/11/', u'id': 11, u'host_label': u'lotus-53vm5.lotus.hpdd.lab.intel.com'}, {u'use': False, u'volume_id': 5, u'primary': False, u'host': u'/api/host/3/', u'path': u'/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_disk2', u'host_id': 3, u'resource_uri': u'/api/volume_node/16/', u'id': 16, u'host_label': u'lotus-53vm7.lotus.hpdd.lab.intel.com'}], u'filesystem_type': None, u'label': u'disk2', u'usable_for_lustre': True, u'resource_uri': u'/api/volume/5/', u'id': 5, u'size': u'10737418240'}
COMMAND 24: FAILED
-----------------------------------------------------------
{u'jobs': [u'/api/job/84/', u'/api/job/85/', u'/api/job/86/', u'/api/job/87/'], u'complete': True, u'created_at': u'2017-07-30T03:45:55.758492', u'message': u'Start file system test-fs', u'cancelled': True, u'errored': False, u'resource_uri': u'/api/command/24/', u'id': 24, u'logs': u'\n'}
--------------------- >> end captured stdout << ----------------------
```
|
test
|
timeout in set value in a test hyphenated named filesystem tests integration shared storage configuration test hyphenated filesystem name testhyphenatedfilesystemname fail fail traceback most recent call last file usr share chroma manager tests integration shared storage configuration test hyphenated filesystem name py line in test hyphenated named filesystem self set state filesystem available file usr share chroma manager tests integration core api testcase with test reset py line in set state msg file usr share chroma manager tests integration core api testcase with test reset py line in set value command self wait for command self chroma manager command id verify successful verify successful self verify success no msg msg timeout long test timeout file usr share chroma manager tests integration core api testcase with test reset py line in wait for command self assertfalse command or command command assertionerror u jobs u complete true u created at u u message u start file system test fs u cancelled true u errored false u resource uri u api command u id u logs u n begin captured stdout u status u configured ha u kind u zfspool u storage resource u api storage resource u volume nodes u filesystem type none u label u zfs pool qemu harddisk u usable for lustre true u resource uri u api volume u id u size u u status u configured ha u kind u scsi device u storage resource u api storage resource u volume nodes u filesystem type none u label u u usable for lustre true u resource uri u api volume u id u size u u status u configured ha u kind u zfspool u storage resource u api storage resource u volume nodes u filesystem type none u label u zfs pool qemu harddisk u usable for lustre true u resource uri u api volume u id u size u u status u configured ha u kind u zfspool u storage resource u api storage resource u volume nodes u filesystem type none u label u zfs pool qemu harddisk u usable for lustre true u resource uri u api volume u id u size u u status u configured ha u kind u scsi device u storage resource u api storage resource u volume nodes u filesystem type none u label u u usable for lustre true u resource uri u api volume u id u size u u status u configured ha u kind u zfspool u storage resource u api storage resource u volume nodes u filesystem type none u label u zfs pool qemu harddisk u usable for lustre true u resource uri u api volume u id u size u u status u configured ha u kind u scsi device u storage resource u api storage resource u volume nodes u filesystem type none u label u u usable for lustre true u resource uri u api volume u id u size u u status u configured ha u kind u zfspool u storage resource u api storage resource u volume nodes u filesystem type none u label u zfs pool qemu harddisk u usable for lustre true u resource uri u api volume u id u size u u status u configured ha u kind u zfspool u storage resource u api storage resource u volume nodes u filesystem type none u label u zfs pool qemu harddisk u usable for lustre true u resource uri u api volume u id u size u u status u configured ha u kind u scsi device u storage resource u api storage resource u volume nodes u filesystem type none u label u u usable for lustre true u resource uri u api volume u id u size u command failed u jobs u complete true u created at u u message u start file system test fs u cancelled true u errored false u resource uri u api command u id u logs u n end captured stdout
| 1
|
309,728
| 26,676,317,340
|
IssuesEvent
|
2023-01-26 14:31:17
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
Path.GetTempFileName() sometimes fails on WASM due to insufficient randomness
|
arch-wasm area-System.IO disabled-test os-windows in-pr test-failure
|
[Build](https://dev.azure.com/dnceng/public/_build/results?buildId=1934834&view=logs&jobId=a07e672c-d717-5036-7657-9f851d8c04dc&j=a07e672c-d717-5036-7657-9f851d8c04dc&t=95c8462d-76dc-506a-26a4-fefb7abc2e07), seen on multiple PRs with no related changes. Which suggests that this is on `main`. Failures like:
```
[19:23:32] info: Starting: System.CodeDom.Tests.dll
[19:23:39] fail: [FAIL] System.CodeDom.Compiler.Tests.CodeCompilerTests.FromSource_ValidSource_ReturnsExpected(source: "")
[19:23:39] info: System.IO.IOException : File exists
[19:23:39] info: at Interop.ThrowExceptionForIoErrno(ErrorInfo errorInfo, String path, Boolean isDirectory)
[19:23:39] info: at Interop.ThrowIOExceptionForLastError()
[19:23:39] info: at System.IO.Directory.CreateTempSubdirectoryCore(String prefix)
[19:23:39] info: at System.IO.Directory.CreateTempSubdirectory(String prefix)
[19:23:39] info: at System.CodeDom.Compiler.TempFileCollection.GetTempDirectory()
[19:23:39] info: at System.CodeDom.Compiler.TempFileCollection.EnsureTempNameCreated()
[19:23:39] info: at System.CodeDom.Compiler.TempFileCollection.get_BasePath()
[19:23:39] info: at System.CodeDom.Compiler.TempFileCollection.AddExtension(String fileExtension, Boolean keepFile)
[19:23:39] info: at System.CodeDom.Compiler.TempFileCollection.AddExtension(String fileExtension)
[19:23:39] info: at System.CodeDom.Compiler.CodeCompiler.FromSourceBatch(CompilerParameters options, String[] sources)
[19:23:39] info: at System.CodeDom.Compiler.CodeCompiler.FromSource(CompilerParameters options, String source)
[19:23:39] info: at System.CodeDom.Compiler.Tests.CodeCompilerTests.Compiler.FromSourceEntryPoint(CompilerParameters options, String source)
[19:23:39] info: at System.CodeDom.Compiler.Tests.CodeCompilerTests.FromSource_ValidSource_ReturnsExpected(String source)
[19:23:39] info: at System.Reflection.MethodInvoker.InterpretedInvoke(Object obj, Span`1 args, BindingFlags invokeAttr)
```
.. and ..
```
[19:27:13] info: Starting: System.Diagnostics.TextWriterTraceListener.Tests.dll
[19:27:14] fail: [FAIL] System.Diagnostics.TextWriterTraceListenerTests.CtorsDelimiterTests.TestConstructorWithFileNameAndName(testName: "")
[19:27:14] info: System.IO.IOException : File exists
[19:27:14] info: at Interop.ThrowExceptionForIoErrno(ErrorInfo errorInfo, String path, Boolean isDirectory)
[19:27:14] info: at Interop.CheckIo(Int64 result, String path, Boolean isDirectory)
[19:27:14] info: at Interop.CheckIo(IntPtr result, String path, Boolean isDirectory)
[19:27:14] info: at System.IO.Path.GetTempFileName()
[19:27:14] info: at System.Diagnostics.TextWriterTraceListenerTests.CtorsDelimiterTests.TestConstructorWithFileNameAndName(String testName)
[19:27:14] info: at System.Reflection.MethodInvoker.InterpretedInvoke(Object obj, Span`1 args, BindingFlags invokeAttr)
[19:27:14] fail: [FAIL] System.Diagnostics.TextWriterTraceListenerTests.CtorsDelimiterTests.TestConstructorWithFileNameAndName(testName: "><&")
[19:27:14] info: System.IO.IOException : File exists
[19:27:14] info: at Interop.ThrowExceptionForIoErrno(ErrorInfo errorInfo, String path, Boolean isDirectory)
[19:27:14] info: at Interop.CheckIo(Int64 result, String path, Boolean isDirectory)
[19:27:14] info: at Interop.CheckIo(IntPtr result, String path, Boolean isDirectory)
[19:27:14] info: at System.IO.Path.GetTempFileName()
[19:27:14] info: at System.Diagnostics.TextWriterTraceListenerTests.CtorsDelimiterTests.TestConstructorWithFileNameAndName(String testName)
[19:27:14] info: at System.Reflection.MethodInvoker.InterpretedInvoke(Object obj, Span`1 args, BindingFlags invokeAttr)
```
Just for people reading this in the future, the list of failed tests:
```
System.CodeDom.Compiler.Tests.CodeCompilerTests.FromDom_ValidCodeCompileUnit_ReturnsExpected(compilationUnit: CodeCompileUnit { AssemblyCustomAttributes = [], EndDirectives = [], Namespaces = [], ReferencedAssemblies = ["assembly1", "assembly2"], StartDirectives = [], ... })
System.CodeDom.Compiler.Tests.CodeCompilerTests.GetResponseFileCmdArgs_ValidCmdArgs_ReturnsExpected(cmdArgs: "")
System.CodeDom.Compiler.Tests.CodeCompilerTests.CompileAssemblyFromSourceBatch_ValidSources_ReturnsExpected(sources: [null])
System.CodeDom.Compiler.Tests.CodeCompilerTests.FromSource_ValidSource_ReturnsExpected(source: "")
System.CodeDom.Compiler.Tests.CodeCompilerTests.FromSourceBatch_ValidSources_ReturnsExpected(sources: [""])
System.CodeDom.Compiler.Tests.CodeCompilerTests.FromSource_ValidSource_ThrowsPlatformNotSupportedException(source: "")
System.CodeDom.Compiler.Tests.CodeCompilerTests.FromDom_ValidCodeCompileUnit_ThrowsPlatformNotSupportedException(compilationUnit: CodeCompileUnit { AssemblyCustomAttributes = [], EndDirectives = [], Namespaces = [], ReferencedAssemblies = ["referenced", "assembly1"], StartDirectives = [], ... })
System.CodeDom.Compiler.Tests.CodeCompilerTests.FromSourceBatch_ValidSources_ThrowsPlatformNotSupportedException(sources: [""])
System.Diagnostics.TextWriterTraceListenerTests.CtorsDelimiterTests.TestConstructorWithFileNameAndName(testName: "aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa"...)
System.Diagnostics.TextWriterTraceListenerTests.CtorsDelimiterTests.TestConstructorWithFileNameAndName(testName: "><&")
System.Diagnostics.TextWriterTraceListenerTests.XmlWriterTraceListenerTests.SingleArgumentConstructorTest
ystem.ServiceModel.Syndication.Tests.BasicScenarioTests.SyndicationFeed_Write_RSS_Atom
```
The full set can be seen with that build url. I think https://github.com/dotnet/runtime/pull/73408 broke these tests.
@eerhardt
|
2.0
|
Path.GetTempFileName() sometimes fails on WASM due to insufficient randomness - [Build](https://dev.azure.com/dnceng/public/_build/results?buildId=1934834&view=logs&jobId=a07e672c-d717-5036-7657-9f851d8c04dc&j=a07e672c-d717-5036-7657-9f851d8c04dc&t=95c8462d-76dc-506a-26a4-fefb7abc2e07), seen on multiple PRs with no related changes. Which suggests that this is on `main`. Failures like:
```
[19:23:32] info: Starting: System.CodeDom.Tests.dll
[19:23:39] fail: [FAIL] System.CodeDom.Compiler.Tests.CodeCompilerTests.FromSource_ValidSource_ReturnsExpected(source: "")
[19:23:39] info: System.IO.IOException : File exists
[19:23:39] info: at Interop.ThrowExceptionForIoErrno(ErrorInfo errorInfo, String path, Boolean isDirectory)
[19:23:39] info: at Interop.ThrowIOExceptionForLastError()
[19:23:39] info: at System.IO.Directory.CreateTempSubdirectoryCore(String prefix)
[19:23:39] info: at System.IO.Directory.CreateTempSubdirectory(String prefix)
[19:23:39] info: at System.CodeDom.Compiler.TempFileCollection.GetTempDirectory()
[19:23:39] info: at System.CodeDom.Compiler.TempFileCollection.EnsureTempNameCreated()
[19:23:39] info: at System.CodeDom.Compiler.TempFileCollection.get_BasePath()
[19:23:39] info: at System.CodeDom.Compiler.TempFileCollection.AddExtension(String fileExtension, Boolean keepFile)
[19:23:39] info: at System.CodeDom.Compiler.TempFileCollection.AddExtension(String fileExtension)
[19:23:39] info: at System.CodeDom.Compiler.CodeCompiler.FromSourceBatch(CompilerParameters options, String[] sources)
[19:23:39] info: at System.CodeDom.Compiler.CodeCompiler.FromSource(CompilerParameters options, String source)
[19:23:39] info: at System.CodeDom.Compiler.Tests.CodeCompilerTests.Compiler.FromSourceEntryPoint(CompilerParameters options, String source)
[19:23:39] info: at System.CodeDom.Compiler.Tests.CodeCompilerTests.FromSource_ValidSource_ReturnsExpected(String source)
[19:23:39] info: at System.Reflection.MethodInvoker.InterpretedInvoke(Object obj, Span`1 args, BindingFlags invokeAttr)
```
.. and ..
```
[19:27:13] info: Starting: System.Diagnostics.TextWriterTraceListener.Tests.dll
[19:27:14] fail: [FAIL] System.Diagnostics.TextWriterTraceListenerTests.CtorsDelimiterTests.TestConstructorWithFileNameAndName(testName: "")
[19:27:14] info: System.IO.IOException : File exists
[19:27:14] info: at Interop.ThrowExceptionForIoErrno(ErrorInfo errorInfo, String path, Boolean isDirectory)
[19:27:14] info: at Interop.CheckIo(Int64 result, String path, Boolean isDirectory)
[19:27:14] info: at Interop.CheckIo(IntPtr result, String path, Boolean isDirectory)
[19:27:14] info: at System.IO.Path.GetTempFileName()
[19:27:14] info: at System.Diagnostics.TextWriterTraceListenerTests.CtorsDelimiterTests.TestConstructorWithFileNameAndName(String testName)
[19:27:14] info: at System.Reflection.MethodInvoker.InterpretedInvoke(Object obj, Span`1 args, BindingFlags invokeAttr)
[19:27:14] fail: [FAIL] System.Diagnostics.TextWriterTraceListenerTests.CtorsDelimiterTests.TestConstructorWithFileNameAndName(testName: "><&")
[19:27:14] info: System.IO.IOException : File exists
[19:27:14] info: at Interop.ThrowExceptionForIoErrno(ErrorInfo errorInfo, String path, Boolean isDirectory)
[19:27:14] info: at Interop.CheckIo(Int64 result, String path, Boolean isDirectory)
[19:27:14] info: at Interop.CheckIo(IntPtr result, String path, Boolean isDirectory)
[19:27:14] info: at System.IO.Path.GetTempFileName()
[19:27:14] info: at System.Diagnostics.TextWriterTraceListenerTests.CtorsDelimiterTests.TestConstructorWithFileNameAndName(String testName)
[19:27:14] info: at System.Reflection.MethodInvoker.InterpretedInvoke(Object obj, Span`1 args, BindingFlags invokeAttr)
```
Just for people reading this in the future, the list of failed tests:
```
System.CodeDom.Compiler.Tests.CodeCompilerTests.FromDom_ValidCodeCompileUnit_ReturnsExpected(compilationUnit: CodeCompileUnit { AssemblyCustomAttributes = [], EndDirectives = [], Namespaces = [], ReferencedAssemblies = ["assembly1", "assembly2"], StartDirectives = [], ... })
System.CodeDom.Compiler.Tests.CodeCompilerTests.GetResponseFileCmdArgs_ValidCmdArgs_ReturnsExpected(cmdArgs: "")
System.CodeDom.Compiler.Tests.CodeCompilerTests.CompileAssemblyFromSourceBatch_ValidSources_ReturnsExpected(sources: [null])
System.CodeDom.Compiler.Tests.CodeCompilerTests.FromSource_ValidSource_ReturnsExpected(source: "")
System.CodeDom.Compiler.Tests.CodeCompilerTests.FromSourceBatch_ValidSources_ReturnsExpected(sources: [""])
System.CodeDom.Compiler.Tests.CodeCompilerTests.FromSource_ValidSource_ThrowsPlatformNotSupportedException(source: "")
System.CodeDom.Compiler.Tests.CodeCompilerTests.FromDom_ValidCodeCompileUnit_ThrowsPlatformNotSupportedException(compilationUnit: CodeCompileUnit { AssemblyCustomAttributes = [], EndDirectives = [], Namespaces = [], ReferencedAssemblies = ["referenced", "assembly1"], StartDirectives = [], ... })
System.CodeDom.Compiler.Tests.CodeCompilerTests.FromSourceBatch_ValidSources_ThrowsPlatformNotSupportedException(sources: [""])
System.Diagnostics.TextWriterTraceListenerTests.CtorsDelimiterTests.TestConstructorWithFileNameAndName(testName: "aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa"...)
System.Diagnostics.TextWriterTraceListenerTests.CtorsDelimiterTests.TestConstructorWithFileNameAndName(testName: "><&")
System.Diagnostics.TextWriterTraceListenerTests.XmlWriterTraceListenerTests.SingleArgumentConstructorTest
ystem.ServiceModel.Syndication.Tests.BasicScenarioTests.SyndicationFeed_Write_RSS_Atom
```
The full set can be seen with that build url. I think https://github.com/dotnet/runtime/pull/73408 broke these tests.
@eerhardt
|
test
|
path gettempfilename sometimes fails on wasm due to insufficient randomness seen on multiple prs with no related changes which suggests that this is on main failures like info starting system codedom tests dll fail system codedom compiler tests codecompilertests fromsource validsource returnsexpected source info system io ioexception file exists info at interop throwexceptionforioerrno errorinfo errorinfo string path boolean isdirectory info at interop throwioexceptionforlasterror info at system io directory createtempsubdirectorycore string prefix info at system io directory createtempsubdirectory string prefix info at system codedom compiler tempfilecollection gettempdirectory info at system codedom compiler tempfilecollection ensuretempnamecreated info at system codedom compiler tempfilecollection get basepath info at system codedom compiler tempfilecollection addextension string fileextension boolean keepfile info at system codedom compiler tempfilecollection addextension string fileextension info at system codedom compiler codecompiler fromsourcebatch compilerparameters options string sources info at system codedom compiler codecompiler fromsource compilerparameters options string source info at system codedom compiler tests codecompilertests compiler fromsourceentrypoint compilerparameters options string source info at system codedom compiler tests codecompilertests fromsource validsource returnsexpected string source info at system reflection methodinvoker interpretedinvoke object obj span args bindingflags invokeattr and info starting system diagnostics textwritertracelistener tests dll fail system diagnostics textwritertracelistenertests ctorsdelimitertests testconstructorwithfilenameandname testname info system io ioexception file exists info at interop throwexceptionforioerrno errorinfo errorinfo string path boolean isdirectory info at interop checkio result string path boolean isdirectory info at interop checkio intptr result string path boolean isdirectory info at system io path gettempfilename info at system diagnostics textwritertracelistenertests ctorsdelimitertests testconstructorwithfilenameandname string testname info at system reflection methodinvoker interpretedinvoke object obj span args bindingflags invokeattr fail system diagnostics textwritertracelistenertests ctorsdelimitertests testconstructorwithfilenameandname testname info system io ioexception file exists info at interop throwexceptionforioerrno errorinfo errorinfo string path boolean isdirectory info at interop checkio result string path boolean isdirectory info at interop checkio intptr result string path boolean isdirectory info at system io path gettempfilename info at system diagnostics textwritertracelistenertests ctorsdelimitertests testconstructorwithfilenameandname string testname info at system reflection methodinvoker interpretedinvoke object obj span args bindingflags invokeattr just for people reading this in the future the list of failed tests system codedom compiler tests codecompilertests fromdom validcodecompileunit returnsexpected compilationunit codecompileunit assemblycustomattributes enddirectives namespaces referencedassemblies startdirectives system codedom compiler tests codecompilertests getresponsefilecmdargs validcmdargs returnsexpected cmdargs system codedom compiler tests codecompilertests compileassemblyfromsourcebatch validsources returnsexpected sources system codedom compiler tests codecompilertests fromsource validsource returnsexpected source system codedom compiler tests codecompilertests fromsourcebatch validsources returnsexpected sources system codedom compiler tests codecompilertests fromsource validsource throwsplatformnotsupportedexception source system codedom compiler tests codecompilertests fromdom validcodecompileunit throwsplatformnotsupportedexception compilationunit codecompileunit assemblycustomattributes enddirectives namespaces referencedassemblies startdirectives system codedom compiler tests codecompilertests fromsourcebatch validsources throwsplatformnotsupportedexception sources system diagnostics textwritertracelistenertests ctorsdelimitertests testconstructorwithfilenameandname testname aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa system diagnostics textwritertracelistenertests ctorsdelimitertests testconstructorwithfilenameandname testname system diagnostics textwritertracelistenertests xmlwritertracelistenertests singleargumentconstructortest ystem servicemodel syndication tests basicscenariotests syndicationfeed write rss atom the full set can be seen with that build url i think broke these tests eerhardt
| 1
|
531,182
| 15,442,473,262
|
IssuesEvent
|
2021-03-08 07:46:58
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
support.mozilla.org - see bug description
|
browser-fenix engine-gecko ml-needsdiagnosis-false ml-probability-high priority-important
|
<!-- @browser: Firefox Mobile 88.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:88.0) Gecko/88.0 Firefox/88.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/67910 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://support.mozilla.org/fr/kb/idees-recues-sur-navigation-privee
**Browser / Version**: Firefox Mobile 88.0
**Operating System**: Android 8.1.0
**Tested Another Browser**: No
**Problem type**: Something else
**Description**: Traduction
**Steps to Reproduce**:
J'ai trouvé traduction pour Firefox nighthl
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/3/ec652403-8dbc-482c-948d-c9d088427daa.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20210302034602</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/3/a3dbf5e0-d71a-4556-acc4-aaee79f8b05e)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
support.mozilla.org - see bug description - <!-- @browser: Firefox Mobile 88.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:88.0) Gecko/88.0 Firefox/88.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/67910 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://support.mozilla.org/fr/kb/idees-recues-sur-navigation-privee
**Browser / Version**: Firefox Mobile 88.0
**Operating System**: Android 8.1.0
**Tested Another Browser**: No
**Problem type**: Something else
**Description**: Traduction
**Steps to Reproduce**:
J'ai trouvé traduction pour Firefox nighthl
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/3/ec652403-8dbc-482c-948d-c9d088427daa.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20210302034602</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/3/a3dbf5e0-d71a-4556-acc4-aaee79f8b05e)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_test
|
support mozilla org see bug description url browser version firefox mobile operating system android tested another browser no problem type something else description traduction steps to reproduce j ai trouvé traduction pour firefox nighthl view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
50,708
| 12,545,275,585
|
IssuesEvent
|
2020-06-05 18:39:18
|
tensorflow/tfjs
|
https://api.github.com/repos/tensorflow/tfjs
|
closed
|
npm WARN tar ENOENT: no such file or directory, open
|
comp:node.js type:build/install
|
To get help from the community, we encourage using Stack Overflow and the [`tensorflow.js`](https://stackoverflow.com/questions/tagged/tensorflow.js) tag.
#### TensorFlow.js version
`v12.13.1`
#### Browser version
`nodejs`
#### Describe the problem or feature request
```
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/lodash-79dc2199/package.json'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/lodash-79dc2199/_apply.js'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/lodash-79dc2199/_arrayAggregator.js'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/nan-a016ca8a/README.md'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/linq-dc98b547/linq.min.js'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/lodash-79dc2199/_arrayEach.js'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/lodash-79dc2199/_arrayEachRight.js'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/lodash-79dc2199/_arrayEvery.js'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/lodash-79dc2199/_arrayFilter.js'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/lodash-79dc2199/_arrayIncludes.js'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/lodash-79dc2199/_arrayIncludesWith.js'
...
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/@tensorflow/tfjs-core-08250c86/dist/tensor_format.js'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/@tensorflow/tfjs-core-08250c86/dist/tensor_format.js.map'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/@tensorflow/tfjs-core-08250c86/dist/tensor_test.d.ts'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/@tensorflow/tfjs-core-08250c86/dist/tensor_test.js'
```
#### Code to reproduce the bug / link to feature request
run `npm install`
|
1.0
|
npm WARN tar ENOENT: no such file or directory, open - To get help from the community, we encourage using Stack Overflow and the [`tensorflow.js`](https://stackoverflow.com/questions/tagged/tensorflow.js) tag.
#### TensorFlow.js version
`v12.13.1`
#### Browser version
`nodejs`
#### Describe the problem or feature request
```
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/lodash-79dc2199/package.json'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/lodash-79dc2199/_apply.js'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/lodash-79dc2199/_arrayAggregator.js'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/nan-a016ca8a/README.md'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/linq-dc98b547/linq.min.js'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/lodash-79dc2199/_arrayEach.js'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/lodash-79dc2199/_arrayEachRight.js'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/lodash-79dc2199/_arrayEvery.js'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/lodash-79dc2199/_arrayFilter.js'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/lodash-79dc2199/_arrayIncludes.js'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/lodash-79dc2199/_arrayIncludesWith.js'
...
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/@tensorflow/tfjs-core-08250c86/dist/tensor_format.js'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/@tensorflow/tfjs-core-08250c86/dist/tensor_format.js.map'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/@tensorflow/tfjs-core-08250c86/dist/tensor_test.d.ts'
npm WARN tar ENOENT: no such file or directory, open '/musixmatch/node_modules/.staging/@tensorflow/tfjs-core-08250c86/dist/tensor_test.js'
```
#### Code to reproduce the bug / link to feature request
run `npm install`
|
non_test
|
npm warn tar enoent no such file or directory open to get help from the community we encourage using stack overflow and the tag tensorflow js version browser version nodejs describe the problem or feature request npm warn tar enoent no such file or directory open musixmatch node modules staging lodash package json npm warn tar enoent no such file or directory open musixmatch node modules staging lodash apply js npm warn tar enoent no such file or directory open musixmatch node modules staging lodash arrayaggregator js npm warn tar enoent no such file or directory open musixmatch node modules staging nan readme md npm warn tar enoent no such file or directory open musixmatch node modules staging linq linq min js npm warn tar enoent no such file or directory open musixmatch node modules staging lodash arrayeach js npm warn tar enoent no such file or directory open musixmatch node modules staging lodash arrayeachright js npm warn tar enoent no such file or directory open musixmatch node modules staging lodash arrayevery js npm warn tar enoent no such file or directory open musixmatch node modules staging lodash arrayfilter js npm warn tar enoent no such file or directory open musixmatch node modules staging lodash arrayincludes js npm warn tar enoent no such file or directory open musixmatch node modules staging lodash arrayincludeswith js npm warn tar enoent no such file or directory open musixmatch node modules staging tensorflow tfjs core dist tensor format js npm warn tar enoent no such file or directory open musixmatch node modules staging tensorflow tfjs core dist tensor format js map npm warn tar enoent no such file or directory open musixmatch node modules staging tensorflow tfjs core dist tensor test d ts npm warn tar enoent no such file or directory open musixmatch node modules staging tensorflow tfjs core dist tensor test js code to reproduce the bug link to feature request run npm install
| 0
|
225,487
| 17,862,077,606
|
IssuesEvent
|
2021-09-06 03:11:12
|
aodn/nrmn-application
|
https://api.github.com/repos/aodn/nrmn-application
|
closed
|
Invert Sizing used on ingest row by row based on column 'Use InvertSizing'
|
ready to test
|
The "Use InvertSizing" column in the datasheet sets which size class to use to ingest the data (i.e. Fish or Invert size Class). The 'Use Invert Sizing" column is not displayed in the UI. This information is essential for ingest of the correct sizes, as sometimes the values are edited as Yes or No for different reasons.
|
1.0
|
Invert Sizing used on ingest row by row based on column 'Use InvertSizing' - The "Use InvertSizing" column in the datasheet sets which size class to use to ingest the data (i.e. Fish or Invert size Class). The 'Use Invert Sizing" column is not displayed in the UI. This information is essential for ingest of the correct sizes, as sometimes the values are edited as Yes or No for different reasons.
|
test
|
invert sizing used on ingest row by row based on column use invertsizing the use invertsizing column in the datasheet sets which size class to use to ingest the data i e fish or invert size class the use invert sizing column is not displayed in the ui this information is essential for ingest of the correct sizes as sometimes the values are edited as yes or no for different reasons
| 1
|
29,455
| 14,121,723,167
|
IssuesEvent
|
2020-11-09 02:54:09
|
simerplaha/SwayDB
|
https://api.github.com/repos/simerplaha/SwayDB
|
closed
|
Remove caching header ListSegment from PersistentSegmentMany
|
Performance
|
`PersistentSegmentMany` caches the header Segment on access which should be disabled in favour of persistent binary-search and hash-index.
Same configurations [binary-search](http://swaydb.io/configuration/binarySearchIndex/?language=java&q=binary) and [hashIndex](http://swaydb.io/configuration/randomKeyIndex/?language=java&q=hashin) should be used for indexing header segment.
|
True
|
Remove caching header ListSegment from PersistentSegmentMany - `PersistentSegmentMany` caches the header Segment on access which should be disabled in favour of persistent binary-search and hash-index.
Same configurations [binary-search](http://swaydb.io/configuration/binarySearchIndex/?language=java&q=binary) and [hashIndex](http://swaydb.io/configuration/randomKeyIndex/?language=java&q=hashin) should be used for indexing header segment.
|
non_test
|
remove caching header listsegment from persistentsegmentmany persistentsegmentmany caches the header segment on access which should be disabled in favour of persistent binary search and hash index same configurations and should be used for indexing header segment
| 0
|
194,551
| 14,681,521,274
|
IssuesEvent
|
2020-12-31 13:30:16
|
mapasculturais/mapasculturais
|
https://api.github.com/repos/mapasculturais/mapasculturais
|
closed
|
Método de avaliação "Critérios e Cotas": formulário de configuração das cotas
|
BACKEND secao:OPORTUNIDADES status:tested tipo:FUNCIONALIDADE
|
formulário de configuração das cotas (tem que mostrar todos os campos de seleção, ou tipo)
depende de #1060 #1061
|
1.0
|
Método de avaliação "Critérios e Cotas": formulário de configuração das cotas - formulário de configuração das cotas (tem que mostrar todos os campos de seleção, ou tipo)
depende de #1060 #1061
|
test
|
método de avaliação critérios e cotas formulário de configuração das cotas formulário de configuração das cotas tem que mostrar todos os campos de seleção ou tipo depende de
| 1
|
94,435
| 27,193,550,984
|
IssuesEvent
|
2023-02-20 01:50:26
|
gitpod-io/gitpod
|
https://api.github.com/repos/gitpod-io/gitpod
|
closed
|
bitbucket server does not install webhooks
|
type: bug git provider: bitbucket meta: stale feature: prebuilds feature: context awareness team: webapp
|
### Bug description
When trying to setup a prebuild hook , prefixing the url with gitpod.io/#prebuild/ *does not* create the BB webhook.
it does create the webhook for BB cloud.
### Steps to reproduce
Create a new repo, and gitpod project. prefix the branch url with `gitpod.io/#prebuild/ `
### Workspace affected
_No response_
### Expected behavior
webhook to be created on the repo
### Example repository
_No response_
### Anything else?
_No response_
|
1.0
|
bitbucket server does not install webhooks - ### Bug description
When trying to setup a prebuild hook , prefixing the url with gitpod.io/#prebuild/ *does not* create the BB webhook.
it does create the webhook for BB cloud.
### Steps to reproduce
Create a new repo, and gitpod project. prefix the branch url with `gitpod.io/#prebuild/ `
### Workspace affected
_No response_
### Expected behavior
webhook to be created on the repo
### Example repository
_No response_
### Anything else?
_No response_
|
non_test
|
bitbucket server does not install webhooks bug description when trying to setup a prebuild hook prefixing the url with gitpod io prebuild does not create the bb webhook it does create the webhook for bb cloud steps to reproduce create a new repo and gitpod project prefix the branch url with gitpod io prebuild workspace affected no response expected behavior webhook to be created on the repo example repository no response anything else no response
| 0
|
21,486
| 3,900,022,788
|
IssuesEvent
|
2016-04-18 02:07:37
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
TestDockerReadWriteReferenceVersion in deployment can't find /cockroach
|
test-failure
|
https://circleci.com/gh/cockroachdb/cockroach/16347 failed during the deploy-time run of the acceptance tests (the previous invocation was successful). Note the `/bin/bash: line 20: /cockroach: No such file or directory` below.
```
--- FAIL: TestDockerReadWriteReferenceVersion (6.32s)
+ echo 'Use the reference binary to write a couple rows, then render its output to a file and shut down.'
Use the reference binary to write a couple rows, then render its output to a file and shut down.
+ /reference-version/cockroach sql -e 'CREATE DATABASE old'
CREATE DATABASE
+ /reference-version/cockroach sql -d old -e 'CREATE TABLE testing (i int primary key, b bool, s string unique, d decimal, f float, t timestamp, v interval, index sb (s, b))'
CREATE TABLE
+ /reference-version/cockroach sql -d old -e 'INSERT INTO testing values (1, true, '\''hello'\'', decimal '\''3.14159'\'', 3.14159, NOW(), interval '\''1h'\'')'
INSERT 1
+ /reference-version/cockroach sql -d old -e 'INSERT INTO testing values (2, false, '\''world'\'', decimal '\''0.14159'\'', 0.14159, NOW(), interval '\''234h45m2s234ms'\'')'
INSERT 1
+ /reference-version/cockroach sql -d old -e 'SELECT * FROM testing'
+ /reference-version/cockroach quit
node drained and shutdown: ok
+ wait
initiating graceful shutdown of server
server drained and shutdown completed
+ bin=/cockroach
+ sleep 1
+ /cockroach start
/bin/bash: line 20: /cockroach: No such file or directory
+ echo 'Read data written by referencce version using new binary'
+ /cockroach sql -d old -e 'SELECT * FROM testing'
Read data written by referencce version using new binary
/bin/bash: line 23: /cockroach: No such file or directory
util_test.go:281: expected success: non-zero exit code: 1
```
|
1.0
|
TestDockerReadWriteReferenceVersion in deployment can't find /cockroach - https://circleci.com/gh/cockroachdb/cockroach/16347 failed during the deploy-time run of the acceptance tests (the previous invocation was successful). Note the `/bin/bash: line 20: /cockroach: No such file or directory` below.
```
--- FAIL: TestDockerReadWriteReferenceVersion (6.32s)
+ echo 'Use the reference binary to write a couple rows, then render its output to a file and shut down.'
Use the reference binary to write a couple rows, then render its output to a file and shut down.
+ /reference-version/cockroach sql -e 'CREATE DATABASE old'
CREATE DATABASE
+ /reference-version/cockroach sql -d old -e 'CREATE TABLE testing (i int primary key, b bool, s string unique, d decimal, f float, t timestamp, v interval, index sb (s, b))'
CREATE TABLE
+ /reference-version/cockroach sql -d old -e 'INSERT INTO testing values (1, true, '\''hello'\'', decimal '\''3.14159'\'', 3.14159, NOW(), interval '\''1h'\'')'
INSERT 1
+ /reference-version/cockroach sql -d old -e 'INSERT INTO testing values (2, false, '\''world'\'', decimal '\''0.14159'\'', 0.14159, NOW(), interval '\''234h45m2s234ms'\'')'
INSERT 1
+ /reference-version/cockroach sql -d old -e 'SELECT * FROM testing'
+ /reference-version/cockroach quit
node drained and shutdown: ok
+ wait
initiating graceful shutdown of server
server drained and shutdown completed
+ bin=/cockroach
+ sleep 1
+ /cockroach start
/bin/bash: line 20: /cockroach: No such file or directory
+ echo 'Read data written by referencce version using new binary'
+ /cockroach sql -d old -e 'SELECT * FROM testing'
Read data written by referencce version using new binary
/bin/bash: line 23: /cockroach: No such file or directory
util_test.go:281: expected success: non-zero exit code: 1
```
|
test
|
testdockerreadwritereferenceversion in deployment can t find cockroach failed during the deploy time run of the acceptance tests the previous invocation was successful note the bin bash line cockroach no such file or directory below fail testdockerreadwritereferenceversion echo use the reference binary to write a couple rows then render its output to a file and shut down use the reference binary to write a couple rows then render its output to a file and shut down reference version cockroach sql e create database old create database reference version cockroach sql d old e create table testing i int primary key b bool s string unique d decimal f float t timestamp v interval index sb s b create table reference version cockroach sql d old e insert into testing values true hello decimal now interval insert reference version cockroach sql d old e insert into testing values false world decimal now interval insert reference version cockroach sql d old e select from testing reference version cockroach quit node drained and shutdown ok wait initiating graceful shutdown of server server drained and shutdown completed bin cockroach sleep cockroach start bin bash line cockroach no such file or directory echo read data written by referencce version using new binary cockroach sql d old e select from testing read data written by referencce version using new binary bin bash line cockroach no such file or directory util test go expected success non zero exit code
| 1
|
18,421
| 4,270,547,981
|
IssuesEvent
|
2016-07-13 07:40:44
|
mathjax/MathJax
|
https://api.github.com/repos/mathjax/MathJax
|
closed
|
mlabeledtr and minlabelspacing implementation
|
MML spec limitation Needs Documentation
|
I'm not sure what the spec wants the effect of `minlabelspacing` to be and I'm not sure what our implementation is trying to do.
More generally, I don't (fully) understand how our `mlabeldtr` implementation works. I think we're following the spec's ["suggestion" (fourth paragraph)](http://www.w3.org/Math/draft-spec/chapter3.html#id.3.5.3.1) -- is that correct @dpvc?
However, I also don't fully understand that "suggestion" (e.g., is the "shift down" just referring to matching the row the label is labelling?). In particular, I don't understand what the spec's "suggestions" would suggest when the row+label are too wide for the mtable's container (overlap? shift away? shift where? outside the table?).
Anyway, it looks like MathJax adds the space from `minlabelspacing` to both sides of the "inner" mtable (by which I mean the mtable/mtr without the label). Naively reading spec I would have expected that `minlabelspacing` defines the minimal amount of space that guaranteed to be added between the label and the row it's labelling. In particular, I would only have expected space to be added one one side label.
|
1.0
|
mlabeledtr and minlabelspacing implementation - I'm not sure what the spec wants the effect of `minlabelspacing` to be and I'm not sure what our implementation is trying to do.
More generally, I don't (fully) understand how our `mlabeldtr` implementation works. I think we're following the spec's ["suggestion" (fourth paragraph)](http://www.w3.org/Math/draft-spec/chapter3.html#id.3.5.3.1) -- is that correct @dpvc?
However, I also don't fully understand that "suggestion" (e.g., is the "shift down" just referring to matching the row the label is labelling?). In particular, I don't understand what the spec's "suggestions" would suggest when the row+label are too wide for the mtable's container (overlap? shift away? shift where? outside the table?).
Anyway, it looks like MathJax adds the space from `minlabelspacing` to both sides of the "inner" mtable (by which I mean the mtable/mtr without the label). Naively reading spec I would have expected that `minlabelspacing` defines the minimal amount of space that guaranteed to be added between the label and the row it's labelling. In particular, I would only have expected space to be added one one side label.
|
non_test
|
mlabeledtr and minlabelspacing implementation i m not sure what the spec wants the effect of minlabelspacing to be and i m not sure what our implementation is trying to do more generally i don t fully understand how our mlabeldtr implementation works i think we re following the spec s is that correct dpvc however i also don t fully understand that suggestion e g is the shift down just referring to matching the row the label is labelling in particular i don t understand what the spec s suggestions would suggest when the row label are too wide for the mtable s container overlap shift away shift where outside the table anyway it looks like mathjax adds the space from minlabelspacing to both sides of the inner mtable by which i mean the mtable mtr without the label naively reading spec i would have expected that minlabelspacing defines the minimal amount of space that guaranteed to be added between the label and the row it s labelling in particular i would only have expected space to be added one one side label
| 0
|
42,103
| 10,819,558,081
|
IssuesEvent
|
2019-11-08 14:38:29
|
raz0red/wii-mednafen
|
https://api.github.com/repos/raz0red/wii-mednafen
|
closed
|
USB drive not mounting
|
Priority-Medium Type-Defect auto-migrated
|
```
What steps will reproduce the problem?
1. Opening the emulator through usb loader gx
2. Selecting to open a ROM from USB drive (which is fat32)
3.
What is the expected output? What do you see instead?
The expected output is for the emulator to recognize the drive and open my pace
CD games. Instead it says it's unable to mount the drive
What version of the product are you using? On what operating system?
0.1.1 on wii
Please provide any additional information below.
```
Original issue reported on code.google.com by `machem...@gmail.com` on 25 Feb 2015 at 2:25
|
1.0
|
USB drive not mounting - ```
What steps will reproduce the problem?
1. Opening the emulator through usb loader gx
2. Selecting to open a ROM from USB drive (which is fat32)
3.
What is the expected output? What do you see instead?
The expected output is for the emulator to recognize the drive and open my pace
CD games. Instead it says it's unable to mount the drive
What version of the product are you using? On what operating system?
0.1.1 on wii
Please provide any additional information below.
```
Original issue reported on code.google.com by `machem...@gmail.com` on 25 Feb 2015 at 2:25
|
non_test
|
usb drive not mounting what steps will reproduce the problem opening the emulator through usb loader gx selecting to open a rom from usb drive which is what is the expected output what do you see instead the expected output is for the emulator to recognize the drive and open my pace cd games instead it says it s unable to mount the drive what version of the product are you using on what operating system on wii please provide any additional information below original issue reported on code google com by machem gmail com on feb at
| 0
|
345,485
| 10,368,339,990
|
IssuesEvent
|
2019-09-07 16:08:41
|
MLP-VectorClub/Winterchilla
|
https://api.github.com/repos/MLP-VectorClub/Winterchilla
|
closed
|
Database unique violation when adding synonym tags
|
Admin Tools Bug Priority: Medium
|
When the `male` tag is present and `stallion` is added this causes a unique violation for some reason. Needs investigating.
|
1.0
|
Database unique violation when adding synonym tags - When the `male` tag is present and `stallion` is added this causes a unique violation for some reason. Needs investigating.
|
non_test
|
database unique violation when adding synonym tags when the male tag is present and stallion is added this causes a unique violation for some reason needs investigating
| 0
|
7,628
| 18,719,045,239
|
IssuesEvent
|
2021-11-03 09:39:24
|
tremor-rs/tremor-www
|
https://api.github.com/repos/tremor-rs/tremor-www
|
closed
|
Move Getting Started section out of community and into docs
|
bug documentation good first issue hacktoberfest information-architecture
|
The new website is AWESOME!! One change I would made:
Move https://www.tremor.rs/community/getting-started/getting-started to https://www.tremor.rs/docs/getting-started/getting-started
This is a very code-and-stuff based set of docs, so it needs to be in with the rest of the docs IMHO.
|
1.0
|
Move Getting Started section out of community and into docs - The new website is AWESOME!! One change I would made:
Move https://www.tremor.rs/community/getting-started/getting-started to https://www.tremor.rs/docs/getting-started/getting-started
This is a very code-and-stuff based set of docs, so it needs to be in with the rest of the docs IMHO.
|
non_test
|
move getting started section out of community and into docs the new website is awesome one change i would made move to this is a very code and stuff based set of docs so it needs to be in with the rest of the docs imho
| 0
|
195,788
| 15,556,280,084
|
IssuesEvent
|
2021-03-16 07:34:11
|
mridulgain/whiteboard
|
https://api.github.com/repos/mridulgain/whiteboard
|
opened
|
getting started with documentation!
|
documentation good first issue
|
Expectations:
1. helping contributors to get started with project
2. though the app is simple, a how to use giving feature details is always good.
|
1.0
|
getting started with documentation! - Expectations:
1. helping contributors to get started with project
2. though the app is simple, a how to use giving feature details is always good.
|
non_test
|
getting started with documentation expectations helping contributors to get started with project though the app is simple a how to use giving feature details is always good
| 0
|
222,880
| 7,440,481,383
|
IssuesEvent
|
2018-03-27 10:13:01
|
duaraghav8/Solium
|
https://api.github.com/repos/duaraghav8/Solium
|
closed
|
Node EventEmitter: set max listeners
|
Fixed in upcoming release Top Priority
|
https://nodejs.org/docs/latest/api/events.html#events_emitter_setmaxlisteners_n
As more rules are added (either core or via plugin), more listeners will spawn up for every AST Node type.
https://github.com/duaraghav8/solium-plugin-security/pull/1 already causes node to warn for `CallExpression` event listener limit exceed.
Need to set max listeners to `INFINITY`
this is not a bug, its just that we need to increase the limit
|
1.0
|
Node EventEmitter: set max listeners - https://nodejs.org/docs/latest/api/events.html#events_emitter_setmaxlisteners_n
As more rules are added (either core or via plugin), more listeners will spawn up for every AST Node type.
https://github.com/duaraghav8/solium-plugin-security/pull/1 already causes node to warn for `CallExpression` event listener limit exceed.
Need to set max listeners to `INFINITY`
this is not a bug, its just that we need to increase the limit
|
non_test
|
node eventemitter set max listeners as more rules are added either core or via plugin more listeners will spawn up for every ast node type already causes node to warn for callexpression event listener limit exceed need to set max listeners to infinity this is not a bug its just that we need to increase the limit
| 0
|
89,703
| 15,835,506,664
|
IssuesEvent
|
2021-04-06 18:06:09
|
idonthaveafifaaddiction/fast-csv
|
https://api.github.com/repos/idonthaveafifaaddiction/fast-csv
|
opened
|
CVE-2021-27292 (High) detected in ua-parser-js-0.7.22.tgz
|
security vulnerability
|
## CVE-2021-27292 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ua-parser-js-0.7.22.tgz</b></p></summary>
<p>Lightweight JavaScript-based user-agent string parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.22.tgz">https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.22.tgz</a></p>
<p>Path to dependency file: fast-csv/documentation/package.json</p>
<p>Path to vulnerable library: fast-csv/documentation/node_modules/ua-parser-js/package.json</p>
<p>
Dependency Hierarchy:
- preset-classic-2.0.0-alpha.66.tgz (Root Library)
- plugin-debug-2.0.0-alpha.66.tgz
- react-json-view-1.19.1.tgz
- flux-3.1.3.tgz
- fbjs-0.8.17.tgz
- :x: **ua-parser-js-0.7.22.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/idonthaveafifaaddiction/fast-csv/commit/4844094d21cb62fe761a99d9a4b23935d27956fb">4844094d21cb62fe761a99d9a4b23935d27956fb</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
ua-parser-js >= 0.7.14, fixed in 0.7.24, uses a regular expression which is vulnerable to denial of service. If an attacker sends a malicious User-Agent header, ua-parser-js will get stuck processing it for an extended period of time.
<p>Publish Date: 2021-03-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-27292>CVE-2021-27292</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/faisalman/ua-parser-js/releases/tag/0.7.24">https://github.com/faisalman/ua-parser-js/releases/tag/0.7.24</a></p>
<p>Release Date: 2021-03-17</p>
<p>Fix Resolution: ua-parser-js - 0.7.24</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"ua-parser-js","packageVersion":"0.7.22","packageFilePaths":["/documentation/package.json"],"isTransitiveDependency":true,"dependencyTree":"@docusaurus/preset-classic:2.0.0-alpha.66;@docusaurus/plugin-debug:2.0.0-alpha.66;react-json-view:1.19.1;flux:3.1.3;fbjs:0.8.17;ua-parser-js:0.7.22","isMinimumFixVersionAvailable":true,"minimumFixVersion":"ua-parser-js - 0.7.24"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-27292","vulnerabilityDetails":"ua-parser-js \u003e\u003d 0.7.14, fixed in 0.7.24, uses a regular expression which is vulnerable to denial of service. If an attacker sends a malicious User-Agent header, ua-parser-js will get stuck processing it for an extended period of time.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-27292","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2021-27292 (High) detected in ua-parser-js-0.7.22.tgz - ## CVE-2021-27292 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ua-parser-js-0.7.22.tgz</b></p></summary>
<p>Lightweight JavaScript-based user-agent string parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.22.tgz">https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.22.tgz</a></p>
<p>Path to dependency file: fast-csv/documentation/package.json</p>
<p>Path to vulnerable library: fast-csv/documentation/node_modules/ua-parser-js/package.json</p>
<p>
Dependency Hierarchy:
- preset-classic-2.0.0-alpha.66.tgz (Root Library)
- plugin-debug-2.0.0-alpha.66.tgz
- react-json-view-1.19.1.tgz
- flux-3.1.3.tgz
- fbjs-0.8.17.tgz
- :x: **ua-parser-js-0.7.22.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/idonthaveafifaaddiction/fast-csv/commit/4844094d21cb62fe761a99d9a4b23935d27956fb">4844094d21cb62fe761a99d9a4b23935d27956fb</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
ua-parser-js >= 0.7.14, fixed in 0.7.24, uses a regular expression which is vulnerable to denial of service. If an attacker sends a malicious User-Agent header, ua-parser-js will get stuck processing it for an extended period of time.
<p>Publish Date: 2021-03-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-27292>CVE-2021-27292</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/faisalman/ua-parser-js/releases/tag/0.7.24">https://github.com/faisalman/ua-parser-js/releases/tag/0.7.24</a></p>
<p>Release Date: 2021-03-17</p>
<p>Fix Resolution: ua-parser-js - 0.7.24</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"ua-parser-js","packageVersion":"0.7.22","packageFilePaths":["/documentation/package.json"],"isTransitiveDependency":true,"dependencyTree":"@docusaurus/preset-classic:2.0.0-alpha.66;@docusaurus/plugin-debug:2.0.0-alpha.66;react-json-view:1.19.1;flux:3.1.3;fbjs:0.8.17;ua-parser-js:0.7.22","isMinimumFixVersionAvailable":true,"minimumFixVersion":"ua-parser-js - 0.7.24"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-27292","vulnerabilityDetails":"ua-parser-js \u003e\u003d 0.7.14, fixed in 0.7.24, uses a regular expression which is vulnerable to denial of service. If an attacker sends a malicious User-Agent header, ua-parser-js will get stuck processing it for an extended period of time.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-27292","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_test
|
cve high detected in ua parser js tgz cve high severity vulnerability vulnerable library ua parser js tgz lightweight javascript based user agent string parser library home page a href path to dependency file fast csv documentation package json path to vulnerable library fast csv documentation node modules ua parser js package json dependency hierarchy preset classic alpha tgz root library plugin debug alpha tgz react json view tgz flux tgz fbjs tgz x ua parser js tgz vulnerable library found in head commit a href found in base branch master vulnerability details ua parser js fixed in uses a regular expression which is vulnerable to denial of service if an attacker sends a malicious user agent header ua parser js will get stuck processing it for an extended period of time publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ua parser js isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree docusaurus preset classic alpha docusaurus plugin debug alpha react json view flux fbjs ua parser js isminimumfixversionavailable true minimumfixversion ua parser js basebranches vulnerabilityidentifier cve vulnerabilitydetails ua parser js fixed in uses a regular expression which is vulnerable to denial of service if an attacker sends a malicious user agent header ua parser js will get stuck processing it for an extended period of time vulnerabilityurl
| 0
|
60,937
| 3,135,912,585
|
IssuesEvent
|
2015-09-10 17:22:39
|
JMurk/OAM-Activity-Manager
|
https://api.github.com/repos/JMurk/OAM-Activity-Manager
|
closed
|
OAM - CW Lookup - More Wild Cards
|
enhancement high priority
|
**Issue:** Amy would like users to have the ability to add wild cards to the house number so that results are returned for a specific block.
For example: "2% N FULTON%" would return:
-264 N FULTON AVE
-2331 N FULTON ST
-21 N FULTON AV

Is this possible? Or perhaps there is a better way for doing this?
|
1.0
|
OAM - CW Lookup - More Wild Cards - **Issue:** Amy would like users to have the ability to add wild cards to the house number so that results are returned for a specific block.
For example: "2% N FULTON%" would return:
-264 N FULTON AVE
-2331 N FULTON ST
-21 N FULTON AV

Is this possible? Or perhaps there is a better way for doing this?
|
non_test
|
oam cw lookup more wild cards issue amy would like users to have the ability to add wild cards to the house number so that results are returned for a specific block for example n fulton would return n fulton ave n fulton st n fulton av is this possible or perhaps there is a better way for doing this
| 0
|
3,788
| 2,692,121,740
|
IssuesEvent
|
2015-04-01 04:05:34
|
GoogleCloudPlatform/kubernetes
|
https://api.github.com/repos/GoogleCloudPlatform/kubernetes
|
closed
|
Integration test is flaky due to time out and pods not started
|
area/test priority/P1 team/testing
|
Once a while, I saw the failure as the below, and re-run works
F0214 00:24:10.780696 700 integration.go:297] FAILED: pods never started running timed out waiting for the condition
!!! Error in hack/test-integration.sh:47
'"${KUBE_OUTPUT_HOSTBIN}/integration" --v=2' exited with status 255
Call stack:
1: hack/test-integration.sh:47 main(...)
Exiting with status 1
+++ Integration test cleanup complete
!!! Error in hack/../build/../build/common.sh:405
'"${docker_cmd[@]}" "$@"' exited with status 1
Call stack:
1: hack/../build/../build/common.sh:405 kube::build::run_build_command(...)
2: hack/../build/release.sh:36 main(...)
Exiting with status 1
|
2.0
|
Integration test is flaky due to time out and pods not started - Once a while, I saw the failure as the below, and re-run works
F0214 00:24:10.780696 700 integration.go:297] FAILED: pods never started running timed out waiting for the condition
!!! Error in hack/test-integration.sh:47
'"${KUBE_OUTPUT_HOSTBIN}/integration" --v=2' exited with status 255
Call stack:
1: hack/test-integration.sh:47 main(...)
Exiting with status 1
+++ Integration test cleanup complete
!!! Error in hack/../build/../build/common.sh:405
'"${docker_cmd[@]}" "$@"' exited with status 1
Call stack:
1: hack/../build/../build/common.sh:405 kube::build::run_build_command(...)
2: hack/../build/release.sh:36 main(...)
Exiting with status 1
|
test
|
integration test is flaky due to time out and pods not started once a while i saw the failure as the below and re run works integration go failed pods never started running timed out waiting for the condition error in hack test integration sh kube output hostbin integration v exited with status call stack hack test integration sh main exiting with status integration test cleanup complete error in hack build build common sh docker cmd exited with status call stack hack build build common sh kube build run build command hack build release sh main exiting with status
| 1
|
65,542
| 16,385,351,959
|
IssuesEvent
|
2021-05-17 09:42:44
|
wso2-attic/ux-font-wso2
|
https://api.github.com/repos/wso2-attic/ux-font-wso2
|
closed
|
Remove fontforge meta information
|
build enhancement
|
Need to remove fontforge local PC meta information which embeds in `font-wso2.svg` on build
e.g.
```xml
<metadata>
Created by FontForge 20170104 at Wed Jul 19 15:41:44 2017
By Sharon David
</metadata>
```
|
1.0
|
Remove fontforge meta information - Need to remove fontforge local PC meta information which embeds in `font-wso2.svg` on build
e.g.
```xml
<metadata>
Created by FontForge 20170104 at Wed Jul 19 15:41:44 2017
By Sharon David
</metadata>
```
|
non_test
|
remove fontforge meta information need to remove fontforge local pc meta information which embeds in font svg on build e g xml created by fontforge at wed jul by sharon david
| 0
|
361,738
| 25,350,204,328
|
IssuesEvent
|
2022-11-19 17:25:50
|
treyburn/routed
|
https://api.github.com/repos/treyburn/routed
|
opened
|
Create a README
|
documentation
|
We could use a readme to convey what this package does, how to use, how to setup your dev environment, how to test, and the basic architecture of the of the code design (use PlantUML)
|
1.0
|
Create a README - We could use a readme to convey what this package does, how to use, how to setup your dev environment, how to test, and the basic architecture of the of the code design (use PlantUML)
|
non_test
|
create a readme we could use a readme to convey what this package does how to use how to setup your dev environment how to test and the basic architecture of the of the code design use plantuml
| 0
|
377,633
| 11,177,380,736
|
IssuesEvent
|
2019-12-30 10:30:39
|
FireCubeStudios/Appraisit
|
https://api.github.com/repos/FireCubeStudios/Appraisit
|
closed
|
Organizing github code
|
Priority bug enhancement good first issue help wanted
|
Xaml code:
- Chnage x:Names to be more unique and easily indentifiable
- Move resources to a resource dictionary or another file in some way
- Move content dialogs to seperate files without affecting anything.
- Move usercontrol content into user control file✔
- Change format to single line
C# code:
- Change names of functions to be more unique and easily identifiable
>> especially between lines 1866 to 2052 in mainpage.xaml.cs
- Get rid of useless code
- Get rid of comments that I keep just in case
- Add comments detailing each section
- Use comments to put code in sections with related functions in different sections
- Move some repeated code to other files
Github project:
- Add a readme with images and links to the app
- Readme shows guidelines for helping with the project
- Add badges
- Change description
|
1.0
|
Organizing github code - Xaml code:
- Chnage x:Names to be more unique and easily indentifiable
- Move resources to a resource dictionary or another file in some way
- Move content dialogs to seperate files without affecting anything.
- Move usercontrol content into user control file✔
- Change format to single line
C# code:
- Change names of functions to be more unique and easily identifiable
>> especially between lines 1866 to 2052 in mainpage.xaml.cs
- Get rid of useless code
- Get rid of comments that I keep just in case
- Add comments detailing each section
- Use comments to put code in sections with related functions in different sections
- Move some repeated code to other files
Github project:
- Add a readme with images and links to the app
- Readme shows guidelines for helping with the project
- Add badges
- Change description
|
non_test
|
organizing github code xaml code chnage x names to be more unique and easily indentifiable move resources to a resource dictionary or another file in some way move content dialogs to seperate files without affecting anything move usercontrol content into user control file✔ change format to single line c code change names of functions to be more unique and easily identifiable especially between lines to in mainpage xaml cs get rid of useless code get rid of comments that i keep just in case add comments detailing each section use comments to put code in sections with related functions in different sections move some repeated code to other files github project add a readme with images and links to the app readme shows guidelines for helping with the project add badges change description
| 0
|
48,822
| 10,281,712,290
|
IssuesEvent
|
2019-08-26 09:11:49
|
mozilla/fxa
|
https://api.github.com/repos/mozilla/fxa
|
closed
|
Simplify fxa-js-client release process
|
cat:code quality help wanted
|
Running `grunt release` in the js client directory still tries to build and deploy the docs, which fails and is redundant anyway because we have the docs being built and deployed automagically in Circle now. We should take that stuff out.
|
1.0
|
Simplify fxa-js-client release process - Running `grunt release` in the js client directory still tries to build and deploy the docs, which fails and is redundant anyway because we have the docs being built and deployed automagically in Circle now. We should take that stuff out.
|
non_test
|
simplify fxa js client release process running grunt release in the js client directory still tries to build and deploy the docs which fails and is redundant anyway because we have the docs being built and deployed automagically in circle now we should take that stuff out
| 0
|
3,537
| 2,692,321,189
|
IssuesEvent
|
2015-04-01 06:43:56
|
angular-ui/ui-router
|
https://api.github.com/repos/angular-ui/ui-router
|
reopened
|
Repeated search parameters introduce unwanted polymorphism
|
design
|
Right now, if I want to pass multiple values into a search parameter, UI-router simply repeats the search parameter for every value that I pass in.
For example:
```js
new UrlMatcher('/users?{id:int}').format({ id:[1,2] }); // '/users?id=1&id=2`
// In a state:
$stateParams.id; // [1,2]
```
However, if I only pass in one user ID:
```js
new UrlMatcher('/users?{id:int}').format({ id:[1] }); // '/users?id=1`
// In a state:
$stateParams.id; // 1
```
This is less-than-ideal, because it makes `$stateParams.id` polymorphic, even when I've specified a typed parameter, and requires a lot of type-checking in controllers.'
At the very minimum, I end up with lots of:
```js
var id = $stateParams.id;
if(!angular.isArray(id)){
id = [id];
}
id.forEach(...)
```
Worse still, *any* state parameter risks being cast into an array if a user makes a typo, and specifies the same parameter twice. This seems likely to cause unpredictable breakages in controllers, given that most developers will not code defensively against this scenario.
IMO, the search parameter grouping introduced by #373 should require an explicit opt-in for each parameter (*especially* if the parameter is typed). If the matcher expects multiple values for a parameter, it should *always* deserialize the value(s) into an array, even if only one value is provided. If the matcher does not expect multiple values, it should return the first one that it finds, and discard subsequent matches (or throw an error).
---
Similarly, I'd also be interested to hear thoughts on introducing support for delimited search/query parameters (eg. `/url?id=1,2`). While UI-router's current implementation seems to be the most common, [there does not appear to be a standard](http://stackoverflow.com/a/24728298), and delimited lists arguably might produce more readable URLs...
Certainly not appropriate for all use-cases, but possibly quite useful...
---
### Summary
* Predictability is good.
* Strong typing improves predictability. It's been an immensely useful addition to ui-router.
* Type-checking is obnoxious, requires lots of boilerplate, and most developers won't do it if they're only expecting a single value.
* If developers *are* expecting multiple values, they shouldn't need to handle a single value as a special case.
* Defining a typed parameter should guarantee that the parameter's value always matches the specified type.
* For the sake of simplicity and compatibility, the current behavior might be appropriate/acceptable for untyped parameters. However, it seems inappropriate for typed parameters.
* Comma-delimited parameters might be cool?
|
1.0
|
Repeated search parameters introduce unwanted polymorphism - Right now, if I want to pass multiple values into a search parameter, UI-router simply repeats the search parameter for every value that I pass in.
For example:
```js
new UrlMatcher('/users?{id:int}').format({ id:[1,2] }); // '/users?id=1&id=2`
// In a state:
$stateParams.id; // [1,2]
```
However, if I only pass in one user ID:
```js
new UrlMatcher('/users?{id:int}').format({ id:[1] }); // '/users?id=1`
// In a state:
$stateParams.id; // 1
```
This is less-than-ideal, because it makes `$stateParams.id` polymorphic, even when I've specified a typed parameter, and requires a lot of type-checking in controllers.'
At the very minimum, I end up with lots of:
```js
var id = $stateParams.id;
if(!angular.isArray(id)){
id = [id];
}
id.forEach(...)
```
Worse still, *any* state parameter risks being cast into an array if a user makes a typo, and specifies the same parameter twice. This seems likely to cause unpredictable breakages in controllers, given that most developers will not code defensively against this scenario.
IMO, the search parameter grouping introduced by #373 should require an explicit opt-in for each parameter (*especially* if the parameter is typed). If the matcher expects multiple values for a parameter, it should *always* deserialize the value(s) into an array, even if only one value is provided. If the matcher does not expect multiple values, it should return the first one that it finds, and discard subsequent matches (or throw an error).
---
Similarly, I'd also be interested to hear thoughts on introducing support for delimited search/query parameters (eg. `/url?id=1,2`). While UI-router's current implementation seems to be the most common, [there does not appear to be a standard](http://stackoverflow.com/a/24728298), and delimited lists arguably might produce more readable URLs...
Certainly not appropriate for all use-cases, but possibly quite useful...
---
### Summary
* Predictability is good.
* Strong typing improves predictability. It's been an immensely useful addition to ui-router.
* Type-checking is obnoxious, requires lots of boilerplate, and most developers won't do it if they're only expecting a single value.
* If developers *are* expecting multiple values, they shouldn't need to handle a single value as a special case.
* Defining a typed parameter should guarantee that the parameter's value always matches the specified type.
* For the sake of simplicity and compatibility, the current behavior might be appropriate/acceptable for untyped parameters. However, it seems inappropriate for typed parameters.
* Comma-delimited parameters might be cool?
|
non_test
|
repeated search parameters introduce unwanted polymorphism right now if i want to pass multiple values into a search parameter ui router simply repeats the search parameter for every value that i pass in for example js new urlmatcher users id int format id users id id in a state stateparams id however if i only pass in one user id js new urlmatcher users id int format id users id in a state stateparams id this is less than ideal because it makes stateparams id polymorphic even when i ve specified a typed parameter and requires a lot of type checking in controllers at the very minimum i end up with lots of js var id stateparams id if angular isarray id id id foreach worse still any state parameter risks being cast into an array if a user makes a typo and specifies the same parameter twice this seems likely to cause unpredictable breakages in controllers given that most developers will not code defensively against this scenario imo the search parameter grouping introduced by should require an explicit opt in for each parameter especially if the parameter is typed if the matcher expects multiple values for a parameter it should always deserialize the value s into an array even if only one value is provided if the matcher does not expect multiple values it should return the first one that it finds and discard subsequent matches or throw an error similarly i d also be interested to hear thoughts on introducing support for delimited search query parameters eg url id while ui router s current implementation seems to be the most common and delimited lists arguably might produce more readable urls certainly not appropriate for all use cases but possibly quite useful summary predictability is good strong typing improves predictability it s been an immensely useful addition to ui router type checking is obnoxious requires lots of boilerplate and most developers won t do it if they re only expecting a single value if developers are expecting multiple values they shouldn t need to handle a single value as a special case defining a typed parameter should guarantee that the parameter s value always matches the specified type for the sake of simplicity and compatibility the current behavior might be appropriate acceptable for untyped parameters however it seems inappropriate for typed parameters comma delimited parameters might be cool
| 0
|
160,889
| 12,520,971,363
|
IssuesEvent
|
2020-06-03 16:43:29
|
aliasrobotics/RVD
|
https://api.github.com/repos/aliasrobotics/RVD
|
opened
|
Using ServerProxy to parse untrusted XML data is known to be vulnerable..., ./src/ros_comm/roslaunch/test/unit/test_core.py:157
|
bandit bug components software robot component: ROS static analysis testing triage version: melodic
|
```yaml
{
"id": 1,
"title": "Using ServerProxy to parse untrusted XML data is known to be vulnerable..., ./src/ros_comm/roslaunch/test/unit/test_core.py:157",
"type": "bug",
"description": "HIGH confidence of HIGH severity bug. Using ServerProxy to parse untrusted XML data is known to be vulnerable to XML attacks. Use defused.xmlrpc.monkey_patch() function to monkey-patch xmlrpclib and mitigate XML vulnerabilities. ./src/ros_comm/roslaunch/test/unit/test_core.py:157. See links for more info on the bug.",
"cwe": "None",
"cve": "None",
"keywords": [
"bandit",
"bug",
"static analysis",
"testing",
"triage",
"bug",
"version: melodic",
"robot component: ROS",
"components software"
],
"system": "",
"vendor": null,
"severity": {
"rvss-score": 0,
"rvss-vector": "",
"severity-description": "",
"cvss-score": 0,
"cvss-vector": ""
},
"links": "",
"flaw": {
"phase": "testing",
"specificity": "subject-specific",
"architectural-location": "application-specific",
"application": "N/A",
"subsystem": "N/A",
"package": "N/A",
"languages": "None",
"date-detected": "2020-06-03 (16:43)",
"detected-by": "Alias Robotics",
"detected-by-method": "testing static",
"date-reported": "2020-06-03 (16:43)",
"reported-by": "Alias Robotics",
"reported-by-relationship": "automatic",
"issue": "",
"reproducibility": "always",
"trace": "./src/ros_comm/roslaunch/test/unit/test_core.py:157",
"reproduction": "See artifacts below (if available)",
"reproduction-image": ""
},
"exploitation": {
"description": "",
"exploitation-image": "",
"exploitation-vector": ""
},
"mitigation": {
"description": "",
"pull-request": "",
"date-mitigation": ""
}
}
```
|
1.0
|
Using ServerProxy to parse untrusted XML data is known to be vulnerable..., ./src/ros_comm/roslaunch/test/unit/test_core.py:157 - ```yaml
{
"id": 1,
"title": "Using ServerProxy to parse untrusted XML data is known to be vulnerable..., ./src/ros_comm/roslaunch/test/unit/test_core.py:157",
"type": "bug",
"description": "HIGH confidence of HIGH severity bug. Using ServerProxy to parse untrusted XML data is known to be vulnerable to XML attacks. Use defused.xmlrpc.monkey_patch() function to monkey-patch xmlrpclib and mitigate XML vulnerabilities. ./src/ros_comm/roslaunch/test/unit/test_core.py:157. See links for more info on the bug.",
"cwe": "None",
"cve": "None",
"keywords": [
"bandit",
"bug",
"static analysis",
"testing",
"triage",
"bug",
"version: melodic",
"robot component: ROS",
"components software"
],
"system": "",
"vendor": null,
"severity": {
"rvss-score": 0,
"rvss-vector": "",
"severity-description": "",
"cvss-score": 0,
"cvss-vector": ""
},
"links": "",
"flaw": {
"phase": "testing",
"specificity": "subject-specific",
"architectural-location": "application-specific",
"application": "N/A",
"subsystem": "N/A",
"package": "N/A",
"languages": "None",
"date-detected": "2020-06-03 (16:43)",
"detected-by": "Alias Robotics",
"detected-by-method": "testing static",
"date-reported": "2020-06-03 (16:43)",
"reported-by": "Alias Robotics",
"reported-by-relationship": "automatic",
"issue": "",
"reproducibility": "always",
"trace": "./src/ros_comm/roslaunch/test/unit/test_core.py:157",
"reproduction": "See artifacts below (if available)",
"reproduction-image": ""
},
"exploitation": {
"description": "",
"exploitation-image": "",
"exploitation-vector": ""
},
"mitigation": {
"description": "",
"pull-request": "",
"date-mitigation": ""
}
}
```
|
test
|
using serverproxy to parse untrusted xml data is known to be vulnerable src ros comm roslaunch test unit test core py yaml id title using serverproxy to parse untrusted xml data is known to be vulnerable src ros comm roslaunch test unit test core py type bug description high confidence of high severity bug using serverproxy to parse untrusted xml data is known to be vulnerable to xml attacks use defused xmlrpc monkey patch function to monkey patch xmlrpclib and mitigate xml vulnerabilities src ros comm roslaunch test unit test core py see links for more info on the bug cwe none cve none keywords bandit bug static analysis testing triage bug version melodic robot component ros components software system vendor null severity rvss score rvss vector severity description cvss score cvss vector links flaw phase testing specificity subject specific architectural location application specific application n a subsystem n a package n a languages none date detected detected by alias robotics detected by method testing static date reported reported by alias robotics reported by relationship automatic issue reproducibility always trace src ros comm roslaunch test unit test core py reproduction see artifacts below if available reproduction image exploitation description exploitation image exploitation vector mitigation description pull request date mitigation
| 1
|
699,855
| 24,034,929,086
|
IssuesEvent
|
2022-09-15 18:13:52
|
bryntum/support
|
https://api.github.com/repos/bryntum/support
|
closed
|
TimeRange disappears when moving it causes scroll
|
bug invalid high-priority bugbash
|
Need to retain its elements to prevent it from being released
|
1.0
|
TimeRange disappears when moving it causes scroll - Need to retain its elements to prevent it from being released
|
non_test
|
timerange disappears when moving it causes scroll need to retain its elements to prevent it from being released
| 0
|
205,164
| 15,964,654,857
|
IssuesEvent
|
2021-04-16 06:36:38
|
hojiefeng/pe
|
https://api.github.com/repos/hojiefeng/pe
|
opened
|
Up command oddness
|
severity.VeryLow type.DocumentationBug
|
up command is described as /up{X}, but the help command has a similar functionality, allowing help 123 with the space. A bit inconsistent.
<!--session: 1618552647352-d6bf609c-8ee0-46b3-83d3-d924c046401a-->
|
1.0
|
Up command oddness - up command is described as /up{X}, but the help command has a similar functionality, allowing help 123 with the space. A bit inconsistent.
<!--session: 1618552647352-d6bf609c-8ee0-46b3-83d3-d924c046401a-->
|
non_test
|
up command oddness up command is described as up x but the help command has a similar functionality allowing help with the space a bit inconsistent
| 0
|
320,199
| 27,425,300,147
|
IssuesEvent
|
2023-03-01 19:48:48
|
Kong/kubernetes-ingress-controller
|
https://api.github.com/repos/Kong/kubernetes-ingress-controller
|
closed
|
Add envtest based tests for `KongAdminAPIServiceReconciler`
|
area/tests
|
### Problem statement
After the introduction of support for Admin API service discovery #702 and hence single controller deployments (partially via `KongAdminAPIServiceReconciler`) we need to add tests for this functionality.
Some of the related functionality is already being tests with
- UTs e.g. https://github.com/Kong/kubernetes-ingress-controller/blob/93fc98efc6249dfaaa65420b76f22e36ddf602ba/internal/adminapi/endpoints.go
- of E2E tests e.g. https://github.com/Kong/kubernetes-ingress-controller/blob/b9b9256fcdc3b9607a3f27aa96d6af052c219af5/test/e2e/all_in_one_test.go#L318-L383
but ideally we would like like to add some integration tests. This unfortunately is not easy because service discovery is being done via EndpointSlice watch which returns cluster internal addresses.
### Proposed solution
Add [`envtest`](https://cluster-api.sigs.k8s.io/developer/testing.html#envtest) based tests to test predicates about `KongAdminAPIServiceReconciler`.
|
1.0
|
Add envtest based tests for `KongAdminAPIServiceReconciler` - ### Problem statement
After the introduction of support for Admin API service discovery #702 and hence single controller deployments (partially via `KongAdminAPIServiceReconciler`) we need to add tests for this functionality.
Some of the related functionality is already being tests with
- UTs e.g. https://github.com/Kong/kubernetes-ingress-controller/blob/93fc98efc6249dfaaa65420b76f22e36ddf602ba/internal/adminapi/endpoints.go
- of E2E tests e.g. https://github.com/Kong/kubernetes-ingress-controller/blob/b9b9256fcdc3b9607a3f27aa96d6af052c219af5/test/e2e/all_in_one_test.go#L318-L383
but ideally we would like like to add some integration tests. This unfortunately is not easy because service discovery is being done via EndpointSlice watch which returns cluster internal addresses.
### Proposed solution
Add [`envtest`](https://cluster-api.sigs.k8s.io/developer/testing.html#envtest) based tests to test predicates about `KongAdminAPIServiceReconciler`.
|
test
|
add envtest based tests for kongadminapiservicereconciler problem statement after the introduction of support for admin api service discovery and hence single controller deployments partially via kongadminapiservicereconciler we need to add tests for this functionality some of the related functionality is already being tests with uts e g of tests e g but ideally we would like like to add some integration tests this unfortunately is not easy because service discovery is being done via endpointslice watch which returns cluster internal addresses proposed solution add based tests to test predicates about kongadminapiservicereconciler
| 1
|
150,774
| 19,619,211,853
|
IssuesEvent
|
2022-01-07 02:39:55
|
snykiotcubedev/arangodb-3.7.6
|
https://api.github.com/repos/snykiotcubedev/arangodb-3.7.6
|
opened
|
CVE-2022-0122 (Medium) detected in node-forge-0.9.0.tgz
|
security vulnerability
|
## CVE-2022-0122 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.9.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz</a></p>
<p>
Dependency Hierarchy:
- webpack-dev-server-3.11.0.tgz (Root Library)
- selfsigned-1.10.7.tgz
- :x: **node-forge-0.9.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
forge is vulnerable to URL Redirection to Untrusted Site
<p>Publish Date: 2022-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0122>CVE-2022-0122</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://huntr.dev/bounties/41852c50-3c6d-4703-8c55-4db27164a4ae/">https://huntr.dev/bounties/41852c50-3c6d-4703-8c55-4db27164a4ae/</a></p>
<p>Release Date: 2022-01-06</p>
<p>Fix Resolution: forge - v1.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-0122 (Medium) detected in node-forge-0.9.0.tgz - ## CVE-2022-0122 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.9.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz</a></p>
<p>
Dependency Hierarchy:
- webpack-dev-server-3.11.0.tgz (Root Library)
- selfsigned-1.10.7.tgz
- :x: **node-forge-0.9.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
forge is vulnerable to URL Redirection to Untrusted Site
<p>Publish Date: 2022-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0122>CVE-2022-0122</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://huntr.dev/bounties/41852c50-3c6d-4703-8c55-4db27164a4ae/">https://huntr.dev/bounties/41852c50-3c6d-4703-8c55-4db27164a4ae/</a></p>
<p>Release Date: 2022-01-06</p>
<p>Fix Resolution: forge - v1.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in node forge tgz cve medium severity vulnerability vulnerable library node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href dependency hierarchy webpack dev server tgz root library selfsigned tgz x node forge tgz vulnerable library found in base branch main vulnerability details forge is vulnerable to url redirection to untrusted site publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution forge step up your open source security game with whitesource
| 0
|
727,242
| 25,027,451,952
|
IssuesEvent
|
2022-11-04 09:22:33
|
battlefield-portal-community/bfportal.gg
|
https://api.github.com/repos/battlefield-portal-community/bfportal.gg
|
opened
|
new experiences not loading when the user has scrolled down
|
Priority- High 🥇 type:frontend 🏞
|
Sometimes when the user scrolls downs all the way the page is stuck on `Load More` and just stays there.
There are **no errors** in the console

|
1.0
|
new experiences not loading when the user has scrolled down - Sometimes when the user scrolls downs all the way the page is stuck on `Load More` and just stays there.
There are **no errors** in the console

|
non_test
|
new experiences not loading when the user has scrolled down sometimes when the user scrolls downs all the way the page is stuck on load more and just stays there there are no errors in the console
| 0
|
28,809
| 8,211,916,254
|
IssuesEvent
|
2018-09-04 14:58:37
|
quicklisp/quicklisp-projects
|
https://api.github.com/repos/quicklisp/quicklisp-projects
|
closed
|
Please add illusion
|
canbuild
|
Illusion is a library for customization and management of Lisp left paren reader.
https://github.com/ailisp/illusion
MIT License
|
1.0
|
Please add illusion - Illusion is a library for customization and management of Lisp left paren reader.
https://github.com/ailisp/illusion
MIT License
|
non_test
|
please add illusion illusion is a library for customization and management of lisp left paren reader mit license
| 0
|
216,715
| 16,666,494,860
|
IssuesEvent
|
2021-06-07 05:10:35
|
gitpod-io/website
|
https://api.github.com/repos/gitpod-io/website
|
opened
|
Provide screenshots with new branding for unlinked docs pages
|
documentation
|
For SEO reasons and to make sure existing links continue to work, we have a number of docs pages in `src/routes/docs` that are not linked in the new navigation. For example, `teams.md`.
Some of these pages contain screenshots with the blue Gitpod branding.
As part of this issue:
- [ ] Review all `src/routes/docs/*.md` files and ensure all screenshots reference the kumquat branding
- [ ] Update the docs navigation to find a place for all `*.md` docs pages
|
1.0
|
Provide screenshots with new branding for unlinked docs pages - For SEO reasons and to make sure existing links continue to work, we have a number of docs pages in `src/routes/docs` that are not linked in the new navigation. For example, `teams.md`.
Some of these pages contain screenshots with the blue Gitpod branding.
As part of this issue:
- [ ] Review all `src/routes/docs/*.md` files and ensure all screenshots reference the kumquat branding
- [ ] Update the docs navigation to find a place for all `*.md` docs pages
|
non_test
|
provide screenshots with new branding for unlinked docs pages for seo reasons and to make sure existing links continue to work we have a number of docs pages in src routes docs that are not linked in the new navigation for example teams md some of these pages contain screenshots with the blue gitpod branding as part of this issue review all src routes docs md files and ensure all screenshots reference the kumquat branding update the docs navigation to find a place for all md docs pages
| 0
|
77,052
| 7,558,961,449
|
IssuesEvent
|
2018-04-20 01:06:51
|
s-newman/skitter
|
https://api.github.com/repos/s-newman/skitter
|
closed
|
Test username changing
|
required test
|
We need to ensure that username changes take place as expected. This will test the `/changeUserName` API endpoint.
|
1.0
|
Test username changing - We need to ensure that username changes take place as expected. This will test the `/changeUserName` API endpoint.
|
test
|
test username changing we need to ensure that username changes take place as expected this will test the changeusername api endpoint
| 1
|
235,768
| 19,427,801,029
|
IssuesEvent
|
2021-12-21 08:26:02
|
Componolit/RecordFlux
|
https://api.github.com/repos/Componolit/RecordFlux
|
opened
|
Nightly tests unreliable due to Memcached server
|
bug testing
|
The nightly tests fail [frequently](https://github.com/Componolit/RecordFlux/actions/workflows/tests.yml?query=event%3Aschedule) because of [connection issues](https://github.com/Componolit/RecordFlux/runs/4590016614?check_suite_focus=true) to the Memcached server.
|
1.0
|
Nightly tests unreliable due to Memcached server - The nightly tests fail [frequently](https://github.com/Componolit/RecordFlux/actions/workflows/tests.yml?query=event%3Aschedule) because of [connection issues](https://github.com/Componolit/RecordFlux/runs/4590016614?check_suite_focus=true) to the Memcached server.
|
test
|
nightly tests unreliable due to memcached server the nightly tests fail because of to the memcached server
| 1
|
68,148
| 7,088,323,770
|
IssuesEvent
|
2018-01-11 21:05:23
|
CoastalResilienceNetwork/GeositeFramework
|
https://api.github.com/repos/CoastalResilienceNetwork/GeositeFramework
|
closed
|
Single vs multiple app mode in develop branch
|
tested/verified
|
Casey - just want to make sure the develop branch can do both single and multiple app mode. I just tried --framework-branch=develop and got this:

I had to deploy the master framework in order to produce the dev oceanwealth region. I suspect we will need both modes working well on the develop branch to test both the super app in Hawaii as well as the in-app print (Jacob will test on dev oceanwealth).
|
1.0
|
Single vs multiple app mode in develop branch - Casey - just want to make sure the develop branch can do both single and multiple app mode. I just tried --framework-branch=develop and got this:

I had to deploy the master framework in order to produce the dev oceanwealth region. I suspect we will need both modes working well on the develop branch to test both the super app in Hawaii as well as the in-app print (Jacob will test on dev oceanwealth).
|
test
|
single vs multiple app mode in develop branch casey just want to make sure the develop branch can do both single and multiple app mode i just tried framework branch develop and got this i had to deploy the master framework in order to produce the dev oceanwealth region i suspect we will need both modes working well on the develop branch to test both the super app in hawaii as well as the in app print jacob will test on dev oceanwealth
| 1
|
46,102
| 9,883,283,854
|
IssuesEvent
|
2019-06-24 19:00:09
|
mozilla-mobile/fenix
|
https://api.github.com/repos/mozilla-mobile/fenix
|
opened
|
Migrate to lib-state in search_fragment
|
🤒 code health
|
With the addition of `lib-state` to android-components we want to migrate the search_fragment and document the process
|
1.0
|
Migrate to lib-state in search_fragment - With the addition of `lib-state` to android-components we want to migrate the search_fragment and document the process
|
non_test
|
migrate to lib state in search fragment with the addition of lib state to android components we want to migrate the search fragment and document the process
| 0
|
210,488
| 16,373,677,882
|
IssuesEvent
|
2021-05-15 17:06:46
|
aaugustin/websockets
|
https://api.github.com/repos/aaugustin/websockets
|
closed
|
Handling disconnected clients and terminating the connection handler
|
documentation question
|
I am a newcomer to asyncio and websockets. I am developing a server where clients connect and from time to time a message is broadcast to all of them.
I am using a connection handler that looks something like this:
```python
clients = set()
async def handle_client(websocket, path):
clients.add(websocket)
while True:
await asyncio.sleep(60)
```
Meanwhile, occasionally I will enumerate over `clients` and call `client.send()`. At this point, `client.send()` may raise a `ConnectionClosed` exception if the corresponding client has disconnected.
First, how do I properly handle this exception to remove the client from the connection set? I was thinking it would be something like this, but it isn't quite right.
```python
done, _ = \
await asyncio.wait([client.send(message) for client in clients])
for task in done:
if isinstance(task.exception(), websockets.ConnectionClosed):
# I don't see such an attribute that I can use to get the associated instance
clients.remove(task.exception().websocket)
```
Secondly, I believe my connection handler routine will forever be in its wait loop even after the client is disconnected. I could change the loop to something like:
```python
while websocket in clients:
await asyncio.sleep(60)
```
Is that the best option available to me? It seems like I could also send occasional pings from the connection handler but I would rather not generate excess traffic if it can be avoided.
The [counter.py example](https://github.com/aaugustin/websockets/blob/master/example/counter.py) is similar but because it calls `send()` in its connection handler, it receives the `ConnectionClosed` exception there.
|
1.0
|
Handling disconnected clients and terminating the connection handler - I am a newcomer to asyncio and websockets. I am developing a server where clients connect and from time to time a message is broadcast to all of them.
I am using a connection handler that looks something like this:
```python
clients = set()
async def handle_client(websocket, path):
clients.add(websocket)
while True:
await asyncio.sleep(60)
```
Meanwhile, occasionally I will enumerate over `clients` and call `client.send()`. At this point, `client.send()` may raise a `ConnectionClosed` exception if the corresponding client has disconnected.
First, how do I properly handle this exception to remove the client from the connection set? I was thinking it would be something like this, but it isn't quite right.
```python
done, _ = \
await asyncio.wait([client.send(message) for client in clients])
for task in done:
if isinstance(task.exception(), websockets.ConnectionClosed):
# I don't see such an attribute that I can use to get the associated instance
clients.remove(task.exception().websocket)
```
Secondly, I believe my connection handler routine will forever be in its wait loop even after the client is disconnected. I could change the loop to something like:
```python
while websocket in clients:
await asyncio.sleep(60)
```
Is that the best option available to me? It seems like I could also send occasional pings from the connection handler but I would rather not generate excess traffic if it can be avoided.
The [counter.py example](https://github.com/aaugustin/websockets/blob/master/example/counter.py) is similar but because it calls `send()` in its connection handler, it receives the `ConnectionClosed` exception there.
|
non_test
|
handling disconnected clients and terminating the connection handler i am a newcomer to asyncio and websockets i am developing a server where clients connect and from time to time a message is broadcast to all of them i am using a connection handler that looks something like this python clients set async def handle client websocket path clients add websocket while true await asyncio sleep meanwhile occasionally i will enumerate over clients and call client send at this point client send may raise a connectionclosed exception if the corresponding client has disconnected first how do i properly handle this exception to remove the client from the connection set i was thinking it would be something like this but it isn t quite right python done await asyncio wait for task in done if isinstance task exception websockets connectionclosed i don t see such an attribute that i can use to get the associated instance clients remove task exception websocket secondly i believe my connection handler routine will forever be in its wait loop even after the client is disconnected i could change the loop to something like python while websocket in clients await asyncio sleep is that the best option available to me it seems like i could also send occasional pings from the connection handler but i would rather not generate excess traffic if it can be avoided the is similar but because it calls send in its connection handler it receives the connectionclosed exception there
| 0
|
12,038
| 7,635,317,532
|
IssuesEvent
|
2018-05-07 02:42:22
|
godotengine/godot
|
https://api.github.com/repos/godotengine/godot
|
closed
|
Feature Request: Allow us to drag assets from desktop folder to folder in FileSystem in godot sidebar
|
enhancement topic:editor usability
|
**Godot version:**
<!-- Specify commit hash if non-official. -->
3.x
**OS/device including version:**
Windows 10
**Suggestion description:**
Allow us to drag an asset from any desktop folder directly into the sidebar of the filesystem. That way say you don't have the current folder open and you just downloaded a new font, and you wanted to quickly drag it to the project without needing to open the folder since you have that folder already opened in godot directly.
**Why would it be useful**
Sometimes you don't always have the project folder open on your desktop and you're already looking through your assets and you just want to drag them directly into the project since godot's already open. We download free assets offline all the time, it'll save us a bit of time being able to do this without an extra step.
|
True
|
Feature Request: Allow us to drag assets from desktop folder to folder in FileSystem in godot sidebar - **Godot version:**
<!-- Specify commit hash if non-official. -->
3.x
**OS/device including version:**
Windows 10
**Suggestion description:**
Allow us to drag an asset from any desktop folder directly into the sidebar of the filesystem. That way say you don't have the current folder open and you just downloaded a new font, and you wanted to quickly drag it to the project without needing to open the folder since you have that folder already opened in godot directly.
**Why would it be useful**
Sometimes you don't always have the project folder open on your desktop and you're already looking through your assets and you just want to drag them directly into the project since godot's already open. We download free assets offline all the time, it'll save us a bit of time being able to do this without an extra step.
|
non_test
|
feature request allow us to drag assets from desktop folder to folder in filesystem in godot sidebar godot version x os device including version windows suggestion description allow us to drag an asset from any desktop folder directly into the sidebar of the filesystem that way say you don t have the current folder open and you just downloaded a new font and you wanted to quickly drag it to the project without needing to open the folder since you have that folder already opened in godot directly why would it be useful sometimes you don t always have the project folder open on your desktop and you re already looking through your assets and you just want to drag them directly into the project since godot s already open we download free assets offline all the time it ll save us a bit of time being able to do this without an extra step
| 0
|
179,261
| 13,853,700,559
|
IssuesEvent
|
2020-10-15 08:30:35
|
tracim/tracim
|
https://api.github.com/repos/tracim/tracim
|
closed
|
Fix linting errors in tracim_frontend_lib
|
frontend manually tested not in changelog
|
## Impacted components
- *frontend*
## Extra information
We will also need to reenable linting checks on tracim_frontend_lib.
|
1.0
|
Fix linting errors in tracim_frontend_lib - ## Impacted components
- *frontend*
## Extra information
We will also need to reenable linting checks on tracim_frontend_lib.
|
test
|
fix linting errors in tracim frontend lib impacted components frontend extra information we will also need to reenable linting checks on tracim frontend lib
| 1
|
75,828
| 7,494,416,080
|
IssuesEvent
|
2018-04-07 09:25:10
|
ring-project/ring
|
https://api.github.com/repos/ring-project/ring
|
closed
|
move "Find someone" search field to Dialer section only
|
next ready for testing
|
- Remove "Find someone" search field from all other sections (bookmarks, contacts and timeline).
- Add "Find someone" to the dialer section.
That makes the dialer section consistent and people aware if they want to search for people or make calls to exisiting people they use the dialer pad or the "find someone" search field above in that section.
|
1.0
|
move "Find someone" search field to Dialer section only - - Remove "Find someone" search field from all other sections (bookmarks, contacts and timeline).
- Add "Find someone" to the dialer section.
That makes the dialer section consistent and people aware if they want to search for people or make calls to exisiting people they use the dialer pad or the "find someone" search field above in that section.
|
test
|
move find someone search field to dialer section only remove find someone search field from all other sections bookmarks contacts and timeline add find someone to the dialer section that makes the dialer section consistent and people aware if they want to search for people or make calls to exisiting people they use the dialer pad or the find someone search field above in that section
| 1
|
101,574
| 21,718,432,425
|
IssuesEvent
|
2022-05-10 20:29:33
|
sourcegraph/sourcegraph
|
https://api.github.com/repos/sourcegraph/sourcegraph
|
opened
|
[Accessibility]: Timeline steps are not numbered
|
team/code-intelligence accessibility wcag/2.1/fixing wcag/2.1
|
### Audit type
Screen reader navigation
### User journey audit issue
https://github.com/sourcegraph/sourcegraph/issues/33513, https://github.com/sourcegraph/sourcegraph/issues/33512
### Problem description
The timeline steps are not part of a list or have any auditory numbering, with the only indication of a list-like element approaching being the word "Timeline" at a stretch. 10:00 in the recording.
### Expected behavior
Timelines should number each step directly (hard-coded numbers) or indirectly (via list elements) to make it clear where in the timeline one is and how much there is to navigate.
### Additional details
_No response_
### Assigning labels
- [X] Please give this issue an estimate by applying a label like `estimate/Xd`, where X is the estimated number of days it will take to complete.
- [X] If this issue is specific to a specific Sourcegraph product, please assign the appropriate team label to this issue.
- [X] If this issue will require input from designers in order to complete, please assign the label `needs-design`.
- [X] If you are confident that this issue should be fixed by GitStart, please assign the label `gitstart`.
### Owner
This issue will be fixed by my team, I have assigned a relevant member to it, or I will do so in the near future.
|
1.0
|
[Accessibility]: Timeline steps are not numbered - ### Audit type
Screen reader navigation
### User journey audit issue
https://github.com/sourcegraph/sourcegraph/issues/33513, https://github.com/sourcegraph/sourcegraph/issues/33512
### Problem description
The timeline steps are not part of a list or have any auditory numbering, with the only indication of a list-like element approaching being the word "Timeline" at a stretch. 10:00 in the recording.
### Expected behavior
Timelines should number each step directly (hard-coded numbers) or indirectly (via list elements) to make it clear where in the timeline one is and how much there is to navigate.
### Additional details
_No response_
### Assigning labels
- [X] Please give this issue an estimate by applying a label like `estimate/Xd`, where X is the estimated number of days it will take to complete.
- [X] If this issue is specific to a specific Sourcegraph product, please assign the appropriate team label to this issue.
- [X] If this issue will require input from designers in order to complete, please assign the label `needs-design`.
- [X] If you are confident that this issue should be fixed by GitStart, please assign the label `gitstart`.
### Owner
This issue will be fixed by my team, I have assigned a relevant member to it, or I will do so in the near future.
|
non_test
|
timeline steps are not numbered audit type screen reader navigation user journey audit issue problem description the timeline steps are not part of a list or have any auditory numbering with the only indication of a list like element approaching being the word timeline at a stretch in the recording expected behavior timelines should number each step directly hard coded numbers or indirectly via list elements to make it clear where in the timeline one is and how much there is to navigate additional details no response assigning labels please give this issue an estimate by applying a label like estimate xd where x is the estimated number of days it will take to complete if this issue is specific to a specific sourcegraph product please assign the appropriate team label to this issue if this issue will require input from designers in order to complete please assign the label needs design if you are confident that this issue should be fixed by gitstart please assign the label gitstart owner this issue will be fixed by my team i have assigned a relevant member to it or i will do so in the near future
| 0
|
355,625
| 25,175,983,346
|
IssuesEvent
|
2022-11-11 09:18:21
|
charleslimjh/pe
|
https://api.github.com/repos/charleslimjh/pe
|
opened
|
Div tag appears in edit command in UG
|
severity.VeryLow type.DocumentationBug
|
Div tag appears in the `notes about patient type` segment. Looks like a minor formatting issue.

<!--session: 1668153050766-4a2f8d15-aea5-4467-8c57-3430516cb045-->
<!--Version: Web v3.4.4-->
|
1.0
|
Div tag appears in edit command in UG - Div tag appears in the `notes about patient type` segment. Looks like a minor formatting issue.

<!--session: 1668153050766-4a2f8d15-aea5-4467-8c57-3430516cb045-->
<!--Version: Web v3.4.4-->
|
non_test
|
div tag appears in edit command in ug div tag appears in the notes about patient type segment looks like a minor formatting issue
| 0
|
237,219
| 19,601,174,148
|
IssuesEvent
|
2022-01-06 01:38:59
|
elisebeall/dnd-spells
|
https://api.github.com/repos/elisebeall/dnd-spells
|
closed
|
[🧑💻] Testing : Create data file
|
testing
|
Create data file for early stages of development, as well as Cypress testing
|
1.0
|
[🧑💻] Testing : Create data file - Create data file for early stages of development, as well as Cypress testing
|
test
|
testing create data file create data file for early stages of development as well as cypress testing
| 1
|
496,991
| 14,360,088,073
|
IssuesEvent
|
2020-11-30 16:24:48
|
okTurtles/group-income-simple
|
https://api.github.com/repos/okTurtles/group-income-simple
|
closed
|
Unable to communicate with server(on trying to log in) using Ngrok
|
App:Backend App:Frontend Kind:Core Level:Advanced Priority:High
|
### Problem
Trying to test `grunt dev` server on mobile phone using [Ngrok](https://ngrok.com/) produces some issues(it appears that it fails to communicate with the server upon requests).
1) When trying to log in, with an account already created by other device(PC).

2) When trying to sign up on mobile device.

P.S.
@taoeffect I tried the other way you advised me first(connecting the phone with the same wifi as the laptop and then trying http://ipv4_address:8000 which worked this time), and also was presented with the same error messages.
### Solution
Close #852
|
1.0
|
Unable to communicate with server(on trying to log in) using Ngrok - ### Problem
Trying to test `grunt dev` server on mobile phone using [Ngrok](https://ngrok.com/) produces some issues(it appears that it fails to communicate with the server upon requests).
1) When trying to log in, with an account already created by other device(PC).

2) When trying to sign up on mobile device.

P.S.
@taoeffect I tried the other way you advised me first(connecting the phone with the same wifi as the laptop and then trying http://ipv4_address:8000 which worked this time), and also was presented with the same error messages.
### Solution
Close #852
|
non_test
|
unable to communicate with server on trying to log in using ngrok problem trying to test grunt dev server on mobile phone using produces some issues it appears that it fails to communicate with the server upon requests when trying to log in with an account already created by other device pc when trying to sign up on mobile device p s taoeffect i tried the other way you advised me first connecting the phone with the same wifi as the laptop and then trying which worked this time and also was presented with the same error messages solution close
| 0
|
227,845
| 18,104,402,879
|
IssuesEvent
|
2021-09-22 17:31:45
|
coyim/coyim
|
https://api.github.com/repos/coyim/coyim
|
closed
|
Add rooms to @coy.im
|
task Priority: Now Estimate - small State: Test in-user-test Release pre-release 0.4
|
We should support MUC on the XMPP server for coy.im, so that we can create #support and #welcome rooms there, at least.
|
2.0
|
Add rooms to @coy.im - We should support MUC on the XMPP server for coy.im, so that we can create #support and #welcome rooms there, at least.
|
test
|
add rooms to coy im we should support muc on the xmpp server for coy im so that we can create support and welcome rooms there at least
| 1
|
154,126
| 12,194,128,359
|
IssuesEvent
|
2020-04-29 15:21:52
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachtest: sqlsmith/setup=empty/setting=no-ddl failed
|
C-test-failure O-roachtest O-robot branch-master release-blocker
|
[(roachtest).sqlsmith/setup=empty/setting=no-ddl failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1904546&tab=buildLog) on [master@c3bb2bce3d69d0ceef20797d856a6ac107a3ef47](https://github.com/cockroachdb/cockroach/commits/c3bb2bce3d69d0ceef20797d856a6ac107a3ef47):
```
The test failed on branch=master, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/sqlsmith/setup=empty/setting=no-ddl/run_1
sqlsmith.go:169,sqlsmith.go:199,test_runner.go:753: error: pq: internal error: runtime error: invalid memory address or nil pointer dereference
stmt:
WITH
with_1907 (col_10268)
AS (
SELECT
*
FROM
(
VALUES
((e'\x1d\x1bi|\x11\x11p\x1b0':::STRING::STRING || '':::STRING::STRING)::STRING),
(e'\x0e0pM':::STRING),
(NULL),
(abbrev('151.127.149.30/25':::INET::INET)::STRING)
)
AS tab_5881 (col_10268)
)
SELECT
tab_5882.col_10269 AS col_10271, tab_5882.col_10270 AS col_10272
FROM
(
VALUES
('0101000020E6100000000000000000F03F000000000000F03F':::GEOGRAPHY(GEOMETRY,4326), '4714-11-24 BC':::DATE),
(st_geogfromgeojson('null':::JSONB::JSONB)::GEOGRAPHY(GEOMETRY,4326), '1984-06-06':::DATE),
('0101000020E6100000000000000000F03F000000000000F03F':::GEOGRAPHY(GEOMETRY,4326), '1996-07-07':::DATE),
('0101000020E6100000000000000000F03F000000000000F03F':::GEOGRAPHY(GEOMETRY,4326), '1984-07-23':::DATE),
('0101000020E6100000000000000000F03F000000000000F03F':::GEOGRAPHY(GEOMETRY,4326), '1986-05-01':::DATE),
('0101000020E6100000000000000000F03F000000000000F03F':::GEOGRAPHY(GEOMETRY,4326), '1971-12-31':::DATE)
)
AS tab_5882 (col_10269, col_10270)
ORDER BY
tab_5882.col_10270 ASC, tab_5882.col_10269 ASC, tab_5882.col_10270 ASC, tab_5882.col_10269;
```
<details><summary>More</summary><p>
Artifacts: [/sqlsmith/setup=empty/setting=no-ddl](https://teamcity.cockroachdb.com/viewLog.html?buildId=1904546&tab=artifacts#/sqlsmith/setup=empty/setting=no-ddl)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Asqlsmith%2Fsetup%3Dempty%2Fsetting%3Dno-ddl.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
2.0
|
roachtest: sqlsmith/setup=empty/setting=no-ddl failed - [(roachtest).sqlsmith/setup=empty/setting=no-ddl failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1904546&tab=buildLog) on [master@c3bb2bce3d69d0ceef20797d856a6ac107a3ef47](https://github.com/cockroachdb/cockroach/commits/c3bb2bce3d69d0ceef20797d856a6ac107a3ef47):
```
The test failed on branch=master, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/sqlsmith/setup=empty/setting=no-ddl/run_1
sqlsmith.go:169,sqlsmith.go:199,test_runner.go:753: error: pq: internal error: runtime error: invalid memory address or nil pointer dereference
stmt:
WITH
with_1907 (col_10268)
AS (
SELECT
*
FROM
(
VALUES
((e'\x1d\x1bi|\x11\x11p\x1b0':::STRING::STRING || '':::STRING::STRING)::STRING),
(e'\x0e0pM':::STRING),
(NULL),
(abbrev('151.127.149.30/25':::INET::INET)::STRING)
)
AS tab_5881 (col_10268)
)
SELECT
tab_5882.col_10269 AS col_10271, tab_5882.col_10270 AS col_10272
FROM
(
VALUES
('0101000020E6100000000000000000F03F000000000000F03F':::GEOGRAPHY(GEOMETRY,4326), '4714-11-24 BC':::DATE),
(st_geogfromgeojson('null':::JSONB::JSONB)::GEOGRAPHY(GEOMETRY,4326), '1984-06-06':::DATE),
('0101000020E6100000000000000000F03F000000000000F03F':::GEOGRAPHY(GEOMETRY,4326), '1996-07-07':::DATE),
('0101000020E6100000000000000000F03F000000000000F03F':::GEOGRAPHY(GEOMETRY,4326), '1984-07-23':::DATE),
('0101000020E6100000000000000000F03F000000000000F03F':::GEOGRAPHY(GEOMETRY,4326), '1986-05-01':::DATE),
('0101000020E6100000000000000000F03F000000000000F03F':::GEOGRAPHY(GEOMETRY,4326), '1971-12-31':::DATE)
)
AS tab_5882 (col_10269, col_10270)
ORDER BY
tab_5882.col_10270 ASC, tab_5882.col_10269 ASC, tab_5882.col_10270 ASC, tab_5882.col_10269;
```
<details><summary>More</summary><p>
Artifacts: [/sqlsmith/setup=empty/setting=no-ddl](https://teamcity.cockroachdb.com/viewLog.html?buildId=1904546&tab=artifacts#/sqlsmith/setup=empty/setting=no-ddl)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Asqlsmith%2Fsetup%3Dempty%2Fsetting%3Dno-ddl.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
test
|
roachtest sqlsmith setup empty setting no ddl failed on the test failed on branch master cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts sqlsmith setup empty setting no ddl run sqlsmith go sqlsmith go test runner go error pq internal error runtime error invalid memory address or nil pointer dereference stmt with with col as select from values e string string string string string e string null abbrev inet inet string as tab col select tab col as col tab col as col from values geography geometry bc date st geogfromgeojson null jsonb jsonb geography geometry date geography geometry date geography geometry date geography geometry date geography geometry date as tab col col order by tab col asc tab col asc tab col asc tab col more artifacts powered by
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.