Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 4 112 | repo_url stringlengths 33 141 | action stringclasses 3 values | title stringlengths 1 1.02k | labels stringlengths 4 1.54k | body stringlengths 1 262k | index stringclasses 17 values | text_combine stringlengths 95 262k | label stringclasses 2 values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
215,570 | 16,679,154,648 | IssuesEvent | 2021-06-07 20:26:23 | uoForms/App-CANBeWell | https://api.github.com/repos/uoForms/App-CANBeWell | closed | 'Family Planning' missing under Sexual and reproductive health in TM | Follow up with client To be tested | Its because of the missing age for 'Family planning'- (Transmasculine) in the excel sheet. | 1.0 | 'Family Planning' missing under Sexual and reproductive health in TM - Its because of the missing age for 'Family planning'- (Transmasculine) in the excel sheet. | test | family planning missing under sexual and reproductive health in tm its because of the missing age for family planning transmasculine in the excel sheet | 1 |
237,164 | 19,596,634,074 | IssuesEvent | 2022-01-05 18:40:45 | Arch666Angel/mods | https://api.github.com/repos/Arch666Angel/mods | closed | No slot for bio-token in T2 exploration lab | Impact: Bug Angels Industries (Technology mode) Angels Unit Tests |

**Describe the bug**
No bio-token slot in science overhaul lab
| 1.0 | No slot for bio-token in T2 exploration lab -

**Describe the bug**
No bio-token slot in science overhaul lab
| test | no slot for bio token in exploration lab describe the bug no bio token slot in science overhaul lab | 1 |
315,964 | 27,122,711,764 | IssuesEvent | 2023-02-16 00:58:39 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | opened | DISABLED test_to_cpu_blocking_by_default (__main__.TestCuda) | module: cuda triaged module: flaky-tests skipped | Platforms: rocm
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_to_cpu_blocking_by_default&suite=TestCuda) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/11365293941).
Over the past 3 hours, it has been determined flaky in 4 workflow(s) with 4 failures and 4 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_to_cpu_blocking_by_default`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
Test file path: `test_cuda.py` | 1.0 | DISABLED test_to_cpu_blocking_by_default (__main__.TestCuda) - Platforms: rocm
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_to_cpu_blocking_by_default&suite=TestCuda) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/11365293941).
Over the past 3 hours, it has been determined flaky in 4 workflow(s) with 4 failures and 4 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_to_cpu_blocking_by_default`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
Test file path: `test_cuda.py` | test | disabled test to cpu blocking by default main testcuda platforms rocm this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not assume things are okay if the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test to cpu blocking by default there should be several instances run as flaky tests are rerun in ci from which you can study the logs test file path test cuda py | 1 |
19,497 | 3,774,165,269 | IssuesEvent | 2016-03-17 07:49:56 | dojo/core | https://api.github.com/repos/dojo/core | closed | Issue with location of helpers... | bug - test case | It appears some of the changes made to Intern (specifically how it deals with dependencies) have caused issues with the location of some files and the [CI is failing](https://travis-ci.org/dojo/core/builds/116074773#L225).
```
Error: Failed to load module leadfoot/helpers/pollUntil from /home/travis/build/dojo/core/leadfoot/helpers/pollUntil.js (parent: tests/functional/text/textPlugin)
at reportModuleLoadError <node_modules/dojo-loader/loader.ts:122:44>
at loadCallback <node_modules/dojo-loader/loader.ts:884:7>
at executeModule <node_modules/dojo-loader/loader.ts:632:52>
at <node_modules/dojo-loader/loader.ts:620:10>
at Array.map <native>
at executeModule <node_modules/dojo-loader/loader.ts:614:43>
at <node_modules/dojo-loader/loader.ts:620:10>
at Array.map <native>
at executeModule <node_modules/dojo-loader/loader.ts:614:43>
at <node_modules/dojo-loader/loader.ts:620:10>
``` | 1.0 | Issue with location of helpers... - It appears some of the changes made to Intern (specifically how it deals with dependencies) have caused issues with the location of some files and the [CI is failing](https://travis-ci.org/dojo/core/builds/116074773#L225).
```
Error: Failed to load module leadfoot/helpers/pollUntil from /home/travis/build/dojo/core/leadfoot/helpers/pollUntil.js (parent: tests/functional/text/textPlugin)
at reportModuleLoadError <node_modules/dojo-loader/loader.ts:122:44>
at loadCallback <node_modules/dojo-loader/loader.ts:884:7>
at executeModule <node_modules/dojo-loader/loader.ts:632:52>
at <node_modules/dojo-loader/loader.ts:620:10>
at Array.map <native>
at executeModule <node_modules/dojo-loader/loader.ts:614:43>
at <node_modules/dojo-loader/loader.ts:620:10>
at Array.map <native>
at executeModule <node_modules/dojo-loader/loader.ts:614:43>
at <node_modules/dojo-loader/loader.ts:620:10>
``` | test | issue with location of helpers it appears some of the changes made to intern specifically how it deals with dependencies have caused issues with the location of some files and the error failed to load module leadfoot helpers polluntil from home travis build dojo core leadfoot helpers polluntil js parent tests functional text textplugin at reportmoduleloaderror at loadcallback at executemodule at at array map at executemodule at at array map at executemodule at | 1 |
397,681 | 11,731,288,149 | IssuesEvent | 2020-03-10 23:37:49 | IS-AgroSmart/AgroSmart-Web | https://api.github.com/repos/IS-AgroSmart/AgroSmart-Web | opened | Refactor Flight datepicker to use BootstrapVue's native dialog | enhancement low-priority | NewFlight.vue currently uses a plain b-form-input with type="date" to select the Flight date. https://bootstrap-vue.js.org/docs/components/form-datepicker is a control with better format and, in theory, the same functionality.
Expected work:
- [ ] Swap [the date input control](https://github.com/IS-AgroSmart/AgroSmart-Web/blob/281e17d495ae1587f7f70ad221667b05c050e79e/frontend/src/components/NewFlight.vue#L9) for `<b-form-datepicker id="input-2" v-model="form.date"></b-form-datepicker>`
- [ ] The new control should work correctly with the form submission. | 1.0 | Refactor Flight datepicker to use BootstrapVue's native dialog - NewFlight.vue currently uses a plain b-form-input with type="date" to select the Flight date. https://bootstrap-vue.js.org/docs/components/form-datepicker is a control with better format and, in theory, the same functionality.
Expected work:
- [ ] Swap [the date input control](https://github.com/IS-AgroSmart/AgroSmart-Web/blob/281e17d495ae1587f7f70ad221667b05c050e79e/frontend/src/components/NewFlight.vue#L9) for `<b-form-datepicker id="input-2" v-model="form.date"></b-form-datepicker>`
- [ ] The new control should work correctly with the form submission. | non_test | refactor flight datepicker to use bootstrapvue s native dialog newflight vue currently uses a plain b form input with type date to select the flight date is a control with better format and in theory the same functionality expected work swap for the new control should work correctly with the form submission | 0 |
248,102 | 18,858,033,735 | IssuesEvent | 2021-11-12 09:18:38 | kaushikkrdy/pe | https://api.github.com/repos/kaushikkrdy/pe | opened | Marking contact as called user story missing | severity.Low type.DocumentationBug | Marking contact as called exists as a feature in UG and code, but not present in DG User Stories
<!--session: 1636703758671-33d508eb-996f-4ff9-b08f-08310e28f766-->
<!--Version: Web v3.4.1--> | 1.0 | Marking contact as called user story missing - Marking contact as called exists as a feature in UG and code, but not present in DG User Stories
<!--session: 1636703758671-33d508eb-996f-4ff9-b08f-08310e28f766-->
<!--Version: Web v3.4.1--> | non_test | marking contact as called user story missing marking contact as called exists as a feature in ug and code but not present in dg user stories | 0 |
37,945 | 12,510,814,472 | IssuesEvent | 2020-06-02 19:21:51 | autoai-org/AID | https://api.github.com/repos/autoai-org/AID | closed | WS-2019-0367 (Medium) detected in angular-1.4.2.min.js | security vulnerability wontfix | ## WS-2019-0367 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>angular-1.4.2.min.js</b></p></summary>
<p>AngularJS is an MVC framework for building web applications. The core features include HTML enhanced with custom component and data-binding capabilities, dependency injection and strong focus on simplicity, testability, maintainability and boiler-plate reduction.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.4.2/angular.min.js">https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.4.2/angular.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/AID/docs/node_modules/autocomplete.js/test/playground_angular.html</p>
<p>Path to vulnerable library: /AID/docs/node_modules/autocomplete.js/test/playground_angular.html,/AID/docs/node_modules/autocomplete.js/examples/basic_angular.html</p>
<p>
Dependency Hierarchy:
- :x: **angular-1.4.2.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/autoai-org/AID/commit/708c0f68e8540b044f2c44c3ef7e7aff2e34dfef">708c0f68e8540b044f2c44c3ef7e7aff2e34dfef</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype Pollution vulnerability found in Angular before 1.7.9.
<p>Publish Date: 2020-01-08
<p>URL: <a href=https://github.com/RetireJS/retire.js/commit/f07a7557d3fc1c26b86fe11a5b33cb1b8f3dcf2f>WS-2019-0367</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/angular/angular.js/blob/master/CHANGELOG.md#179-pollution-eradication-2019-11-19">https://github.com/angular/angular.js/blob/master/CHANGELOG.md#179-pollution-eradication-2019-11-19</a></p>
<p>Release Date: 2020-01-08</p>
<p>Fix Resolution: angular - 1.7.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2019-0367 (Medium) detected in angular-1.4.2.min.js - ## WS-2019-0367 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>angular-1.4.2.min.js</b></p></summary>
<p>AngularJS is an MVC framework for building web applications. The core features include HTML enhanced with custom component and data-binding capabilities, dependency injection and strong focus on simplicity, testability, maintainability and boiler-plate reduction.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.4.2/angular.min.js">https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.4.2/angular.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/AID/docs/node_modules/autocomplete.js/test/playground_angular.html</p>
<p>Path to vulnerable library: /AID/docs/node_modules/autocomplete.js/test/playground_angular.html,/AID/docs/node_modules/autocomplete.js/examples/basic_angular.html</p>
<p>
Dependency Hierarchy:
- :x: **angular-1.4.2.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/autoai-org/AID/commit/708c0f68e8540b044f2c44c3ef7e7aff2e34dfef">708c0f68e8540b044f2c44c3ef7e7aff2e34dfef</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype Pollution vulnerability found in Angular before 1.7.9.
<p>Publish Date: 2020-01-08
<p>URL: <a href=https://github.com/RetireJS/retire.js/commit/f07a7557d3fc1c26b86fe11a5b33cb1b8f3dcf2f>WS-2019-0367</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/angular/angular.js/blob/master/CHANGELOG.md#179-pollution-eradication-2019-11-19">https://github.com/angular/angular.js/blob/master/CHANGELOG.md#179-pollution-eradication-2019-11-19</a></p>
<p>Release Date: 2020-01-08</p>
<p>Fix Resolution: angular - 1.7.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | ws medium detected in angular min js ws medium severity vulnerability vulnerable library angular min js angularjs is an mvc framework for building web applications the core features include html enhanced with custom component and data binding capabilities dependency injection and strong focus on simplicity testability maintainability and boiler plate reduction library home page a href path to dependency file tmp ws scm aid docs node modules autocomplete js test playground angular html path to vulnerable library aid docs node modules autocomplete js test playground angular html aid docs node modules autocomplete js examples basic angular html dependency hierarchy x angular min js vulnerable library found in head commit a href vulnerability details prototype pollution vulnerability found in angular before publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution angular step up your open source security game with whitesource | 0 |
42,204 | 17,082,734,911 | IssuesEvent | 2021-07-08 07:57:15 | NewTec-GmbH/ZumoComSystem | https://api.github.com/repos/NewTec-GmbH/ZumoComSystem | closed | Implement InitAPMode | SystemServices | **Acceptance Criteria**
- [ ] Architecture designed
- [ ] AP is spawned when WiFiKey is pressed during boot-up
- [ ] AP is spawned when there are no network credentials available for AP mode
- [ ] Transferred data is encrypted
- [ ] IP clients cannot communicate with each other
- [ ] Uses default SSID and PSK
- [ ] DHCP service available
- [ ] DNS service available
- [ ] ComPlatform reachable at "https://complatform.local" | 1.0 | Implement InitAPMode - **Acceptance Criteria**
- [ ] Architecture designed
- [ ] AP is spawned when WiFiKey is pressed during boot-up
- [ ] AP is spawned when there are no network credentials available for AP mode
- [ ] Transferred data is encrypted
- [ ] IP clients cannot communicate with each other
- [ ] Uses default SSID and PSK
- [ ] DHCP service available
- [ ] DNS service available
- [ ] ComPlatform reachable at "https://complatform.local" | non_test | implement initapmode acceptance criteria architecture designed ap is spawned when wifikey is pressed during boot up ap is spawned when there are no network credentials available for ap mode transferred data is encrypted ip clients cannot communicate with each other uses default ssid and psk dhcp service available dns service available complatform reachable at | 0 |
234,465 | 19,184,124,741 | IssuesEvent | 2021-12-04 22:54:51 | Hamlib/Hamlib | https://api.github.com/repos/Hamlib/Hamlib | closed | Kenwood mode set problems | bug needs test critical WSJTX JTDX | TS590S is setting mode when mode does not need to be set causing lots of relay chatter when transmitting with WSJTX.
| 1.0 | Kenwood mode set problems - TS590S is setting mode when mode does not need to be set causing lots of relay chatter when transmitting with WSJTX.
| test | kenwood mode set problems is setting mode when mode does not need to be set causing lots of relay chatter when transmitting with wsjtx | 1 |
31,790 | 5,997,050,905 | IssuesEvent | 2017-06-03 19:58:37 | networkupstools/nut | https://api.github.com/repos/networkupstools/nut | opened | Developer manual: fix typo and formatting in subdriver commands | documentation | http://buildbot.networkupstools.org/~buildbot/docker-debian-jessie/docs/latest//docs/developer-guide.chunked/ar01s04.html#_writing_a_subdriver has a command that ends in `auto` - should be `-a auto`. Also fix formatting of surrounding paragraphs with commands. | 1.0 | Developer manual: fix typo and formatting in subdriver commands - http://buildbot.networkupstools.org/~buildbot/docker-debian-jessie/docs/latest//docs/developer-guide.chunked/ar01s04.html#_writing_a_subdriver has a command that ends in `auto` - should be `-a auto`. Also fix formatting of surrounding paragraphs with commands. | non_test | developer manual fix typo and formatting in subdriver commands has a command that ends in auto should be a auto also fix formatting of surrounding paragraphs with commands | 0 |
59,606 | 17,023,174,963 | IssuesEvent | 2021-07-03 00:42:49 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | Name finder loses cities west of Greenwich | Component: namefinder Priority: major Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 1.39pm, Friday, 24th August 2007]**
-------- Original Message --------
Subject: Re: [OSM-dev] Name finder for the main OSM page?
Date: Fri, 24 Aug 2007 00:21:06 +0100
From: Tom Hughes <tom@compton.nu>
To: dev@openstreetmap.org
I've been playing with this using the three queries that your new
version uses and I've found a couple of issues...
The "towns near" and "places near" queries generally behave fairly
sensibly, and give things near "requestedplace" which has the lat
and lon I supplied.
The "cities near" query does not do this however, and gives me all
sorts of wierd results which are relative to other places and not
the lat and lon I gave. My test case of 51.76,0.0 which is just north
of London gives me Bristol as the first result!
A second problem is that it doesn't cope with wrapping around the
zero meridian - as you can probably guess from that test case I live
about half a mile or so from the meridian and I find that if I'm to
the east of it I only find towns to the east and vice verse when I'm
to the west of it.
| 1.0 | Name finder loses cities west of Greenwich - **[Submitted to the original trac issue database at 1.39pm, Friday, 24th August 2007]**
-------- Original Message --------
Subject: Re: [OSM-dev] Name finder for the main OSM page?
Date: Fri, 24 Aug 2007 00:21:06 +0100
From: Tom Hughes <tom@compton.nu>
To: dev@openstreetmap.org
I've been playing with this using the three queries that your new
version uses and I've found a couple of issues...
The "towns near" and "places near" queries generally behave fairly
sensibly, and give things near "requestedplace" which has the lat
and lon I supplied.
The "cities near" query does not do this however, and gives me all
sorts of wierd results which are relative to other places and not
the lat and lon I gave. My test case of 51.76,0.0 which is just north
of London gives me Bristol as the first result!
A second problem is that it doesn't cope with wrapping around the
zero meridian - as you can probably guess from that test case I live
about half a mile or so from the meridian and I find that if I'm to
the east of it I only find towns to the east and vice verse when I'm
to the west of it.
| non_test | name finder loses cities west of greenwich original message subject re name finder for the main osm page date fri aug from tom hughes to dev openstreetmap org i ve been playing with this using the three queries that your new version uses and i ve found a couple of issues the towns near and places near queries generally behave fairly sensibly and give things near requestedplace which has the lat and lon i supplied the cities near query does not do this however and gives me all sorts of wierd results which are relative to other places and not the lat and lon i gave my test case of which is just north of london gives me bristol as the first result a second problem is that it doesn t cope with wrapping around the zero meridian as you can probably guess from that test case i live about half a mile or so from the meridian and i find that if i m to the east of it i only find towns to the east and vice verse when i m to the west of it | 0 |
65,889 | 6,977,748,798 | IssuesEvent | 2017-12-12 15:32:01 | GoogleCloudPlatform/google-cloud-eclipse | https://api.github.com/repos/GoogleCloudPlatform/google-cloud-eclipse | closed | Does Kokoro only build on googler commits? | high priority testing | Quickly reviewing my PR queue, I don't see any build from Kokoro. Yet I have seen Kokoro builds on PRs from @chanseokoh and @elharo. We should verify that Kokoro is doing builds for external contributors. | 1.0 | Does Kokoro only build on googler commits? - Quickly reviewing my PR queue, I don't see any build from Kokoro. Yet I have seen Kokoro builds on PRs from @chanseokoh and @elharo. We should verify that Kokoro is doing builds for external contributors. | test | does kokoro only build on googler commits quickly reviewing my pr queue i don t see any build from kokoro yet i have seen kokoro builds on prs from chanseokoh and elharo we should verify that kokoro is doing builds for external contributors | 1 |
74,505 | 7,431,472,167 | IssuesEvent | 2018-03-25 15:03:06 | phil-r/react-native-looped-carousel | https://api.github.com/repos/phil-r/react-native-looped-carousel | closed | Fix test error | tests | Tests work, but here is an error in the output:
```
console.error node_modules/fbjs/lib/warning.js:33
Warning: Can only update a mounted or mounting component. This usually means you called setState, replaceState, or forceUpdate on an unmounted component. This is a no-op.
Please check the code for the Carousel component.
``` | 1.0 | Fix test error - Tests work, but here is an error in the output:
```
console.error node_modules/fbjs/lib/warning.js:33
Warning: Can only update a mounted or mounting component. This usually means you called setState, replaceState, or forceUpdate on an unmounted component. This is a no-op.
Please check the code for the Carousel component.
``` | test | fix test error tests work but here is an error in the output console error node modules fbjs lib warning js warning can only update a mounted or mounting component this usually means you called setstate replacestate or forceupdate on an unmounted component this is a no op please check the code for the carousel component | 1 |
82,885 | 7,855,097,180 | IssuesEvent | 2018-06-20 23:36:22 | att/ast | https://api.github.com/repos/att/ast | opened | Switch (or add) one of the Travis environments to 32-bits | compatibility enhancement testing | I've been focusing this past week on making ksh free of compiler lint warnings. It now builds with no compiler warnings on macOS, OpenSuse, and FreeBSD. On Ubuntu 32-bit there is this warning:
```
../src/cmd/ksh93/edit/edit.c: In function ‘sh_ioctl’:
../src/cmd/ksh93/edit/edit.c:1383:22: warning: cast from pointer to integer of different size [-Wpointer-to-int-cast]
Sflong_t l = (Sflong_t)val;
^
```
We should add, or switch one of the existing, Travis test environments to 32-bits to help catch code that incorrectly assumes the length of a `long int` is the same as a pointer. | 1.0 | Switch (or add) one of the Travis environments to 32-bits - I've been focusing this past week on making ksh free of compiler lint warnings. It now builds with no compiler warnings on macOS, OpenSuse, and FreeBSD. On Ubuntu 32-bit there is this warning:
```
../src/cmd/ksh93/edit/edit.c: In function ‘sh_ioctl’:
../src/cmd/ksh93/edit/edit.c:1383:22: warning: cast from pointer to integer of different size [-Wpointer-to-int-cast]
Sflong_t l = (Sflong_t)val;
^
```
We should add, or switch one of the existing, Travis test environments to 32-bits to help catch code that incorrectly assumes the length of a `long int` is the same as a pointer. | test | switch or add one of the travis environments to bits i ve been focusing this past week on making ksh free of compiler lint warnings it now builds with no compiler warnings on macos opensuse and freebsd on ubuntu bit there is this warning src cmd edit edit c in function ‘sh ioctl’ src cmd edit edit c warning cast from pointer to integer of different size sflong t l sflong t val we should add or switch one of the existing travis test environments to bits to help catch code that incorrectly assumes the length of a long int is the same as a pointer | 1 |
19,192 | 10,335,227,139 | IssuesEvent | 2019-09-03 10:04:42 | anhnongdan/Spark1.6_Problems | https://api.github.com/repos/anhnongdan/Spark1.6_Problems | closed | Repartition with and without caching | In/Out Performance Platform best-practice enhancement | **Context: many 'where' look up on a big DF, the result has just 345 records but take 9k steps to calc.**
```
start_d = 20171001
if "hist_df" in locals():
hist_df.unpersist()
hist_df = sqlContext.read.parquet("{}/{}/update_date={}".format(WARE_HOUSE_PATH, HIS, start_d))
hist_df.repartition(100, 'sub_encrypted_phone_number')
#smpl = df.where(df.sub_encrypted_phone_number == 0)
if 'sampled_calls' in locals():
sampled_calls.unpersist()
sampled_calls = sqlContext.createDataFrame([] ,hist_df.schema, samplingRatio=0)
# append call list of visited pn and return cont_list as Pandas DF
def read_connections(pn, df):
print 'reading connection ...'
global sampled_calls
# g is a graph data structure
# how can I improve this where?
calls_list = df.where('sub_encrypted_phone_number = {}'.format(pn))
sampled_calls = sampled_calls.unionAll(calls_list)
# avoid counting g here
#print g.count()
#return g
return calls_list
'''
taboo_list is a python list
df is the df of big tc_call_history tables
'''
def snowball_sampling(center, df, max_depth = 1, current_depth = 0, taboo_list = []):
print center, current_depth, max_depth, taboo_list
if current_depth == max_depth:
print 'out of depth'
return taboo_list
if center in taboo_list and len(taboo_list)>0:
# Visited this person -- exit
return taboo_list
else:
# New person! Don't visit again
taboo_list.append(center)
calls_list = read_connections(center, df)
#print 'fetched list: ', g.count()
#print g.count()
# This command seems slow
cont_list = calls_list.select('contact_encrypted_phone_number').distinct().toPandas()
for index, call in cont_list.iterrows():
# Iterate through all friends of the central node, and
# recursively call snowball sampling
taboo_list = snowball_sampling(call['contact_encrypted_phone_number'], df, current_depth = current_depth + 1,
max_depth = max_depth, taboo_list = taboo_list)
return taboo_list
def run_ss(pn, df, max_depth = 1):
#print smpl.show()
snowball_sampling(pn, df, max_depth)
```
| True | Repartition with and without caching - **Context: many 'where' look up on a big DF, the result has just 345 records but take 9k steps to calc.**
```
start_d = 20171001
if "hist_df" in locals():
hist_df.unpersist()
hist_df = sqlContext.read.parquet("{}/{}/update_date={}".format(WARE_HOUSE_PATH, HIS, start_d))
hist_df.repartition(100, 'sub_encrypted_phone_number')
#smpl = df.where(df.sub_encrypted_phone_number == 0)
if 'sampled_calls' in locals():
sampled_calls.unpersist()
sampled_calls = sqlContext.createDataFrame([] ,hist_df.schema, samplingRatio=0)
# append call list of visited pn and return cont_list as Pandas DF
def read_connections(pn, df):
print 'reading connection ...'
global sampled_calls
# g is a graph data structure
# how can I improve this where?
calls_list = df.where('sub_encrypted_phone_number = {}'.format(pn))
sampled_calls = sampled_calls.unionAll(calls_list)
# avoid counting g here
#print g.count()
#return g
return calls_list
'''
taboo_list is a python list
df is the df of big tc_call_history tables
'''
def snowball_sampling(center, df, max_depth = 1, current_depth = 0, taboo_list = []):
print center, current_depth, max_depth, taboo_list
if current_depth == max_depth:
print 'out of depth'
return taboo_list
if center in taboo_list and len(taboo_list)>0:
# Visited this person -- exit
return taboo_list
else:
# New person! Don't visit again
taboo_list.append(center)
calls_list = read_connections(center, df)
#print 'fetched list: ', g.count()
#print g.count()
# This command seems slow
cont_list = calls_list.select('contact_encrypted_phone_number').distinct().toPandas()
for index, call in cont_list.iterrows():
# Iterate through all friends of the central node, and
# recursively call snowball sampling
taboo_list = snowball_sampling(call['contact_encrypted_phone_number'], df, current_depth = current_depth + 1,
max_depth = max_depth, taboo_list = taboo_list)
return taboo_list
def run_ss(pn, df, max_depth = 1):
#print smpl.show()
snowball_sampling(pn, df, max_depth)
```
| non_test | repartition with and without caching context many where look up on a big df the result has just records but take steps to calc start d if hist df in locals hist df unpersist hist df sqlcontext read parquet update date format ware house path his start d hist df repartition sub encrypted phone number smpl df where df sub encrypted phone number if sampled calls in locals sampled calls unpersist sampled calls sqlcontext createdataframe hist df schema samplingratio append call list of visited pn and return cont list as pandas df def read connections pn df print reading connection global sampled calls g is a graph data structure how can i improve this where calls list df where sub encrypted phone number format pn sampled calls sampled calls unionall calls list avoid counting g here print g count return g return calls list taboo list is a python list df is the df of big tc call history tables def snowball sampling center df max depth current depth taboo list print center current depth max depth taboo list if current depth max depth print out of depth return taboo list if center in taboo list and len taboo list visited this person exit return taboo list else new person don t visit again taboo list append center calls list read connections center df print fetched list g count print g count this command seems slow cont list calls list select contact encrypted phone number distinct topandas for index call in cont list iterrows iterate through all friends of the central node and recursively call snowball sampling taboo list snowball sampling call df current depth current depth max depth max depth taboo list taboo list return taboo list def run ss pn df max depth print smpl show snowball sampling pn df max depth | 0 |
78,890 | 7,680,924,717 | IssuesEvent | 2018-05-16 04:45:30 | adobe/brackets | https://api.github.com/repos/adobe/brackets | closed | [Brackets auto-update Mac] Update scenario fails on Mac. | Testing | ### Description
Update scenario fails on Mac.
### Steps to Reproduce
1. Launch brackets 1.13.
2. Click on Update Notification Button.
3. Click on Get it now button.
4. Click on Restart button.
5. Brackets gets updated.
6. Launch brackets.
7. Red colored update info bar comes up on launching brackets after it has been updated.
This bar does not display any message.
**Expected behavior:** Brackets should get updated to 1.14 and launch.
**Actual behavior:** Update to 1.14 fails as download failed.
### Versions
Mac 10.13
Release 1.13 build 1.13.0-17639 | 1.0 | [Brackets auto-update Mac] Update scenario fails on Mac. - ### Description
Update scenario fails on Mac.
### Steps to Reproduce
1. Launch brackets 1.13.
2. Click on Update Notification Button.
3. Click on Get it now button.
4. Click on Restart button.
5. Brackets gets updated.
6. Launch brackets.
7. Red colored update info bar comes up on launching brackets after it has been updated.
This bar does not display any message.
**Expected behavior:** Brackets should get updated to 1.14 and launch.
**Actual behavior:** Update to 1.14 fails as download failed.
### Versions
Mac 10.13
Release 1.13 build 1.13.0-17639 | test | update scenario fails on mac description update scenario fails on mac steps to reproduce launch brackets click on update notification button click on get it now button click on restart button brackets gets updated launch brackets red colored update info bar comes up on launching brackets after it has been updated this bar does not display any message expected behavior brackets should get updated to and launch actual behavior update to fails as download failed versions mac release build | 1 |
68,027 | 3,283,808,137 | IssuesEvent | 2015-10-28 14:22:45 | leeensminger/OED_Wetlands | https://api.github.com/repos/leeensminger/OED_Wetlands | closed | Delineation - Inspections - Cannot edit inspection where Inspection Date = current system date | bug - high priority | Within the Inspections tab of an asset, the system disables Editing when the Inspection Date is the same as the current system date.
To recreate:
1. Create New Inspection, with inspection date of today
2. Click Save
3. Click the feature record in the Feature Manager to refresh
4. The Inspection fields and associated Photo tab are disabled

| 1.0 | Delineation - Inspections - Cannot edit inspection where Inspection Date = current system date - Within the Inspections tab of an asset, the system disables Editing when the Inspection Date is the same as the current system date.
To recreate:
1. Create New Inspection, with inspection date of today
2. Click Save
3. Click the feature record in the Feature Manager to refresh
4. The Inspection fields and associated Photo tab are disabled

| non_test | delineation inspections cannot edit inspection where inspection date current system date within the inspections tab of an asset the system disables editing when the inspection date is the same as the current system date to recreate create new inspection with inspection date of today click save click the feature record in the feature manager to refresh the inspection fields and associated photo tab are disabled | 0 |
120,846 | 10,136,797,415 | IssuesEvent | 2019-08-02 13:51:15 | elastic/elasticsearch | https://api.github.com/repos/elastic/elasticsearch | closed | [ML] DataFrameTaskFailedStateIT.testForceStartFailedTransform failure | :ml :ml/Data Frame >test-failure | Build link:
https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+master+multijob+fast+part2/394/console
Reproduce with:
```
./gradlew :x-pack:plugin:data-frame:qa:single-node-tests:integTestRunner --tests "org.elasticsearch.xpack.dataframe.integration.DataFrameTaskFailedStateIT.testForceStartFailedTransform" -Dtests.seed=2FFED17786A4BC97 -Dtests.security.manager=true -Dtests.locale=ar-LB -Dtests.timezone=Africa/Gaborone -Dcompiler.java=12 -Druntime.java=11
```
Failure:
```
java.lang.AssertionError:
Expected: <1>
but: was <2>
at __randomizedtesting.SeedInfo.seed([2FFED17786A4BC97:3E73F40F72853770]:0)
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18)
at org.junit.Assert.assertThat(Assert.java:956)
at org.junit.Assert.assertThat(Assert.java:923)
at org.elasticsearch.xpack.dataframe.integration.DataFrameTaskFailedStateIT.testForceStartFailedTransform(DataFrameTaskFailedStateIT.java:111)
```
I could not reproduce locally.
It seems that somehow the test may get to have 2 index failures rather than 1. | 1.0 | [ML] DataFrameTaskFailedStateIT.testForceStartFailedTransform failure - Build link:
https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+master+multijob+fast+part2/394/console
Reproduce with:
```
./gradlew :x-pack:plugin:data-frame:qa:single-node-tests:integTestRunner --tests "org.elasticsearch.xpack.dataframe.integration.DataFrameTaskFailedStateIT.testForceStartFailedTransform" -Dtests.seed=2FFED17786A4BC97 -Dtests.security.manager=true -Dtests.locale=ar-LB -Dtests.timezone=Africa/Gaborone -Dcompiler.java=12 -Druntime.java=11
```
Failure:
```
java.lang.AssertionError:
Expected: <1>
but: was <2>
at __randomizedtesting.SeedInfo.seed([2FFED17786A4BC97:3E73F40F72853770]:0)
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18)
at org.junit.Assert.assertThat(Assert.java:956)
at org.junit.Assert.assertThat(Assert.java:923)
at org.elasticsearch.xpack.dataframe.integration.DataFrameTaskFailedStateIT.testForceStartFailedTransform(DataFrameTaskFailedStateIT.java:111)
```
I could not reproduce locally.
It seems that somehow the test may get to have 2 index failures rather than 1. | test | dataframetaskfailedstateit testforcestartfailedtransform failure build link reproduce with gradlew x pack plugin data frame qa single node tests integtestrunner tests org elasticsearch xpack dataframe integration dataframetaskfailedstateit testforcestartfailedtransform dtests seed dtests security manager true dtests locale ar lb dtests timezone africa gaborone dcompiler java druntime java failure java lang assertionerror expected but was at randomizedtesting seedinfo seed at org hamcrest matcherassert assertthat matcherassert java at org junit assert assertthat assert java at org junit assert assertthat assert java at org elasticsearch xpack dataframe integration dataframetaskfailedstateit testforcestartfailedtransform dataframetaskfailedstateit java i could not reproduce locally it seems that somehow the test may get to have index failures rather than | 1 |
9,742 | 3,069,226,046 | IssuesEvent | 2015-08-18 19:25:58 | Darthpbal/homeTests | https://api.github.com/repos/Darthpbal/homeTests | closed | Who is the primary users of the app | test | ## gameplan
Read through some of the code that would be identifying of an owner department
read through some of the database tables that I have to identify an owner.
**report** back with my findings. | 1.0 | Who is the primary users of the app - ## gameplan
Read through some of the code that would be identifying of an owner department
read through some of the database tables that I have to identify an owner.
**report** back with my findings. | test | who is the primary users of the app gameplan read through some of the code that would be identifying of an owner department read through some of the database tables that i have to identify an owner report back with my findings | 1 |
97,094 | 8,644,293,707 | IssuesEvent | 2018-11-26 01:52:06 | vmware/vic | https://api.github.com/repos/vmware/vic | closed | [Scenario]Update 5-4-High-Availability steps | component/test/scenario source/scenario | 5-4-High-Availability is not stable and sometimes get failure. The root cause is that our test scripts does not differentiate where the tested container vm is. When the container vm is in the same ESXi with VCH and scripts triggers ESXi shutdown, the container VM is not moved to other hosts yet and then hit the issue of 'Error response from daemon: Server error from portlayer: Couldn't remove container.The ESX host is temporarily disconnected.', when calling docker rm <container>. Details info is in https://github.com/vmware/vic/issues/6667
At case layer, we will check if tested container vm is in the same as VCH, scripts will retry a few times to wait container vm move to other available host and then call docker rm <container id>. | 1.0 | [Scenario]Update 5-4-High-Availability steps - 5-4-High-Availability is not stable and sometimes get failure. The root cause is that our test scripts does not differentiate where the tested container vm is. When the container vm is in the same ESXi with VCH and scripts triggers ESXi shutdown, the container VM is not moved to other hosts yet and then hit the issue of 'Error response from daemon: Server error from portlayer: Couldn't remove container.The ESX host is temporarily disconnected.', when calling docker rm <container>. Details info is in https://github.com/vmware/vic/issues/6667
At case layer, we will check if tested container vm is in the same as VCH, scripts will retry a few times to wait container vm move to other available host and then call docker rm <container id>. | test | update high availability steps high availability is not stable and sometimes get failure the root cause is that our test scripts does not differentiate where the tested container vm is when the container vm is in the same esxi with vch and scripts triggers esxi shutdown the container vm is not moved to other hosts yet and then hit the issue of error response from daemon server error from portlayer couldn t remove container the esx host is temporarily disconnected when calling docker rm details info is in at case layer we will check if tested container vm is in the same as vch scripts will retry a few times to wait container vm move to other available host and then call docker rm | 1 |
122,504 | 10,225,324,477 | IssuesEvent | 2019-08-16 14:54:00 | ValveSoftware/steam-for-linux | https://api.github.com/repos/ValveSoftware/steam-for-linux | closed | Bluetooth module is detected as a gamepad in Overlord games | Need Retest reviewed | CPU : Intel I7 3820 3.6GHz
GPU : Nvidia Gtx 1070 Gigabyte G1 gaming (drivers 367.35)
RAM : 16GB DDR 2133MHz
DM : Gnome 3.20
Bluetooth : Atheros Communications, Inc. AR3011 Bluetooth
- Steam client version: 1468023329 (Latest Stable)
- Distribution (e.g. Ubuntu): Archlinux
- Opted into Steam client beta?: No
- Have you checked for system updates?: Yes
I played Overlord that just released yesterday on Steam on linux. I have this issue when the bluetooth is enabled the game camera moves itselfs. Seems that the bluetooth module is detected as a steam controller by Steam
Virtual-programming helped me with a workaround for their game but it would be nice to be resolved in Steam.
I link the issue on VP github https://github.com/virtual-programming/overlord-linux/issues/6
Here is my log file from Overlord game it could be useful .
[eon.txt](https://github.com/ValveSoftware/steam-for-linux/files/385972/eon.txt)
1. Enable Bluetooth module
2. Launch Overlord or Overlord II game with steam
3. start a new game or load game
4. Camera move itself while playing
Thank you very much
| 1.0 | Bluetooth module is detected as a gamepad in Overlord games - CPU : Intel I7 3820 3.6GHz
GPU : Nvidia Gtx 1070 Gigabyte G1 gaming (drivers 367.35)
RAM : 16GB DDR 2133MHz
DM : Gnome 3.20
Bluetooth : Atheros Communications, Inc. AR3011 Bluetooth
- Steam client version: 1468023329 (Latest Stable)
- Distribution (e.g. Ubuntu): Archlinux
- Opted into Steam client beta?: No
- Have you checked for system updates?: Yes
I played Overlord that just released yesterday on Steam on linux. I have this issue when the bluetooth is enabled the game camera moves itselfs. Seems that the bluetooth module is detected as a steam controller by Steam
Virtual-programming helped me with a workaround for their game but it would be nice to be resolved in Steam.
I link the issue on VP github https://github.com/virtual-programming/overlord-linux/issues/6
Here is my log file from Overlord game it could be useful .
[eon.txt](https://github.com/ValveSoftware/steam-for-linux/files/385972/eon.txt)
1. Enable Bluetooth module
2. Launch Overlord or Overlord II game with steam
3. start a new game or load game
4. Camera move itself while playing
Thank you very much
| test | bluetooth module is detected as a gamepad in overlord games cpu intel gpu nvidia gtx gigabyte gaming drivers ram ddr dm gnome bluetooth atheros communications inc bluetooth steam client version latest stable distribution e g ubuntu archlinux opted into steam client beta no have you checked for system updates yes i played overlord that just released yesterday on steam on linux i have this issue when the bluetooth is enabled the game camera moves itselfs seems that the bluetooth module is detected as a steam controller by steam virtual programming helped me with a workaround for their game but it would be nice to be resolved in steam i link the issue on vp github here is my log file from overlord game it could be useful enable bluetooth module launch overlord or overlord ii game with steam start a new game or load game camera move itself while playing thank you very much | 1 |
568,836 | 16,989,590,401 | IssuesEvent | 2021-06-30 18:34:08 | IdoSagiv/CommuniDog | https://api.github.com/repos/IdoSagiv/CommuniDog | closed | Logout button doesn't showed right xml tag not closed correctly | Mid Priority bug | Logout button doesn't showed right because it is not closed correctly
picture of the bug:
It's in info_menu_bar.xml under menu directory.


Enviroment :
* emulator Pixel 2
* Android api 29 | 1.0 | Logout button doesn't showed right xml tag not closed correctly - Logout button doesn't showed right because it is not closed correctly
picture of the bug:
It's in info_menu_bar.xml under menu directory.


Enviroment :
* emulator Pixel 2
* Android api 29 | non_test | logout button doesn t showed right xml tag not closed correctly logout button doesn t showed right because it is not closed correctly picture of the bug it s in info menu bar xml under menu directory enviroment emulator pixel android api | 0 |
39,050 | 5,215,678,234 | IssuesEvent | 2017-01-26 06:50:38 | mjs7231/python-plexapi | https://api.github.com/repos/mjs7231/python-plexapi | closed | Automated testing | testing | I'd like to incorporate some automated tests into the project, especially for the code submitted for #7
In the past, i've used **py.test** but id be open to another framework if you are partial to one.
| 1.0 | Automated testing - I'd like to incorporate some automated tests into the project, especially for the code submitted for #7
In the past, i've used **py.test** but id be open to another framework if you are partial to one.
| test | automated testing i d like to incorporate some automated tests into the project especially for the code submitted for in the past i ve used py test but id be open to another framework if you are partial to one | 1 |
89,445 | 8,203,896,154 | IssuesEvent | 2018-09-03 02:36:24 | backdrop/backdrop-issues | https://api.github.com/repos/backdrop/backdrop-issues | closed | [UX] Bring back node preview | pr - reviewed & tested by the community status - has pull request type - feature request | We removed node edit previews early on, but haven't agreed on how to re-introduce it. I think this is expected functionality.
Lets decide how to bring it back.
---
~PR: https://github.com/backdrop/backdrop/pull/2174~
PR: https://github.com/backdrop/backdrop/pull/2290 | 1.0 | [UX] Bring back node preview - We removed node edit previews early on, but haven't agreed on how to re-introduce it. I think this is expected functionality.
Lets decide how to bring it back.
---
~PR: https://github.com/backdrop/backdrop/pull/2174~
PR: https://github.com/backdrop/backdrop/pull/2290 | test | bring back node preview we removed node edit previews early on but haven t agreed on how to re introduce it i think this is expected functionality lets decide how to bring it back pr pr | 1 |
259,405 | 22,472,679,375 | IssuesEvent | 2022-06-22 09:28:11 | kyma-project/kyma | https://api.github.com/repos/kyma-project/kyma | closed | Worker reconciling Rafter occasionally fails with context deadline timeout | kind/failing-test release blocker | **Description**
See occurences:
- https://storage.googleapis.com/kyma-prow-logs/logs/kyma-upgrade-gardener-kyma2-to-main-reconciler-main/1528707388486455296/build-log.txt
- https://storage.googleapis.com/kyma-prow-logs/pr-logs/pull/kyma-project_kyma/14334/pre-main-kyma-gardener-azure-alpha-prod/1531219067166265344/build-log.txt
**Expected result**
Worker installing/updating rafter deployment should finish within desired context timeout
**Actual result**
Fails to finish ( gets stuck ) and timeout
| 1.0 | Worker reconciling Rafter occasionally fails with context deadline timeout - **Description**
See occurences:
- https://storage.googleapis.com/kyma-prow-logs/logs/kyma-upgrade-gardener-kyma2-to-main-reconciler-main/1528707388486455296/build-log.txt
- https://storage.googleapis.com/kyma-prow-logs/pr-logs/pull/kyma-project_kyma/14334/pre-main-kyma-gardener-azure-alpha-prod/1531219067166265344/build-log.txt
**Expected result**
Worker installing/updating rafter deployment should finish within desired context timeout
**Actual result**
Fails to finish ( gets stuck ) and timeout
| test | worker reconciling rafter occasionally fails with context deadline timeout description see occurences expected result worker installing updating rafter deployment should finish within desired context timeout actual result fails to finish gets stuck and timeout | 1 |
586,748 | 17,596,197,968 | IssuesEvent | 2021-08-17 05:39:47 | wso2/product-apim | https://api.github.com/repos/wso2/product-apim | opened | Message under API application quota | Type/Improvement Priority/Normal | ### Describe your problem(s)
<!-- Describe why you think this project needs this feature -->
When we are creating a new application and choose Quota for a token, under the dropdown filed the message is
"....**Assign API request quota per access token**. Allocated quota will be shared among all the subscribed APIs of the application...."
I get feedback that this message is confusing. Indicate that EACH token will have a chosen token, but in reality, a quota will be shared among all tokens.
can we change the message or explain it differently to avoid any confusion.
| 1.0 | Message under API application quota - ### Describe your problem(s)
<!-- Describe why you think this project needs this feature -->
When we are creating a new application and choose Quota for a token, under the dropdown filed the message is
"....**Assign API request quota per access token**. Allocated quota will be shared among all the subscribed APIs of the application...."
I get feedback that this message is confusing. Indicate that EACH token will have a chosen token, but in reality, a quota will be shared among all tokens.
can we change the message or explain it differently to avoid any confusion.
| non_test | message under api application quota describe your problem s when we are creating a new application and choose quota for a token under the dropdown filed the message is assign api request quota per access token allocated quota will be shared among all the subscribed apis of the application i get feedback that this message is confusing indicate that each token will have a chosen token but in reality a quota will be shared among all tokens can we change the message or explain it differently to avoid any confusion | 0 |
398,635 | 27,205,364,246 | IssuesEvent | 2023-02-20 12:41:53 | feelpp/book.feelpp.org | https://api.github.com/repos/feelpp/book.feelpp.org | closed | All pyfeel++ toolboxes examples point to "Fluid" | domain:documentation no-issue-activity | Could you provide examples for toolboxes other than "Fluid"?
Or at least change the links in the docs... | 1.0 | All pyfeel++ toolboxes examples point to "Fluid" - Could you provide examples for toolboxes other than "Fluid"?
Or at least change the links in the docs... | non_test | all pyfeel toolboxes examples point to fluid could you provide examples for toolboxes other than fluid or at least change the links in the docs | 0 |
449,004 | 12,961,903,651 | IssuesEvent | 2020-07-20 16:19:39 | monarch-initiative/mondo | https://api.github.com/repos/monarch-initiative/mondo | closed | Large drop in monogenic diseases after DO-MOndo mapping | high priority obsolete |
The following 210 genes used to be annotated to "monogenic disease"
https://www.pombase.org/results/from/id/707b3611-0b71-4ebd-9ee1-9326e44d76c6
Possibly not all are, but most do appear to be from spot checking
For example
SPAC31A2.05c | mis4 | cohesin loading factor (adherin) Mis4/Scc2
SPBC776.13 | cnd1 | condensin complex non-SMC subunit Cnd1
SPCC306.03c | cnd2 | condensin complex non-SMC subunit Cnd2
all have causative mutations for Cornelia de Lange syndrome
ark1 | aurora-B kinase Ark1
spermatogenic failure 5 (DOID:0070183) -has_material_basis_in mutation in the AURKC gene on chromosome 19q13
alg14 |
SPAC5D6.06c | alg14 | UDP-GlcNAc transferase associated protein Alg14
congenital myasthenic syndrome 15
compound heterozygous mutation in the ALG14 gene on chromosome 1p21.
SPAC18G6.10 | lem2 | LEM domain nuclear inner membrane protein Heh1/Lem2
cataract 46 juvenile-onset (DOID:0110243)
A cataract that has_material_basis_in homozygous mutation in the LEMD2 gene on chromosome 6p21.
| 1.0 | Large drop in monogenic diseases after DO-MOndo mapping -
The following 210 genes used to be annotated to "monogenic disease"
https://www.pombase.org/results/from/id/707b3611-0b71-4ebd-9ee1-9326e44d76c6
Possibly not all are, but most do appear to be from spot checking
For example
SPAC31A2.05c | mis4 | cohesin loading factor (adherin) Mis4/Scc2
SPBC776.13 | cnd1 | condensin complex non-SMC subunit Cnd1
SPCC306.03c | cnd2 | condensin complex non-SMC subunit Cnd2
all have causative mutations for Cornelia de Lange syndrome
ark1 | aurora-B kinase Ark1
spermatogenic failure 5 (DOID:0070183) -has_material_basis_in mutation in the AURKC gene on chromosome 19q13
alg14 |
SPAC5D6.06c | alg14 | UDP-GlcNAc transferase associated protein Alg14
congenital myasthenic syndrome 15
compound heterozygous mutation in the ALG14 gene on chromosome 1p21.
SPAC18G6.10 | lem2 | LEM domain nuclear inner membrane protein Heh1/Lem2
cataract 46 juvenile-onset (DOID:0110243)
A cataract that has_material_basis_in homozygous mutation in the LEMD2 gene on chromosome 6p21.
| non_test | large drop in monogenic diseases after do mondo mapping the following genes used to be annotated to monogenic disease possibly not all are but most do appear to be from spot checking for example cohesin loading factor adherin condensin complex non smc subunit condensin complex non smc subunit all have causative mutations for cornelia de lange syndrome aurora b kinase spermatogenic failure doid has material basis in mutation in the aurkc gene on chromosome udp glcnac transferase associated protein congenital myasthenic syndrome compound heterozygous mutation in the gene on chromosome lem domain nuclear inner membrane protein cataract juvenile onset doid a cataract that has material basis in homozygous mutation in the gene on chromosome | 0 |
232,150 | 18,847,409,815 | IssuesEvent | 2021-11-11 16:24:08 | ContinualAI/avalanche | https://api.github.com/repos/ContinualAI/avalanche | closed | Creation of a new envirnment failed for python 3.9 | test Continuous integration | Here are the differences between the last working environment and the new one that I tried to run:
```
9c9
< - absl-py=0.13.0=pyhd8ed1ab_0
---
> - absl-py=0.15.0=pyhd8ed1ab_0
16c16
< - brotlipy=0.7.0=py39h3811e60_1001
---
> - brotlipy=0.7.0=py39h3811e60_1003
18,22c18,22
< - c-ares=1.17.2=h7f98852_0
< - ca-certificates=2021.5.30=ha878542_0
< - cachetools=4.2.2=pyhd8ed1ab_0
< - certifi=2021.5.30=py39hf3d152e_0
< - cffi=1.14.6=py39he32792d_0
---
> - c-ares=1.18.1=h7f98852_0
> - ca-certificates=2021.10.8=ha878542_0
> - cachetools=4.2.4=pyhd8ed1ab_0
> - certifi=2021.10.8=py39hf3d152e_1
> - cffi=1.14.6=py39h4bc2ebd_2
24c24
< - click=8.0.1=py39hf3d152e_0
---
> - click=8.0.3=py39hf3d152e_1
27,30c27,30
< - cpuonly=1.0=0
< - cryptography=3.4.7=py39hbca0aa6_0
< - cycler=0.10.0=py_2
< - cython=0.29.24=py39he80948d_0
---
> - cpuonly=2.0=0
> - cryptography=35.0.0=py39h95dcef6_1
> - cycler=0.11.0=pyhd8ed1ab_0
> - cython=0.29.24=py39he80948d_1
32d31
< - dbus=1.13.6=he372182_0
34d32
< - expat=2.4.1=h9c3ff4c_0
36,40c34,37
< - fontconfig=2.13.1=he4413a7_1000
< - freetype=2.10.4=h5ab3b9f_0
< - gitdb=4.0.7=pyhd8ed1ab_0
< - gitpython=3.1.18=pyhd8ed1ab_0
< - glib=2.69.1=h5202010_0
---
> - freetype=2.11.0=h70c0345_0
> - giflib=5.2.1=h7b6447c_0
> - gitdb=4.0.9=pyhd8ed1ab_0
> - gitpython=3.1.24=pyhd8ed1ab_0
44c41
< - google-auth-oauthlib=0.4.5=pyhd8ed1ab_0
---
> - google-auth-oauthlib=0.4.6=pyhd8ed1ab_0
48,51c45
< - grpcio=1.38.1=py39hff7568b_0
< - gst-plugins-base=1.14.0=hbbd80ab_1
< - gstreamer=1.14.0=h28cd5cc_2
< - icu=58.2=hf484d3e_1000
---
> - grpcio=1.41.1=py39hff7568b_1
53,57c47,51
< - importlib-metadata=4.7.1=py39hf3d152e_1
< - intel-openmp=2021.3.0=h06a4308_3350
< - joblib=1.0.1=pyhd8ed1ab_0
< - jpeg=9b=h024ee3a_2
< - kiwisolver=1.3.2=py39h1a9c180_0
---
> - importlib-metadata=4.8.1=py39hf3d152e_1
> - intel-openmp=2021.4.0=h06a4308_3561
> - joblib=1.1.0=pyhd8ed1ab_0
> - jpeg=9d=h7f8727e_0
> - kiwisolver=1.3.2=py39h1a9c180_1
61,67c55,61
< - libblas=3.9.0=11_linux64_mkl
< - libcblas=3.9.0=11_linux64_mkl
< - libffi=3.3=h58526e2_2
< - libgcc-ng=11.1.0=hc902ee8_8
< - libgfortran-ng=11.1.0=h69a702a_8
< - libgfortran5=11.1.0=h6c583b3_8
< - libgomp=11.1.0=hc902ee8_8
---
> - libblas=3.9.0=12_linux64_mkl
> - libcblas=3.9.0=12_linux64_mkl
> - libffi=3.4.2=h9c3ff4c_4
> - libgcc-ng=11.2.0=h1d223b6_11
> - libgfortran-ng=11.2.0=h69a702a_11
> - libgfortran5=11.2.0=h5c6108e_11
> - libgomp=11.2.0=h1d223b6_11
70c64
< - liblapack=3.9.0=11_linux64_mkl
---
> - liblapack=3.9.0=12_linux64_mkl
72,73c66,67
< - libprotobuf=3.17.2=h780b84a_1
< - libstdcxx-ng=11.1.0=h56837e0_8
---
> - libprotobuf=3.19.1=h780b84a_0
> - libstdcxx-ng=11.2.0=he4da1e4_11
77d70
< - libuuid=2.32.1=h7f98852_1000
78a72
> - libwebp=1.2.0=h89dd481_0
80,81c74
< - libxcb=1.13=h7f98852_1003
< - libxml2=2.9.9=h13577e0_2
---
> - libzlib=1.2.11=h36c2ea0_1013
84,87c77,83
< - matplotlib=3.4.3=py39hf3d152e_0
< - matplotlib-base=3.4.3=py39h2fa2bec_0
< - mkl=2021.3.0=h06a4308_520
< - multidict=5.1.0=py39h3811e60_1
---
> - matplotlib=3.3.2=0
> - matplotlib-base=3.3.2=py39h98787fa_1
> - mkl=2021.4.0=h06a4308_640
> - mkl-service=2.4.0=py39h7f8727e_0
> - mkl_fft=1.3.1=py39hd3c417c_0
> - mkl_random=1.2.2=py39h51133e4_0
> - multidict=5.2.0=py39h3811e60_1
90,91c86,87
< - ninja=1.10.2=hff7bd54_1
< - numpy=1.21.2=py39hdbf815f_0
---
> - numpy=1.21.2=py39h20f2e39_0
> - numpy-base=1.21.2=py39h79a1101_0
94c90
< - olefile=0.46=py_0
---
> - olefile=0.46=pyhd3eb1b0_0
96,97c92
< - openjpeg=2.4.0=h3ad879b_0
< - openssl=1.1.1k=h7f98852_1
---
> - openssl=1.1.1l=h7f98852_0
99,105c94,98
< - pcre=8.45=h9c3ff4c_0
< - pillow=8.3.1=py39h2c7a002_0
< - pip=21.2.4=pyhd8ed1ab_0
< - promise=2.3=py39hf3d152e_3
< - protobuf=3.17.2=py39he80948d_0
< - psutil=5.8.0=py39h3811e60_1
< - pthread-stubs=0.4=h36c2ea0_1001
---
> - pillow=8.4.0=py39h5aabda8_0
> - pip=21.3.1=pyhd8ed1ab_0
> - promise=2.3=py39hf3d152e_4
> - protobuf=3.19.1=py39he80948d_1
> - psutil=5.8.0=py39h3811e60_2
110,115c103,107
< - pyjwt=2.1.0=pyhd8ed1ab_0
< - pyopenssl=20.0.1=pyhd8ed1ab_0
< - pyparsing=2.4.7=pyh9f0ad1d_0
< - pyqt=5.9.2=py39h2531618_6
< - pysocks=1.7.1=py39hf3d152e_3
< - python=3.9.6=h49503c6_1_cpython
---
> - pyjwt=2.3.0=pyhd8ed1ab_0
> - pyopenssl=21.0.0=pyhd8ed1ab_0
> - pyparsing=3.0.4=pyhd8ed1ab_0
> - pysocks=1.7.1=py39hf3d152e_4
> - python=3.9.7=hb7a2778_3_cpython
118c110,111
< - pytorch=1.9.0=py3.9_cpu_0
---
> - pytorch=1.10.0=py3.9_cpu_0
> - pytorch-mutex=1.0=cpu
120,122c113,114
< - pyyaml=5.4.1=py39h3811e60_1
< - qt=5.9.7=h5867ecd_1
< - quadprog=0.1.8=py39h1a9c180_2
---
> - pyyaml=6.0=py39h3811e60_2
> - quadprog=0.1.10=py39h1a9c180_0
127c119
< - scikit-learn=0.24.2=py39h4dfa638_1
---
> - scikit-learn=1.0.1=py39h7c5d8c9_1
129,130c121,122
< - sentry-sdk=1.3.1=pyhd8ed1ab_0
< - setuptools=57.4.0=py39hf3d152e_0
---
> - sentry-sdk=1.4.3=pyhd8ed1ab_0
> - setuptools=58.5.2=py39hf3d152e_0
132,133c124
< - sip=4.19.13=py39h2531618_0
< - six=1.16.0=pyh6c4a22f_0
---
> - six=1.16.0=pyhd3eb1b0_0
135c126
< - sqlite=3.36.0=h9cd32fc_0
---
> - sqlite=3.36.0=h9cd32fc_2
140,148c131,139
< - threadpoolctl=2.2.0=pyh8a188c0_0
< - tk=8.6.11=h21135ba_0
< - torchvision=0.10.0=py39_cpu
< - tornado=6.1=py39h3811e60_1
< - tqdm=4.62.2=pyhd8ed1ab_0
< - typing_extensions=3.10.0.0=pyh06a4308_0
< - tzdata=2021a=he74cb21_1
< - urllib3=1.26.6=pyhd8ed1ab_0
< - wandb=0.11.2=pyhd8ed1ab_0
---
> - threadpoolctl=3.0.0=pyh8a188c0_0
> - tk=8.6.11=h27826a3_1
> - torchvision=0.11.1=py39_cpu
> - tornado=6.1=py39h3811e60_2
> - tqdm=4.62.3=pyhd8ed1ab_0
> - typing_extensions=3.10.0.2=pyh06a4308_0
> - tzdata=2021e=he74cb21_0
> - urllib3=1.26.7=pyhd8ed1ab_0
> - wandb=0.12.1=pyhd8ed1ab_0
152,153d142
< - xorg-libxau=1.0.9=h7f98852_0
< - xorg-libxdmcp=1.1.3=h7f98852_0
156,158c145,147
< - yarl=1.6.3=py39h3811e60_2
< - zipp=3.5.0=pyhd8ed1ab_0
< - zlib=1.2.11=h516909a_1010
---
> - yarl=1.7.0=py39h3811e60_0
> - zipp=3.6.0=pyhd8ed1ab_0
> - zlib=1.2.11=h36c2ea0_1013
161,163c150,156
< - filelock==3.0.12
< - gdown==3.13.0
< - pytorchcv==0.0.66
---
> - beautifulsoup4==4.10.0
> - cloudpickle==2.0.0
> - filelock==3.3.2
> - gdown==4.2.0
> - gym==0.21.0
> - pytorchcv==0.0.67
> - soupsieve==2.3
``` | 1.0 | Creation of a new envirnment failed for python 3.9 - Here are the differences between the last working environment and the new one that I tried to run:
```
9c9
< - absl-py=0.13.0=pyhd8ed1ab_0
---
> - absl-py=0.15.0=pyhd8ed1ab_0
16c16
< - brotlipy=0.7.0=py39h3811e60_1001
---
> - brotlipy=0.7.0=py39h3811e60_1003
18,22c18,22
< - c-ares=1.17.2=h7f98852_0
< - ca-certificates=2021.5.30=ha878542_0
< - cachetools=4.2.2=pyhd8ed1ab_0
< - certifi=2021.5.30=py39hf3d152e_0
< - cffi=1.14.6=py39he32792d_0
---
> - c-ares=1.18.1=h7f98852_0
> - ca-certificates=2021.10.8=ha878542_0
> - cachetools=4.2.4=pyhd8ed1ab_0
> - certifi=2021.10.8=py39hf3d152e_1
> - cffi=1.14.6=py39h4bc2ebd_2
24c24
< - click=8.0.1=py39hf3d152e_0
---
> - click=8.0.3=py39hf3d152e_1
27,30c27,30
< - cpuonly=1.0=0
< - cryptography=3.4.7=py39hbca0aa6_0
< - cycler=0.10.0=py_2
< - cython=0.29.24=py39he80948d_0
---
> - cpuonly=2.0=0
> - cryptography=35.0.0=py39h95dcef6_1
> - cycler=0.11.0=pyhd8ed1ab_0
> - cython=0.29.24=py39he80948d_1
32d31
< - dbus=1.13.6=he372182_0
34d32
< - expat=2.4.1=h9c3ff4c_0
36,40c34,37
< - fontconfig=2.13.1=he4413a7_1000
< - freetype=2.10.4=h5ab3b9f_0
< - gitdb=4.0.7=pyhd8ed1ab_0
< - gitpython=3.1.18=pyhd8ed1ab_0
< - glib=2.69.1=h5202010_0
---
> - freetype=2.11.0=h70c0345_0
> - giflib=5.2.1=h7b6447c_0
> - gitdb=4.0.9=pyhd8ed1ab_0
> - gitpython=3.1.24=pyhd8ed1ab_0
44c41
< - google-auth-oauthlib=0.4.5=pyhd8ed1ab_0
---
> - google-auth-oauthlib=0.4.6=pyhd8ed1ab_0
48,51c45
< - grpcio=1.38.1=py39hff7568b_0
< - gst-plugins-base=1.14.0=hbbd80ab_1
< - gstreamer=1.14.0=h28cd5cc_2
< - icu=58.2=hf484d3e_1000
---
> - grpcio=1.41.1=py39hff7568b_1
53,57c47,51
< - importlib-metadata=4.7.1=py39hf3d152e_1
< - intel-openmp=2021.3.0=h06a4308_3350
< - joblib=1.0.1=pyhd8ed1ab_0
< - jpeg=9b=h024ee3a_2
< - kiwisolver=1.3.2=py39h1a9c180_0
---
> - importlib-metadata=4.8.1=py39hf3d152e_1
> - intel-openmp=2021.4.0=h06a4308_3561
> - joblib=1.1.0=pyhd8ed1ab_0
> - jpeg=9d=h7f8727e_0
> - kiwisolver=1.3.2=py39h1a9c180_1
61,67c55,61
< - libblas=3.9.0=11_linux64_mkl
< - libcblas=3.9.0=11_linux64_mkl
< - libffi=3.3=h58526e2_2
< - libgcc-ng=11.1.0=hc902ee8_8
< - libgfortran-ng=11.1.0=h69a702a_8
< - libgfortran5=11.1.0=h6c583b3_8
< - libgomp=11.1.0=hc902ee8_8
---
> - libblas=3.9.0=12_linux64_mkl
> - libcblas=3.9.0=12_linux64_mkl
> - libffi=3.4.2=h9c3ff4c_4
> - libgcc-ng=11.2.0=h1d223b6_11
> - libgfortran-ng=11.2.0=h69a702a_11
> - libgfortran5=11.2.0=h5c6108e_11
> - libgomp=11.2.0=h1d223b6_11
70c64
< - liblapack=3.9.0=11_linux64_mkl
---
> - liblapack=3.9.0=12_linux64_mkl
72,73c66,67
< - libprotobuf=3.17.2=h780b84a_1
< - libstdcxx-ng=11.1.0=h56837e0_8
---
> - libprotobuf=3.19.1=h780b84a_0
> - libstdcxx-ng=11.2.0=he4da1e4_11
77d70
< - libuuid=2.32.1=h7f98852_1000
78a72
> - libwebp=1.2.0=h89dd481_0
80,81c74
< - libxcb=1.13=h7f98852_1003
< - libxml2=2.9.9=h13577e0_2
---
> - libzlib=1.2.11=h36c2ea0_1013
84,87c77,83
< - matplotlib=3.4.3=py39hf3d152e_0
< - matplotlib-base=3.4.3=py39h2fa2bec_0
< - mkl=2021.3.0=h06a4308_520
< - multidict=5.1.0=py39h3811e60_1
---
> - matplotlib=3.3.2=0
> - matplotlib-base=3.3.2=py39h98787fa_1
> - mkl=2021.4.0=h06a4308_640
> - mkl-service=2.4.0=py39h7f8727e_0
> - mkl_fft=1.3.1=py39hd3c417c_0
> - mkl_random=1.2.2=py39h51133e4_0
> - multidict=5.2.0=py39h3811e60_1
90,91c86,87
< - ninja=1.10.2=hff7bd54_1
< - numpy=1.21.2=py39hdbf815f_0
---
> - numpy=1.21.2=py39h20f2e39_0
> - numpy-base=1.21.2=py39h79a1101_0
94c90
< - olefile=0.46=py_0
---
> - olefile=0.46=pyhd3eb1b0_0
96,97c92
< - openjpeg=2.4.0=h3ad879b_0
< - openssl=1.1.1k=h7f98852_1
---
> - openssl=1.1.1l=h7f98852_0
99,105c94,98
< - pcre=8.45=h9c3ff4c_0
< - pillow=8.3.1=py39h2c7a002_0
< - pip=21.2.4=pyhd8ed1ab_0
< - promise=2.3=py39hf3d152e_3
< - protobuf=3.17.2=py39he80948d_0
< - psutil=5.8.0=py39h3811e60_1
< - pthread-stubs=0.4=h36c2ea0_1001
---
> - pillow=8.4.0=py39h5aabda8_0
> - pip=21.3.1=pyhd8ed1ab_0
> - promise=2.3=py39hf3d152e_4
> - protobuf=3.19.1=py39he80948d_1
> - psutil=5.8.0=py39h3811e60_2
110,115c103,107
< - pyjwt=2.1.0=pyhd8ed1ab_0
< - pyopenssl=20.0.1=pyhd8ed1ab_0
< - pyparsing=2.4.7=pyh9f0ad1d_0
< - pyqt=5.9.2=py39h2531618_6
< - pysocks=1.7.1=py39hf3d152e_3
< - python=3.9.6=h49503c6_1_cpython
---
> - pyjwt=2.3.0=pyhd8ed1ab_0
> - pyopenssl=21.0.0=pyhd8ed1ab_0
> - pyparsing=3.0.4=pyhd8ed1ab_0
> - pysocks=1.7.1=py39hf3d152e_4
> - python=3.9.7=hb7a2778_3_cpython
118c110,111
< - pytorch=1.9.0=py3.9_cpu_0
---
> - pytorch=1.10.0=py3.9_cpu_0
> - pytorch-mutex=1.0=cpu
120,122c113,114
< - pyyaml=5.4.1=py39h3811e60_1
< - qt=5.9.7=h5867ecd_1
< - quadprog=0.1.8=py39h1a9c180_2
---
> - pyyaml=6.0=py39h3811e60_2
> - quadprog=0.1.10=py39h1a9c180_0
127c119
< - scikit-learn=0.24.2=py39h4dfa638_1
---
> - scikit-learn=1.0.1=py39h7c5d8c9_1
129,130c121,122
< - sentry-sdk=1.3.1=pyhd8ed1ab_0
< - setuptools=57.4.0=py39hf3d152e_0
---
> - sentry-sdk=1.4.3=pyhd8ed1ab_0
> - setuptools=58.5.2=py39hf3d152e_0
132,133c124
< - sip=4.19.13=py39h2531618_0
< - six=1.16.0=pyh6c4a22f_0
---
> - six=1.16.0=pyhd3eb1b0_0
135c126
< - sqlite=3.36.0=h9cd32fc_0
---
> - sqlite=3.36.0=h9cd32fc_2
140,148c131,139
< - threadpoolctl=2.2.0=pyh8a188c0_0
< - tk=8.6.11=h21135ba_0
< - torchvision=0.10.0=py39_cpu
< - tornado=6.1=py39h3811e60_1
< - tqdm=4.62.2=pyhd8ed1ab_0
< - typing_extensions=3.10.0.0=pyh06a4308_0
< - tzdata=2021a=he74cb21_1
< - urllib3=1.26.6=pyhd8ed1ab_0
< - wandb=0.11.2=pyhd8ed1ab_0
---
> - threadpoolctl=3.0.0=pyh8a188c0_0
> - tk=8.6.11=h27826a3_1
> - torchvision=0.11.1=py39_cpu
> - tornado=6.1=py39h3811e60_2
> - tqdm=4.62.3=pyhd8ed1ab_0
> - typing_extensions=3.10.0.2=pyh06a4308_0
> - tzdata=2021e=he74cb21_0
> - urllib3=1.26.7=pyhd8ed1ab_0
> - wandb=0.12.1=pyhd8ed1ab_0
152,153d142
< - xorg-libxau=1.0.9=h7f98852_0
< - xorg-libxdmcp=1.1.3=h7f98852_0
156,158c145,147
< - yarl=1.6.3=py39h3811e60_2
< - zipp=3.5.0=pyhd8ed1ab_0
< - zlib=1.2.11=h516909a_1010
---
> - yarl=1.7.0=py39h3811e60_0
> - zipp=3.6.0=pyhd8ed1ab_0
> - zlib=1.2.11=h36c2ea0_1013
161,163c150,156
< - filelock==3.0.12
< - gdown==3.13.0
< - pytorchcv==0.0.66
---
> - beautifulsoup4==4.10.0
> - cloudpickle==2.0.0
> - filelock==3.3.2
> - gdown==4.2.0
> - gym==0.21.0
> - pytorchcv==0.0.67
> - soupsieve==2.3
``` | test | creation of a new envirnment failed for python here are the differences between the last working environment and the new one that i tried to run absl py absl py brotlipy brotlipy c ares ca certificates cachetools certifi cffi c ares ca certificates cachetools certifi cffi click click cpuonly cryptography cycler py cython cpuonly cryptography cycler cython dbus expat fontconfig freetype gitdb gitpython glib freetype giflib gitdb gitpython google auth oauthlib google auth oauthlib grpcio gst plugins base gstreamer icu grpcio importlib metadata intel openmp joblib jpeg kiwisolver importlib metadata intel openmp joblib jpeg kiwisolver libblas mkl libcblas mkl libffi libgcc ng libgfortran ng libgomp libblas mkl libcblas mkl libffi libgcc ng libgfortran ng libgomp liblapack mkl liblapack mkl libprotobuf libstdcxx ng libprotobuf libstdcxx ng libuuid libwebp libxcb libzlib matplotlib matplotlib base mkl multidict matplotlib matplotlib base mkl mkl service mkl fft mkl random multidict ninja numpy numpy numpy base olefile py olefile openjpeg openssl openssl pcre pillow pip promise protobuf psutil pthread stubs pillow pip promise protobuf psutil pyjwt pyopenssl pyparsing pyqt pysocks python cpython pyjwt pyopenssl pyparsing pysocks python cpython pytorch cpu pytorch cpu pytorch mutex cpu pyyaml qt quadprog pyyaml quadprog scikit learn scikit learn sentry sdk setuptools sentry sdk setuptools sip six six sqlite sqlite threadpoolctl tk torchvision cpu tornado tqdm typing extensions tzdata wandb threadpoolctl tk torchvision cpu tornado tqdm typing extensions tzdata wandb xorg libxau xorg libxdmcp yarl zipp zlib yarl zipp zlib filelock gdown pytorchcv cloudpickle filelock gdown gym pytorchcv soupsieve | 1 |
73,610 | 7,346,013,008 | IssuesEvent | 2018-03-07 19:17:15 | istio/test-infra | https://api.github.com/repos/istio/test-infra | closed | Update pipeline artifacts to also include docker images | test-infra | Artifacts link is defined here
https://github.com/istio/test-infra/blob/master/src/org/istio/testutils/GitUtilities.groovy#L152
It should be updated to list all docker images created and pushlished. | 1.0 | Update pipeline artifacts to also include docker images - Artifacts link is defined here
https://github.com/istio/test-infra/blob/master/src/org/istio/testutils/GitUtilities.groovy#L152
It should be updated to list all docker images created and pushlished. | test | update pipeline artifacts to also include docker images artifacts link is defined here it should be updated to list all docker images created and pushlished | 1 |
192,657 | 6,876,394,460 | IssuesEvent | 2017-11-20 00:04:49 | OpenNebula/one | https://api.github.com/repos/OpenNebula/one | opened | Add new columns to oneacct that require some processing | Category: CLI Priority: Normal Status: Pending Tracker: Backlog | ---
Author Name: **Carlos Martín** (Carlos Martín)
Original Redmine Issue: 1491, https://dev.opennebula.org/issues/1491
Original Date: 2012-09-21
---
For example, the running time.
Requested by "Jan Benadik":http://lists.opennebula.org/pipermail/users-opennebula.org/2012-September/020320.html
| 1.0 | Add new columns to oneacct that require some processing - ---
Author Name: **Carlos Martín** (Carlos Martín)
Original Redmine Issue: 1491, https://dev.opennebula.org/issues/1491
Original Date: 2012-09-21
---
For example, the running time.
Requested by "Jan Benadik":http://lists.opennebula.org/pipermail/users-opennebula.org/2012-September/020320.html
| non_test | add new columns to oneacct that require some processing author name carlos martín carlos martín original redmine issue original date for example the running time requested by jan benadik | 0 |
8,474 | 22,615,319,146 | IssuesEvent | 2022-06-29 21:19:34 | Quran-Journey/backend | https://api.github.com/repos/Quran-Journey/backend | opened | Create db Schema for backend | architecture | Include all of the tables that are laid out in our ERD. You can see the ERD in the README.md | 1.0 | Create db Schema for backend - Include all of the tables that are laid out in our ERD. You can see the ERD in the README.md | non_test | create db schema for backend include all of the tables that are laid out in our erd you can see the erd in the readme md | 0 |
56,279 | 3,078,753,124 | IssuesEvent | 2015-08-21 12:36:06 | pavel-pimenov/flylinkdc-r5xx | https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx | opened | Падение при удалении запроса на скачивание файл-листа юзера по DHT из Очереди скачивания | bug Component-Logic imported Priority-High | _From [reaor...@gmail.com](https://code.google.com/u/102418317896447533964/) on August 29, 2011 18:59:00_
1) Ставим на скачивание файл большого размера и довольно популярный (при этом лучше, чтобы локальные хабы отсутствовали; чтобы хабов было мало)
2)DHT включён, соединений по DHT установлено много; включено Автоматическое добавление в очередь.
3)Ждём.
4)Видим то, что на 1.png
5)Пытаемся удалить этот проблемный файллист. Крах. В принципе тоже самое получим, если попытаемся грохнуть всю папку File Lists. http://narod.ru/disk/23329906001/crash-r501-beta31-build-8044.7z.html
**Attachment:** [1.png 2.png](http://code.google.com/p/flylinkdc/issues/detail?id=552)
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=552_ | 1.0 | Падение при удалении запроса на скачивание файл-листа юзера по DHT из Очереди скачивания - _From [reaor...@gmail.com](https://code.google.com/u/102418317896447533964/) on August 29, 2011 18:59:00_
1) Ставим на скачивание файл большого размера и довольно популярный (при этом лучше, чтобы локальные хабы отсутствовали; чтобы хабов было мало)
2)DHT включён, соединений по DHT установлено много; включено Автоматическое добавление в очередь.
3)Ждём.
4)Видим то, что на 1.png
5)Пытаемся удалить этот проблемный файллист. Крах. В принципе тоже самое получим, если попытаемся грохнуть всю папку File Lists. http://narod.ru/disk/23329906001/crash-r501-beta31-build-8044.7z.html
**Attachment:** [1.png 2.png](http://code.google.com/p/flylinkdc/issues/detail?id=552)
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=552_ | non_test | падение при удалении запроса на скачивание файл листа юзера по dht из очереди скачивания from on august ставим на скачивание файл большого размера и довольно популярный при этом лучше чтобы локальные хабы отсутствовали чтобы хабов было мало dht включён соединений по dht установлено много включено автоматическое добавление в очередь ждём видим то что на png пытаемся удалить этот проблемный файллист крах в принципе тоже самое получим если попытаемся грохнуть всю папку file lists attachment original issue | 0 |
793,449 | 27,997,531,323 | IssuesEvent | 2023-03-27 09:28:13 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | m.facebook.com - site is not usable | priority-critical browser-fenix engine-gecko | <!-- @browser: Firefox Mobile 112.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 12; Mobile; rv:109.0) Gecko/112.0 Firefox/112.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/120102 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://m.facebook.com/
**Browser / Version**: Firefox Mobile 112.0
**Operating System**: Android 12
**Tested Another Browser**: Yes Other
**Problem type**: Site is not usable
**Description**: Buttons or links not working
**Steps to Reproduce**:
I cant rightclick, select, copy text nor can i open links
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20230323181038</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2023/3/bbc7221b-6317-4a7a-b63f-e540dec5cd05)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | m.facebook.com - site is not usable - <!-- @browser: Firefox Mobile 112.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 12; Mobile; rv:109.0) Gecko/112.0 Firefox/112.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/120102 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://m.facebook.com/
**Browser / Version**: Firefox Mobile 112.0
**Operating System**: Android 12
**Tested Another Browser**: Yes Other
**Problem type**: Site is not usable
**Description**: Buttons or links not working
**Steps to Reproduce**:
I cant rightclick, select, copy text nor can i open links
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20230323181038</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2023/3/bbc7221b-6317-4a7a-b63f-e540dec5cd05)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_test | m facebook com site is not usable url browser version firefox mobile operating system android tested another browser yes other problem type site is not usable description buttons or links not working steps to reproduce i cant rightclick select copy text nor can i open links browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 0 |
508,018 | 14,688,401,780 | IssuesEvent | 2021-01-02 02:29:29 | leanprover/lean4 | https://api.github.com/repos/leanprover/lean4 | opened | store partial mutual block information in .olean file | feature low priority | Given a partial/unsafe definition, we don't store the list of mutually recursive definitions in the same mutual block.
Note that we store this information for mutually inductive datatypes.
https://github.com/leanprover/lean4/blob/57e05af9da8899ff62d37e30cf8afc3cea4c7cc6/src/Lean/Declaration.lean#L179
We should add a similar list to `DefinitionVal`.
This information may be useful for users that want to export .olean files.
In principle, users can retrieve the mutual block by traversing definitions, but it is not convenient.
Note that, users that want to ensure the logical soundness of their theorems may ignore partial mutual blocks.
| 1.0 | store partial mutual block information in .olean file - Given a partial/unsafe definition, we don't store the list of mutually recursive definitions in the same mutual block.
Note that we store this information for mutually inductive datatypes.
https://github.com/leanprover/lean4/blob/57e05af9da8899ff62d37e30cf8afc3cea4c7cc6/src/Lean/Declaration.lean#L179
We should add a similar list to `DefinitionVal`.
This information may be useful for users that want to export .olean files.
In principle, users can retrieve the mutual block by traversing definitions, but it is not convenient.
Note that, users that want to ensure the logical soundness of their theorems may ignore partial mutual blocks.
| non_test | store partial mutual block information in olean file given a partial unsafe definition we don t store the list of mutually recursive definitions in the same mutual block note that we store this information for mutually inductive datatypes we should add a similar list to definitionval this information may be useful for users that want to export olean files in principle users can retrieve the mutual block by traversing definitions but it is not convenient note that users that want to ensure the logical soundness of their theorems may ignore partial mutual blocks | 0 |
79,699 | 7,723,731,504 | IssuesEvent | 2018-05-24 13:19:10 | italia/spid | https://api.github.com/repos/italia/spid | closed | Controllo metadata Comune di Lacchiarella | metadata nuovo md test | Buongiorno,
Per conto del comune di Lacchiarella, abbiamo predisposto i metadata e pubblicati nella cartella
https://lacchiarella.comune-online.it/serviziSPID/metadata.xml
cordiali saluti
Facondini Stefano
Maggioli spa
[metadata_lacchiarella-signed.zip](https://github.com/italia/spid/files/2028056/metadata_lacchiarella-signed.zip)
| 1.0 | Controllo metadata Comune di Lacchiarella - Buongiorno,
Per conto del comune di Lacchiarella, abbiamo predisposto i metadata e pubblicati nella cartella
https://lacchiarella.comune-online.it/serviziSPID/metadata.xml
cordiali saluti
Facondini Stefano
Maggioli spa
[metadata_lacchiarella-signed.zip](https://github.com/italia/spid/files/2028056/metadata_lacchiarella-signed.zip)
| test | controllo metadata comune di lacchiarella buongiorno per conto del comune di lacchiarella abbiamo predisposto i metadata e pubblicati nella cartella cordiali saluti facondini stefano maggioli spa | 1 |
335,149 | 30,014,673,540 | IssuesEvent | 2023-06-26 17:44:58 | eclipse-openj9/openj9 | https://api.github.com/repos/eclipse-openj9/openj9 | closed | Crashes in object allocation path with SPECjbb2015 on PPCLE | comp:gc test failure | Seen in internal testing:
Original failed job:
http://vmfarm.rtp.raleigh.ibm.com/job_output.php?id=57852605
```
Unhandled exception
Type=Segmentation error vmState=0x00020019
J9Generic_Signal_Number=00000018 Signal_Number=0000000b Error_Value=00000000 Signal_Code=00000001
Handler1=00007D5F5A671460 Handler2=00007D5F5A484100
R0=00007D5F5905EC40 R1=00007D5F408AB360 R2=00007D5F591B5500 R3=00007D5F5406BDA8
R4=00000000FBFC4698 R5=00000000FBFC4698 R6=00007D5F408AB5B8 R7=00007D5F54045B50
R8=00000000000B5900 R9=0000000000000000 R10=0000000000000000 R11=0000000000000000
R12=00007D5F590E6D40 R13=00007D5F408B68F0 R14=0000000000992E30 R15=000000000078DD00
R16=00007D5F23D00038 R17=0000000068BE7AE8 R18=00000000FBF09728 R19=00000000FBFC26F0
R20=0000000000000000 R21=000000007F445850 R22=0000000000000000 R23=00000000FBFC26F0
R24=0000000000000000 R25=0000000000000720 R26=00007D5F408AB5D0 R27=00007D5F5406BCB0
R28=00000000FBFC4698 R29=00007D5ED41CD4E8 R30=00007D5F408AB5B8 R31=00000000FBFC4698
NIP=00007D5F590E1B68 MSR=800000000280F033 ORIG_GPR3=00007D5F59482F5C CTR=00007D5F590E6D40
LINK=00007D5F5905EC40 XER=0000000020000000 CCR=0000000022004822 SOFTE=0000000000000001
TRAP=0000000000000300 DAR=0000000000000018 dsisr=0000000040000000 RESULT=0000000000000000
Module=/bluebird/builds/bld_49023/sdk/xl6480/jre/lib/ppc64le/compressedrefs/libj9gc29.so
Module_base_address=00007D5F58E90000
Target=2_90_20230412_49023 (Linux 5.4.0-144-generic)
CPU=ppc64le (4 logical CPUs) (0x1fcc00000 RAM)
----------- Stack Backtrace -----------
_ZN33MM_IndexableObjectAllocationModel25initializeIndexableObjectEP18MM_EnvironmentBasePv+0x198 (0x00007D5F590E1B68 [libj9gc29.so+0x251b68])
_ZN22GC_ObjectModelDelegate20initializeAllocationEP18MM_EnvironmentBasePvP25MM_AllocateInitialization+0x30 (0x00007D5F5905EC40 [libj9gc29.so+0x1cec40])
_Z21OMR_GC_AllocateObjectP12OMR_VMThreadP25MM_AllocateInitialization+0x178 (0x00007D5F59064D38 [libj9gc29.so+0x1d4d38])
J9AllocateIndexableObjectNoGC+0x1f8 (0x00007D5F58EEBDE8 [libj9gc29.so+0x5bde8])
fast_jitNewArray+0x5c (0x00007D5F59BC359C [libj9jit29.so+0xa0359c])
(0x00007D5F59BD8A1C [libj9jit29.so+0xa18a1c])
runJavaThread+0x3b4 (0x00007D5F5A650374 [libj9vm29.so+0xc0374])
javaProtectedThreadProc+0xdc (0x00007D5F5A6C5FAC [libj9vm29.so+0x135fac])
omrsig_protect+0x358 (0x00007D5F5A485508 [libj9prt29.so+0x35508])
javaThreadProc+0x64 (0x00007D5F5A6C1684 [libj9vm29.so+0x131684])
thread_wrapper+0x1a8 (0x00007D5F5A5613C8 [libj9thr29.so+0x113c8])
start_thread+0xe8 (0x00007D5F5B4D8838 [libpthread.so.0+0x8838])
clone+0x74 (0x00007D5F5B35B884 [libc.so.6+0x14b884])
```
Reproduced in the grinder 2 of 10:
```
Unhandled exception
Type=Segmentation error vmState=0x00000000
J9Generic_Signal_Number=00000018 Signal_Number=0000000b Error_Value=00000000 Signal_Code=00000001
Handler1=000073CDF4F21460 Handler2=000073CDF4D34100
R0=000073CDF449B308 R1=000073CDDB43B5A0 R2=000073CDF4765500 R3=00000000FAEDBAF8
R4=0000000000000000 R5=000073CDDB43B618 R6=000073CDF00BF490 R7=000073CDDB43B618
R8=0000000000000000 R9=000073CDF0045B50 R10=0000000000000000 R11=0000000000000008
R12=000073CDF45B8650 R13=000073CDDB4468F0 R14=000000000095C240 R15=0000000000A11900
R16=000073CDDA620038 R17=00000000FADA85B0 R18=FFFFFFFFFFFFFFFF R19=0000000068A45310
R20=00000000FADA85D8 R21=00000000FAE5A448 R22=00000000FADA85F8 R23=0000000068A45250
R24=0000000068A61CA0 R25=000073CDF006BCB0 R26=FFFFFFFFFFFFFFFF R27=00000018BFE3A31A
R28=00000000FAEDBAF8 R29=0000000000A11900 R30=00000000000DB600 R31=000073CDBC066F08
NIP=000073CDF449B330 MSR=800000000280F033 ORIG_GPR3=000073CDDC55D340 CTR=000073CDF45B8650
LINK=000073CDF449B308 XER=0000000020000000 CCR=0000000044004844 SOFTE=0000000000000001
TRAP=0000000000000300 DAR=0000000000000018 dsisr=0000000040000000 RESULT=0000000000000000
Module=/bluebird/builds/bld_49023/sdk/xl6480/jre/lib/ppc64le/compressedrefs/libj9gc29.so
Module_base_address=000073CDF4440000
Target=2_90_20230412_49023 (Linux 5.4.0-144-generic)
CPU=ppc64le (4 logical CPUs) (0x1fcc00000 RAM)
----------- Stack Backtrace -----------
J9AllocateObjectNoGC+0x1f0 (0x000073CDF449B330 [libj9gc29.so+0x5b330])
fast_jitNewObject+0x9c (0x000073CDEFD03C5C [libj9jit29.so+0xa03c5c])
(0x000073CDEFD1865C [libj9jit29.so+0xa1865c])
runJavaThread+0x3b4 (0x000073CDF4F00374 [libj9vm29.so+0xc0374])
javaProtectedThreadProc+0xdc (0x000073CDF4F75FAC [libj9vm29.so+0x135fac])
omrsig_protect+0x358 (0x000073CDF4D35508 [libj9prt29.so+0x35508])
javaThreadProc+0x64 (0x000073CDF4F71684 [libj9vm29.so+0x131684])
thread_wrapper+0x1a8 (0x000073CDF4E113C8 [libj9thr29.so+0x113c8])
start_thread+0xe8 (0x000073CDF5D88838 [libpthread.so.0+0x8838])
clone+0x74 (0x000073CDF5C0B884 [libc.so.6+0x14b884])
Unhandled exception
Type=Segmentation error vmState=0x00000000
J9Generic_Signal_Number=00000018 Signal_Number=0000000b Error_Value=00000000 Signal_Code=00000001
Handler1=00007D1B15961460 Handler2=00007D1B15774100
R0=00007D1B141DBDE8 R1=00007D1A79E7B540 R2=00007D1B144A5500 R3=00000000FB6876E8
R4=0000000000000000 R5=00007D1A79E7B5C0 R6=00007D1B100B99F0 R7=0000000000000000
R8=000000002595BF00 R9=00007D1B10045B50 R10=0000000000000000 R11=0000000000000138
R12=00007D1B142F8650 R13=00007D1A79E868F0 R14=0000000000B124F0 R15=00000000009EE200
R16=00007D1AFC140038 R17=00000000BCB45438 R18=00000000FB6773B0 R19=00000000BCB573C0
R20=00000000BCB573C0 R21=0000000000000000 R22=00000000FB686108 R23=00000000FB6773F0
R24=00007D1B1006BCB0 R25=00000000FB686108 R26=00000000000C5900 R27=00007D1AF816A778
R28=00000000FB6876E8 R29=000000000000000A R30=0000000000000001 R31=00000000009EE200
NIP=00007D1B141DBE0C MSR=800000000280F033 ORIG_GPR3=00007D1B1591E804 CTR=00007D1B142F8650
LINK=00007D1B141DBDE8 XER=0000000020000000 CCR=0000000044004444 SOFTE=0000000000000001
TRAP=0000000000000300 DAR=0000000000000018 dsisr=0000000040000000 RESULT=0000000000000000
Module=/bluebird/builds/bld_49023/sdk/xl6480/jre/lib/ppc64le/compressedrefs/libj9gc29.so
Module_base_address=00007D1B14180000
Target=2_90_20230412_49023 (Linux 5.4.0-144-generic)
CPU=ppc64le (4 logical CPUs) (0x1fcc00000 RAM)
----------- Stack Backtrace -----------
J9AllocateIndexableObjectNoGC+0x21c (0x00007D1B141DBE0C [libj9gc29.so+0x5be0c])
fast_jitANewArray+0x64 (0x00007D1B14EB33E4 [libj9jit29.so+0xa033e4])
(0x00007D1B14EC883C [libj9jit29.so+0xa1883c])
runJavaThread+0x3b4 (0x00007D1B15940374 [libj9vm29.so+0xc0374])
javaProtectedThreadProc+0xdc (0x00007D1B159B5FAC [libj9vm29.so+0x135fac])
omrsig_protect+0x358 (0x00007D1B15775508 [libj9prt29.so+0x35508])
javaThreadProc+0x64 (0x00007D1B159B1684 [libj9vm29.so+0x131684])
thread_wrapper+0x1a8 (0x00007D1B158513C8 [libj9thr29.so+0x113c8])
start_thread+0xe8 (0x00007D1B167C8838 [libpthread.so.0+0x8838])
clone+0x74 (0x00007D1B1664B884 [libc.so.6+0x14b884])
---------------------------------------
```
Unfortunately there is no stored results due infra issue | 1.0 | Crashes in object allocation path with SPECjbb2015 on PPCLE - Seen in internal testing:
Original failed job:
http://vmfarm.rtp.raleigh.ibm.com/job_output.php?id=57852605
```
Unhandled exception
Type=Segmentation error vmState=0x00020019
J9Generic_Signal_Number=00000018 Signal_Number=0000000b Error_Value=00000000 Signal_Code=00000001
Handler1=00007D5F5A671460 Handler2=00007D5F5A484100
R0=00007D5F5905EC40 R1=00007D5F408AB360 R2=00007D5F591B5500 R3=00007D5F5406BDA8
R4=00000000FBFC4698 R5=00000000FBFC4698 R6=00007D5F408AB5B8 R7=00007D5F54045B50
R8=00000000000B5900 R9=0000000000000000 R10=0000000000000000 R11=0000000000000000
R12=00007D5F590E6D40 R13=00007D5F408B68F0 R14=0000000000992E30 R15=000000000078DD00
R16=00007D5F23D00038 R17=0000000068BE7AE8 R18=00000000FBF09728 R19=00000000FBFC26F0
R20=0000000000000000 R21=000000007F445850 R22=0000000000000000 R23=00000000FBFC26F0
R24=0000000000000000 R25=0000000000000720 R26=00007D5F408AB5D0 R27=00007D5F5406BCB0
R28=00000000FBFC4698 R29=00007D5ED41CD4E8 R30=00007D5F408AB5B8 R31=00000000FBFC4698
NIP=00007D5F590E1B68 MSR=800000000280F033 ORIG_GPR3=00007D5F59482F5C CTR=00007D5F590E6D40
LINK=00007D5F5905EC40 XER=0000000020000000 CCR=0000000022004822 SOFTE=0000000000000001
TRAP=0000000000000300 DAR=0000000000000018 dsisr=0000000040000000 RESULT=0000000000000000
Module=/bluebird/builds/bld_49023/sdk/xl6480/jre/lib/ppc64le/compressedrefs/libj9gc29.so
Module_base_address=00007D5F58E90000
Target=2_90_20230412_49023 (Linux 5.4.0-144-generic)
CPU=ppc64le (4 logical CPUs) (0x1fcc00000 RAM)
----------- Stack Backtrace -----------
_ZN33MM_IndexableObjectAllocationModel25initializeIndexableObjectEP18MM_EnvironmentBasePv+0x198 (0x00007D5F590E1B68 [libj9gc29.so+0x251b68])
_ZN22GC_ObjectModelDelegate20initializeAllocationEP18MM_EnvironmentBasePvP25MM_AllocateInitialization+0x30 (0x00007D5F5905EC40 [libj9gc29.so+0x1cec40])
_Z21OMR_GC_AllocateObjectP12OMR_VMThreadP25MM_AllocateInitialization+0x178 (0x00007D5F59064D38 [libj9gc29.so+0x1d4d38])
J9AllocateIndexableObjectNoGC+0x1f8 (0x00007D5F58EEBDE8 [libj9gc29.so+0x5bde8])
fast_jitNewArray+0x5c (0x00007D5F59BC359C [libj9jit29.so+0xa0359c])
(0x00007D5F59BD8A1C [libj9jit29.so+0xa18a1c])
runJavaThread+0x3b4 (0x00007D5F5A650374 [libj9vm29.so+0xc0374])
javaProtectedThreadProc+0xdc (0x00007D5F5A6C5FAC [libj9vm29.so+0x135fac])
omrsig_protect+0x358 (0x00007D5F5A485508 [libj9prt29.so+0x35508])
javaThreadProc+0x64 (0x00007D5F5A6C1684 [libj9vm29.so+0x131684])
thread_wrapper+0x1a8 (0x00007D5F5A5613C8 [libj9thr29.so+0x113c8])
start_thread+0xe8 (0x00007D5F5B4D8838 [libpthread.so.0+0x8838])
clone+0x74 (0x00007D5F5B35B884 [libc.so.6+0x14b884])
```
Reproduced in the grinder 2 of 10:
```
Unhandled exception
Type=Segmentation error vmState=0x00000000
J9Generic_Signal_Number=00000018 Signal_Number=0000000b Error_Value=00000000 Signal_Code=00000001
Handler1=000073CDF4F21460 Handler2=000073CDF4D34100
R0=000073CDF449B308 R1=000073CDDB43B5A0 R2=000073CDF4765500 R3=00000000FAEDBAF8
R4=0000000000000000 R5=000073CDDB43B618 R6=000073CDF00BF490 R7=000073CDDB43B618
R8=0000000000000000 R9=000073CDF0045B50 R10=0000000000000000 R11=0000000000000008
R12=000073CDF45B8650 R13=000073CDDB4468F0 R14=000000000095C240 R15=0000000000A11900
R16=000073CDDA620038 R17=00000000FADA85B0 R18=FFFFFFFFFFFFFFFF R19=0000000068A45310
R20=00000000FADA85D8 R21=00000000FAE5A448 R22=00000000FADA85F8 R23=0000000068A45250
R24=0000000068A61CA0 R25=000073CDF006BCB0 R26=FFFFFFFFFFFFFFFF R27=00000018BFE3A31A
R28=00000000FAEDBAF8 R29=0000000000A11900 R30=00000000000DB600 R31=000073CDBC066F08
NIP=000073CDF449B330 MSR=800000000280F033 ORIG_GPR3=000073CDDC55D340 CTR=000073CDF45B8650
LINK=000073CDF449B308 XER=0000000020000000 CCR=0000000044004844 SOFTE=0000000000000001
TRAP=0000000000000300 DAR=0000000000000018 dsisr=0000000040000000 RESULT=0000000000000000
Module=/bluebird/builds/bld_49023/sdk/xl6480/jre/lib/ppc64le/compressedrefs/libj9gc29.so
Module_base_address=000073CDF4440000
Target=2_90_20230412_49023 (Linux 5.4.0-144-generic)
CPU=ppc64le (4 logical CPUs) (0x1fcc00000 RAM)
----------- Stack Backtrace -----------
J9AllocateObjectNoGC+0x1f0 (0x000073CDF449B330 [libj9gc29.so+0x5b330])
fast_jitNewObject+0x9c (0x000073CDEFD03C5C [libj9jit29.so+0xa03c5c])
(0x000073CDEFD1865C [libj9jit29.so+0xa1865c])
runJavaThread+0x3b4 (0x000073CDF4F00374 [libj9vm29.so+0xc0374])
javaProtectedThreadProc+0xdc (0x000073CDF4F75FAC [libj9vm29.so+0x135fac])
omrsig_protect+0x358 (0x000073CDF4D35508 [libj9prt29.so+0x35508])
javaThreadProc+0x64 (0x000073CDF4F71684 [libj9vm29.so+0x131684])
thread_wrapper+0x1a8 (0x000073CDF4E113C8 [libj9thr29.so+0x113c8])
start_thread+0xe8 (0x000073CDF5D88838 [libpthread.so.0+0x8838])
clone+0x74 (0x000073CDF5C0B884 [libc.so.6+0x14b884])
Unhandled exception
Type=Segmentation error vmState=0x00000000
J9Generic_Signal_Number=00000018 Signal_Number=0000000b Error_Value=00000000 Signal_Code=00000001
Handler1=00007D1B15961460 Handler2=00007D1B15774100
R0=00007D1B141DBDE8 R1=00007D1A79E7B540 R2=00007D1B144A5500 R3=00000000FB6876E8
R4=0000000000000000 R5=00007D1A79E7B5C0 R6=00007D1B100B99F0 R7=0000000000000000
R8=000000002595BF00 R9=00007D1B10045B50 R10=0000000000000000 R11=0000000000000138
R12=00007D1B142F8650 R13=00007D1A79E868F0 R14=0000000000B124F0 R15=00000000009EE200
R16=00007D1AFC140038 R17=00000000BCB45438 R18=00000000FB6773B0 R19=00000000BCB573C0
R20=00000000BCB573C0 R21=0000000000000000 R22=00000000FB686108 R23=00000000FB6773F0
R24=00007D1B1006BCB0 R25=00000000FB686108 R26=00000000000C5900 R27=00007D1AF816A778
R28=00000000FB6876E8 R29=000000000000000A R30=0000000000000001 R31=00000000009EE200
NIP=00007D1B141DBE0C MSR=800000000280F033 ORIG_GPR3=00007D1B1591E804 CTR=00007D1B142F8650
LINK=00007D1B141DBDE8 XER=0000000020000000 CCR=0000000044004444 SOFTE=0000000000000001
TRAP=0000000000000300 DAR=0000000000000018 dsisr=0000000040000000 RESULT=0000000000000000
Module=/bluebird/builds/bld_49023/sdk/xl6480/jre/lib/ppc64le/compressedrefs/libj9gc29.so
Module_base_address=00007D1B14180000
Target=2_90_20230412_49023 (Linux 5.4.0-144-generic)
CPU=ppc64le (4 logical CPUs) (0x1fcc00000 RAM)
----------- Stack Backtrace -----------
J9AllocateIndexableObjectNoGC+0x21c (0x00007D1B141DBE0C [libj9gc29.so+0x5be0c])
fast_jitANewArray+0x64 (0x00007D1B14EB33E4 [libj9jit29.so+0xa033e4])
(0x00007D1B14EC883C [libj9jit29.so+0xa1883c])
runJavaThread+0x3b4 (0x00007D1B15940374 [libj9vm29.so+0xc0374])
javaProtectedThreadProc+0xdc (0x00007D1B159B5FAC [libj9vm29.so+0x135fac])
omrsig_protect+0x358 (0x00007D1B15775508 [libj9prt29.so+0x35508])
javaThreadProc+0x64 (0x00007D1B159B1684 [libj9vm29.so+0x131684])
thread_wrapper+0x1a8 (0x00007D1B158513C8 [libj9thr29.so+0x113c8])
start_thread+0xe8 (0x00007D1B167C8838 [libpthread.so.0+0x8838])
clone+0x74 (0x00007D1B1664B884 [libc.so.6+0x14b884])
---------------------------------------
```
Unfortunately there is no stored results due infra issue | test | crashes in object allocation path with on ppcle seen in internal testing original failed job unhandled exception type segmentation error vmstate signal number signal number error value signal code nip msr orig ctr link xer ccr softe trap dar dsisr result module bluebird builds bld sdk jre lib compressedrefs so module base address target linux generic cpu logical cpus ram stack backtrace environmentbasepv allocateinitialization gc allocateinitialization fast jitnewarray runjavathread javaprotectedthreadproc omrsig protect javathreadproc thread wrapper start thread clone reproduced in the grinder of unhandled exception type segmentation error vmstate signal number signal number error value signal code ffffffffffffffff ffffffffffffffff nip msr orig ctr link xer ccr softe trap dar dsisr result module bluebird builds bld sdk jre lib compressedrefs so module base address target linux generic cpu logical cpus ram stack backtrace fast jitnewobject runjavathread javaprotectedthreadproc omrsig protect javathreadproc thread wrapper start thread clone unhandled exception type segmentation error vmstate signal number signal number error value signal code nip msr orig ctr link xer ccr softe trap dar dsisr result module bluebird builds bld sdk jre lib compressedrefs so module base address target linux generic cpu logical cpus ram stack backtrace fast jitanewarray runjavathread javaprotectedthreadproc omrsig protect javathreadproc thread wrapper start thread clone unfortunately there is no stored results due infra issue | 1 |
286,479 | 8,788,873,561 | IssuesEvent | 2018-12-21 00:15:56 | ACEmulator/ACE | https://api.github.com/repos/ACEmulator/ACE | closed | Critical: Redesign Sequence system | Architectural priority:high type:Bug type:Enhancement | <!--
IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION YOUR ISSUE MIGHT BE CLOSED WITHOUT INVESTIGATING
-->
### Bug Report or Feature Request (mark with an `x`)
```
- [ x ] bug report -> please search issues before submitting
- [ ] feature request
```
Sequence system needs to be redesigned to be 1:1 with all of our Properties defined in enums
Overview is that each object gets a Sequences object that is keyed based on PropertyType + PropertyIndex
Each Private and Public Update for those properties would then reference this sequence in the same manner.
in ACLogView, StatType enum could be useful for the Property Type part, and we have all the properties in enums already.
| 1.0 | Critical: Redesign Sequence system - <!--
IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION YOUR ISSUE MIGHT BE CLOSED WITHOUT INVESTIGATING
-->
### Bug Report or Feature Request (mark with an `x`)
```
- [ x ] bug report -> please search issues before submitting
- [ ] feature request
```
Sequence system needs to be redesigned to be 1:1 with all of our Properties defined in enums
Overview is that each object gets a Sequences object that is keyed based on PropertyType + PropertyIndex
Each Private and Public Update for those properties would then reference this sequence in the same manner.
in ACLogView, StatType enum could be useful for the Property Type part, and we have all the properties in enums already.
| non_test | critical redesign sequence system if you don t fill out the following information your issue might be closed without investigating bug report or feature request mark with an x bug report please search issues before submitting feature request sequence system needs to be redesigned to be with all of our properties defined in enums overview is that each object gets a sequences object that is keyed based on propertytype propertyindex each private and public update for those properties would then reference this sequence in the same manner in aclogview stattype enum could be useful for the property type part and we have all the properties in enums already | 0 |
37,599 | 10,038,911,916 | IssuesEvent | 2019-07-18 16:06:07 | scality/metalk8s | https://api.github.com/repos/scality/metalk8s | opened | Properly integrate chart rendering in buildchain | build moonshot | https://github.com/scality/metalk8s/pull/1445 introduces a Helm chart that's rendered using a simple script and then checked into the repo. Instead, this should be done by the buildchain (creating the `chart.sls` file), such that this functionality can easily be reused for other charts. | 1.0 | Properly integrate chart rendering in buildchain - https://github.com/scality/metalk8s/pull/1445 introduces a Helm chart that's rendered using a simple script and then checked into the repo. Instead, this should be done by the buildchain (creating the `chart.sls` file), such that this functionality can easily be reused for other charts. | non_test | properly integrate chart rendering in buildchain introduces a helm chart that s rendered using a simple script and then checked into the repo instead this should be done by the buildchain creating the chart sls file such that this functionality can easily be reused for other charts | 0 |
31,546 | 4,264,432,266 | IssuesEvent | 2016-07-12 07:08:03 | RickyV33/Jama-Software | https://api.github.com/repos/RickyV33/Jama-Software | opened | Session expiration and logout handling | Design enhancement | *Chances mentioned the need to deal with session timeout and logging out of JamaTrace events | 1.0 | Session expiration and logout handling - *Chances mentioned the need to deal with session timeout and logging out of JamaTrace events | non_test | session expiration and logout handling chances mentioned the need to deal with session timeout and logging out of jamatrace events | 0 |
96,513 | 20,026,252,823 | IssuesEvent | 2022-02-01 21:41:17 | tconbeer/sqlfmt | https://api.github.com/repos/tconbeer/sqlfmt | closed | Fix "complex_logic" issue in src/sqlfmt/splitter.py | code climate | Consider simplifying this complex logical expression.
https://codeclimate.com/github/tconbeer/sqlfmt/src/sqlfmt/splitter.py#issue_61f46a9f597a090001000042 | 1.0 | Fix "complex_logic" issue in src/sqlfmt/splitter.py - Consider simplifying this complex logical expression.
https://codeclimate.com/github/tconbeer/sqlfmt/src/sqlfmt/splitter.py#issue_61f46a9f597a090001000042 | non_test | fix complex logic issue in src sqlfmt splitter py consider simplifying this complex logical expression | 0 |
348,537 | 24,916,250,845 | IssuesEvent | 2022-10-30 12:40:38 | hashicorp/terraform | https://api.github.com/repos/hashicorp/terraform | opened | New developer.hashicorp.com has no way to list built-in functions | documentation new | ### Terraform Version
```shell
Doesn't matter. This is an issue with https://developer.hashicorp.com.
```
### Affected Pages
https://developer.hashicorp.com/terraform/language/functions/coalescelist (as one example)
### What is the docs issue?
Example - I'm searching for a function. I Google around thinking coalescelist might be my thing and I'm taken to https://developer.hashicorp.com/terraform/language/functions/coalescelist. It turns out that's not what I want, so I need to keep looking. But...
There's nowhere in https://developer.hashicorp.com (either on this specific page, in TOC or anywhere I can find) that takes me to a list of all the available functions in TF. Maybe there is, but it's certainly hidden quite well.
The old docs stuff had a bookmarkable page that allowed me to find and tool around in all the available functions.
### Proposal
Provide easy, in-your-face links to a single index of all available functions. Just like in the old docs. If this exists today in developer.hashicorp.com it is quite well hidden.
### References
_No response_ | 1.0 | New developer.hashicorp.com has no way to list built-in functions - ### Terraform Version
```shell
Doesn't matter. This is an issue with https://developer.hashicorp.com.
```
### Affected Pages
https://developer.hashicorp.com/terraform/language/functions/coalescelist (as one example)
### What is the docs issue?
Example - I'm searching for a function. I Google around thinking coalescelist might be my thing and I'm taken to https://developer.hashicorp.com/terraform/language/functions/coalescelist. It turns out that's not what I want, so I need to keep looking. But...
There's nowhere in https://developer.hashicorp.com (either on this specific page, in TOC or anywhere I can find) that takes me to a list of all the available functions in TF. Maybe there is, but it's certainly hidden quite well.
The old docs stuff had a bookmarkable page that allowed me to find and tool around in all the available functions.
### Proposal
Provide easy, in-your-face links to a single index of all available functions. Just like in the old docs. If this exists today in developer.hashicorp.com it is quite well hidden.
### References
_No response_ | non_test | new developer hashicorp com has no way to list built in functions terraform version shell doesn t matter this is an issue with affected pages as one example what is the docs issue example i m searching for a function i google around thinking coalescelist might be my thing and i m taken to it turns out that s not what i want so i need to keep looking but there s nowhere in either on this specific page in toc or anywhere i can find that takes me to a list of all the available functions in tf maybe there is but it s certainly hidden quite well the old docs stuff had a bookmarkable page that allowed me to find and tool around in all the available functions proposal provide easy in your face links to a single index of all available functions just like in the old docs if this exists today in developer hashicorp com it is quite well hidden references no response | 0 |
61,494 | 6,741,649,909 | IssuesEvent | 2017-10-20 02:21:35 | MikeSmvl/classifieds | https://api.github.com/repos/MikeSmvl/classifieds | closed | Setting Build and test Scripts | Points: 3 Priority: Medium Type: Test | For every pull request raised against any branch, we want to have integrated scripts that test the functionality of the website and provide us with the results . | 1.0 | Setting Build and test Scripts - For every pull request raised against any branch, we want to have integrated scripts that test the functionality of the website and provide us with the results . | test | setting build and test scripts for every pull request raised against any branch we want to have integrated scripts that test the functionality of the website and provide us with the results | 1 |
279,684 | 24,247,439,557 | IssuesEvent | 2022-09-27 11:46:25 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | opened | Claim adaptive captcha UI is broken on staging env | bug feature/rewards QA/Yes QA/Test-Plan-Specified OS/Desktop | <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
Claim adaptive captcha UI is broken on staging env
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Clean profile 1.46.14 staging env
2. Enable rewards and view an ad
3. Schedule a adaptive captcha
4. Click on adaptive captcha notification to solve the captcha
5. Captcha claim UI is broken
## Actual result:
<!--Please add screenshots if needed-->

## Expected result:
Should be abl eto clear the captcha
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
Easy
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.46.14 Chromium: 106.0.5249.61 (Official Build) nightly (64-bit)
-- | --
Revision | 511755355844955cd3e264779baf0dd38212a4d0-refs/branch-heads/5249@{#569}
OS | Windows 10 Version 21H2 (Build 19044.2006)
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? NA
- Can you reproduce this issue with the beta channel? NA
- Can you reproduce this issue with the nightly channel? Yes
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields? NA
- Does the issue resolve itself when disabling Brave Rewards? NA
- Is the issue reproducible on the latest version of Chrome? NA
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
cc: @Miyayes @gpestana @brave/qa-team @zenparsing | 1.0 | Claim adaptive captcha UI is broken on staging env - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
Claim adaptive captcha UI is broken on staging env
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Clean profile 1.46.14 staging env
2. Enable rewards and view an ad
3. Schedule a adaptive captcha
4. Click on adaptive captcha notification to solve the captcha
5. Captcha claim UI is broken
## Actual result:
<!--Please add screenshots if needed-->

## Expected result:
Should be abl eto clear the captcha
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
Easy
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.46.14 Chromium: 106.0.5249.61 (Official Build) nightly (64-bit)
-- | --
Revision | 511755355844955cd3e264779baf0dd38212a4d0-refs/branch-heads/5249@{#569}
OS | Windows 10 Version 21H2 (Build 19044.2006)
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? NA
- Can you reproduce this issue with the beta channel? NA
- Can you reproduce this issue with the nightly channel? Yes
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields? NA
- Does the issue resolve itself when disabling Brave Rewards? NA
- Is the issue reproducible on the latest version of Chrome? NA
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
cc: @Miyayes @gpestana @brave/qa-team @zenparsing | test | claim adaptive captcha ui is broken on staging env have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description claim adaptive captcha ui is broken on staging env steps to reproduce clean profile staging env enable rewards and view an ad schedule a adaptive captcha click on adaptive captcha notification to solve the captcha captcha claim ui is broken actual result expected result should be abl eto clear the captcha reproduces how often easy brave version brave version info brave chromium official build nightly bit revision refs branch heads os windows version build version channel information can you reproduce this issue with the current release na can you reproduce this issue with the beta channel na can you reproduce this issue with the nightly channel yes other additional information does the issue resolve itself when disabling brave shields na does the issue resolve itself when disabling brave rewards na is the issue reproducible on the latest version of chrome na miscellaneous information cc miyayes gpestana brave qa team zenparsing | 1 |
733,867 | 25,326,516,926 | IssuesEvent | 2022-11-18 09:44:36 | hashicorp/terraform-cdk | https://api.github.com/repos/hashicorp/terraform-cdk | closed | Azurerm | LinuxWebAppSiteConfigIpRestriction + `Token.null_value()` not working | bug priority/important-longterm feature/tokens | <!--- Please keep this note for the community --->
### Community Note
- Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
- Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
- If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### cdktf & Language Versions
<img width="360" alt="image" src="https://user-images.githubusercontent.com/51444462/202291362-9f2d77a7-5db8-4b88-bbc0-0c3205b1a25d.png">
### Affected Resource(s)
LinuxWebAppSiteConfigIpRestriction
### Debug Output
### Expected Behavior
Token.null_value should be processed to null
### Actual Behavior
<img width="987" alt="image" src="https://user-images.githubusercontent.com/51444462/202291940-28d9a39c-8c54-4764-8fac-0d5240fb38fe.png">
### Steps to Reproduce
<img width="337" alt="image" src="https://user-images.githubusercontent.com/51444462/202291525-ec7016ea-9926-43df-ac92-f9a613c5f739.png">
### Important Factoids
<!--- Are there anything atypical about your accounts that we should know? --->
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor documentation? For example:
--->
- #0000
| 1.0 | Azurerm | LinuxWebAppSiteConfigIpRestriction + `Token.null_value()` not working - <!--- Please keep this note for the community --->
### Community Note
- Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
- Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
- If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### cdktf & Language Versions
<img width="360" alt="image" src="https://user-images.githubusercontent.com/51444462/202291362-9f2d77a7-5db8-4b88-bbc0-0c3205b1a25d.png">
### Affected Resource(s)
LinuxWebAppSiteConfigIpRestriction
### Debug Output
### Expected Behavior
Token.null_value should be processed to null
### Actual Behavior
<img width="987" alt="image" src="https://user-images.githubusercontent.com/51444462/202291940-28d9a39c-8c54-4764-8fac-0d5240fb38fe.png">
### Steps to Reproduce
<img width="337" alt="image" src="https://user-images.githubusercontent.com/51444462/202291525-ec7016ea-9926-43df-ac92-f9a613c5f739.png">
### Important Factoids
<!--- Are there anything atypical about your accounts that we should know? --->
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor documentation? For example:
--->
- #0000
| non_test | azurerm linuxwebappsiteconfigiprestriction token null value not working community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment cdktf language versions img width alt image src affected resource s linuxwebappsiteconfigiprestriction debug output expected behavior token null value should be processed to null actual behavior img width alt image src steps to reproduce img width alt image src important factoids references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor documentation for example | 0 |
27,183 | 13,195,065,913 | IssuesEvent | 2020-08-13 17:57:55 | OpenLiberty/open-liberty | https://api.github.com/repos/OpenLiberty/open-liberty | closed | Microprofile GraphQL Performance Regression | in:MicroProfile/GraphQL performance regression | We have a MPGraphQL regression in 20.0.0.9. I see a new builder factory being created.
```
Parent 0 0.00 17.10 0 430 J:io/smallrye/graphql/execution/context/SmallRyeContext.<init>(Ljavax/json/JsonObject;)V
Self 0 0.00 17.10 0 430 J:javax/json/Json.createBuilderFactory(Ljava/util/Map;)Ljavax/json/JsonBuilderFactory;
Child 0 0.00 17.10 0 430 J:javax/json/spi/JsonProvider.provider()Ljavax/json/spi/JsonProvider;
```
This looks to be a new class: https://github.com/smallrye/smallrye-graphql/blob/master/server/implementation/src/main/java/io/smallrye/graphql/execution/context/SmallRyeContext.java
Need a cache fix like: https://github.com/smallrye/smallrye-graphql/pull/186 | True | Microprofile GraphQL Performance Regression - We have a MPGraphQL regression in 20.0.0.9. I see a new builder factory being created.
```
Parent 0 0.00 17.10 0 430 J:io/smallrye/graphql/execution/context/SmallRyeContext.<init>(Ljavax/json/JsonObject;)V
Self 0 0.00 17.10 0 430 J:javax/json/Json.createBuilderFactory(Ljava/util/Map;)Ljavax/json/JsonBuilderFactory;
Child 0 0.00 17.10 0 430 J:javax/json/spi/JsonProvider.provider()Ljavax/json/spi/JsonProvider;
```
This looks to be a new class: https://github.com/smallrye/smallrye-graphql/blob/master/server/implementation/src/main/java/io/smallrye/graphql/execution/context/SmallRyeContext.java
Need a cache fix like: https://github.com/smallrye/smallrye-graphql/pull/186 | non_test | microprofile graphql performance regression we have a mpgraphql regression in i see a new builder factory being created parent j io smallrye graphql execution context smallryecontext ljavax json jsonobject v self j javax json json createbuilderfactory ljava util map ljavax json jsonbuilderfactory child j javax json spi jsonprovider provider ljavax json spi jsonprovider this looks to be a new class need a cache fix like | 0 |
33,100 | 6,155,977,864 | IssuesEvent | 2017-06-28 15:45:02 | aurelia/i18n | https://api.github.com/repos/aurelia/i18n | closed | "Unable to find module with ID: aurelia-i18n" (Webpack) | documentation | **I'm submitting a bug report**
* **Library Version:**
1.6.1
**Please tell us about your environment:**
* **Operating System:**
Windows 10
* **Node Version:**
7.3.0
* **NPM Version:**
4.5.0
* **JSPM OR Webpack AND Version**
webpack 3.0.0
* **Browser:**
all
* **Language:**
TypeScript 2.4.1
**Current behavior:**
As soon as I include `.plugin('aurelia-i18n', ...` I get this error at runtime, no errors during typescript compilation:
```
bluebird.js:5253 Error: Unable to find module with ID: aurelia-i18n
at WebpackLoader.<anonymous> (aurelia-loader-webpack.js:187)
at step (aurelia-loader-webpack.js:36)
at Object.next (aurelia-loader-webpack.js:17)
at aurelia-loader-webpack.js:11
at Promise._execute (bluebird.js:968)
```
**Instructions to reproduce:**
I have not created any gist (at least not yet) but I reproduced it like this, based on the `skeleton-navigation repo` and the i18n guide from the documentation: http://aurelia.io/hub.html#/doc/article/aurelia/i18n/latest/i18n-with-aurelia
1. clone the aurelia "skeleton-navigation" report**
2. `npm install aureila-i18n --save`
3. create `locales\en\translation.json` and `locales\sv\translation.json`
4. add the following to `src\main.ts`:
* `import {I18N, Backend} from 'aurelia-i18n';`
* `.plugin('aurelia-i18n', (instance) => {})`
* **What is the expected behavior?**
That aureila-i18n is loaded and that the translations works as expected.
Am I missing something vital? I've tried to include the plugin initialization code within the brackets in `.plugin('aurelia-i18n...`, I tried to install and use `i18next-xhr-backend` and some other changes but nothing seems to help. Any ideas on what I'm missing? | 1.0 | "Unable to find module with ID: aurelia-i18n" (Webpack) - **I'm submitting a bug report**
* **Library Version:**
1.6.1
**Please tell us about your environment:**
* **Operating System:**
Windows 10
* **Node Version:**
7.3.0
* **NPM Version:**
4.5.0
* **JSPM OR Webpack AND Version**
webpack 3.0.0
* **Browser:**
all
* **Language:**
TypeScript 2.4.1
**Current behavior:**
As soon as I include `.plugin('aurelia-i18n', ...` I get this error at runtime, no errors during typescript compilation:
```
bluebird.js:5253 Error: Unable to find module with ID: aurelia-i18n
at WebpackLoader.<anonymous> (aurelia-loader-webpack.js:187)
at step (aurelia-loader-webpack.js:36)
at Object.next (aurelia-loader-webpack.js:17)
at aurelia-loader-webpack.js:11
at Promise._execute (bluebird.js:968)
```
**Instructions to reproduce:**
I have not created any gist (at least not yet) but I reproduced it like this, based on the `skeleton-navigation repo` and the i18n guide from the documentation: http://aurelia.io/hub.html#/doc/article/aurelia/i18n/latest/i18n-with-aurelia
1. clone the aurelia "skeleton-navigation" report**
2. `npm install aureila-i18n --save`
3. create `locales\en\translation.json` and `locales\sv\translation.json`
4. add the following to `src\main.ts`:
* `import {I18N, Backend} from 'aurelia-i18n';`
* `.plugin('aurelia-i18n', (instance) => {})`
* **What is the expected behavior?**
That aureila-i18n is loaded and that the translations works as expected.
Am I missing something vital? I've tried to include the plugin initialization code within the brackets in `.plugin('aurelia-i18n...`, I tried to install and use `i18next-xhr-backend` and some other changes but nothing seems to help. Any ideas on what I'm missing? | non_test | unable to find module with id aurelia webpack i m submitting a bug report library version please tell us about your environment operating system windows node version npm version jspm or webpack and version webpack browser all language typescript current behavior as soon as i include plugin aurelia i get this error at runtime no errors during typescript compilation bluebird js error unable to find module with id aurelia at webpackloader aurelia loader webpack js at step aurelia loader webpack js at object next aurelia loader webpack js at aurelia loader webpack js at promise execute bluebird js instructions to reproduce i have not created any gist at least not yet but i reproduced it like this based on the skeleton navigation repo and the guide from the documentation clone the aurelia skeleton navigation report npm install aureila save create locales en translation json and locales sv translation json add the following to src main ts import backend from aurelia plugin aurelia instance what is the expected behavior that aureila is loaded and that the translations works as expected am i missing something vital i ve tried to include the plugin initialization code within the brackets in plugin aurelia i tried to install and use xhr backend and some other changes but nothing seems to help any ideas on what i m missing | 0 |
62,136 | 25,883,649,708 | IssuesEvent | 2022-12-14 13:06:04 | kyma-project/istio | https://api.github.com/repos/kyma-project/istio | closed | Istio installation with istio-operator api | area/service-mesh | <!-- Thank you for your contribution. Before you submit the issue:
1. Search open and closed issues for duplicates.
2. Read the contributing guidelines.
-->
**Description**
Implement installation method using imported istio-operator install method. First implementation should support installation of istio in version 1.13.x on remote cluster.
ACs:
- [ ] uses `install` method of imported istio operator sources
- [ ] supports passing kubeconfig as a string
- [ ] installs istio 1.13.x
**Reasons**
Modularisation of kyma components | 1.0 | Istio installation with istio-operator api - <!-- Thank you for your contribution. Before you submit the issue:
1. Search open and closed issues for duplicates.
2. Read the contributing guidelines.
-->
**Description**
Implement installation method using imported istio-operator install method. First implementation should support installation of istio in version 1.13.x on remote cluster.
ACs:
- [ ] uses `install` method of imported istio operator sources
- [ ] supports passing kubeconfig as a string
- [ ] installs istio 1.13.x
**Reasons**
Modularisation of kyma components | non_test | istio installation with istio operator api thank you for your contribution before you submit the issue search open and closed issues for duplicates read the contributing guidelines description implement installation method using imported istio operator install method first implementation should support installation of istio in version x on remote cluster acs uses install method of imported istio operator sources supports passing kubeconfig as a string installs istio x reasons modularisation of kyma components | 0 |
251,421 | 21,480,701,915 | IssuesEvent | 2022-04-26 17:26:15 | Scotsguy/now-playing | https://api.github.com/repos/Scotsguy/now-playing | closed | [Bug?] Repeated toast dialogs | bug branch:needs-testing needs-repro | Found this happen when testing. Player was spammed with end boss music notifications upon joining the end for the first time. Haven't been able to reproduce since.
**Version**: Running version 1.1.4
**Video link**: https://streamable.com/ygjo9q | 1.0 | [Bug?] Repeated toast dialogs - Found this happen when testing. Player was spammed with end boss music notifications upon joining the end for the first time. Haven't been able to reproduce since.
**Version**: Running version 1.1.4
**Video link**: https://streamable.com/ygjo9q | test | repeated toast dialogs found this happen when testing player was spammed with end boss music notifications upon joining the end for the first time haven t been able to reproduce since version running version video link | 1 |
264,327 | 23,111,823,536 | IssuesEvent | 2022-07-27 13:38:11 | ethereumjs/ethereumjs-monorepo | https://api.github.com/repos/ethereumjs/ethereumjs-monorepo | closed | Merge tracking issue | type: meta type: tests package: vm | Tracking issue for things which need to change which I noticed in #2064
- [ ] Rename `td` to `ttd` where appropriate. TD: total difficulty. TTD: total terminal difficulty. In `merge` hardfork this should be `ttd`, it is currently `td` (which is confusing). Inputs to most (if not all) methods in common should remain `td` #2075
- [ ] `td` is currently of type `number` in Common, I think we should change to either string or bigint #2075
Mainnet TD is currently larger than max safe number, so we should definitely change the type of `number` of `ttd` to either `string` or `bigint` (probably string for JSON) | 1.0 | Merge tracking issue - Tracking issue for things which need to change which I noticed in #2064
- [ ] Rename `td` to `ttd` where appropriate. TD: total difficulty. TTD: total terminal difficulty. In `merge` hardfork this should be `ttd`, it is currently `td` (which is confusing). Inputs to most (if not all) methods in common should remain `td` #2075
- [ ] `td` is currently of type `number` in Common, I think we should change to either string or bigint #2075
Mainnet TD is currently larger than max safe number, so we should definitely change the type of `number` of `ttd` to either `string` or `bigint` (probably string for JSON) | test | merge tracking issue tracking issue for things which need to change which i noticed in rename td to ttd where appropriate td total difficulty ttd total terminal difficulty in merge hardfork this should be ttd it is currently td which is confusing inputs to most if not all methods in common should remain td td is currently of type number in common i think we should change to either string or bigint mainnet td is currently larger than max safe number so we should definitely change the type of number of ttd to either string or bigint probably string for json | 1 |
170,584 | 14,264,718,899 | IssuesEvent | 2020-11-20 16:07:00 | Revolutionary-Games/Thrive | https://api.github.com/repos/Revolutionary-Games/Thrive | opened | Clarify TranslationServer.Translate usage in translation guide | documentation | add localization note that `TranslationServer.Translate("just a string");` only works. If some variable is used, it won't work as the text is extracted by an analysis tool.
So all text that needs to be detected, needs to be a basic string inside the `Translate` call. | 1.0 | Clarify TranslationServer.Translate usage in translation guide - add localization note that `TranslationServer.Translate("just a string");` only works. If some variable is used, it won't work as the text is extracted by an analysis tool.
So all text that needs to be detected, needs to be a basic string inside the `Translate` call. | non_test | clarify translationserver translate usage in translation guide add localization note that translationserver translate just a string only works if some variable is used it won t work as the text is extracted by an analysis tool so all text that needs to be detected needs to be a basic string inside the translate call | 0 |
172,107 | 14,350,332,429 | IssuesEvent | 2020-11-29 20:25:06 | andrewdimmer/csi-3370-software-project | https://api.github.com/repos/andrewdimmer/csi-3370-software-project | closed | [Display_01] Define Testing Plan for Display_01 Production Classes | Display_01 documentation | **What is the use case? Please describe.**
For test driven development, it is important to determine a test plan before just jumping in and writing tests and code. For this sprint task, you'll be developing a testing plan for all of the classes in the `edu.oakland.production.display01` package. Note: This issue is similar to #51, but is tracked differently as helper and production classes have different deliverable dates.
**Prerequisites**
Please makes sure that #38 is closed and merged to `main` before starting this, to ensure that all of the documentation you need for this is in the correct spot.
**Sprint Task Setup**
To be done by @EC01EC , @stevecatherman , or @casher-ou before the rest of the team starts working on it:
1. Use the documentation at https://github.com/andrewdimmer/csi-3370-software-project/tree/main/docs/uml-class-diagrams/display01/production to identify the classes that require a test plan.
2. Edit this issue message to add a checkbox for each class, and assign it to the correct team member(s). Be sure to balance your team between this issue at #51 . Note that the teams assigned here should also be the same teams in charge of building the appropriate classes in #67 .
**Software Setup**
For this task, you'll be updating PlantUML diagrams. To preview them on your local machine, set up:
- [Visual Studio Code](https://code.visualstudio.com/)
- [PlantUML Visual Studio Code Extension](https://marketplace.visualstudio.com/items?itemName=jebbs.plantuml)
- [Graphviz](https://graphviz.org/)
After that, you can open a .puml document, right click on it, and "Preview Current Diagram".
If you don't want to set that up though, you don't have to. When you commit, GitHub has been configured to automatically update all the diagrams, and you can push to GitHub then view the diagrams from there and/or pull them back to your current machine.
**Steps**
To be done by each team member for each class they are developing a test plan for:
1. Each team member should watch the first ~9 minutes of @andrewdimmer 's [Test Driven Development Crash Course](https://youtu.be/eSc0aY4GhPs)
2. One person should create a branch off of main called `define-Name_Of_Test_Class` (ex. the `Demo` test plan gets defined on a branch called `define-DemoTests`). Everyone else in the programming pair or group should then checkout that branch.
3. Open the Test Class UML Diagram at `/docs/uml-class-diagrams/display01/test/Name_Of_Test_Class/Name_Of_Test_Class.puml`.
4. Using the information in the video, record your tests on the Test Class UML Diagram you opened in step 3. Need help with the PlantUML Diagram itself? Check out https://plantuml.com/class-diagram. Feel free to use the existing class diagrams already provided at https://github.com/andrewdimmer/csi-3370-software-project/tree/main/docs/uml-class-diagrams/display01/production to help guide you in creating the test classes!
5. Commit your changes with the message `[Display_01] Define Name_Of_Test_Class`
6. Push your changes
7. Create a pull request with the name `[Display_01] Define Name_Of_Test_Class` and the rest of the information like description, tags, etc. For reference, the settings on the right hand side should match this issue, with the exception that the assignees on the PR are just those who are pushing code.
8. Once your test plan is complete, add @andrewdimmer as a reviewer.
**Tasks to Complete**
- [x] Allocate Team Members to Classes - @EC01EC , @stevecatherman , or @casher-ou
- [x] Define DisplayGpsInterfaceImplementationTests - @reedsikorskiOU or @DLidz
- [x] Define DisplayGpsManagerImplementationTests - @atpoynter or @elilley15
- [x] Define DisplayGpsReceiverImplementationTests - @EC01EC or @casher-ou
**Additional context**
Note: if for some reason you need to change a public interface or shared data type, reach out to @andrewdimmer who can help you manage making that BREAKING CHANGE.
Note: if you run into merge conflicts, let @andrewdimmer know and he can help you resolve them. | 1.0 | [Display_01] Define Testing Plan for Display_01 Production Classes - **What is the use case? Please describe.**
For test driven development, it is important to determine a test plan before just jumping in and writing tests and code. For this sprint task, you'll be developing a testing plan for all of the classes in the `edu.oakland.production.display01` package. Note: This issue is similar to #51, but is tracked differently as helper and production classes have different deliverable dates.
**Prerequisites**
Please makes sure that #38 is closed and merged to `main` before starting this, to ensure that all of the documentation you need for this is in the correct spot.
**Sprint Task Setup**
To be done by @EC01EC , @stevecatherman , or @casher-ou before the rest of the team starts working on it:
1. Use the documentation at https://github.com/andrewdimmer/csi-3370-software-project/tree/main/docs/uml-class-diagrams/display01/production to identify the classes that require a test plan.
2. Edit this issue message to add a checkbox for each class, and assign it to the correct team member(s). Be sure to balance your team between this issue at #51 . Note that the teams assigned here should also be the same teams in charge of building the appropriate classes in #67 .
**Software Setup**
For this task, you'll be updating PlantUML diagrams. To preview them on your local machine, set up:
- [Visual Studio Code](https://code.visualstudio.com/)
- [PlantUML Visual Studio Code Extension](https://marketplace.visualstudio.com/items?itemName=jebbs.plantuml)
- [Graphviz](https://graphviz.org/)
After that, you can open a .puml document, right click on it, and "Preview Current Diagram".
If you don't want to set that up though, you don't have to. When you commit, GitHub has been configured to automatically update all the diagrams, and you can push to GitHub then view the diagrams from there and/or pull them back to your current machine.
**Steps**
To be done by each team member for each class they are developing a test plan for:
1. Each team member should watch the first ~9 minutes of @andrewdimmer 's [Test Driven Development Crash Course](https://youtu.be/eSc0aY4GhPs)
2. One person should create a branch off of main called `define-Name_Of_Test_Class` (ex. the `Demo` test plan gets defined on a branch called `define-DemoTests`). Everyone else in the programming pair or group should then checkout that branch.
3. Open the Test Class UML Diagram at `/docs/uml-class-diagrams/display01/test/Name_Of_Test_Class/Name_Of_Test_Class.puml`.
4. Using the information in the video, record your tests on the Test Class UML Diagram you opened in step 3. Need help with the PlantUML Diagram itself? Check out https://plantuml.com/class-diagram. Feel free to use the existing class diagrams already provided at https://github.com/andrewdimmer/csi-3370-software-project/tree/main/docs/uml-class-diagrams/display01/production to help guide you in creating the test classes!
5. Commit your changes with the message `[Display_01] Define Name_Of_Test_Class`
6. Push your changes
7. Create a pull request with the name `[Display_01] Define Name_Of_Test_Class` and the rest of the information like description, tags, etc. For reference, the settings on the right hand side should match this issue, with the exception that the assignees on the PR are just those who are pushing code.
8. Once your test plan is complete, add @andrewdimmer as a reviewer.
**Tasks to Complete**
- [x] Allocate Team Members to Classes - @EC01EC , @stevecatherman , or @casher-ou
- [x] Define DisplayGpsInterfaceImplementationTests - @reedsikorskiOU or @DLidz
- [x] Define DisplayGpsManagerImplementationTests - @atpoynter or @elilley15
- [x] Define DisplayGpsReceiverImplementationTests - @EC01EC or @casher-ou
**Additional context**
Note: if for some reason you need to change a public interface or shared data type, reach out to @andrewdimmer who can help you manage making that BREAKING CHANGE.
Note: if you run into merge conflicts, let @andrewdimmer know and he can help you resolve them. | non_test | define testing plan for display production classes what is the use case please describe for test driven development it is important to determine a test plan before just jumping in and writing tests and code for this sprint task you ll be developing a testing plan for all of the classes in the edu oakland production package note this issue is similar to but is tracked differently as helper and production classes have different deliverable dates prerequisites please makes sure that is closed and merged to main before starting this to ensure that all of the documentation you need for this is in the correct spot sprint task setup to be done by stevecatherman or casher ou before the rest of the team starts working on it use the documentation at to identify the classes that require a test plan edit this issue message to add a checkbox for each class and assign it to the correct team member s be sure to balance your team between this issue at note that the teams assigned here should also be the same teams in charge of building the appropriate classes in software setup for this task you ll be updating plantuml diagrams to preview them on your local machine set up after that you can open a puml document right click on it and preview current diagram if you don t want to set that up though you don t have to when you commit github has been configured to automatically update all the diagrams and you can push to github then view the diagrams from there and or pull them back to your current machine steps to be done by each team member for each class they are developing a test plan for each team member should watch the first minutes of andrewdimmer s one person should create a branch off of main called define name of test class ex the demo test plan gets defined on a branch called define demotests everyone else in the programming pair or group should then checkout that branch open the test class uml diagram at docs uml class diagrams test name of test class name of test class puml using the information in the video record your tests on the test class uml diagram you opened in step need help with the plantuml diagram itself check out feel free to use the existing class diagrams already provided at to help guide you in creating the test classes commit your changes with the message define name of test class push your changes create a pull request with the name define name of test class and the rest of the information like description tags etc for reference the settings on the right hand side should match this issue with the exception that the assignees on the pr are just those who are pushing code once your test plan is complete add andrewdimmer as a reviewer tasks to complete allocate team members to classes stevecatherman or casher ou define displaygpsinterfaceimplementationtests reedsikorskiou or dlidz define displaygpsmanagerimplementationtests atpoynter or define displaygpsreceiverimplementationtests or casher ou additional context note if for some reason you need to change a public interface or shared data type reach out to andrewdimmer who can help you manage making that breaking change note if you run into merge conflicts let andrewdimmer know and he can help you resolve them | 0 |
264,375 | 20,017,312,389 | IssuesEvent | 2022-02-01 13:23:03 | redhat-documentation/vale-at-red-hat | https://api.github.com/repos/redhat-documentation/vale-at-red-hat | closed | Create a reference guide | documentation | Create a reference guide containing a page for each rule, that we could use as a link in the rule.
* Generate with a script new reference pages (only if they don't already exist).
* Each new page displays the name of the rule as a title, the message of the rule as main content, and the content of the rule file in a source block. Manual authoring necessary to complete the information.
* Generate the navigation from the pages list.
* Generate an overview with some helpers to navigate the pages.
Review the Reference guide pages to make them meaningful:
* When applicable, copy over the word usage description from https://redhat-documentation.github.io/supplementary-style-guide/
* When necessary, complete the usage description using the rule source file.
| 1.0 | Create a reference guide - Create a reference guide containing a page for each rule, that we could use as a link in the rule.
* Generate with a script new reference pages (only if they don't already exist).
* Each new page displays the name of the rule as a title, the message of the rule as main content, and the content of the rule file in a source block. Manual authoring necessary to complete the information.
* Generate the navigation from the pages list.
* Generate an overview with some helpers to navigate the pages.
Review the Reference guide pages to make them meaningful:
* When applicable, copy over the word usage description from https://redhat-documentation.github.io/supplementary-style-guide/
* When necessary, complete the usage description using the rule source file.
| non_test | create a reference guide create a reference guide containing a page for each rule that we could use as a link in the rule generate with a script new reference pages only if they don t already exist each new page displays the name of the rule as a title the message of the rule as main content and the content of the rule file in a source block manual authoring necessary to complete the information generate the navigation from the pages list generate an overview with some helpers to navigate the pages review the reference guide pages to make them meaningful when applicable copy over the word usage description from when necessary complete the usage description using the rule source file | 0 |
125,738 | 10,351,549,548 | IssuesEvent | 2019-09-05 07:11:23 | LiskHQ/lisk-sdk | https://api.github.com/repos/LiskHQ/lisk-sdk | closed | Unit tests for eviction of peers from peer buckets | elements/P2P type: test | ### Description
After the implementation of peer buckets its important to have an extensive test for peer eviction from the buckets of each peer lists that have different strategies of eviction.
### Motivation
Since each of these peer list's have different eviction strategy on their buckets we need to test it by filling up the buckets with valid IPs and test the strategy based on time or random evictions.
### Acceptance Criteria
- Should test eviction based on time in `newPeers` and random eviction if no old peers
- Should test eviction based on random when the bucket is full in `triedPeers` and moved to `newPeers`
| 1.0 | Unit tests for eviction of peers from peer buckets - ### Description
After the implementation of peer buckets its important to have an extensive test for peer eviction from the buckets of each peer lists that have different strategies of eviction.
### Motivation
Since each of these peer list's have different eviction strategy on their buckets we need to test it by filling up the buckets with valid IPs and test the strategy based on time or random evictions.
### Acceptance Criteria
- Should test eviction based on time in `newPeers` and random eviction if no old peers
- Should test eviction based on random when the bucket is full in `triedPeers` and moved to `newPeers`
| test | unit tests for eviction of peers from peer buckets description after the implementation of peer buckets its important to have an extensive test for peer eviction from the buckets of each peer lists that have different strategies of eviction motivation since each of these peer list s have different eviction strategy on their buckets we need to test it by filling up the buckets with valid ips and test the strategy based on time or random evictions acceptance criteria should test eviction based on time in newpeers and random eviction if no old peers should test eviction based on random when the bucket is full in triedpeers and moved to newpeers | 1 |
483,110 | 13,918,632,961 | IssuesEvent | 2020-10-21 07:59:05 | redhat-developer/vscode-openshift-tools | https://api.github.com/repos/redhat-developer/vscode-openshift-tools | closed | Show devfile component type quickpick items first | demo kind/enhancement priority/major | Currently quick pick is loaded directly from odo json output that contains two parts s2i type list and devfile type list. That leads to appearance below.

Instead devfile components should be showed first

| 1.0 | Show devfile component type quickpick items first - Currently quick pick is loaded directly from odo json output that contains two parts s2i type list and devfile type list. That leads to appearance below.

Instead devfile components should be showed first

| non_test | show devfile component type quickpick items first currently quick pick is loaded directly from odo json output that contains two parts type list and devfile type list that leads to appearance below instead devfile components should be showed first | 0 |
217,820 | 16,739,586,912 | IssuesEvent | 2021-06-11 08:11:15 | gen-classroom/projet-oe | https://api.github.com/repos/gen-classroom/projet-oe | closed | S16 - Code coverage | documentation | ##### Description :
Configurer Jacoco et ajouter les données au rapport.
##### Temps :
estimé : 20 min
min : 10 min
max : 40 min | 1.0 | S16 - Code coverage - ##### Description :
Configurer Jacoco et ajouter les données au rapport.
##### Temps :
estimé : 20 min
min : 10 min
max : 40 min | non_test | code coverage description configurer jacoco et ajouter les données au rapport temps estimé min min min max min | 0 |
278,672 | 24,167,330,563 | IssuesEvent | 2022-09-22 16:02:07 | kyma-project/lifecycle-manager | https://api.github.com/repos/kyma-project/lifecycle-manager | closed | Create smoke-test pipeline for lifecycle-manager | area/tests area/reconciler 2022-Q3 | **Description**
A pipeline for running smoke tests has to be configured in Prow. The pipeline has to execute test cases (e.g. by using Ginko, EnvTest or Go Test) which verify the correct behaviour of the lifecycle manager within a K8s runtime context (e.g. by using via K3d or EnvTest). The executed tests don't have to be comprehensive, but have to run fast and should give a clear indication whether the lifecycle manager works as expected.
AC:
- [ ] Pipeline is created in Prow and configured as quality gate in Github for PRs to the lifecycle-manager repository
- [ ] Pipeline triggers the smoke test and allows merging of PRs only if the test was passed successfully
| 1.0 | Create smoke-test pipeline for lifecycle-manager - **Description**
A pipeline for running smoke tests has to be configured in Prow. The pipeline has to execute test cases (e.g. by using Ginko, EnvTest or Go Test) which verify the correct behaviour of the lifecycle manager within a K8s runtime context (e.g. by using via K3d or EnvTest). The executed tests don't have to be comprehensive, but have to run fast and should give a clear indication whether the lifecycle manager works as expected.
AC:
- [ ] Pipeline is created in Prow and configured as quality gate in Github for PRs to the lifecycle-manager repository
- [ ] Pipeline triggers the smoke test and allows merging of PRs only if the test was passed successfully
| test | create smoke test pipeline for lifecycle manager description a pipeline for running smoke tests has to be configured in prow the pipeline has to execute test cases e g by using ginko envtest or go test which verify the correct behaviour of the lifecycle manager within a runtime context e g by using via or envtest the executed tests don t have to be comprehensive but have to run fast and should give a clear indication whether the lifecycle manager works as expected ac pipeline is created in prow and configured as quality gate in github for prs to the lifecycle manager repository pipeline triggers the smoke test and allows merging of prs only if the test was passed successfully | 1 |
312,048 | 26,833,547,215 | IssuesEvent | 2023-02-02 17:39:27 | privacy-scaling-explorations/zkevm-circuits | https://api.github.com/repos/privacy-scaling-explorations/zkevm-circuits | closed | Replace `assert!(prover.verify(), Ok(()))` by `prover.assert_satisfied()` for positive tests | T-refactor crate-zkevm-circuits T-test | As discussed with @str4d during the last Halo2 office hours call on Jan 24th, `prover.verify()` should only be used to test expected circuit errors.
Instead, for any other test, `prover.assert_satisfied` is much more nicer as it provides a way more info and also prints annotations for more cases.
So we should make that change. | 1.0 | Replace `assert!(prover.verify(), Ok(()))` by `prover.assert_satisfied()` for positive tests - As discussed with @str4d during the last Halo2 office hours call on Jan 24th, `prover.verify()` should only be used to test expected circuit errors.
Instead, for any other test, `prover.assert_satisfied` is much more nicer as it provides a way more info and also prints annotations for more cases.
So we should make that change. | test | replace assert prover verify ok by prover assert satisfied for positive tests as discussed with during the last office hours call on jan prover verify should only be used to test expected circuit errors instead for any other test prover assert satisfied is much more nicer as it provides a way more info and also prints annotations for more cases so we should make that change | 1 |
674,740 | 23,064,544,253 | IssuesEvent | 2022-07-25 12:54:58 | episphere/connect | https://api.github.com/repos/episphere/connect | closed | Biospec - Dshbrd - Clinical Specimen Link Screen Discussion | Biospecimen Dashboard Clinical Priority 1 Dev Ops Feedback Backend | We are rethinking having a table on the Clinical specimen link page. Instead, we would like to discuss if the following is possible:
User scans Blood Accession ID into the system on the Specimen Link Page
System checks to see if that Accession ID is already linked to a Collection ID
2 scenarios at this point:
1 - The Accession ID entered had not been previously linked to a Collection ID, page remains as is and user can continue with linking a Collection ID. No Modal will pop up.
2 - The Accession ID entered has previously been linked to a Collection ID, a modal appears stating "Accession ID entered is already associated to Collection ID CXA#######. Choose an action:". Two action buttons should be present. First one should state "Link New Collection ID" and the other should state "Add Specimens to existing Collection ID"
This page would no longer need a delete function for the Collection IDs.
Thought behind this change - down the road when Follow-up collections occur, the table may get very large and difficult to view. Also, instead of relying on the user to perform an Accession ID check, the system would preform the check helping to reduce human error.
As stated, we would like to have a discussion about this to figure out if this is possible or if there is better approach. An alternate screen slide has been added to the Clinical Dashboard Navigation Updates file in Box. | 1.0 | Biospec - Dshbrd - Clinical Specimen Link Screen Discussion - We are rethinking having a table on the Clinical specimen link page. Instead, we would like to discuss if the following is possible:
User scans Blood Accession ID into the system on the Specimen Link Page
System checks to see if that Accession ID is already linked to a Collection ID
2 scenarios at this point:
1 - The Accession ID entered had not been previously linked to a Collection ID, page remains as is and user can continue with linking a Collection ID. No Modal will pop up.
2 - The Accession ID entered has previously been linked to a Collection ID, a modal appears stating "Accession ID entered is already associated to Collection ID CXA#######. Choose an action:". Two action buttons should be present. First one should state "Link New Collection ID" and the other should state "Add Specimens to existing Collection ID"
This page would no longer need a delete function for the Collection IDs.
Thought behind this change - down the road when Follow-up collections occur, the table may get very large and difficult to view. Also, instead of relying on the user to perform an Accession ID check, the system would preform the check helping to reduce human error.
As stated, we would like to have a discussion about this to figure out if this is possible or if there is better approach. An alternate screen slide has been added to the Clinical Dashboard Navigation Updates file in Box. | non_test | biospec dshbrd clinical specimen link screen discussion we are rethinking having a table on the clinical specimen link page instead we would like to discuss if the following is possible user scans blood accession id into the system on the specimen link page system checks to see if that accession id is already linked to a collection id scenarios at this point the accession id entered had not been previously linked to a collection id page remains as is and user can continue with linking a collection id no modal will pop up the accession id entered has previously been linked to a collection id a modal appears stating accession id entered is already associated to collection id cxa choose an action two action buttons should be present first one should state link new collection id and the other should state add specimens to existing collection id this page would no longer need a delete function for the collection ids thought behind this change down the road when follow up collections occur the table may get very large and difficult to view also instead of relying on the user to perform an accession id check the system would preform the check helping to reduce human error as stated we would like to have a discussion about this to figure out if this is possible or if there is better approach an alternate screen slide has been added to the clinical dashboard navigation updates file in box | 0 |
248,122 | 26,784,775,120 | IssuesEvent | 2023-02-01 01:14:21 | ChoeMinji/react-17.0.2 | https://api.github.com/repos/ChoeMinji/react-17.0.2 | closed | CVE-2022-23529 (High) detected in jsonwebtoken-8.5.1.tgz - autoclosed | security vulnerability | ## CVE-2022-23529 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jsonwebtoken-8.5.1.tgz</b></p></summary>
<p>JSON Web Token implementation (symmetric and asymmetric)</p>
<p>Library home page: <a href="https://registry.npmjs.org/jsonwebtoken/-/jsonwebtoken-8.5.1.tgz">https://registry.npmjs.org/jsonwebtoken/-/jsonwebtoken-8.5.1.tgz</a></p>
<p>
Dependency Hierarchy:
- danger-9.2.10.tgz (Root Library)
- :x: **jsonwebtoken-8.5.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ChoeMinji/react-17.0.2/commit/4669645897ed4ebcd4ee037f4dabb509ed4754c7">4669645897ed4ebcd4ee037f4dabb509ed4754c7</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
node-jsonwebtoken is a JsonWebToken implementation for node.js. For versions `<= 8.5.1` of `jsonwebtoken` library, if a malicious actor has the ability to modify the key retrieval parameter (referring to the `secretOrPublicKey` argument from the readme link of the `jwt.verify()` function, they can write arbitrary files on the host machine. Users are affected only if untrusted entities are allowed to modify the key retrieval parameter of the `jwt.verify()` on a host that you control. This issue has been fixed, please update to version 9.0.0.
<p>Publish Date: 2022-12-21
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23529>CVE-2022-23529</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/auth0/node-jsonwebtoken/security/advisories/GHSA-27h2-hvpr-p74q">https://github.com/auth0/node-jsonwebtoken/security/advisories/GHSA-27h2-hvpr-p74q</a></p>
<p>Release Date: 2022-12-21</p>
<p>Fix Resolution: jsonwebtoken - 9.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-23529 (High) detected in jsonwebtoken-8.5.1.tgz - autoclosed - ## CVE-2022-23529 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jsonwebtoken-8.5.1.tgz</b></p></summary>
<p>JSON Web Token implementation (symmetric and asymmetric)</p>
<p>Library home page: <a href="https://registry.npmjs.org/jsonwebtoken/-/jsonwebtoken-8.5.1.tgz">https://registry.npmjs.org/jsonwebtoken/-/jsonwebtoken-8.5.1.tgz</a></p>
<p>
Dependency Hierarchy:
- danger-9.2.10.tgz (Root Library)
- :x: **jsonwebtoken-8.5.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ChoeMinji/react-17.0.2/commit/4669645897ed4ebcd4ee037f4dabb509ed4754c7">4669645897ed4ebcd4ee037f4dabb509ed4754c7</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
node-jsonwebtoken is a JsonWebToken implementation for node.js. For versions `<= 8.5.1` of `jsonwebtoken` library, if a malicious actor has the ability to modify the key retrieval parameter (referring to the `secretOrPublicKey` argument from the readme link of the `jwt.verify()` function, they can write arbitrary files on the host machine. Users are affected only if untrusted entities are allowed to modify the key retrieval parameter of the `jwt.verify()` on a host that you control. This issue has been fixed, please update to version 9.0.0.
<p>Publish Date: 2022-12-21
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23529>CVE-2022-23529</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/auth0/node-jsonwebtoken/security/advisories/GHSA-27h2-hvpr-p74q">https://github.com/auth0/node-jsonwebtoken/security/advisories/GHSA-27h2-hvpr-p74q</a></p>
<p>Release Date: 2022-12-21</p>
<p>Fix Resolution: jsonwebtoken - 9.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve high detected in jsonwebtoken tgz autoclosed cve high severity vulnerability vulnerable library jsonwebtoken tgz json web token implementation symmetric and asymmetric library home page a href dependency hierarchy danger tgz root library x jsonwebtoken tgz vulnerable library found in head commit a href found in base branch master vulnerability details node jsonwebtoken is a jsonwebtoken implementation for node js for versions of jsonwebtoken library if a malicious actor has the ability to modify the key retrieval parameter referring to the secretorpublickey argument from the readme link of the jwt verify function they can write arbitrary files on the host machine users are affected only if untrusted entities are allowed to modify the key retrieval parameter of the jwt verify on a host that you control this issue has been fixed please update to version publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jsonwebtoken step up your open source security game with mend | 0 |
27,219 | 4,288,157,578 | IssuesEvent | 2016-07-17 08:29:55 | bitcoin/bitcoin | https://api.github.com/repos/bitcoin/bitcoin | opened | Enable rpcbind_test again | Easy to implement Tests | This involves some more work but should be generally easy to implement.
https://github.com/bitcoin/bitcoin/blob/248b8b8390e5d5caedc9fcb82d98e90dc5f1fa47/qa/rpc-tests/rpcbind_test.py#L9 | 1.0 | Enable rpcbind_test again - This involves some more work but should be generally easy to implement.
https://github.com/bitcoin/bitcoin/blob/248b8b8390e5d5caedc9fcb82d98e90dc5f1fa47/qa/rpc-tests/rpcbind_test.py#L9 | test | enable rpcbind test again this involves some more work but should be generally easy to implement | 1 |
7,371 | 2,898,223,584 | IssuesEvent | 2015-06-17 00:33:29 | codalab/codalab | https://api.github.com/repos/codalab/codalab | reopened | Add score display | Competition Coopetition P1 Test | A score should be located on the public submissions panel, so users will know how good it is, and can verify it. | 1.0 | Add score display - A score should be located on the public submissions panel, so users will know how good it is, and can verify it. | test | add score display a score should be located on the public submissions panel so users will know how good it is and can verify it | 1 |
25,794 | 4,446,085,861 | IssuesEvent | 2016-08-20 12:59:26 | bridgedotnet/Bridge | https://api.github.com/repos/bridgedotnet/Bridge | opened | Reflection doesn't work for native javascript objects even the type is included by reflection filter | defect | For example, reflection doesn't work for console
### Steps To Reproduce
```csharp
[assembly:Reflectable("System.Console")]
namespace Demo
{
public class App
{
public static void Main()
{
typeof(System.Console).GetMethod("WriteLine", new[] { typeof(string) }).Invoke(null, new[] { "Hello" });
}
}
}
``` | 1.0 | Reflection doesn't work for native javascript objects even the type is included by reflection filter - For example, reflection doesn't work for console
### Steps To Reproduce
```csharp
[assembly:Reflectable("System.Console")]
namespace Demo
{
public class App
{
public static void Main()
{
typeof(System.Console).GetMethod("WriteLine", new[] { typeof(string) }).Invoke(null, new[] { "Hello" });
}
}
}
``` | non_test | reflection doesn t work for native javascript objects even the type is included by reflection filter for example reflection doesn t work for console steps to reproduce csharp namespace demo public class app public static void main typeof system console getmethod writeline new typeof string invoke null new hello | 0 |
79,775 | 7,724,915,120 | IssuesEvent | 2018-05-24 16:19:34 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | tests/lib/c_lib fails on native_posix on FC28 | area: C Library area: Testing Suite area: native port bug priority: low | native_posix builds with the host compiler, and on Fedora 28 the tests/lib/c_lib fails sanitycheck as follows:
```
/home/apboie/projects/zephyr/tests/lib/c_lib/src/main.c: In function ‘test_strncpy’:
/home/apboie/projects/zephyr/tests/lib/c_lib/src/main.c:196:2: error: ‘strncpy’ output truncated copying 10 bytes from a string of length 26 [-Werror=stringop-truncation]
strncpy(buffer, "This is over 10 characters", BUFSIZE);
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
cc1: all warnings being treated as errors
```
| 1.0 | tests/lib/c_lib fails on native_posix on FC28 - native_posix builds with the host compiler, and on Fedora 28 the tests/lib/c_lib fails sanitycheck as follows:
```
/home/apboie/projects/zephyr/tests/lib/c_lib/src/main.c: In function ‘test_strncpy’:
/home/apboie/projects/zephyr/tests/lib/c_lib/src/main.c:196:2: error: ‘strncpy’ output truncated copying 10 bytes from a string of length 26 [-Werror=stringop-truncation]
strncpy(buffer, "This is over 10 characters", BUFSIZE);
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
cc1: all warnings being treated as errors
```
| test | tests lib c lib fails on native posix on native posix builds with the host compiler and on fedora the tests lib c lib fails sanitycheck as follows home apboie projects zephyr tests lib c lib src main c in function ‘test strncpy’ home apboie projects zephyr tests lib c lib src main c error ‘strncpy’ output truncated copying bytes from a string of length strncpy buffer this is over characters bufsize all warnings being treated as errors | 1 |
156,907 | 24,626,204,432 | IssuesEvent | 2022-10-16 14:56:03 | dotnet/efcore | https://api.github.com/repos/dotnet/efcore | reopened | efCore won't consider backing field type when initializing IEnumerable property | closed-by-design customer-reported | this issue is closely related to #15506, if not the same. Please read and run tests attached.
i have a private `HashSet` backing field for public `IEnumerable`. if i didnt initialize the hashset myself (on the field declaration). ef core would initialize it with some `ICollection` type.
```
private HashSet<VehicleFeature> _features;
public IEnumerable<VehicleFeature> Features
{
get => _features;
private set => _features = new HashSet<VehicleFeature>(value);
}
```
with above code, when `ef` initializes the entity. The HashSet `_features` would gladly add duplicate values. `_features` is no longer a `hashSet`! But if `_features` initialization be:
`private HashSet<VehicleFeature> _features = new HashSet<VehicleFeature>();`
it works just fine.
More over, i created a test entity to make sure it's not something that i am missing.
```
private HashSet<string> _strings;
public IEnumerable<string> Strings
{
get => _strings;
private set => _strings = new HashSet<string>(value);
}
```
But this time i initialize the `_strings` or not, i receive
> System.InvalidCastException
> HResult=0x80004002
> Message=Unable to cast object of type 'System.Collections.Generic.List`1[System.String]' to type 'System.Collections.Generic.HashSet`1[System.String]'.
> Source=Microsoft.EntityFrameworkCore.Relational
> StackTrace:
> at Microsoft.EntityFrameworkCore.Query.RelationalShapedQueryCompilingExpressionVisitor.QueryingEnumerable`1.Enumerator.MoveNext()
> at System.Linq.Enumerable.Single[TSource](IEnumerable`1 source)
> at Microsoft.EntityFrameworkCore.Query.Internal.QueryCompiler.Execute[TResult](Expression query)
> at Microsoft.EntityFrameworkCore.Query.Internal.EntityQueryProvider.Execute[TResult](Expression expression)
> at System.Linq.Queryable.First[TSource](IQueryable`1 source)
> at EF.Tests.StringsTest.Can_EfCore_Initialize_The_Strings_IEnumerable_Properly()
[efHashsetIssue.zip](https://github.com/aspnet/EntityFrameworkCore/files/3846519/efHashsetIssue.zip)
//cc @ajcvickers
ef 3 with Sqlite 3
.netstandard 2.1 | 1.0 | efCore won't consider backing field type when initializing IEnumerable property - this issue is closely related to #15506, if not the same. Please read and run tests attached.
i have a private `HashSet` backing field for public `IEnumerable`. if i didnt initialize the hashset myself (on the field declaration). ef core would initialize it with some `ICollection` type.
```
private HashSet<VehicleFeature> _features;
public IEnumerable<VehicleFeature> Features
{
get => _features;
private set => _features = new HashSet<VehicleFeature>(value);
}
```
with above code, when `ef` initializes the entity. The HashSet `_features` would gladly add duplicate values. `_features` is no longer a `hashSet`! But if `_features` initialization be:
`private HashSet<VehicleFeature> _features = new HashSet<VehicleFeature>();`
it works just fine.
More over, i created a test entity to make sure it's not something that i am missing.
```
private HashSet<string> _strings;
public IEnumerable<string> Strings
{
get => _strings;
private set => _strings = new HashSet<string>(value);
}
```
But this time i initialize the `_strings` or not, i receive
> System.InvalidCastException
> HResult=0x80004002
> Message=Unable to cast object of type 'System.Collections.Generic.List`1[System.String]' to type 'System.Collections.Generic.HashSet`1[System.String]'.
> Source=Microsoft.EntityFrameworkCore.Relational
> StackTrace:
> at Microsoft.EntityFrameworkCore.Query.RelationalShapedQueryCompilingExpressionVisitor.QueryingEnumerable`1.Enumerator.MoveNext()
> at System.Linq.Enumerable.Single[TSource](IEnumerable`1 source)
> at Microsoft.EntityFrameworkCore.Query.Internal.QueryCompiler.Execute[TResult](Expression query)
> at Microsoft.EntityFrameworkCore.Query.Internal.EntityQueryProvider.Execute[TResult](Expression expression)
> at System.Linq.Queryable.First[TSource](IQueryable`1 source)
> at EF.Tests.StringsTest.Can_EfCore_Initialize_The_Strings_IEnumerable_Properly()
[efHashsetIssue.zip](https://github.com/aspnet/EntityFrameworkCore/files/3846519/efHashsetIssue.zip)
//cc @ajcvickers
ef 3 with Sqlite 3
.netstandard 2.1 | non_test | efcore won t consider backing field type when initializing ienumerable property this issue is closely related to if not the same please read and run tests attached i have a private hashset backing field for public ienumerable if i didnt initialize the hashset myself on the field declaration ef core would initialize it with some icollection type private hashset features public ienumerable features get features private set features new hashset value with above code when ef initializes the entity the hashset features would gladly add duplicate values features is no longer a hashset but if features initialization be private hashset features new hashset it works just fine more over i created a test entity to make sure it s not something that i am missing private hashset strings public ienumerable strings get strings private set strings new hashset value but this time i initialize the strings or not i receive system invalidcastexception hresult message unable to cast object of type system collections generic list to type system collections generic hashset source microsoft entityframeworkcore relational stacktrace at microsoft entityframeworkcore query relationalshapedquerycompilingexpressionvisitor queryingenumerable enumerator movenext at system linq enumerable single ienumerable source at microsoft entityframeworkcore query internal querycompiler execute expression query at microsoft entityframeworkcore query internal entityqueryprovider execute expression expression at system linq queryable first iqueryable source at ef tests stringstest can efcore initialize the strings ienumerable properly cc ajcvickers ef with sqlite netstandard | 0 |
112,087 | 17,068,116,174 | IssuesEvent | 2021-07-07 09:52:03 | AlexRogalskiy/java4you | https://api.github.com/repos/AlexRogalskiy/java4you | closed | CVE-2020-25638 (High) detected in hibernate-core-5.4.2.Final.jar - autoclosed | security vulnerability | ## CVE-2020-25638 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hibernate-core-5.4.2.Final.jar</b></p></summary>
<p>Hibernate's core ORM functionality</p>
<p>Library home page: <a href="http://hibernate.org/orm">http://hibernate.org/orm</a></p>
<p>Path to dependency file: java4you/pom.xml</p>
<p>Path to vulnerable library: 20210131175039_ODYKXS/downloadResource_HOMULE/20210131175852/hibernate-core-5.4.2.Final.jar</p>
<p>
Dependency Hierarchy:
- :x: **hibernate-core-5.4.2.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/java4you/commit/d7e409b73576f85ef958df67617ad1342a9e53d1">d7e409b73576f85ef958df67617ad1342a9e53d1</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in hibernate-core in versions prior to and including 5.4.23.Final. A SQL injection in the implementation of the JPA Criteria API can permit unsanitized literals when a literal is used in the SQL comments of the query. This flaw could allow an attacker to access unauthorized information or possibly conduct further attacks. The highest threat from this vulnerability is to data confidentiality and integrity.
<p>Publish Date: 2020-12-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25638>CVE-2020-25638</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://in.relation.to/2020/11/19/hibernate-orm-5424-final-release/">https://in.relation.to/2020/11/19/hibernate-orm-5424-final-release/</a></p>
<p>Release Date: 2020-12-02</p>
<p>Fix Resolution: org.hibernate:hibernate-core:5.3.20.Final,5.4.24.Final</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-25638 (High) detected in hibernate-core-5.4.2.Final.jar - autoclosed - ## CVE-2020-25638 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hibernate-core-5.4.2.Final.jar</b></p></summary>
<p>Hibernate's core ORM functionality</p>
<p>Library home page: <a href="http://hibernate.org/orm">http://hibernate.org/orm</a></p>
<p>Path to dependency file: java4you/pom.xml</p>
<p>Path to vulnerable library: 20210131175039_ODYKXS/downloadResource_HOMULE/20210131175852/hibernate-core-5.4.2.Final.jar</p>
<p>
Dependency Hierarchy:
- :x: **hibernate-core-5.4.2.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/java4you/commit/d7e409b73576f85ef958df67617ad1342a9e53d1">d7e409b73576f85ef958df67617ad1342a9e53d1</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in hibernate-core in versions prior to and including 5.4.23.Final. A SQL injection in the implementation of the JPA Criteria API can permit unsanitized literals when a literal is used in the SQL comments of the query. This flaw could allow an attacker to access unauthorized information or possibly conduct further attacks. The highest threat from this vulnerability is to data confidentiality and integrity.
<p>Publish Date: 2020-12-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25638>CVE-2020-25638</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://in.relation.to/2020/11/19/hibernate-orm-5424-final-release/">https://in.relation.to/2020/11/19/hibernate-orm-5424-final-release/</a></p>
<p>Release Date: 2020-12-02</p>
<p>Fix Resolution: org.hibernate:hibernate-core:5.3.20.Final,5.4.24.Final</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve high detected in hibernate core final jar autoclosed cve high severity vulnerability vulnerable library hibernate core final jar hibernate s core orm functionality library home page a href path to dependency file pom xml path to vulnerable library odykxs downloadresource homule hibernate core final jar dependency hierarchy x hibernate core final jar vulnerable library found in head commit a href found in base branch master vulnerability details a flaw was found in hibernate core in versions prior to and including final a sql injection in the implementation of the jpa criteria api can permit unsanitized literals when a literal is used in the sql comments of the query this flaw could allow an attacker to access unauthorized information or possibly conduct further attacks the highest threat from this vulnerability is to data confidentiality and integrity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org hibernate hibernate core final final step up your open source security game with whitesource | 0 |
287,710 | 21,671,614,670 | IssuesEvent | 2022-05-08 03:05:05 | mosn/layotto | https://api.github.com/repos/mosn/layotto | closed | doc: Translate "Explain Layotto's github workflow" into Chinese | documentation enhancement stale translation | <!-- Please only use this template for submitting enhancement requests -->
**What would you like to be added**:
Translate ["Explain Layotto's github workflow"](https://mosn.io/layotto/#/zh/development/explanation-for-github-workflow) into Chinese.
**Why is this needed**:
Make it easier for chinese contributors to understand the CI pipeline | 1.0 | doc: Translate "Explain Layotto's github workflow" into Chinese - <!-- Please only use this template for submitting enhancement requests -->
**What would you like to be added**:
Translate ["Explain Layotto's github workflow"](https://mosn.io/layotto/#/zh/development/explanation-for-github-workflow) into Chinese.
**Why is this needed**:
Make it easier for chinese contributors to understand the CI pipeline | non_test | doc translate explain layotto s github workflow into chinese what would you like to be added translate into chinese why is this needed make it easier for chinese contributors to understand the ci pipeline | 0 |
56,452 | 6,520,013,402 | IssuesEvent | 2017-08-28 14:57:00 | mozilla/activity-stream | https://api.github.com/repos/mozilla/activity-stream | opened | On about:home clicking the search button should trigger a search | fx57 Parity Tests | ```
waitForDocLoadAndStopIt: The expected URL was loaded - Got about:home, expected http://mochi.test:8888/
```
in `browser/base/content/test/about/browser_aboutHome.js`
For some reason doing `.click()` on the search button does not trigger the search. | 1.0 | On about:home clicking the search button should trigger a search - ```
waitForDocLoadAndStopIt: The expected URL was loaded - Got about:home, expected http://mochi.test:8888/
```
in `browser/base/content/test/about/browser_aboutHome.js`
For some reason doing `.click()` on the search button does not trigger the search. | test | on about home clicking the search button should trigger a search waitfordocloadandstopit the expected url was loaded got about home expected in browser base content test about browser abouthome js for some reason doing click on the search button does not trigger the search | 1 |
129,137 | 10,564,226,974 | IssuesEvent | 2019-10-05 00:13:25 | sass/libsass | https://api.github.com/repos/sass/libsass | closed | selector-unify: Wrong result if selector contains relationship combinator | Bug - Confirmed Compatibility - P3 Dev - Test Written | selector-unify seems unable to handle selectors that contain relationship combinators such as '+', '~' or '>'. It doesn't matter if that's the case for just one of the selectors or both.
## input.scss
```scss
$sel1: '.something__child + .something__child--mod1';
$sel2: '.something__child ~ .something__child--mod2';
$result1: selector-unify($sel1, $sel2);
#{$result1} {
/* nothing */
}
```
## Actual results
[libsass 3.5.2][1]
```css
+ .something__child--mod1 .something__child ~ .something__child--mod2 .something__child--mod1.something__child--mod2, ~ .something__child--mod2 .something__child--mod1.something__child--mod2 .something__child + .something__child--mod1 {
/* nothing */
}
```
## Expected result
ruby sass 3.4.21
```css
.something__child + .something__child--mod1.something__child--mod2 {
/* nothing */
}
```
version info:
```cmd
$ node-sass --version
node-sass 4.8.3 (Wrapper) [JavaScript]
libsass 3.5.2 (Sass Compiler) [C/C++]
```
[1]: http://libsass.ocbnet.ch/srcmap/#JHNlbDE6ICcuc29tZXRoaW5nX19jaGlsZCArIC5zb21ldGhpbmdfX2NoaWxkLS1tb2QxJzsKJHNlbDI6ICcuc29tZXRoaW5nX19jaGlsZCB+IC5zb21ldGhpbmdfX2NoaWxkLS1tb2QyJzsKJHJlc3VsdDE6IHNlbGVjdG9yLXVuaWZ5KCRzZWwxLCAkc2VsMik7CgojeyRyZXN1bHQxfSB7CiAgLyogbm90aGluZyAqLwp9Cg==
| 1.0 | selector-unify: Wrong result if selector contains relationship combinator - selector-unify seems unable to handle selectors that contain relationship combinators such as '+', '~' or '>'. It doesn't matter if that's the case for just one of the selectors or both.
## input.scss
```scss
$sel1: '.something__child + .something__child--mod1';
$sel2: '.something__child ~ .something__child--mod2';
$result1: selector-unify($sel1, $sel2);
#{$result1} {
/* nothing */
}
```
## Actual results
[libsass 3.5.2][1]
```css
+ .something__child--mod1 .something__child ~ .something__child--mod2 .something__child--mod1.something__child--mod2, ~ .something__child--mod2 .something__child--mod1.something__child--mod2 .something__child + .something__child--mod1 {
/* nothing */
}
```
## Expected result
ruby sass 3.4.21
```css
.something__child + .something__child--mod1.something__child--mod2 {
/* nothing */
}
```
version info:
```cmd
$ node-sass --version
node-sass 4.8.3 (Wrapper) [JavaScript]
libsass 3.5.2 (Sass Compiler) [C/C++]
```
[1]: http://libsass.ocbnet.ch/srcmap/#JHNlbDE6ICcuc29tZXRoaW5nX19jaGlsZCArIC5zb21ldGhpbmdfX2NoaWxkLS1tb2QxJzsKJHNlbDI6ICcuc29tZXRoaW5nX19jaGlsZCB+IC5zb21ldGhpbmdfX2NoaWxkLS1tb2QyJzsKJHJlc3VsdDE6IHNlbGVjdG9yLXVuaWZ5KCRzZWwxLCAkc2VsMik7CgojeyRyZXN1bHQxfSB7CiAgLyogbm90aGluZyAqLwp9Cg==
| test | selector unify wrong result if selector contains relationship combinator selector unify seems unable to handle selectors that contain relationship combinators such as or it doesn t matter if that s the case for just one of the selectors or both input scss scss something child something child something child something child selector unify nothing actual results css something child something child something child something child something child something child something child something child something child something child nothing expected result ruby sass css something child something child something child nothing version info cmd node sass version node sass wrapper libsass sass compiler | 1 |
550,448 | 16,112,871,914 | IssuesEvent | 2021-04-28 01:01:44 | uwblueprint/shoe-project | https://api.github.com/repos/uwblueprint/shoe-project | opened | Adding Tags in Upload/Edit story lack design prompts | bug priority: high | From our current implementation we have this:

This is our Figma design:

As a new user, idk how to add my own tags until I've played around. | 1.0 | Adding Tags in Upload/Edit story lack design prompts - From our current implementation we have this:

This is our Figma design:

As a new user, idk how to add my own tags until I've played around. | non_test | adding tags in upload edit story lack design prompts from our current implementation we have this this is our figma design as a new user idk how to add my own tags until i ve played around | 0 |
229,511 | 18,359,246,372 | IssuesEvent | 2021-10-09 00:39:34 | backend-br/vagas | https://api.github.com/repos/backend-br/vagas | closed | [RJ / Remoto] Back-end / Full Stack Developer [Pleno / Senior] @ Nuuvem | PJ Pleno Senior Junior JavaScript Python TDD Ruby Remoto MongoDB Redis PostgreSQL Testes Unitários Git Linux Scala HTML CSS FullStack Stale | ## Nossa empresa
Greetings, gamers!
[Nuuvem.com](https://www.nuuvem.com/) é uma das maiores plataformas de distribuição digital da América Latina. Estamos trabalhando para nos tornarmos uma plataforma universal para comprar, discutir e interagir sobre jogos no mundo todo.
Estamos em busca de **Desenvolvedores Back-end / Full Stack com foco em Ruby de diferentes senioridades**, para criação e implementação de interfaces web, modernas, resilientes e fáceis de usar e para nos ajudar a manter e desenvolver nossa atual plataforma.
Você se juntará a uma equipe de entusiastas da indústria de jogos altamente qualificados, criando soluções em um ambiente ágil e apaixonado. **O trabalho em equipe é essencial**. Buscamos alguém obcecado em entregar trechos de código de alta qualidade de maneira rápida e estável, imaginando qualquer recurso como parte do valor total da plataforma, sem perder o foco em fazer as coisas. Segurança e desempenho devem ser suas prioridades.
## Local
Apesar do nosso escritório ficar no Rio de Janeiro, esta oportunidade de trabalho é válida para trabalho **remoto** e está disponível para qualquer lugar do Brasil.
## Responsabilidades
- Projetar, desenvolver e testar novos recursos para nossa plataforma;
- Liderar pesquisas e experimentos ao lidar com novos requisitos;
- Manter e melhorar a base de código atual corrigindo bugs rapidamente e refatorando códigos problemáticos quando necessário;
- Escrever testes unitários e integrar novos recursos e correções de bugs;
## Requisitos
- Usuário autossuficiente de Linux ou macOS e não teme o Terminal;
- Boa experiência em Ruby on Rails com TDD / BDD (unidade e integração), Git, revisão de código, integração contínua e ambientes de desenvolvimento / teste;
- Conhecimento de PostgreSQL, MongoDB e Redis;
- Experiência de construção e consumo de APIs e manutenção de aplicativos baseados em microsserviços;
- Alguma experiência com outras linguagens de programação como Python, Node.js, Scala, Go;
- Boa experiência com HTML, CSS, JavaScript;
**Perfil:**
- Proficiência e boa comunicação em inglês verbal e escrito;
- Ser apaixonado por tecnologia (e jogos, obivamente :P);
- Entender os requisitos de negócios e traduzi-los em definições técnicas;
- Fazer perguntas poderosas - a partir disso, iremos (re)definir histórias de usuários, critérios de aceitação, especificações. Sempre questionando (sem perder o foco);
- Ser obcecado por detalhes (relevantes);
- Orientado por resultados / entrega e solucionador de problemas;
## Nossas Crenças
- Pair Programming é sempre bem vindo;
- Qualidade e quantidade vivem em equilíbrio - nosso trabalho é encontrar esse equilíbrio.
- Um ótimo produto tem a ver com uma ótima experiência e valor agregado.
- Ouvir é mais importante do que falar. Mostrar é melhor do que contar.
## Benefícios
- Plano de Saúde e Odontológico (AMIL);
- Retirada do Jogo Do Mês (um jogo por mês gratuitamente em nossa plataforma);
- 10 cupons de 20% de desconto na nossa plataforma por ano;
- Saldo semestral para cursos de aperfeiçoamento e livros;
Além dos benefícios, oferecemos um ambiente extremamente colaborativo, com alta transparência, flexibilidade de horário e muito espaço para inovação e participação em novas soluções;
## Contratação
Nossa contratação é feita em cima de modelo PJ e a faixa salarial varia conforme senioridade:
**Nível Pleno**: De R$6.500,00 e R$9.000,00
**Nível Senior**: De R$9.500,00 à R$12.000,00
## Como se candidatar
Bata se candidatar em nosso site https://careers.nuuvem.com/o/fullstack-developer-ruby ou enviar um email para odara.menezes@nuuvem.com com seu CV ou perfil do linkedin atualizados e GG!
Se você se identificou conosco, mas ainda não se enquadra no perfil Pleno / Senior, dá uma olhadinha na nossa vaga de Junior Dev ;)
https://careers.nuuvem.com/o/junior-fullstack-developer-ruby
## Labels
- Remoto
- PJ
- Sênior
- Pleno
- Ruby
- Javascript
| 1.0 | [RJ / Remoto] Back-end / Full Stack Developer [Pleno / Senior] @ Nuuvem - ## Nossa empresa
Greetings, gamers!
[Nuuvem.com](https://www.nuuvem.com/) é uma das maiores plataformas de distribuição digital da América Latina. Estamos trabalhando para nos tornarmos uma plataforma universal para comprar, discutir e interagir sobre jogos no mundo todo.
Estamos em busca de **Desenvolvedores Back-end / Full Stack com foco em Ruby de diferentes senioridades**, para criação e implementação de interfaces web, modernas, resilientes e fáceis de usar e para nos ajudar a manter e desenvolver nossa atual plataforma.
Você se juntará a uma equipe de entusiastas da indústria de jogos altamente qualificados, criando soluções em um ambiente ágil e apaixonado. **O trabalho em equipe é essencial**. Buscamos alguém obcecado em entregar trechos de código de alta qualidade de maneira rápida e estável, imaginando qualquer recurso como parte do valor total da plataforma, sem perder o foco em fazer as coisas. Segurança e desempenho devem ser suas prioridades.
## Local
Apesar do nosso escritório ficar no Rio de Janeiro, esta oportunidade de trabalho é válida para trabalho **remoto** e está disponível para qualquer lugar do Brasil.
## Responsabilidades
- Projetar, desenvolver e testar novos recursos para nossa plataforma;
- Liderar pesquisas e experimentos ao lidar com novos requisitos;
- Manter e melhorar a base de código atual corrigindo bugs rapidamente e refatorando códigos problemáticos quando necessário;
- Escrever testes unitários e integrar novos recursos e correções de bugs;
## Requisitos
- Usuário autossuficiente de Linux ou macOS e não teme o Terminal;
- Boa experiência em Ruby on Rails com TDD / BDD (unidade e integração), Git, revisão de código, integração contínua e ambientes de desenvolvimento / teste;
- Conhecimento de PostgreSQL, MongoDB e Redis;
- Experiência de construção e consumo de APIs e manutenção de aplicativos baseados em microsserviços;
- Alguma experiência com outras linguagens de programação como Python, Node.js, Scala, Go;
- Boa experiência com HTML, CSS, JavaScript;
**Perfil:**
- Proficiência e boa comunicação em inglês verbal e escrito;
- Ser apaixonado por tecnologia (e jogos, obivamente :P);
- Entender os requisitos de negócios e traduzi-los em definições técnicas;
- Fazer perguntas poderosas - a partir disso, iremos (re)definir histórias de usuários, critérios de aceitação, especificações. Sempre questionando (sem perder o foco);
- Ser obcecado por detalhes (relevantes);
- Orientado por resultados / entrega e solucionador de problemas;
## Nossas Crenças
- Pair Programming é sempre bem vindo;
- Qualidade e quantidade vivem em equilíbrio - nosso trabalho é encontrar esse equilíbrio.
- Um ótimo produto tem a ver com uma ótima experiência e valor agregado.
- Ouvir é mais importante do que falar. Mostrar é melhor do que contar.
## Benefícios
- Plano de Saúde e Odontológico (AMIL);
- Retirada do Jogo Do Mês (um jogo por mês gratuitamente em nossa plataforma);
- 10 cupons de 20% de desconto na nossa plataforma por ano;
- Saldo semestral para cursos de aperfeiçoamento e livros;
Além dos benefícios, oferecemos um ambiente extremamente colaborativo, com alta transparência, flexibilidade de horário e muito espaço para inovação e participação em novas soluções;
## Contratação
Nossa contratação é feita em cima de modelo PJ e a faixa salarial varia conforme senioridade:
**Nível Pleno**: De R$6.500,00 e R$9.000,00
**Nível Senior**: De R$9.500,00 à R$12.000,00
## Como se candidatar
Bata se candidatar em nosso site https://careers.nuuvem.com/o/fullstack-developer-ruby ou enviar um email para odara.menezes@nuuvem.com com seu CV ou perfil do linkedin atualizados e GG!
Se você se identificou conosco, mas ainda não se enquadra no perfil Pleno / Senior, dá uma olhadinha na nossa vaga de Junior Dev ;)
https://careers.nuuvem.com/o/junior-fullstack-developer-ruby
## Labels
- Remoto
- PJ
- Sênior
- Pleno
- Ruby
- Javascript
| test | back end full stack developer nuuvem nossa empresa greetings gamers é uma das maiores plataformas de distribuição digital da américa latina estamos trabalhando para nos tornarmos uma plataforma universal para comprar discutir e interagir sobre jogos no mundo todo estamos em busca de desenvolvedores back end full stack com foco em ruby de diferentes senioridades para criação e implementação de interfaces web modernas resilientes e fáceis de usar e para nos ajudar a manter e desenvolver nossa atual plataforma você se juntará a uma equipe de entusiastas da indústria de jogos altamente qualificados criando soluções em um ambiente ágil e apaixonado o trabalho em equipe é essencial buscamos alguém obcecado em entregar trechos de código de alta qualidade de maneira rápida e estável imaginando qualquer recurso como parte do valor total da plataforma sem perder o foco em fazer as coisas segurança e desempenho devem ser suas prioridades local apesar do nosso escritório ficar no rio de janeiro esta oportunidade de trabalho é válida para trabalho remoto e está disponível para qualquer lugar do brasil responsabilidades projetar desenvolver e testar novos recursos para nossa plataforma liderar pesquisas e experimentos ao lidar com novos requisitos manter e melhorar a base de código atual corrigindo bugs rapidamente e refatorando códigos problemáticos quando necessário escrever testes unitários e integrar novos recursos e correções de bugs requisitos usuário autossuficiente de linux ou macos e não teme o terminal boa experiência em ruby on rails com tdd bdd unidade e integração git revisão de código integração contínua e ambientes de desenvolvimento teste conhecimento de postgresql mongodb e redis experiência de construção e consumo de apis e manutenção de aplicativos baseados em microsserviços alguma experiência com outras linguagens de programação como python node js scala go boa experiência com html css javascript perfil proficiência e boa comunicação em inglês verbal e escrito ser apaixonado por tecnologia e jogos obivamente p entender os requisitos de negócios e traduzi los em definições técnicas fazer perguntas poderosas a partir disso iremos re definir histórias de usuários critérios de aceitação especificações sempre questionando sem perder o foco ser obcecado por detalhes relevantes orientado por resultados entrega e solucionador de problemas nossas crenças pair programming é sempre bem vindo qualidade e quantidade vivem em equilíbrio nosso trabalho é encontrar esse equilíbrio um ótimo produto tem a ver com uma ótima experiência e valor agregado ouvir é mais importante do que falar mostrar é melhor do que contar benefícios plano de saúde e odontológico amil retirada do jogo do mês um jogo por mês gratuitamente em nossa plataforma cupons de de desconto na nossa plataforma por ano saldo semestral para cursos de aperfeiçoamento e livros além dos benefícios oferecemos um ambiente extremamente colaborativo com alta transparência flexibilidade de horário e muito espaço para inovação e participação em novas soluções contratação nossa contratação é feita em cima de modelo pj e a faixa salarial varia conforme senioridade nível pleno de r e r nível senior de r à r como se candidatar bata se candidatar em nosso site ou enviar um email para odara menezes nuuvem com com seu cv ou perfil do linkedin atualizados e gg se você se identificou conosco mas ainda não se enquadra no perfil pleno senior dá uma olhadinha na nossa vaga de junior dev labels remoto pj sênior pleno ruby javascript | 1 |
54,159 | 3,060,449,638 | IssuesEvent | 2015-08-14 20:45:50 | google/paco | https://api.github.com/repos/google/paco | closed | report of total pings submitted each day as well as total for all of study period [moved] | Component-Server enhancement Priority-Critical | original issue: https://code.google.com/p/paco/issues/detail?id=99 | 1.0 | report of total pings submitted each day as well as total for all of study period [moved] - original issue: https://code.google.com/p/paco/issues/detail?id=99 | non_test | report of total pings submitted each day as well as total for all of study period original issue | 0 |
223,079 | 17,539,543,029 | IssuesEvent | 2021-08-12 10:18:04 | tarantool/cartridge | https://api.github.com/repos/tarantool/cartridge | closed | Adapt nightly tests to tarantool >2.8 | testing cartridge teamS | Currently nightly tests for tarantool 2.8 and 2.9 are red. Mostly because of changed "duplicate key" message. It doesn't affect any logic and can be fixed by taking into account tarantool version.
Affected tests:
- `integration/issues_test.lua`
- `integration/stateboard_test.lua` | 1.0 | Adapt nightly tests to tarantool >2.8 - Currently nightly tests for tarantool 2.8 and 2.9 are red. Mostly because of changed "duplicate key" message. It doesn't affect any logic and can be fixed by taking into account tarantool version.
Affected tests:
- `integration/issues_test.lua`
- `integration/stateboard_test.lua` | test | adapt nightly tests to tarantool currently nightly tests for tarantool and are red mostly because of changed duplicate key message it doesn t affect any logic and can be fixed by taking into account tarantool version affected tests integration issues test lua integration stateboard test lua | 1 |
30,325 | 7,188,841,771 | IssuesEvent | 2018-02-02 11:40:15 | mozilla/addons-server | https://api.github.com/repos/mozilla/addons-server | closed | Avoid xpi extraction when reading manifest.json information | component: code quality component: devhub priority: p3 triaged | Reading through our code I always assumed we extract the XPI for more than just reading the manifest.json but it appears we don't. When I implemented the `_locale` messages parsing it uses the zipped data directly and the linter does that as well.
So there seems to be no reason to ever extract the XPI, especially not for reading the manifest.json. Some add-ons contain quite a lot of files and extracting that on our EFS can be quite slow and unnecessary.
Let's remove the `extract_xpi` functionality and make sure it's never used during our upload process.
Refs #4438 | 1.0 | Avoid xpi extraction when reading manifest.json information - Reading through our code I always assumed we extract the XPI for more than just reading the manifest.json but it appears we don't. When I implemented the `_locale` messages parsing it uses the zipped data directly and the linter does that as well.
So there seems to be no reason to ever extract the XPI, especially not for reading the manifest.json. Some add-ons contain quite a lot of files and extracting that on our EFS can be quite slow and unnecessary.
Let's remove the `extract_xpi` functionality and make sure it's never used during our upload process.
Refs #4438 | non_test | avoid xpi extraction when reading manifest json information reading through our code i always assumed we extract the xpi for more than just reading the manifest json but it appears we don t when i implemented the locale messages parsing it uses the zipped data directly and the linter does that as well so there seems to be no reason to ever extract the xpi especially not for reading the manifest json some add ons contain quite a lot of files and extracting that on our efs can be quite slow and unnecessary let s remove the extract xpi functionality and make sure it s never used during our upload process refs | 0 |
340,400 | 10,271,935,151 | IssuesEvent | 2019-08-23 15:12:43 | SparkDevNetwork/Rock | https://api.github.com/repos/SparkDevNetwork/Rock | closed | [EN] Sending System Emails should resolve FromAddress/FromName per-recipient | Priority: Low Status: Confirmed Topic: Communications Type: Enhancement | ### Prerequisites
* [X] Put an X between the brackets on this line if you have done all of the following:
* Can you reproduce the problem on a fresh install or the [demo site ](http://rock.rocksolidchurchdemo.com/)?
* Did you include your Rock version number and [client culture ](https://github.com/SparkDevNetwork/Rock/wiki/Environment-and-Diagnostics-Information) setting?
* Did you [%3Aissue+user%3ASparkDevNetwork+-repo%3ASparkDevNetwork%2FSlack perform a cursory search ](https://github.com/issues?q=is) to see if your bug or enhancement is already reported?
### Description
Currently, the `From Address` and `From Name` fields in a system e-mail are Lava aware, but they are resolved too early in the process and do not have access to the recipient merge data. Suggesting that this be moved later in the process of `SMTPComponent` so that we have access to the recipient merge fields as well. For performance we can do an initial check to see if those fields have lava so that we only resolve the lava (in the loop) if it actually exists.
Specific use case, I want to be able to send a system e-mail to a group of staff members and have the from address come out like `pod-12345@mg.sheph...rch.com` and then setup a rule in Mailgun to catch `pod-####@` addresses and send them to a special handler that will parse the number and do a specific task on that entity along with the body of the e-mail message.
Example: `pod-{{ MyEntity.Id }}@mg.shep...rch.com`
### Versions
* **Rock Version:** 6.8
* **Client Culture Setting:** en-US
| 1.0 | [EN] Sending System Emails should resolve FromAddress/FromName per-recipient - ### Prerequisites
* [X] Put an X between the brackets on this line if you have done all of the following:
* Can you reproduce the problem on a fresh install or the [demo site ](http://rock.rocksolidchurchdemo.com/)?
* Did you include your Rock version number and [client culture ](https://github.com/SparkDevNetwork/Rock/wiki/Environment-and-Diagnostics-Information) setting?
* Did you [%3Aissue+user%3ASparkDevNetwork+-repo%3ASparkDevNetwork%2FSlack perform a cursory search ](https://github.com/issues?q=is) to see if your bug or enhancement is already reported?
### Description
Currently, the `From Address` and `From Name` fields in a system e-mail are Lava aware, but they are resolved too early in the process and do not have access to the recipient merge data. Suggesting that this be moved later in the process of `SMTPComponent` so that we have access to the recipient merge fields as well. For performance we can do an initial check to see if those fields have lava so that we only resolve the lava (in the loop) if it actually exists.
Specific use case, I want to be able to send a system e-mail to a group of staff members and have the from address come out like `pod-12345@mg.sheph...rch.com` and then setup a rule in Mailgun to catch `pod-####@` addresses and send them to a special handler that will parse the number and do a specific task on that entity along with the body of the e-mail message.
Example: `pod-{{ MyEntity.Id }}@mg.shep...rch.com`
### Versions
* **Rock Version:** 6.8
* **Client Culture Setting:** en-US
| non_test | sending system emails should resolve fromaddress fromname per recipient prerequisites put an x between the brackets on this line if you have done all of the following can you reproduce the problem on a fresh install or the did you include your rock version number and setting did you to see if your bug or enhancement is already reported description currently the from address and from name fields in a system e mail are lava aware but they are resolved too early in the process and do not have access to the recipient merge data suggesting that this be moved later in the process of smtpcomponent so that we have access to the recipient merge fields as well for performance we can do an initial check to see if those fields have lava so that we only resolve the lava in the loop if it actually exists specific use case i want to be able to send a system e mail to a group of staff members and have the from address come out like pod mg sheph rch com and then setup a rule in mailgun to catch pod addresses and send them to a special handler that will parse the number and do a specific task on that entity along with the body of the e mail message example pod myentity id mg shep rch com versions rock version client culture setting en us | 0 |
152,033 | 12,069,251,390 | IssuesEvent | 2020-04-16 15:49:53 | paritytech/polkassembly | https://api.github.com/repos/paritytech/polkassembly | closed | Add tests for resetPassword mutation | i-4-tests s-2-auth-server | After https://github.com/paritytech/polkassembly/pull/606 is merged Add tests for resetPassword mutation as password reset token moved to redis | 1.0 | Add tests for resetPassword mutation - After https://github.com/paritytech/polkassembly/pull/606 is merged Add tests for resetPassword mutation as password reset token moved to redis | test | add tests for resetpassword mutation after is merged add tests for resetpassword mutation as password reset token moved to redis | 1 |
124,919 | 10,329,806,431 | IssuesEvent | 2019-09-02 13:09:07 | iotaledger/iri | https://api.github.com/repos/iotaledger/iri | opened | Generate new DBs for the spent addresses test to pass | C-RegressionTests C-Spent Addresses L-Groom | ### Description
Our regression tests with local snapshot files. The way the files work is now changing with in #1584, and we need to create new DB(s) for the tests to work.
We can use Brord's tool in tangle tools, which might not work anymore, so might need adjustments, as per:
https://github.com/iotaledger/iri/pull/1584#issuecomment-526273324
### Motivation
Regression tests pass.
### Requirements
- [ ] Regression tests pass with the DB as is defined #1584 | 1.0 | Generate new DBs for the spent addresses test to pass - ### Description
Our regression tests with local snapshot files. The way the files work is now changing with in #1584, and we need to create new DB(s) for the tests to work.
We can use Brord's tool in tangle tools, which might not work anymore, so might need adjustments, as per:
https://github.com/iotaledger/iri/pull/1584#issuecomment-526273324
### Motivation
Regression tests pass.
### Requirements
- [ ] Regression tests pass with the DB as is defined #1584 | test | generate new dbs for the spent addresses test to pass description our regression tests with local snapshot files the way the files work is now changing with in and we need to create new db s for the tests to work we can use brord s tool in tangle tools which might not work anymore so might need adjustments as per motivation regression tests pass requirements regression tests pass with the db as is defined | 1 |
463,623 | 13,284,988,132 | IssuesEvent | 2020-08-24 07:20:38 | benjaminortizulloa/ExploreGitAPI | https://api.github.com/repos/benjaminortizulloa/ExploreGitAPI | opened | A much longer title to test the new line capabilities | Complexity_Medium Priority_Low | Let me tell you a story about a long description to test out whether this works or not, this should work
[impact: This will let me know if the reallignment is good]
[timeline: short]
[originally proposed by @benjaminortizulloa]
[suggested repo: Do not know]
[additional notes: Successfully changed review page] | 1.0 | A much longer title to test the new line capabilities - Let me tell you a story about a long description to test out whether this works or not, this should work
[impact: This will let me know if the reallignment is good]
[timeline: short]
[originally proposed by @benjaminortizulloa]
[suggested repo: Do not know]
[additional notes: Successfully changed review page] | non_test | a much longer title to test the new line capabilities let me tell you a story about a long description to test out whether this works or not this should work | 0 |
532,406 | 15,555,734,320 | IssuesEvent | 2021-03-16 06:40:14 | wso2/product-apim | https://api.github.com/repos/wso2/product-apim | opened | Changing deployments DisplayOnDevportal field needs to do a redeployment issue | API-M 4.0.0 Priority/Normal Type/Bug | ### Description:
Currently, in order to simply change the display on devportal field user needs to do a redeployment. Need to support a separate REST api to handle this.
### Steps to reproduce:
Create a revision and deploy with display on devportal as true.
Try to change the display on devportal value to false.
Currently, needs to undeploy and redeploy with value set as false. | 1.0 | Changing deployments DisplayOnDevportal field needs to do a redeployment issue - ### Description:
Currently, in order to simply change the display on devportal field user needs to do a redeployment. Need to support a separate REST api to handle this.
### Steps to reproduce:
Create a revision and deploy with display on devportal as true.
Try to change the display on devportal value to false.
Currently, needs to undeploy and redeploy with value set as false. | non_test | changing deployments displayondevportal field needs to do a redeployment issue description currently in order to simply change the display on devportal field user needs to do a redeployment need to support a separate rest api to handle this steps to reproduce create a revision and deploy with display on devportal as true try to change the display on devportal value to false currently needs to undeploy and redeploy with value set as false | 0 |
311,487 | 26,795,244,719 | IssuesEvent | 2023-02-01 11:22:46 | finos/waltz | https://api.github.com/repos/finos/waltz | closed | Upgrade Liquibase | fixed (test & close) task | ### Description
Looks like our version does not fully support h2 `2.x`.
(see https://github.com/liquibase/liquibase/pull/2285 )
We use h2 for integration tests, all of which are currently failing (and have been disable by omitting the `ci` maven profile.
### Resourcing
We intend to implement this task | 1.0 | Upgrade Liquibase - ### Description
Looks like our version does not fully support h2 `2.x`.
(see https://github.com/liquibase/liquibase/pull/2285 )
We use h2 for integration tests, all of which are currently failing (and have been disable by omitting the `ci` maven profile.
### Resourcing
We intend to implement this task | test | upgrade liquibase description looks like our version does not fully support x see we use for integration tests all of which are currently failing and have been disable by omitting the ci maven profile resourcing we intend to implement this task | 1 |
241,140 | 20,102,436,967 | IssuesEvent | 2022-02-07 06:44:00 | kyma-project/busola | https://api.github.com/repos/kyma-project/busola | closed | [SPIKE] Dev tools in tests | test-enhancement | <!-- Thank you for your contribution. Before you submit the issue:
1. Search open and closed issues for duplicates.
2. Read the contributing guidelines.
-->
**Description**
Try to add dev tools to the tests to have more control over what caused the error.
<!-- Provide a clear and concise description of the feature. -->
**Reasons**
When the test is failing we want to have dev tools on the record for better control over what caused the error.
<!-- Explain why we should add this feature. Provide use cases to illustrate its benefits. -->
**TIME**
1 PD
<!-- Attach any files, links, code samples, or screenshots that will convince us to your idea. -->
| 1.0 | [SPIKE] Dev tools in tests - <!-- Thank you for your contribution. Before you submit the issue:
1. Search open and closed issues for duplicates.
2. Read the contributing guidelines.
-->
**Description**
Try to add dev tools to the tests to have more control over what caused the error.
<!-- Provide a clear and concise description of the feature. -->
**Reasons**
When the test is failing we want to have dev tools on the record for better control over what caused the error.
<!-- Explain why we should add this feature. Provide use cases to illustrate its benefits. -->
**TIME**
1 PD
<!-- Attach any files, links, code samples, or screenshots that will convince us to your idea. -->
| test | dev tools in tests thank you for your contribution before you submit the issue search open and closed issues for duplicates read the contributing guidelines description try to add dev tools to the tests to have more control over what caused the error reasons when the test is failing we want to have dev tools on the record for better control over what caused the error time pd | 1 |
169,463 | 13,148,347,603 | IssuesEvent | 2020-08-08 20:58:26 | ansible/ansible | https://api.github.com/repos/ansible/ansible | closed | Mysql_db import failed with more than 5GB database. | affects_2.4 bug collection collection:community.general database module mysql needs_collection_redirect needs_info support:community test | <!---
Verify first that your issue/request is not already reported on GitHub.
Also test if the latest release, and devel branch are affected too.
Always add information AFTER of these html comments.
-->
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
<!---
-->
mysql_db
##### ANSIBLE VERSION
<!--- Paste verbatim output from "ansible --version" between quotes below -->
```
ansible 2.4.2.0
```
##### CONFIGURATION
<!---
-->
DEFAULT_HOST_LIST(/etc/ansible/ansible.cfg) = [u'/etc/ansible/hosts']
DEFAULT_LOG_PATH(/etc/ansible/ansible.cfg) = /var/log/ansible.log
DEFAULT_REMOTE_TMP(/etc/ansible/ansible.cfg) = ~/ansible/tmp
DEPRECATION_WARNINGS(/etc/ansible/ansible.cfg) = False
HOST_KEY_CHECKING(/etc/ansible/ansible.cfg) = False
##### OS / ENVIRONMENT
<!---
-->
Ansible server= Red Hat Enterprise Linux Server release 7.4
Host: Red Hat Enterprise Linux Server release 6.8 with mysql-community-server-5.7.21-1.el6.x86_64
##### SUMMARY
I'm trying to upgrade mysql from 5.1 to 5.7.
built 3 ansible playbook as follow.
1.backup the mysql
2.install and intialize mysql-5.7
3. import mysql-5.1 data dump to mysql-5.7 database.
Playbooks 1 & 3 are successfully completed, but playbook 3 is not exiting from the playbook after import the databases. I could see that all the databases are successfully imported,but the problem is playbook is not marking the tasks as finished and not proceeding with next tasks.
i tried with ignore_error: yes also doesn't help.
NOTE: the problem is reoccurring only for the database has more than 5GB. below 5GB databases are successfully imported.
##### STEPS TO REPRODUCE
<!---
-->
Backup database has more than 5GB from mysql-5.1
restore it to mysql-5.7 using mysql_db module.
<!--- Paste example playbooks or commands between quotes below -->
```yaml
---
- hosts: testvm
gather_facts: no
remote_user: itops
become_user: root
become_method: sudo
become: yes
vars:
backup_path: /srv/backups/mysql_DB_Upgrade/DB_Dump
tasks:
- name: Importing DB backup to MySQL-5.7
mysql_db:
name: all
state: import
target: "{{ backup_path }}/All_Databases_Backup_Except_default_databases_{{ inventory_hostname }}.sql.gz"
login_user: root
login_password: ***
- name: User_Grant_backup_after_upgrade
shell: mysql -uroot -p*** -e source "{{ backup_path }}/First_User_Grant_backup_after_upgrade_to_execute.txt" >> "{{ backup_path }}/Final_Import_Log.txt"
- name: mysql_upgrade command execution
shell: mysql_upgrade -uroot -p*** --socket=/var/lib/mysql/mysql.sock >> "{{ backup_path }}/Final_Import_Log.txt"
- name: Running mysqlcheck for all databases
shell: mysqlcheck -uroot -p*** --all-databases >> "{{ backup_path }}/Final_Import_Log.txt"
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
dump should be completed and go proceeding with next tasks.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
```
[itops@sn1ylvx1f0001 kalai]$ ansible-playbook -i iplist mysql_import.yml --ask-pass --ask-sudo-pass
SSH password:
SUDO password[defaults to SSH password]:
PLAY [testvm] ******************************************************************************************************************************
TASK [Importing DB backup to MySQL-5.7] ****************************************************************************************************
fatal: [10.28.51.203]: UNREACHABLE! => {"changed": false, "msg": "Failed to connect to the host via ssh: Shared connection to 10.28.51.203 closed.\r\n", "unreachable": true}
to retry, use: --limit @/home/itops/kalai/mysql_import.retry
PLAY RECAP *********************************************************************************************************************************
10.28.51.203 : ok=0 changed=0 unreachable=1 failed=0
[itops@sn1ylvx1f0001 kalai]$
``` | 1.0 | Mysql_db import failed with more than 5GB database. - <!---
Verify first that your issue/request is not already reported on GitHub.
Also test if the latest release, and devel branch are affected too.
Always add information AFTER of these html comments.
-->
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
<!---
-->
mysql_db
##### ANSIBLE VERSION
<!--- Paste verbatim output from "ansible --version" between quotes below -->
```
ansible 2.4.2.0
```
##### CONFIGURATION
<!---
-->
DEFAULT_HOST_LIST(/etc/ansible/ansible.cfg) = [u'/etc/ansible/hosts']
DEFAULT_LOG_PATH(/etc/ansible/ansible.cfg) = /var/log/ansible.log
DEFAULT_REMOTE_TMP(/etc/ansible/ansible.cfg) = ~/ansible/tmp
DEPRECATION_WARNINGS(/etc/ansible/ansible.cfg) = False
HOST_KEY_CHECKING(/etc/ansible/ansible.cfg) = False
##### OS / ENVIRONMENT
<!---
-->
Ansible server= Red Hat Enterprise Linux Server release 7.4
Host: Red Hat Enterprise Linux Server release 6.8 with mysql-community-server-5.7.21-1.el6.x86_64
##### SUMMARY
I'm trying to upgrade mysql from 5.1 to 5.7.
built 3 ansible playbook as follow.
1.backup the mysql
2.install and intialize mysql-5.7
3. import mysql-5.1 data dump to mysql-5.7 database.
Playbooks 1 & 3 are successfully completed, but playbook 3 is not exiting from the playbook after import the databases. I could see that all the databases are successfully imported,but the problem is playbook is not marking the tasks as finished and not proceeding with next tasks.
i tried with ignore_error: yes also doesn't help.
NOTE: the problem is reoccurring only for the database has more than 5GB. below 5GB databases are successfully imported.
##### STEPS TO REPRODUCE
<!---
-->
Backup database has more than 5GB from mysql-5.1
restore it to mysql-5.7 using mysql_db module.
<!--- Paste example playbooks or commands between quotes below -->
```yaml
---
- hosts: testvm
gather_facts: no
remote_user: itops
become_user: root
become_method: sudo
become: yes
vars:
backup_path: /srv/backups/mysql_DB_Upgrade/DB_Dump
tasks:
- name: Importing DB backup to MySQL-5.7
mysql_db:
name: all
state: import
target: "{{ backup_path }}/All_Databases_Backup_Except_default_databases_{{ inventory_hostname }}.sql.gz"
login_user: root
login_password: ***
- name: User_Grant_backup_after_upgrade
shell: mysql -uroot -p*** -e source "{{ backup_path }}/First_User_Grant_backup_after_upgrade_to_execute.txt" >> "{{ backup_path }}/Final_Import_Log.txt"
- name: mysql_upgrade command execution
shell: mysql_upgrade -uroot -p*** --socket=/var/lib/mysql/mysql.sock >> "{{ backup_path }}/Final_Import_Log.txt"
- name: Running mysqlcheck for all databases
shell: mysqlcheck -uroot -p*** --all-databases >> "{{ backup_path }}/Final_Import_Log.txt"
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
dump should be completed and go proceeding with next tasks.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
```
[itops@sn1ylvx1f0001 kalai]$ ansible-playbook -i iplist mysql_import.yml --ask-pass --ask-sudo-pass
SSH password:
SUDO password[defaults to SSH password]:
PLAY [testvm] ******************************************************************************************************************************
TASK [Importing DB backup to MySQL-5.7] ****************************************************************************************************
fatal: [10.28.51.203]: UNREACHABLE! => {"changed": false, "msg": "Failed to connect to the host via ssh: Shared connection to 10.28.51.203 closed.\r\n", "unreachable": true}
to retry, use: --limit @/home/itops/kalai/mysql_import.retry
PLAY RECAP *********************************************************************************************************************************
10.28.51.203 : ok=0 changed=0 unreachable=1 failed=0
[itops@sn1ylvx1f0001 kalai]$
``` | test | mysql db import failed with more than database verify first that your issue request is not already reported on github also test if the latest release and devel branch are affected too always add information after of these html comments issue type bug report component name mysql db ansible version ansible configuration default host list etc ansible ansible cfg default log path etc ansible ansible cfg var log ansible log default remote tmp etc ansible ansible cfg ansible tmp deprecation warnings etc ansible ansible cfg false host key checking etc ansible ansible cfg false os environment ansible server red hat enterprise linux server release host red hat enterprise linux server release with mysql community server summary i m trying to upgrade mysql from to built ansible playbook as follow backup the mysql install and intialize mysql import mysql data dump to mysql database playbooks are successfully completed but playbook is not exiting from the playbook after import the databases i could see that all the databases are successfully imported but the problem is playbook is not marking the tasks as finished and not proceeding with next tasks i tried with ignore error yes also doesn t help note the problem is reoccurring only for the database has more than below databases are successfully imported steps to reproduce backup database has more than from mysql restore it to mysql using mysql db module yaml hosts testvm gather facts no remote user itops become user root become method sudo become yes vars backup path srv backups mysql db upgrade db dump tasks name importing db backup to mysql mysql db name all state import target backup path all databases backup except default databases inventory hostname sql gz login user root login password name user grant backup after upgrade shell mysql uroot p e source backup path first user grant backup after upgrade to execute txt backup path final import log txt name mysql upgrade command execution shell mysql upgrade uroot p socket var lib mysql mysql sock backup path final import log txt name running mysqlcheck for all databases shell mysqlcheck uroot p all databases backup path final import log txt expected results dump should be completed and go proceeding with next tasks actual results ansible playbook i iplist mysql import yml ask pass ask sudo pass ssh password sudo password play task fatal unreachable changed false msg failed to connect to the host via ssh shared connection to closed r n unreachable true to retry use limit home itops kalai mysql import retry play recap ok changed unreachable failed | 1 |
101,269 | 8,783,618,463 | IssuesEvent | 2018-12-20 06:51:42 | appium/appium | https://api.github.com/repos/appium/appium | closed | On appium 1.11.0-beta.2 is failing to fetch dependencies for Appium iOS inspector | NeedsInfo ThirdParty XCUITest | ## The problem
After going to the correct location, creating the dir as in the documentation if we run `./Scripts/bootstrap.sh -d` we get this output:
> Fetching dependencies
Only this. No progress at all and we get our command line back to us, immediately.
As far as I understand this is what is responsible for letting us access Appium iOS inspector (which on this version has stopped working, because the server keeps returning 304 as response)
Reverting to 1.8.1 and the output for the same command is what is expected. Also, Appium iOS inspector is working fine there.
## Environment
* Appium version (or git revision) that exhibits the issue: 1.11.0-beta.2
* Last Appium version that did not exhibit the issue (if applicable): 1.8.1
* Desktop OS/version used to run Appium: Mac OS High Sierra
* Node.js version (unless using Appium.app|exe): v10.8.0
* Mobile platform/version under test: iOS 11.4.1 (iPhone 5S)
## Link to Appium logs
https://gist.github.com/notihs/2a59bc22020e082956dfd703d778461b | 1.0 | On appium 1.11.0-beta.2 is failing to fetch dependencies for Appium iOS inspector - ## The problem
After going to the correct location, creating the dir as in the documentation if we run `./Scripts/bootstrap.sh -d` we get this output:
> Fetching dependencies
Only this. No progress at all and we get our command line back to us, immediately.
As far as I understand this is what is responsible for letting us access Appium iOS inspector (which on this version has stopped working, because the server keeps returning 304 as response)
Reverting to 1.8.1 and the output for the same command is what is expected. Also, Appium iOS inspector is working fine there.
## Environment
* Appium version (or git revision) that exhibits the issue: 1.11.0-beta.2
* Last Appium version that did not exhibit the issue (if applicable): 1.8.1
* Desktop OS/version used to run Appium: Mac OS High Sierra
* Node.js version (unless using Appium.app|exe): v10.8.0
* Mobile platform/version under test: iOS 11.4.1 (iPhone 5S)
## Link to Appium logs
https://gist.github.com/notihs/2a59bc22020e082956dfd703d778461b | test | on appium beta is failing to fetch dependencies for appium ios inspector the problem after going to the correct location creating the dir as in the documentation if we run scripts bootstrap sh d we get this output fetching dependencies only this no progress at all and we get our command line back to us immediately as far as i understand this is what is responsible for letting us access appium ios inspector which on this version has stopped working because the server keeps returning as response reverting to and the output for the same command is what is expected also appium ios inspector is working fine there environment appium version or git revision that exhibits the issue beta last appium version that did not exhibit the issue if applicable desktop os version used to run appium mac os high sierra node js version unless using appium app exe mobile platform version under test ios iphone link to appium logs | 1 |
201,428 | 15,195,806,814 | IssuesEvent | 2021-02-16 07:06:45 | WeiXian042901/fyp_repository | https://api.github.com/repos/WeiXian042901/fyp_repository | opened | FU-003-Quest search(non-existing quest) | Acceptance Test Quest User | **Test Case**
- Check if the quest search function will filter out the quests if a non-existing quest is searched for
**Pre-Conditions**
- User must be logged in
- User should be on the quest page
- At least one quest should have been created. (e.g. Learning how to earn Passive Income)
**Test Steps**
1. Click on the search bar.
2. Type in “Hello World”
**Test Data**
- Search input: “Hello World”
**Expected Results**
- The user will be shown the text “No Results..”.
**Actual Results**
- The user is shown the text “No Results..”.
**Pass/Fail**
- Pass
**Tested by**
- Jun Liang
| 1.0 | FU-003-Quest search(non-existing quest) - **Test Case**
- Check if the quest search function will filter out the quests if a non-existing quest is searched for
**Pre-Conditions**
- User must be logged in
- User should be on the quest page
- At least one quest should have been created. (e.g. Learning how to earn Passive Income)
**Test Steps**
1. Click on the search bar.
2. Type in “Hello World”
**Test Data**
- Search input: “Hello World”
**Expected Results**
- The user will be shown the text “No Results..”.
**Actual Results**
- The user is shown the text “No Results..”.
**Pass/Fail**
- Pass
**Tested by**
- Jun Liang
| test | fu quest search non existing quest test case check if the quest search function will filter out the quests if a non existing quest is searched for pre conditions user must be logged in user should be on the quest page at least one quest should have been created e g learning how to earn passive income test steps click on the search bar type in “hello world” test data search input “hello world” expected results the user will be shown the text “no results ” actual results the user is shown the text “no results ” pass fail pass tested by jun liang | 1 |
127,117 | 10,452,867,126 | IssuesEvent | 2019-09-19 15:26:30 | ashadabali/BP-SQUAD3-IM12 | https://api.github.com/repos/ashadabali/BP-SQUAD3-IM12 | closed | suvadeep - Next Gen Pricer completion | Samples - TEST | * cost Rates are being pulled.
* FP Prem% are auto calculated with separate Prem Value.
* Test 4 sent for action. | 1.0 | suvadeep - Next Gen Pricer completion - * cost Rates are being pulled.
* FP Prem% are auto calculated with separate Prem Value.
* Test 4 sent for action. | test | suvadeep next gen pricer completion cost rates are being pulled fp prem are auto calculated with separate prem value test sent for action | 1 |
34,081 | 4,890,526,780 | IssuesEvent | 2016-11-18 14:10:39 | TransNexus/NexOSS | https://api.github.com/repos/TransNexus/NexOSS | closed | New special CDR Details page for Registration Messages | enhancement needs testing | #### Page Path
#### Description of Issue
@fenichelar We create CDRs for Registration messages. It is convenient, but fields in Registration messages do not map exactly to CDRs. So when the details are viewed, there are column titles that do not make sense. Edit registrationCDR page

| 1.0 | New special CDR Details page for Registration Messages - #### Page Path
#### Description of Issue
@fenichelar We create CDRs for Registration messages. It is convenient, but fields in Registration messages do not map exactly to CDRs. So when the details are viewed, there are column titles that do not make sense. Edit registrationCDR page

| test | new special cdr details page for registration messages page path description of issue fenichelar we create cdrs for registration messages it is convenient but fields in registration messages do not map exactly to cdrs so when the details are viewed there are column titles that do not make sense edit registrationcdr page | 1 |
347,383 | 31,161,169,901 | IssuesEvent | 2023-08-16 16:05:51 | elastic/kibana | https://api.github.com/repos/elastic/kibana | opened | Failing test: X-Pack Saved Object API Integration Tests -- security_and_spaces.x-pack/test/saved_object_api_integration/security_and_spaces/apis/bulk_create·ts - saved objects security and spaces enabled _bulk_create legacy user within the default space should return 403 forbidden [isolatedtype/defaultspace-isolatedtype-id,isolatedtype/space1-isolatedtype-id,isolatedtype/space2-isolatedtype-id,sharedtype/all_spaces,sharedtype/default_and_space_1,sharedtype/only_space_1,sharedtype/only_space_2,sharecapabletype/only_default_space,sharecapabletype/only_space_1,globaltype/globaltype-id,dashboard/new-dashboard-id,sharedtype/new-sharedtype-id,globaltype/new-globaltype-id,resolvetype/alias-match,hiddentype/any,isolatedtype/new-other-space-id,sharecapabletype/new-other-space-id,resolvetype/alias-match,isolatedtype/new-other-space-id,sharecapabletype/new-other-space-id,sharedtype/new-each-space-id,sharedtype/new-all-spaces-id] | failed-test | A test failed on a tracked branch
```
Error: read ECONNRESET
at TCP.onStreamRead (node:internal/stream_base_commons:217:20) {
errno: -104,
code: 'ECONNRESET',
syscall: 'read',
response: undefined
}
```
First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/34055#0189fef4-eae1-4423-8394-43024ed9ab8c)
<!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Saved Object API Integration Tests -- security_and_spaces.x-pack/test/saved_object_api_integration/security_and_spaces/apis/bulk_create·ts","test.name":"saved objects security and spaces enabled _bulk_create legacy user within the default space should return 403 forbidden [isolatedtype/defaultspace-isolatedtype-id,isolatedtype/space1-isolatedtype-id,isolatedtype/space2-isolatedtype-id,sharedtype/all_spaces,sharedtype/default_and_space_1,sharedtype/only_space_1,sharedtype/only_space_2,sharecapabletype/only_default_space,sharecapabletype/only_space_1,globaltype/globaltype-id,dashboard/new-dashboard-id,sharedtype/new-sharedtype-id,globaltype/new-globaltype-id,resolvetype/alias-match,hiddentype/any,isolatedtype/new-other-space-id,sharecapabletype/new-other-space-id,resolvetype/alias-match,isolatedtype/new-other-space-id,sharecapabletype/new-other-space-id,sharedtype/new-each-space-id,sharedtype/new-all-spaces-id]","test.failCount":1}} --> | 1.0 | Failing test: X-Pack Saved Object API Integration Tests -- security_and_spaces.x-pack/test/saved_object_api_integration/security_and_spaces/apis/bulk_create·ts - saved objects security and spaces enabled _bulk_create legacy user within the default space should return 403 forbidden [isolatedtype/defaultspace-isolatedtype-id,isolatedtype/space1-isolatedtype-id,isolatedtype/space2-isolatedtype-id,sharedtype/all_spaces,sharedtype/default_and_space_1,sharedtype/only_space_1,sharedtype/only_space_2,sharecapabletype/only_default_space,sharecapabletype/only_space_1,globaltype/globaltype-id,dashboard/new-dashboard-id,sharedtype/new-sharedtype-id,globaltype/new-globaltype-id,resolvetype/alias-match,hiddentype/any,isolatedtype/new-other-space-id,sharecapabletype/new-other-space-id,resolvetype/alias-match,isolatedtype/new-other-space-id,sharecapabletype/new-other-space-id,sharedtype/new-each-space-id,sharedtype/new-all-spaces-id] - A test failed on a tracked branch
```
Error: read ECONNRESET
at TCP.onStreamRead (node:internal/stream_base_commons:217:20) {
errno: -104,
code: 'ECONNRESET',
syscall: 'read',
response: undefined
}
```
First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/34055#0189fef4-eae1-4423-8394-43024ed9ab8c)
<!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Saved Object API Integration Tests -- security_and_spaces.x-pack/test/saved_object_api_integration/security_and_spaces/apis/bulk_create·ts","test.name":"saved objects security and spaces enabled _bulk_create legacy user within the default space should return 403 forbidden [isolatedtype/defaultspace-isolatedtype-id,isolatedtype/space1-isolatedtype-id,isolatedtype/space2-isolatedtype-id,sharedtype/all_spaces,sharedtype/default_and_space_1,sharedtype/only_space_1,sharedtype/only_space_2,sharecapabletype/only_default_space,sharecapabletype/only_space_1,globaltype/globaltype-id,dashboard/new-dashboard-id,sharedtype/new-sharedtype-id,globaltype/new-globaltype-id,resolvetype/alias-match,hiddentype/any,isolatedtype/new-other-space-id,sharecapabletype/new-other-space-id,resolvetype/alias-match,isolatedtype/new-other-space-id,sharecapabletype/new-other-space-id,sharedtype/new-each-space-id,sharedtype/new-all-spaces-id]","test.failCount":1}} --> | test | failing test x pack saved object api integration tests security and spaces x pack test saved object api integration security and spaces apis bulk create·ts saved objects security and spaces enabled bulk create legacy user within the default space should return forbidden a test failed on a tracked branch error read econnreset at tcp onstreamread node internal stream base commons errno code econnreset syscall read response undefined first failure | 1 |
127,987 | 27,170,713,871 | IssuesEvent | 2023-02-17 19:08:46 | files-community/Files | https://api.github.com/repos/files-community/Files | closed | Enforce Xaml indentation in EditorConfig | codebase quality | ### Description
Similar to what is currently there for C# source files, it would make Visual Studio automatically use tabs instead of the default spaces
### Concerned code
All Xaml files
### Gains
A more seamless contribution experience by having code automatically comply with style guidelines
### Requirements
_No response_
### Comments
_No response_ | 1.0 | Enforce Xaml indentation in EditorConfig - ### Description
Similar to what is currently there for C# source files, it would make Visual Studio automatically use tabs instead of the default spaces
### Concerned code
All Xaml files
### Gains
A more seamless contribution experience by having code automatically comply with style guidelines
### Requirements
_No response_
### Comments
_No response_ | non_test | enforce xaml indentation in editorconfig description similar to what is currently there for c source files it would make visual studio automatically use tabs instead of the default spaces concerned code all xaml files gains a more seamless contribution experience by having code automatically comply with style guidelines requirements no response comments no response | 0 |
235,224 | 19,309,010,148 | IssuesEvent | 2021-12-13 14:30:33 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: sqlsmith/setup=rand-tables/setting=no-mutations failed | C-test-failure O-robot O-roachtest release-blocker T-sql-queries branch-release-21.2 | roachtest.sqlsmith/setup=rand-tables/setting=no-mutations [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3889466&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3889466&tab=artifacts#/sqlsmith/setup=rand-tables/setting=no-mutations) on release-21.2 @ [3cf7ac914c42bb024c688564983445f3eb5da58a](https://github.com/cockroachdb/cockroach/commits/3cf7ac914c42bb024c688564983445f3eb5da58a):
```
AS (
SELECT
*
FROM
(
VALUES
(((-0.13295479970525648):::FLOAT8::FLOAT8 + '+Inf':::FLOAT8::FLOAT8)::FLOAT8),
(
((st_frechetdistance('01020000000200000000C0267361603BC126AD370198C8E0C188AD4DEC7E78EF414C6973E3CF9FDDC1':::GEOMETRY::GEOMETRY, '0106000080050000000103000080010000000A000000C837BB203324D2C15455BC630548EAC1E0176809970DEF4150AD52BE3D44F7411054F5A1A20002C28022E388A09893C1B45512ECE28BFB41E53C492358C001C2BC7CDC56439AF241FAD433BEF1C900421816DA6E9E25ECC18C64D9C7EF33F5C1ACDDA25B1EBAF141909DB721AD45C5412016B254109CE14100AFC1686246BD417C786A5158BF0142C36722B1C1C9F4C1D04E8608D2E9C84140F44C58E387F241EC98417033C2E0417A8416E68A2BEAC1E0E0497BCA6401428EA6DA1CA4E0F6414BCDA6AE6AAAF4C1B0282F631261D4417086872081CAFF41C837BB203324D2C15455BC630548EAC1E0176809970DEF410103000080010000000A000000834A1E742E65FBC1DCFF5512AC9FEAC182AD09493F2A02425E71E120E92FFDC1A172A77DF20AF4C1B0FEE7402B6BF241E82CE1A2A92EF8C1C00E3D5270FBFAC18C13776502300042C21E0D794243F2C1E93FD368DB13F7C130E8F05D9455FD41507842C894B7CA41CD574F2C0E1BF1C1985889E05F83DA41B6985CB81E37FA4126AA25774BFEF4C1788A04BC802AF841684508C2C00DFD414890C6347295D3C1B8A3FE6E2957FC41284CF2DB083CFE41D05F6C7CFCFEED41005F87D22E2FC0C136C0051CAA60E8C1484327F16C05D44120D708D8AEDCCAC1834A1E742E65FBC1DCFF5512AC9FEAC182AD09493F2A02420103000080010000000A000000A80B98DA1D2F00C29AC3E885301BE3C114EB23BB987BFAC1082A788CF08AE4C1DC09B425F349E9C1DD939CBD5B90FEC1768D466A82AFF6C150842322FDBB00C204CDB2A211CBE5412079211F12BCBF41E827B49F9AC0F1C17012B9BACD6CC9C1C0BC719BB0E0E841762EB926D04401C2D2811B01264DE2C116EE3F8184AD01428633E6E02120E8C1AC4E43F1AF9EEC41D8E4F13FC33CD341683F58DE3386EF412CF8552B7E01E341700C069D07DFDD41B41CD75760D2FB41E060636CDB21EF41BA34B34980F5F1C1306310091F12CD41BC7B5EC8E25EE941A80B98DA1D2F00C29AC3E885301BE3C114EB23BB987BFAC10103000080010000000D000000C043BDEA241EE5C1489AB9EC9C1BE64158F13FCC2CA6D5416F8142E4E92A00C250CD6C60B181CDC184989D3640D3D8C1D673DAEF6215E4C16861428796CBD541D4423586A54EF841C2E19F996BC00042017E9F85AE4FF1C160751A401155E941701DC0646A53DF4138376763C82AD141D3146411B96801C240E5E1C4EDB6E741D45C7BC7CA02E6419464A4F0E719ECC1EC085E22E712F7416C9EB765C8E0E741D8155DC9674100C2D85CC65E34CEF4416E0BFAEB066201425F39B5C955B7F4C134F211C3D920F4C1C0EAECDF466BF941B013E1962412D2412A938674B59302C2FE15AE33CCCDF941A4C650FFADDFE44178DF3D8D370DF4C1C0C646F2E305F241C8EC934FCA44E84190DDD919003102C20C412045B0B4F0411868FE98B5FBEB41C043BDEA241EE5C1489AB9EC9C1BE64158F13FCC2CA6D54101030000800100000008000000DFE0CD994261FDC1397195B9A600FBC1D60898144F8EFFC1322010437AE1F4C1A48C867C8BC1F9C18AD475019E2A0242963DD214F157F0C1CCC22589E3B1F5C160E6742EE842E1414E4D2469F77AF04148D0C1BAAF0DFA41880689DF8463D1C18E1589FDEA98E3C180CECCE40F8DF941FC25616D8124EA41EA7E82DB6CCCECC1508D85AAE72EDA411603DC03D54AF6411064A8E6EADFFAC1F845D9CEECFBEA41086CF7ACF719D441DFE0CD994261FDC1397195B9A600FBC1D60898144F8EFFC1':::GEOMETRY::GEOMETRY, 0.0006136327746866233:::FLOAT8::FLOAT8)::FLOAT8::FLOAT8 / (-0.061237652166769685):::FLOAT8::FLOAT8)::FLOAT8::FLOAT8 + NULL::FLOAT8)::FLOAT8
),
(COALESCE(0.9259235719109529:::FLOAT8, 0.12205111988676354:::FLOAT8))
)
AS tab_178906 (col_418134)
),
with_73130 (col_418135)
AS (
SELECT
*
FROM
(
VALUES
('54 years 7 mons 924 days 00:21:06.625493':::INTERVAL),
('-29 years -3 mons -29 days -04:27:05.503828':::INTERVAL),
('54 years 7 mons 286 days 22:39:32.140595':::INTERVAL)
)
AS tab_178907 (col_418135)
)
SELECT
tab_178908.col2_8 AS col_418136, '15:18:24.715944+08:54:00':::TIMETZ AS col_418137, tab_178911.col2_2 AS col_418138
FROM
defaultdb.public.table2@table2_col2_0_col2_4_key AS tab_178908
JOIN defaultdb.public.table2@[0] AS tab_178909 ON
(tab_178908.crdb_internal_idx_expr_2) = (tab_178909.crdb_internal_idx_expr_2)
AND (tab_178908.col2_10) = (tab_178909.col2_3)
AND (tab_178908.col2_7) = (tab_178909.tableoid)
AND (tab_178908.crdb_internal_idx_expr_1) = (tab_178909.crdb_internal_idx_expr_2)
AND (tab_178908.col2_6) = (tab_178909.col2_6)
AND (tab_178908.crdb_internal_idx_expr) = (tab_178909.col2_10)
AND (tab_178908.crdb_internal_idx_expr_1) = (tab_178909.col2_10)
AND (tab_178908.col2_10) = (tab_178909.crdb_internal_idx_expr_1),
defaultdb.public.table2 AS tab_178910
JOIN defaultdb.public.table2 AS tab_178911 ON
(tab_178910.col2_3) = (tab_178911.col2_3)
AND (tab_178910.col2_6) = (tab_178911.col2_6)
AND (tab_178910.col2_9) = (tab_178911.col2_9)
AND (tab_178910.col2_10) = (tab_178911.col2_10)
AND (tab_178910.col2_0) = (tab_178911.col2_0),
with_73130 AS cte_ref_21137
LIMIT
43:::INT8;
```
<details><summary>Reproduce</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
</p>
</details>
/cc @cockroachdb/sql-queries
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sqlsmith/setup=rand-tables/setting=no-mutations.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| 2.0 | roachtest: sqlsmith/setup=rand-tables/setting=no-mutations failed - roachtest.sqlsmith/setup=rand-tables/setting=no-mutations [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3889466&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3889466&tab=artifacts#/sqlsmith/setup=rand-tables/setting=no-mutations) on release-21.2 @ [3cf7ac914c42bb024c688564983445f3eb5da58a](https://github.com/cockroachdb/cockroach/commits/3cf7ac914c42bb024c688564983445f3eb5da58a):
```
AS (
SELECT
*
FROM
(
VALUES
(((-0.13295479970525648):::FLOAT8::FLOAT8 + '+Inf':::FLOAT8::FLOAT8)::FLOAT8),
(
((st_frechetdistance('01020000000200000000C0267361603BC126AD370198C8E0C188AD4DEC7E78EF414C6973E3CF9FDDC1':::GEOMETRY::GEOMETRY, '0106000080050000000103000080010000000A000000C837BB203324D2C15455BC630548EAC1E0176809970DEF4150AD52BE3D44F7411054F5A1A20002C28022E388A09893C1B45512ECE28BFB41E53C492358C001C2BC7CDC56439AF241FAD433BEF1C900421816DA6E9E25ECC18C64D9C7EF33F5C1ACDDA25B1EBAF141909DB721AD45C5412016B254109CE14100AFC1686246BD417C786A5158BF0142C36722B1C1C9F4C1D04E8608D2E9C84140F44C58E387F241EC98417033C2E0417A8416E68A2BEAC1E0E0497BCA6401428EA6DA1CA4E0F6414BCDA6AE6AAAF4C1B0282F631261D4417086872081CAFF41C837BB203324D2C15455BC630548EAC1E0176809970DEF410103000080010000000A000000834A1E742E65FBC1DCFF5512AC9FEAC182AD09493F2A02425E71E120E92FFDC1A172A77DF20AF4C1B0FEE7402B6BF241E82CE1A2A92EF8C1C00E3D5270FBFAC18C13776502300042C21E0D794243F2C1E93FD368DB13F7C130E8F05D9455FD41507842C894B7CA41CD574F2C0E1BF1C1985889E05F83DA41B6985CB81E37FA4126AA25774BFEF4C1788A04BC802AF841684508C2C00DFD414890C6347295D3C1B8A3FE6E2957FC41284CF2DB083CFE41D05F6C7CFCFEED41005F87D22E2FC0C136C0051CAA60E8C1484327F16C05D44120D708D8AEDCCAC1834A1E742E65FBC1DCFF5512AC9FEAC182AD09493F2A02420103000080010000000A000000A80B98DA1D2F00C29AC3E885301BE3C114EB23BB987BFAC1082A788CF08AE4C1DC09B425F349E9C1DD939CBD5B90FEC1768D466A82AFF6C150842322FDBB00C204CDB2A211CBE5412079211F12BCBF41E827B49F9AC0F1C17012B9BACD6CC9C1C0BC719BB0E0E841762EB926D04401C2D2811B01264DE2C116EE3F8184AD01428633E6E02120E8C1AC4E43F1AF9EEC41D8E4F13FC33CD341683F58DE3386EF412CF8552B7E01E341700C069D07DFDD41B41CD75760D2FB41E060636CDB21EF41BA34B34980F5F1C1306310091F12CD41BC7B5EC8E25EE941A80B98DA1D2F00C29AC3E885301BE3C114EB23BB987BFAC10103000080010000000D000000C043BDEA241EE5C1489AB9EC9C1BE64158F13FCC2CA6D5416F8142E4E92A00C250CD6C60B181CDC184989D3640D3D8C1D673DAEF6215E4C16861428796CBD541D4423586A54EF841C2E19F996BC00042017E9F85AE4FF1C160751A401155E941701DC0646A53DF4138376763C82AD141D3146411B96801C240E5E1C4EDB6E741D45C7BC7CA02E6419464A4F0E719ECC1EC085E22E712F7416C9EB765C8E0E741D8155DC9674100C2D85CC65E34CEF4416E0BFAEB066201425F39B5C955B7F4C134F211C3D920F4C1C0EAECDF466BF941B013E1962412D2412A938674B59302C2FE15AE33CCCDF941A4C650FFADDFE44178DF3D8D370DF4C1C0C646F2E305F241C8EC934FCA44E84190DDD919003102C20C412045B0B4F0411868FE98B5FBEB41C043BDEA241EE5C1489AB9EC9C1BE64158F13FCC2CA6D54101030000800100000008000000DFE0CD994261FDC1397195B9A600FBC1D60898144F8EFFC1322010437AE1F4C1A48C867C8BC1F9C18AD475019E2A0242963DD214F157F0C1CCC22589E3B1F5C160E6742EE842E1414E4D2469F77AF04148D0C1BAAF0DFA41880689DF8463D1C18E1589FDEA98E3C180CECCE40F8DF941FC25616D8124EA41EA7E82DB6CCCECC1508D85AAE72EDA411603DC03D54AF6411064A8E6EADFFAC1F845D9CEECFBEA41086CF7ACF719D441DFE0CD994261FDC1397195B9A600FBC1D60898144F8EFFC1':::GEOMETRY::GEOMETRY, 0.0006136327746866233:::FLOAT8::FLOAT8)::FLOAT8::FLOAT8 / (-0.061237652166769685):::FLOAT8::FLOAT8)::FLOAT8::FLOAT8 + NULL::FLOAT8)::FLOAT8
),
(COALESCE(0.9259235719109529:::FLOAT8, 0.12205111988676354:::FLOAT8))
)
AS tab_178906 (col_418134)
),
with_73130 (col_418135)
AS (
SELECT
*
FROM
(
VALUES
('54 years 7 mons 924 days 00:21:06.625493':::INTERVAL),
('-29 years -3 mons -29 days -04:27:05.503828':::INTERVAL),
('54 years 7 mons 286 days 22:39:32.140595':::INTERVAL)
)
AS tab_178907 (col_418135)
)
SELECT
tab_178908.col2_8 AS col_418136, '15:18:24.715944+08:54:00':::TIMETZ AS col_418137, tab_178911.col2_2 AS col_418138
FROM
defaultdb.public.table2@table2_col2_0_col2_4_key AS tab_178908
JOIN defaultdb.public.table2@[0] AS tab_178909 ON
(tab_178908.crdb_internal_idx_expr_2) = (tab_178909.crdb_internal_idx_expr_2)
AND (tab_178908.col2_10) = (tab_178909.col2_3)
AND (tab_178908.col2_7) = (tab_178909.tableoid)
AND (tab_178908.crdb_internal_idx_expr_1) = (tab_178909.crdb_internal_idx_expr_2)
AND (tab_178908.col2_6) = (tab_178909.col2_6)
AND (tab_178908.crdb_internal_idx_expr) = (tab_178909.col2_10)
AND (tab_178908.crdb_internal_idx_expr_1) = (tab_178909.col2_10)
AND (tab_178908.col2_10) = (tab_178909.crdb_internal_idx_expr_1),
defaultdb.public.table2 AS tab_178910
JOIN defaultdb.public.table2 AS tab_178911 ON
(tab_178910.col2_3) = (tab_178911.col2_3)
AND (tab_178910.col2_6) = (tab_178911.col2_6)
AND (tab_178910.col2_9) = (tab_178911.col2_9)
AND (tab_178910.col2_10) = (tab_178911.col2_10)
AND (tab_178910.col2_0) = (tab_178911.col2_0),
with_73130 AS cte_ref_21137
LIMIT
43:::INT8;
```
<details><summary>Reproduce</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
</p>
</details>
/cc @cockroachdb/sql-queries
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sqlsmith/setup=rand-tables/setting=no-mutations.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| test | roachtest sqlsmith setup rand tables setting no mutations failed roachtest sqlsmith setup rand tables setting no mutations with on release as select from values inf st frechetdistance geometry geometry geometry geometry null coalesce as tab col with col as select from values years mons days interval years mons days interval years mons days interval as tab col select tab as col timetz as col tab as col from defaultdb public key as tab join defaultdb public as tab on tab crdb internal idx expr tab crdb internal idx expr and tab tab and tab tab tableoid and tab crdb internal idx expr tab crdb internal idx expr and tab tab and tab crdb internal idx expr tab and tab crdb internal idx expr tab and tab tab crdb internal idx expr defaultdb public as tab join defaultdb public as tab on tab tab and tab tab and tab tab and tab tab and tab tab with as cte ref limit reproduce see cc cockroachdb sql queries | 1 |
56,118 | 6,961,536,757 | IssuesEvent | 2017-12-08 09:51:59 | vaadin/flow | https://api.github.com/repos/vaadin/flow | closed | Enable DataProvider sorting via backend Comparators | DX enhancement feedback needs design | When dealing with data providers, I wan to be able to sort the returned stream via my one comparators, which are special implementations for backend optimization.
At the moment, the `Query` object provides this sorting data:
1. a list of `SortOrder<String>` via `getSortOrders()`
2. a `Comparator<T>` via `getInMemorySorting()`
The main problem with (1) is that it assumes backend sorting can be done with a string, which is very common but excludes modern abstractions such as custom database-backed streams which needs special implementation of `Comparator` to sort.
Getting the `Comparator<T>` with `Query::getInMemorySorting()` though, returns an instance built from comparators set via `Column::setComparator`, which immediately re-assigns the comparator instance passed as argument via `this.comparator = comparator::compare` to make is serializable, thus discarding the original implementation which may have been backed-aware.
What I suggest is having `Column::setComparator` to keep the original argument beside the serializable version, and add a `Query::getBackendSorting` returning a `Stream<Comparator<T>>` of all the original instances of column comparators. This way, it would be simple for a data provider to sort the stream with custom comparators, e.g.
```java
@Override
protected Stream<T> fetchFromBackEnd(Query<T, Predicate<T>> query) {
Stream<T> stream = this.repository.stream();
if (query.getFilter().isPresent()) {
stream = stream.filter(query.getFilter().get());
}
Stream<Comparator<T>> comparators = query.getBackendSorting();
Optional<Comparator<T>> comparator = comparators.reduce(Comparator::thenComparing);
if (comparator.isPresent()) {
stream = stream.sorted(comparator.get());
}
return stream.skip(query.getOffset()).limit(query.getLimit());
}
``` | 1.0 | Enable DataProvider sorting via backend Comparators - When dealing with data providers, I wan to be able to sort the returned stream via my one comparators, which are special implementations for backend optimization.
At the moment, the `Query` object provides this sorting data:
1. a list of `SortOrder<String>` via `getSortOrders()`
2. a `Comparator<T>` via `getInMemorySorting()`
The main problem with (1) is that it assumes backend sorting can be done with a string, which is very common but excludes modern abstractions such as custom database-backed streams which needs special implementation of `Comparator` to sort.
Getting the `Comparator<T>` with `Query::getInMemorySorting()` though, returns an instance built from comparators set via `Column::setComparator`, which immediately re-assigns the comparator instance passed as argument via `this.comparator = comparator::compare` to make is serializable, thus discarding the original implementation which may have been backed-aware.
What I suggest is having `Column::setComparator` to keep the original argument beside the serializable version, and add a `Query::getBackendSorting` returning a `Stream<Comparator<T>>` of all the original instances of column comparators. This way, it would be simple for a data provider to sort the stream with custom comparators, e.g.
```java
@Override
protected Stream<T> fetchFromBackEnd(Query<T, Predicate<T>> query) {
Stream<T> stream = this.repository.stream();
if (query.getFilter().isPresent()) {
stream = stream.filter(query.getFilter().get());
}
Stream<Comparator<T>> comparators = query.getBackendSorting();
Optional<Comparator<T>> comparator = comparators.reduce(Comparator::thenComparing);
if (comparator.isPresent()) {
stream = stream.sorted(comparator.get());
}
return stream.skip(query.getOffset()).limit(query.getLimit());
}
``` | non_test | enable dataprovider sorting via backend comparators when dealing with data providers i wan to be able to sort the returned stream via my one comparators which are special implementations for backend optimization at the moment the query object provides this sorting data a list of sortorder via getsortorders a comparator via getinmemorysorting the main problem with is that it assumes backend sorting can be done with a string which is very common but excludes modern abstractions such as custom database backed streams which needs special implementation of comparator to sort getting the comparator with query getinmemorysorting though returns an instance built from comparators set via column setcomparator which immediately re assigns the comparator instance passed as argument via this comparator comparator compare to make is serializable thus discarding the original implementation which may have been backed aware what i suggest is having column setcomparator to keep the original argument beside the serializable version and add a query getbackendsorting returning a stream of all the original instances of column comparators this way it would be simple for a data provider to sort the stream with custom comparators e g java override protected stream fetchfrombackend query query stream stream this repository stream if query getfilter ispresent stream stream filter query getfilter get stream comparators query getbackendsorting optional comparator comparators reduce comparator thencomparing if comparator ispresent stream stream sorted comparator get return stream skip query getoffset limit query getlimit | 0 |
259,373 | 22,470,926,043 | IssuesEvent | 2022-06-22 08:03:18 | tracim/tracim | https://api.github.com/repos/tracim/tracim | opened | Misc: Add tests for to do feature | frontend missing tests cypress | ## Description and expectations
Add unit and functional tests for the to do feature.
Note: first PR about the feature #5716
| 1.0 | Misc: Add tests for to do feature - ## Description and expectations
Add unit and functional tests for the to do feature.
Note: first PR about the feature #5716
| test | misc add tests for to do feature description and expectations add unit and functional tests for the to do feature note first pr about the feature | 1 |
23,371 | 10,881,875,861 | IssuesEvent | 2019-11-17 20:35:33 | MirkoV1987/P5-PHPBlog | https://api.github.com/repos/MirkoV1987/P5-PHPBlog | opened | CVE-2018-19826 (Medium) detected in node-sass-v4.11.0 | security vulnerability | ## CVE-2018-19826 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.11.0</b></p></summary>
<p>
<p>:rainbow: Node.js bindings to libsass</p>
<p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/MirkoV1987/P5-PHPBlog/commit/272c4b6213ea4cc72fed4b0e3ea3c3ceaccf5ba0">272c4b6213ea4cc72fed4b0e3ea3c3ceaccf5ba0</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (66)</summary>
<p></p>
<p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p>
<p>
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/expand.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/expand.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/factory.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/operators.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/boolean.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/util.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/value.h
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/emitter.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/callback_bridge.h
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/file.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/sass.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/operation.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/operators.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/constants.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/error_handling.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/custom_importer_bridge.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/parser.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/constants.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/list.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/cssize.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/functions.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/util.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/custom_function_bridge.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/custom_importer_bridge.h
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/bind.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/eval.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/inspect.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/backtrace.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/extend.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_context_wrapper.h
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/sass_value_wrapper.h
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/error_handling.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/parser.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/debugger.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/emitter.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/number.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/color.h
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/sass_values.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/ast.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/output.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/check_nesting.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/null.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/ast_def_macros.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/functions.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/cssize.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/prelexer.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/ast.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/to_c.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/to_value.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/ast_fwd_decl.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/inspect.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/color.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/values.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_context_wrapper.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/list.h
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/check_nesting.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/map.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/to_value.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/context.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/binding.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/string.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/sass_context.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/prelexer.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/context.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/boolean.h
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/eval.cpp
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
** DISPUTED ** In inspect.cpp in LibSass 3.5.5, a high memory footprint caused by an endless loop (containing a Sass::Inspect::operator()(Sass::String_Quoted*) stack frame) may cause a Denial of Service via crafted sass input files with stray '&' or '/' characters. NOTE: Upstream comments indicate this issue is closed as "won't fix" and "works as intended" by design.
<p>Publish Date: 2018-12-03
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19826>CVE-2018-19826</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19826">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19826</a></p>
<p>Release Date: 2019-09-01</p>
<p>Fix Resolution: Replace or update the following file: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-19826 (Medium) detected in node-sass-v4.11.0 - ## CVE-2018-19826 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.11.0</b></p></summary>
<p>
<p>:rainbow: Node.js bindings to libsass</p>
<p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/MirkoV1987/P5-PHPBlog/commit/272c4b6213ea4cc72fed4b0e3ea3c3ceaccf5ba0">272c4b6213ea4cc72fed4b0e3ea3c3ceaccf5ba0</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (66)</summary>
<p></p>
<p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p>
<p>
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/expand.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/expand.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/factory.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/operators.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/boolean.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/util.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/value.h
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/emitter.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/callback_bridge.h
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/file.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/sass.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/operation.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/operators.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/constants.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/error_handling.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/custom_importer_bridge.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/parser.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/constants.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/list.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/cssize.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/functions.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/util.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/custom_function_bridge.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/custom_importer_bridge.h
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/bind.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/eval.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/inspect.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/backtrace.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/extend.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_context_wrapper.h
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/sass_value_wrapper.h
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/error_handling.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/parser.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/debugger.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/emitter.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/number.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/color.h
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/sass_values.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/ast.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/output.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/check_nesting.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/null.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/ast_def_macros.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/functions.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/cssize.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/prelexer.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/ast.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/to_c.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/to_value.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/ast_fwd_decl.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/inspect.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/color.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/values.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_context_wrapper.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/list.h
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/check_nesting.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/map.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/to_value.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/context.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/binding.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/string.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/sass_context.cpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/prelexer.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/context.hpp
- /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/boolean.h
- /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/eval.cpp
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
** DISPUTED ** In inspect.cpp in LibSass 3.5.5, a high memory footprint caused by an endless loop (containing a Sass::Inspect::operator()(Sass::String_Quoted*) stack frame) may cause a Denial of Service via crafted sass input files with stray '&' or '/' characters. NOTE: Upstream comments indicate this issue is closed as "won't fix" and "works as intended" by design.
<p>Publish Date: 2018-12-03
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19826>CVE-2018-19826</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19826">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19826</a></p>
<p>Release Date: 2019-09-01</p>
<p>Fix Resolution: Replace or update the following file: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve medium detected in node sass cve medium severity vulnerability vulnerable library node rainbow node js bindings to libsass library home page a href found in head commit a href library source files the source files were matched to this source library based on a best effort match source libraries are selected from a list of probable public libraries phpblog public node modules node sass src libsass src expand hpp phpblog public node modules node sass src libsass src expand cpp phpblog public node modules node sass src sass types factory cpp phpblog public node modules node sass src libsass src operators cpp phpblog public node modules node sass src sass types boolean cpp phpblog public node modules node sass src libsass src util hpp phpblog public node modules node sass src sass types value h phpblog public node modules node sass src libsass src emitter hpp phpblog public node modules node sass src callback bridge h phpblog public node modules node sass src libsass src file cpp phpblog public node modules node sass src libsass src sass cpp phpblog public node modules node sass src libsass src operation hpp phpblog public node modules node sass src libsass src operators hpp phpblog public node modules node sass src libsass src constants hpp phpblog public node modules node sass src libsass src error handling hpp phpblog public node modules node sass src custom importer bridge cpp phpblog public node modules node sass src libsass src parser hpp phpblog public node modules node sass src libsass src constants cpp phpblog public node modules node sass src sass types list cpp phpblog public node modules node sass src libsass src cssize cpp phpblog public node modules node sass src libsass src functions hpp phpblog public node modules node sass src libsass src util cpp phpblog public node modules node sass src custom function bridge cpp phpblog public node modules node sass src custom importer bridge h phpblog public node modules node sass src libsass src bind cpp phpblog public node modules node sass src libsass src eval hpp phpblog public node modules node sass src libsass src inspect cpp phpblog public node modules node sass src libsass src backtrace cpp phpblog public node modules node sass src libsass src extend cpp phpblog public node modules node sass src sass context wrapper h phpblog public node modules node sass src sass types sass value wrapper h phpblog public node modules node sass src libsass src error handling cpp phpblog public node modules node sass src libsass src parser cpp phpblog public node modules node sass src libsass src debugger hpp phpblog public node modules node sass src libsass src emitter cpp phpblog public node modules node sass src sass types number cpp phpblog public node modules node sass src sass types color h phpblog public node modules node sass src libsass src sass values cpp phpblog public node modules node sass src libsass src ast hpp phpblog public node modules node sass src libsass src output cpp phpblog public node modules node sass src libsass src check nesting cpp phpblog public node modules node sass src sass types null cpp phpblog public node modules node sass src libsass src ast def macros hpp phpblog public node modules node sass src libsass src functions cpp phpblog public node modules node sass src libsass src cssize hpp phpblog public node modules node sass src libsass src prelexer cpp phpblog public node modules node sass src libsass src ast cpp phpblog public node modules node sass src libsass src to c cpp phpblog public node modules node sass src libsass src to value hpp phpblog public node modules node sass src libsass src ast fwd decl hpp phpblog public node modules node sass src libsass src inspect hpp phpblog public node modules node sass src sass types color cpp phpblog public node modules node sass src libsass src values cpp phpblog public node modules node sass src sass context wrapper cpp phpblog public node modules node sass src sass types list h phpblog public node modules node sass src libsass src check nesting hpp phpblog public node modules node sass src sass types map cpp phpblog public node modules node sass src libsass src to value cpp phpblog public node modules node sass src libsass src context cpp phpblog public node modules node sass src binding cpp phpblog public node modules node sass src sass types string cpp phpblog public node modules node sass src libsass src sass context cpp phpblog public node modules node sass src libsass src prelexer hpp phpblog public node modules node sass src libsass src context hpp phpblog public node modules node sass src sass types boolean h phpblog public node modules node sass src libsass src eval cpp vulnerability details disputed in inspect cpp in libsass a high memory footprint caused by an endless loop containing a sass inspect operator sass string quoted stack frame may cause a denial of service via crafted sass input files with stray or characters note upstream comments indicate this issue is closed as won t fix and works as intended by design publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type change files origin a href release date fix resolution replace or update the following file libsass step up your open source security game with whitesource | 0 |
7,818 | 2,935,158,230 | IssuesEvent | 2015-06-30 13:13:09 | fireball-x/firedoc | https://api.github.com/repos/fireball-x/firedoc | closed | rebuild test suite and remove YUI library | test v1.8 | The YUITest and YUI sucks in some ways, take away from them | 1.0 | rebuild test suite and remove YUI library - The YUITest and YUI sucks in some ways, take away from them | test | rebuild test suite and remove yui library the yuitest and yui sucks in some ways take away from them | 1 |
259,814 | 22,553,123,817 | IssuesEvent | 2022-06-27 07:53:18 | mozilla-mobile/fenix | https://api.github.com/repos/mozilla-mobile/fenix | closed | Intermittent UI test failure - < TopSitesTest.verifyRenameTopSite> | b:crash eng:intermittent-test needs:triage eng:ui-test | ### Firebase Test Run: [Firebase link](https://console.firebase.google.com/u/0/project/moz-fenix/testlab/histories/bh.66b7091e15d53d45/matrices/7551035073070944308/executions/bs.26808ce0c47fdca4/testcases/1/test-cases)
### Stacktrace:
java.lang.RuntimeException: Error while connecting UiAutomation@83d4027[id=-1, flags=0]
at android.app.UiAutomation.connect(UiAutomation.java:259)
at android.app.Instrumentation.getUiAutomation(Instrumentation.java:2176)
at androidx.test.uiautomator.UiDevice.getUiAutomation(UiDevice.java:1129)
at androidx.test.uiautomator.QueryController.<init>(QueryController.java:95)
at androidx.test.uiautomator.UiDevice.<init>(UiDevice.java:109)
at androidx.test.uiautomator.UiDevice.getInstance(UiDevice.java:261)
at org.mozilla.fenix.ui.TopSitesTest.<init>(TopSitesTest.kt:35)
### Build: 5/30 Main
### Notes: Similar with: #25416 #25414 #25342 #25341
┆Issue is synchronized with this [Jira Task](https://mozilla-hub.atlassian.net/browse/FNXV2-20601)
| 2.0 | Intermittent UI test failure - < TopSitesTest.verifyRenameTopSite> - ### Firebase Test Run: [Firebase link](https://console.firebase.google.com/u/0/project/moz-fenix/testlab/histories/bh.66b7091e15d53d45/matrices/7551035073070944308/executions/bs.26808ce0c47fdca4/testcases/1/test-cases)
### Stacktrace:
java.lang.RuntimeException: Error while connecting UiAutomation@83d4027[id=-1, flags=0]
at android.app.UiAutomation.connect(UiAutomation.java:259)
at android.app.Instrumentation.getUiAutomation(Instrumentation.java:2176)
at androidx.test.uiautomator.UiDevice.getUiAutomation(UiDevice.java:1129)
at androidx.test.uiautomator.QueryController.<init>(QueryController.java:95)
at androidx.test.uiautomator.UiDevice.<init>(UiDevice.java:109)
at androidx.test.uiautomator.UiDevice.getInstance(UiDevice.java:261)
at org.mozilla.fenix.ui.TopSitesTest.<init>(TopSitesTest.kt:35)
### Build: 5/30 Main
### Notes: Similar with: #25416 #25414 #25342 #25341
┆Issue is synchronized with this [Jira Task](https://mozilla-hub.atlassian.net/browse/FNXV2-20601)
| test | intermittent ui test failure firebase test run stacktrace java lang runtimeexception error while connecting uiautomation at android app uiautomation connect uiautomation java at android app instrumentation getuiautomation instrumentation java at androidx test uiautomator uidevice getuiautomation uidevice java at androidx test uiautomator querycontroller querycontroller java at androidx test uiautomator uidevice uidevice java at androidx test uiautomator uidevice getinstance uidevice java at org mozilla fenix ui topsitestest topsitestest kt build main notes similar with ┆issue is synchronized with this | 1 |
351,906 | 32,034,907,043 | IssuesEvent | 2023-09-22 14:42:04 | elastic/kibana | https://api.github.com/repos/elastic/kibana | opened | Failing test: Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/observability_log_explorer/dataset_selector·ts - Observability Log Explorer Dataset Selector with installed integrations and uncategorized data streams when open on the integrations tab should display a list of installed integrations | failed-test | A test failed on a tracked branch
```
Error: expected 2 to equal 3
at Assertion.assert (expect.js:100:11)
at Assertion.apply (expect.js:227:8)
at Assertion.be (expect.js:69:22)
at Context.<anonymous> (dataset_selector.ts:194:42)
at processTicksAndRejections (node:internal/process/task_queues:95:5)
at Object.apply (wrap_function.js:73:16)
```
First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/35735#018abd22-c283-475e-afbc-d907fc9bfa8d)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/observability_log_explorer/dataset_selector·ts","test.name":"Observability Log Explorer Dataset Selector with installed integrations and uncategorized data streams when open on the integrations tab should display a list of installed integrations","test.failCount":1}} --> | 1.0 | Failing test: Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/observability_log_explorer/dataset_selector·ts - Observability Log Explorer Dataset Selector with installed integrations and uncategorized data streams when open on the integrations tab should display a list of installed integrations - A test failed on a tracked branch
```
Error: expected 2 to equal 3
at Assertion.assert (expect.js:100:11)
at Assertion.apply (expect.js:227:8)
at Assertion.be (expect.js:69:22)
at Context.<anonymous> (dataset_selector.ts:194:42)
at processTicksAndRejections (node:internal/process/task_queues:95:5)
at Object.apply (wrap_function.js:73:16)
```
First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/35735#018abd22-c283-475e-afbc-d907fc9bfa8d)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/observability_log_explorer/dataset_selector·ts","test.name":"Observability Log Explorer Dataset Selector with installed integrations and uncategorized data streams when open on the integrations tab should display a list of installed integrations","test.failCount":1}} --> | test | failing test chrome x pack ui functional tests x pack test functional apps observability log explorer dataset selector·ts observability log explorer dataset selector with installed integrations and uncategorized data streams when open on the integrations tab should display a list of installed integrations a test failed on a tracked branch error expected to equal at assertion assert expect js at assertion apply expect js at assertion be expect js at context dataset selector ts at processticksandrejections node internal process task queues at object apply wrap function js first failure | 1 |
249,787 | 21,191,729,675 | IssuesEvent | 2022-04-08 18:14:41 | Azure/azure-sdk-for-js | https://api.github.com/repos/Azure/azure-sdk-for-js | closed | [Recorder] Encoding ~ character in recordings | test-utils-recorder | @joheredi reached out with a problem where the playback was failing upon rerecording a certain test.
The reason being the generated recording was not replaced as per the replaceable variables provided.
The reason being the secret had ~ character, and encodeRFC3986 method doesn't encode the character.
RFC 3986 standard does not encode the character, but the recordings have that encoded.
To do - add a new method encodeSpecialCharacters in the recorder to encode ~ character to allow the replacements to happen smoothly.
| 1.0 | [Recorder] Encoding ~ character in recordings - @joheredi reached out with a problem where the playback was failing upon rerecording a certain test.
The reason being the generated recording was not replaced as per the replaceable variables provided.
The reason being the secret had ~ character, and encodeRFC3986 method doesn't encode the character.
RFC 3986 standard does not encode the character, but the recordings have that encoded.
To do - add a new method encodeSpecialCharacters in the recorder to encode ~ character to allow the replacements to happen smoothly.
| test | encoding character in recordings joheredi reached out with a problem where the playback was failing upon rerecording a certain test the reason being the generated recording was not replaced as per the replaceable variables provided the reason being the secret had character and method doesn t encode the character rfc standard does not encode the character but the recordings have that encoded to do add a new method encodespecialcharacters in the recorder to encode character to allow the replacements to happen smoothly | 1 |
26,704 | 4,777,611,048 | IssuesEvent | 2016-10-27 16:46:25 | wheeler-microfluidics/microdrop | https://api.github.com/repos/wheeler-microfluidics/microdrop | closed | Enabling video_recorder_plugin in Windows causes exception (Trac #31) | defect microdrop Migrated from Trac | Enabling video_recorder_plugin in Windows causes the following exception (regardless of whether or not the webcam is plugged in):
[Errno 9] Bad file descriptor
Migrated from http://microfluidics.utoronto.ca/ticket/31
```json
{
"status": "closed",
"changetime": "2014-04-17T19:39:01",
"description": "Enabling video_recorder_plugin in Windows causes the following exception (regardless of whether or not the webcam is plugged in):\n\n [Errno 9] Bad file descriptor",
"reporter": "cfobel",
"cc": "",
"resolution": "fixed",
"_ts": "1397763541728826",
"component": "microdrop",
"summary": "Enabling video_recorder_plugin in Windows causes exception",
"priority": "major",
"keywords": "",
"version": "0.1",
"time": "2012-01-04T23:24:05",
"milestone": "Microdrop 1.0",
"owner": "cfobel",
"type": "defect"
}
```
| 1.0 | Enabling video_recorder_plugin in Windows causes exception (Trac #31) - Enabling video_recorder_plugin in Windows causes the following exception (regardless of whether or not the webcam is plugged in):
[Errno 9] Bad file descriptor
Migrated from http://microfluidics.utoronto.ca/ticket/31
```json
{
"status": "closed",
"changetime": "2014-04-17T19:39:01",
"description": "Enabling video_recorder_plugin in Windows causes the following exception (regardless of whether or not the webcam is plugged in):\n\n [Errno 9] Bad file descriptor",
"reporter": "cfobel",
"cc": "",
"resolution": "fixed",
"_ts": "1397763541728826",
"component": "microdrop",
"summary": "Enabling video_recorder_plugin in Windows causes exception",
"priority": "major",
"keywords": "",
"version": "0.1",
"time": "2012-01-04T23:24:05",
"milestone": "Microdrop 1.0",
"owner": "cfobel",
"type": "defect"
}
```
| non_test | enabling video recorder plugin in windows causes exception trac enabling video recorder plugin in windows causes the following exception regardless of whether or not the webcam is plugged in bad file descriptor migrated from json status closed changetime description enabling video recorder plugin in windows causes the following exception regardless of whether or not the webcam is plugged in n n bad file descriptor reporter cfobel cc resolution fixed ts component microdrop summary enabling video recorder plugin in windows causes exception priority major keywords version time milestone microdrop owner cfobel type defect | 0 |
59,586 | 14,618,323,164 | IssuesEvent | 2020-12-22 16:04:02 | tensorflow/tensorflow | https://api.github.com/repos/tensorflow/tensorflow | closed | Failed to load native tensorflow runtime | TF 2.2 stalled stat:awaiting response subtype:windows type:build/install | <em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution : Windows Server 2012 R2 Standard
- TensorFlow installed from : Binary (pip)
- TensorFlow version: 2.2.0
- Python version: 3.7.2
- Installed using : pip
- CUDA/cuDNN version: NA
- GPU model and memory: NA
Tensorflow is installed without any issues. But when running my program I am getting the below error.
> D:\Tetherfi\TRS\TFaceAuthServer>python tfacecompare.py
> Traceback (most recent call last):
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\site-packag
> es\tensorflow\python\pywrap_tensorflow.py", line 58, in <module>
> from tensorflow.python.pywrap_tensorflow_internal import *
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\site-packag
> es\tensorflow\python\pywrap_tensorflow_internal.py", line 28, in <module>
> _pywrap_tensorflow_internal = swig_import_helper()
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\site-packag
> es\tensorflow\python\pywrap_tensorflow_internal.py", line 24, in swig_import_hel
> per
> _mod = imp.load_module('_pywrap_tensorflow_internal', fp, pathname, descript
> ion)
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\imp.py", li
> ne 242, in load_module
> return load_dynamic(name, filename, file)
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\imp.py", li
> ne 342, in load_dynamic
> return _load(spec)
> ImportError: DLL load failed: The specified module could not be found.
>
> During handling of the above exception, another exception occurred:
>
> Traceback (most recent call last):
> File "tfacecompare.py", line 20, in <module>
> import tensorflow.keras
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\site-packag
> es\tensorflow\__init__.py", line 41, in <module>
> from tensorflow.python.tools import module_util as _module_util
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\site-packag
> es\tensorflow\python\__init__.py", line 50, in <module>
> from tensorflow.python import pywrap_tensorflow
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\site-packag
> es\tensorflow\python\pywrap_tensorflow.py", line 69, in <module>
> raise ImportError(msg)
> ImportError: Traceback (most recent call last):
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\site-packag
> es\tensorflow\python\pywrap_tensorflow.py", line 58, in <module>
> from tensorflow.python.pywrap_tensorflow_internal import *
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\site-packag
> es\tensorflow\python\pywrap_tensorflow_internal.py", line 28, in <module>
> _pywrap_tensorflow_internal = swig_import_helper()
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\site-packag
> es\tensorflow\python\pywrap_tensorflow_internal.py", line 24, in swig_import_hel
> per
> _mod = imp.load_module('_pywrap_tensorflow_internal', fp, pathname, descript
> ion)
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\imp.py", li
> ne 242, in load_module
> return load_dynamic(name, filename, file)
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\imp.py", li
> ne 342, in load_dynamic
> return _load(spec)
> ImportError: DLL load failed: The specified module could not be found.
>
>
> Failed to load the native TensorFlow runtime.
>
> See https://www.tensorflow.org/install/errors
>
> for some common reasons and solutions. Include the entire stack trace
> above this error message when asking for help.
I have tried installing visual c++ redistributable for 2015 (64 and 32 bit), but still the same. Reinstalling of python also didn't help.
Please guide and help to solve this.
| 1.0 | Failed to load native tensorflow runtime - <em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution : Windows Server 2012 R2 Standard
- TensorFlow installed from : Binary (pip)
- TensorFlow version: 2.2.0
- Python version: 3.7.2
- Installed using : pip
- CUDA/cuDNN version: NA
- GPU model and memory: NA
Tensorflow is installed without any issues. But when running my program I am getting the below error.
> D:\Tetherfi\TRS\TFaceAuthServer>python tfacecompare.py
> Traceback (most recent call last):
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\site-packag
> es\tensorflow\python\pywrap_tensorflow.py", line 58, in <module>
> from tensorflow.python.pywrap_tensorflow_internal import *
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\site-packag
> es\tensorflow\python\pywrap_tensorflow_internal.py", line 28, in <module>
> _pywrap_tensorflow_internal = swig_import_helper()
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\site-packag
> es\tensorflow\python\pywrap_tensorflow_internal.py", line 24, in swig_import_hel
> per
> _mod = imp.load_module('_pywrap_tensorflow_internal', fp, pathname, descript
> ion)
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\imp.py", li
> ne 242, in load_module
> return load_dynamic(name, filename, file)
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\imp.py", li
> ne 342, in load_dynamic
> return _load(spec)
> ImportError: DLL load failed: The specified module could not be found.
>
> During handling of the above exception, another exception occurred:
>
> Traceback (most recent call last):
> File "tfacecompare.py", line 20, in <module>
> import tensorflow.keras
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\site-packag
> es\tensorflow\__init__.py", line 41, in <module>
> from tensorflow.python.tools import module_util as _module_util
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\site-packag
> es\tensorflow\python\__init__.py", line 50, in <module>
> from tensorflow.python import pywrap_tensorflow
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\site-packag
> es\tensorflow\python\pywrap_tensorflow.py", line 69, in <module>
> raise ImportError(msg)
> ImportError: Traceback (most recent call last):
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\site-packag
> es\tensorflow\python\pywrap_tensorflow.py", line 58, in <module>
> from tensorflow.python.pywrap_tensorflow_internal import *
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\site-packag
> es\tensorflow\python\pywrap_tensorflow_internal.py", line 28, in <module>
> _pywrap_tensorflow_internal = swig_import_helper()
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\site-packag
> es\tensorflow\python\pywrap_tensorflow_internal.py", line 24, in swig_import_hel
> per
> _mod = imp.load_module('_pywrap_tensorflow_internal', fp, pathname, descript
> ion)
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\imp.py", li
> ne 242, in load_module
> return load_dynamic(name, filename, file)
> File "C:\Users\kiolocal\AppData\Local\Programs\Python\Python37\lib\imp.py", li
> ne 342, in load_dynamic
> return _load(spec)
> ImportError: DLL load failed: The specified module could not be found.
>
>
> Failed to load the native TensorFlow runtime.
>
> See https://www.tensorflow.org/install/errors
>
> for some common reasons and solutions. Include the entire stack trace
> above this error message when asking for help.
I have tried installing visual c++ redistributable for 2015 (64 and 32 bit), but still the same. Reinstalling of python also didn't help.
Please guide and help to solve this.
| non_test | failed to load native tensorflow runtime please make sure that this is a build installation issue as per our we only address code doc bugs performance issues feature requests and build installation issues on github tag build template system information os platform and distribution windows server standard tensorflow installed from binary pip tensorflow version python version installed using pip cuda cudnn version na gpu model and memory na tensorflow is installed without any issues but when running my program i am getting the below error d tetherfi trs tfaceauthserver python tfacecompare py traceback most recent call last file c users kiolocal appdata local programs python lib site packag es tensorflow python pywrap tensorflow py line in from tensorflow python pywrap tensorflow internal import file c users kiolocal appdata local programs python lib site packag es tensorflow python pywrap tensorflow internal py line in pywrap tensorflow internal swig import helper file c users kiolocal appdata local programs python lib site packag es tensorflow python pywrap tensorflow internal py line in swig import hel per mod imp load module pywrap tensorflow internal fp pathname descript ion file c users kiolocal appdata local programs python lib imp py li ne in load module return load dynamic name filename file file c users kiolocal appdata local programs python lib imp py li ne in load dynamic return load spec importerror dll load failed the specified module could not be found during handling of the above exception another exception occurred traceback most recent call last file tfacecompare py line in import tensorflow keras file c users kiolocal appdata local programs python lib site packag es tensorflow init py line in from tensorflow python tools import module util as module util file c users kiolocal appdata local programs python lib site packag es tensorflow python init py line in from tensorflow python import pywrap tensorflow file c users kiolocal appdata local programs python lib site packag es tensorflow python pywrap tensorflow py line in raise importerror msg importerror traceback most recent call last file c users kiolocal appdata local programs python lib site packag es tensorflow python pywrap tensorflow py line in from tensorflow python pywrap tensorflow internal import file c users kiolocal appdata local programs python lib site packag es tensorflow python pywrap tensorflow internal py line in pywrap tensorflow internal swig import helper file c users kiolocal appdata local programs python lib site packag es tensorflow python pywrap tensorflow internal py line in swig import hel per mod imp load module pywrap tensorflow internal fp pathname descript ion file c users kiolocal appdata local programs python lib imp py li ne in load module return load dynamic name filename file file c users kiolocal appdata local programs python lib imp py li ne in load dynamic return load spec importerror dll load failed the specified module could not be found failed to load the native tensorflow runtime see for some common reasons and solutions include the entire stack trace above this error message when asking for help i have tried installing visual c redistributable for and bit but still the same reinstalling of python also didn t help please guide and help to solve this | 0 |
129,445 | 10,575,008,059 | IssuesEvent | 2019-10-07 14:59:07 | kcigeospatial/Fred_Co_Land-Management | https://api.github.com/repos/kcigeospatial/Fred_Co_Land-Management | closed | Appliance Master License | Ready for Test Env. Retest | please make the section complete on the applicant master details page a drop down - yes.. like all the others

| 2.0 | Appliance Master License - please make the section complete on the applicant master details page a drop down - yes.. like all the others

| test | appliance master license please make the section complete on the applicant master details page a drop down yes like all the others | 1 |
257,561 | 22,194,747,015 | IssuesEvent | 2022-06-07 05:24:40 | windingtree/lpms-server | https://api.github.com/repos/windingtree/lpms-server | opened | Test coverage for repository | A-tests D-chore P-normal C-database | **Problem**
As we build on top of the repository layer that we wrote, it is critical that this layer performs as expected.
**Solution**
Unit tests to cover basic functionality, but particularly, we require:
- [ ] Testing all getters to make sure that they return a consistent value (suggest `type | null`).
- [ ] Testing of all indices to make sure that they are being managed correctly (add a stub, adds it to the indices., remove a stub, removes from the indices).
| 1.0 | Test coverage for repository - **Problem**
As we build on top of the repository layer that we wrote, it is critical that this layer performs as expected.
**Solution**
Unit tests to cover basic functionality, but particularly, we require:
- [ ] Testing all getters to make sure that they return a consistent value (suggest `type | null`).
- [ ] Testing of all indices to make sure that they are being managed correctly (add a stub, adds it to the indices., remove a stub, removes from the indices).
| test | test coverage for repository problem as we build on top of the repository layer that we wrote it is critical that this layer performs as expected solution unit tests to cover basic functionality but particularly we require testing all getters to make sure that they return a consistent value suggest type null testing of all indices to make sure that they are being managed correctly add a stub adds it to the indices remove a stub removes from the indices | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.