Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 1k | labels stringlengths 4 1.38k | body stringlengths 1 262k | index stringclasses 16
values | text_combine stringlengths 96 262k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
638,483 | 20,728,989,136 | IssuesEvent | 2022-03-14 07:22:06 | akayangel218/Project-Bite | https://api.github.com/repos/akayangel218/Project-Bite | closed | Publish App (5) | Priority: Low Epic: Logistics | * Clean up code
* [x] Remove unnecessary `console.log()`s
* [x] Remove unused imports
* [x] Remove commented-out code
* [x] Address compile-time warnings
* Host backend app on google app engine (or similar)
* Get new url to access backend app
* Configure frontend app to send requests to that url
* Create a production version of frontend app
* Host frontend prod version on netlify (or similar) | 1.0 | Publish App (5) - * Clean up code
* [x] Remove unnecessary `console.log()`s
* [x] Remove unused imports
* [x] Remove commented-out code
* [x] Address compile-time warnings
* Host backend app on google app engine (or similar)
* Get new url to access backend app
* Configure frontend app to send requests to that url
* Create a production version of frontend app
* Host frontend prod version on netlify (or similar) | priority | publish app clean up code remove unnecessary console log s remove unused imports remove commented out code address compile time warnings host backend app on google app engine or similar get new url to access backend app configure frontend app to send requests to that url create a production version of frontend app host frontend prod version on netlify or similar | 1 |
304,293 | 9,330,261,784 | IssuesEvent | 2019-03-28 06:14:30 | pingcap/tidb | https://api.github.com/repos/pingcap/tidb | closed | `NODE_ID` is mistaken as a full keyword | component/parser priority/release-blocker type/bug | ## Bug Report
Please answer these questions before submitting your issue. Thanks!
1. What did you do?
Execute the DDL
```sql
CREATE TABLE a (node_id INT);
```
2. What did you expect to see?
The table successfully created.
3. What did you see instead?
Syntax error:
```
ERROR 1064 (42000): You have an error in your SQL syntax; check the manual that corresponds to your TiDB version for the right syntax to use line 1 column 22 near "node_id int)"
```
4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)?
```
Release Version: v3.0.0-beta-242-g0b28f30cf
Git Commit Hash: 0b28f30cfdef7a396c7d5474d5536787e2d00bd0
Git Branch: master
UTC Build Time: 2019-03-19 08:25:08
GoVersion: go version go1.12 darwin/amd64
Race Enabled: false
TiKV Min Version: 2.1.0-alpha.1-ff3dd160846b7d1aed9079c389fc188f7f5ea13e
Check Table Before Drop: false
``` | 1.0 | `NODE_ID` is mistaken as a full keyword - ## Bug Report
Please answer these questions before submitting your issue. Thanks!
1. What did you do?
Execute the DDL
```sql
CREATE TABLE a (node_id INT);
```
2. What did you expect to see?
The table successfully created.
3. What did you see instead?
Syntax error:
```
ERROR 1064 (42000): You have an error in your SQL syntax; check the manual that corresponds to your TiDB version for the right syntax to use line 1 column 22 near "node_id int)"
```
4. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)?
```
Release Version: v3.0.0-beta-242-g0b28f30cf
Git Commit Hash: 0b28f30cfdef7a396c7d5474d5536787e2d00bd0
Git Branch: master
UTC Build Time: 2019-03-19 08:25:08
GoVersion: go version go1.12 darwin/amd64
Race Enabled: false
TiKV Min Version: 2.1.0-alpha.1-ff3dd160846b7d1aed9079c389fc188f7f5ea13e
Check Table Before Drop: false
``` | priority | node id is mistaken as a full keyword bug report please answer these questions before submitting your issue thanks what did you do execute the ddl sql create table a node id int what did you expect to see the table successfully created what did you see instead syntax error error you have an error in your sql syntax check the manual that corresponds to your tidb version for the right syntax to use line column near node id int what version of tidb are you using tidb server v or run select tidb version on tidb release version beta git commit hash git branch master utc build time goversion go version darwin race enabled false tikv min version alpha check table before drop false | 1 |
272,218 | 8,506,236,114 | IssuesEvent | 2018-10-30 16:06:19 | mozilla/addons-frontend | https://api.github.com/repos/mozilla/addons-frontend | closed | Always render AddonsByAuthorsCard | component: user profile priority: p3 state: pull request ready | ### Describe the problem and steps to reproduce it:
1. go to https://addons-dev.allizom.org/en-US/firefox/user/abine/
2. refresh
3. observe
### What happened?
The left card is rendered but the right side is empty. Then the right side starts to render some stuff, then it is fully loaded with the add-ons by the user.
Current state in -dev:

### What did you expect to happen?
The full layout should be rendered in a "loading" state.
Expected state locally (hence the no-css render at the beginning):

### Anything else we should know?
<!-- Please include a link to the page, screenshots and any relevant files. -->
| 1.0 | Always render AddonsByAuthorsCard - ### Describe the problem and steps to reproduce it:
1. go to https://addons-dev.allizom.org/en-US/firefox/user/abine/
2. refresh
3. observe
### What happened?
The left card is rendered but the right side is empty. Then the right side starts to render some stuff, then it is fully loaded with the add-ons by the user.
Current state in -dev:

### What did you expect to happen?
The full layout should be rendered in a "loading" state.
Expected state locally (hence the no-css render at the beginning):

### Anything else we should know?
<!-- Please include a link to the page, screenshots and any relevant files. -->
| priority | always render addonsbyauthorscard describe the problem and steps to reproduce it go to refresh observe what happened the left card is rendered but the right side is empty then the right side starts to render some stuff then it is fully loaded with the add ons by the user current state in dev what did you expect to happen the full layout should be rendered in a loading state expected state locally hence the no css render at the beginning anything else we should know | 1 |
172,918 | 21,075,847,973 | IssuesEvent | 2022-04-02 05:50:29 | mheob/mheob-ui | https://api.github.com/repos/mheob/mheob-ui | opened | CVE-2021-44907 (High) detected in qs-6.5.2.tgz | security vulnerability | ## CVE-2021-44907 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-6.5.2.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.5.2.tgz">https://registry.npmjs.org/qs/-/qs-6.5.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- lerna-4.0.0.tgz (Root Library)
- version-4.0.0.tgz
- run-lifecycle-4.0.0.tgz
- npm-lifecycle-3.1.5.tgz
- node-gyp-5.1.1.tgz
- request-2.88.2.tgz
- :x: **qs-6.5.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mheob/mheob-ui/commit/c05a502457c1f3d1cdb2894d3a9cabf1bd2f793e">c05a502457c1f3d1cdb2894d3a9cabf1bd2f793e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Denial of Service vulnerability exists in qs up to 6.8.0 due to insufficient sanitization of property in the gs.parse function. The merge() function allows the assignment of properties on an array in the query. For any property being assigned, a value in the array is converted to an object containing these properties. Essentially, this means that the property whose expected type is Array always has to be checked with Array.isArray() by the user. This may not be obvious to the user and can cause unexpected behavior.
<p>Publish Date: 2022-03-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-44907>CVE-2021-44907</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-44907">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-44907</a></p>
<p>Release Date: 2022-03-17</p>
<p>Fix Resolution: qs - 6.8.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-44907 (High) detected in qs-6.5.2.tgz - ## CVE-2021-44907 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-6.5.2.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.5.2.tgz">https://registry.npmjs.org/qs/-/qs-6.5.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- lerna-4.0.0.tgz (Root Library)
- version-4.0.0.tgz
- run-lifecycle-4.0.0.tgz
- npm-lifecycle-3.1.5.tgz
- node-gyp-5.1.1.tgz
- request-2.88.2.tgz
- :x: **qs-6.5.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mheob/mheob-ui/commit/c05a502457c1f3d1cdb2894d3a9cabf1bd2f793e">c05a502457c1f3d1cdb2894d3a9cabf1bd2f793e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Denial of Service vulnerability exists in qs up to 6.8.0 due to insufficient sanitization of property in the gs.parse function. The merge() function allows the assignment of properties on an array in the query. For any property being assigned, a value in the array is converted to an object containing these properties. Essentially, this means that the property whose expected type is Array always has to be checked with Array.isArray() by the user. This may not be obvious to the user and can cause unexpected behavior.
<p>Publish Date: 2022-03-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-44907>CVE-2021-44907</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-44907">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-44907</a></p>
<p>Release Date: 2022-03-17</p>
<p>Fix Resolution: qs - 6.8.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in qs tgz cve high severity vulnerability vulnerable library qs tgz a querystring parser that supports nesting and arrays with a depth limit library home page a href path to dependency file package json path to vulnerable library node modules qs package json dependency hierarchy lerna tgz root library version tgz run lifecycle tgz npm lifecycle tgz node gyp tgz request tgz x qs tgz vulnerable library found in head commit a href found in base branch main vulnerability details a denial of service vulnerability exists in qs up to due to insufficient sanitization of property in the gs parse function the merge function allows the assignment of properties on an array in the query for any property being assigned a value in the array is converted to an object containing these properties essentially this means that the property whose expected type is array always has to be checked with array isarray by the user this may not be obvious to the user and can cause unexpected behavior publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution qs step up your open source security game with whitesource | 0 |
185,381 | 14,351,617,263 | IssuesEvent | 2020-11-30 01:51:08 | github-vet/rangeclosure-findings | https://api.github.com/repos/github-vet/rangeclosure-findings | opened | weliang1/upstream: Godeps/_workspace/src/k8s.io/kubernetes/pkg/kubectl/resource/builder_test.go; 3 LoC | fresh test tiny |
Found a possible issue in [weliang1/upstream](https://www.github.com/weliang1/upstream) at [Godeps/_workspace/src/k8s.io/kubernetes/pkg/kubectl/resource/builder_test.go](https://github.com/weliang1/upstream/blob/1532bfcc115ccbd74a084568bbc55d131f6e9bf2/Godeps/_workspace/src/k8s.io/kubernetes/pkg/kubectl/resource/builder_test.go#L55-L57)
The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements
which capture loop variables.
[Click here to see the code in its original context.](https://github.com/weliang1/upstream/blob/1532bfcc115ccbd74a084568bbc55d131f6e9bf2/Godeps/_workspace/src/k8s.io/kubernetes/pkg/kubectl/resource/builder_test.go#L55-L57)
<details>
<summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary>
```go
for _, e := range events {
enc.Encode(&e)
}
```
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to e at line 56 may start a goroutine
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 1532bfcc115ccbd74a084568bbc55d131f6e9bf2
| 1.0 | weliang1/upstream: Godeps/_workspace/src/k8s.io/kubernetes/pkg/kubectl/resource/builder_test.go; 3 LoC -
Found a possible issue in [weliang1/upstream](https://www.github.com/weliang1/upstream) at [Godeps/_workspace/src/k8s.io/kubernetes/pkg/kubectl/resource/builder_test.go](https://github.com/weliang1/upstream/blob/1532bfcc115ccbd74a084568bbc55d131f6e9bf2/Godeps/_workspace/src/k8s.io/kubernetes/pkg/kubectl/resource/builder_test.go#L55-L57)
The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements
which capture loop variables.
[Click here to see the code in its original context.](https://github.com/weliang1/upstream/blob/1532bfcc115ccbd74a084568bbc55d131f6e9bf2/Godeps/_workspace/src/k8s.io/kubernetes/pkg/kubectl/resource/builder_test.go#L55-L57)
<details>
<summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary>
```go
for _, e := range events {
enc.Encode(&e)
}
```
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to e at line 56 may start a goroutine
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 1532bfcc115ccbd74a084568bbc55d131f6e9bf2
| non_priority | upstream godeps workspace src io kubernetes pkg kubectl resource builder test go loc found a possible issue in at the below snippet of go code triggered static analysis which searches for goroutines and or defer statements which capture loop variables click here to show the line s of go which triggered the analyzer go for e range events enc encode e below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to e at line may start a goroutine leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id | 0 |
425,094 | 12,335,592,570 | IssuesEvent | 2020-05-14 12:15:36 | SforAiDl/Playground | https://api.github.com/repos/SforAiDl/Playground | closed | Resolving FPS issue | help wanted priority:High | Currently, there's some redundant/unnecessary code because of which FPS of video detector is very poor. Making optimizations to improve FPS is necessary. | 1.0 | Resolving FPS issue - Currently, there's some redundant/unnecessary code because of which FPS of video detector is very poor. Making optimizations to improve FPS is necessary. | priority | resolving fps issue currently there s some redundant unnecessary code because of which fps of video detector is very poor making optimizations to improve fps is necessary | 1 |
737,231 | 25,507,125,718 | IssuesEvent | 2022-11-28 10:16:58 | bounswe/bounswe2022group7 | https://api.github.com/repos/bounswe/bounswe2022group7 | opened | Discussion Post Upvote Downvote Transform as In Comment | Status: In Progress Priority: High Difficulty: Medium Type: Implementation Target: Backend | Now, for discussion post, we still just have counts for upvote downvote. We should convert it to lists of upvoters downvoters. | 1.0 | Discussion Post Upvote Downvote Transform as In Comment - Now, for discussion post, we still just have counts for upvote downvote. We should convert it to lists of upvoters downvoters. | priority | discussion post upvote downvote transform as in comment now for discussion post we still just have counts for upvote downvote we should convert it to lists of upvoters downvoters | 1 |
758,060 | 26,542,266,117 | IssuesEvent | 2023-01-19 20:18:34 | craftercms/craftercms | https://api.github.com/repos/craftercms/craftercms | closed | [kubernetes-deployments] Update kubernetes deployment files | priority: high task validate | Please update
- the cluster files for primary/replica clustering
- the ssh config changes for both simple and cluster deployment files (Related to https://github.com/craftercms/craftercms/issues/5257)
| 1.0 | [kubernetes-deployments] Update kubernetes deployment files - Please update
- the cluster files for primary/replica clustering
- the ssh config changes for both simple and cluster deployment files (Related to https://github.com/craftercms/craftercms/issues/5257)
| priority | update kubernetes deployment files please update the cluster files for primary replica clustering the ssh config changes for both simple and cluster deployment files related to | 1 |
287,771 | 8,820,280,630 | IssuesEvent | 2019-01-01 10:20:04 | TwinePlatform/twine-api | https://api.github.com/repos/TwinePlatform/twine-api | closed | Rewrite database scripts in typescript | 4️⃣ priority-4 | > the database scripts are in JS and are not transpiled/moved into the build folder by the build script, they cannot refer to the typescript-version of the config module, and must instead point to the built JS version of it (in the build folder). This is fine as long as the build command has been run before hand.
#58 | 1.0 | Rewrite database scripts in typescript - > the database scripts are in JS and are not transpiled/moved into the build folder by the build script, they cannot refer to the typescript-version of the config module, and must instead point to the built JS version of it (in the build folder). This is fine as long as the build command has been run before hand.
#58 | priority | rewrite database scripts in typescript the database scripts are in js and are not transpiled moved into the build folder by the build script they cannot refer to the typescript version of the config module and must instead point to the built js version of it in the build folder this is fine as long as the build command has been run before hand | 1 |
817,219 | 30,631,495,677 | IssuesEvent | 2023-07-24 14:48:51 | sebastien-d-me/ToDo-Co | https://api.github.com/repos/sebastien-d-me/ToDo-Co | opened | Audit report | Priority: Medium Status: Not started | #### Description:
Creation of code quality audit report and performance report.
------------
###### Estimated time: 2 day(s)
###### Difficulty: ⭐
| 1.0 | Audit report - #### Description:
Creation of code quality audit report and performance report.
------------
###### Estimated time: 2 day(s)
###### Difficulty: ⭐
| priority | audit report description creation of code quality audit report and performance report estimated time day s difficulty ⭐ | 1 |
80,720 | 7,755,013,223 | IssuesEvent | 2018-05-31 08:50:11 | zeroc-ice/ice | https://api.github.com/repos/zeroc-ice/ice | opened | sporadic Ice/ami test failure with Android emulator | android java testsuite | This occurs when running the android tests on macOS.
```
[ running client/server test - 05/31/18 08:37:14 ]
- Config: ssl,ipv6,mx,androidemulator
testing begin/end invocation... ok
testing local exceptions... ok
testing sent callback... ok
testing sent async callback... ok
testing unexpected exceptions... ok
testing batch requests with proxy... ok
testing batch requests with connection... ok
testing batch requests with communicator... ok
testing future operations... ok
testing graceful close connection with wait... null: unknown exception
java.util.concurrent.CompletionException: com.zeroc.Ice.ConnectionLostException
error = 0
at java.util.concurrent.CompletableFuture.reportJoin(CompletableFuture.java:348)
at java.util.concurrent.CompletableFuture.join(CompletableFuture.java:1931)
at test.Ice.ami.AllTests.allTests(AllTests.java:995)
at test.Ice.ami.Client.run(Client.java:17)
at test.Util.Application.main(Application.java:100)
at test.Util.Application.main(Application.java:42)
at com.zeroc.testcontroller.ControllerApp$MainHelperI.run(ControllerApp.java:366)
Caused by: com.zeroc.Ice.ConnectionLostException
error = 0
at com.zeroc.IceInternal.StreamSocket.read(StreamSocket.java:193)
at com.zeroc.IceInternal.StreamSocket.read(StreamSocket.java:154)
at com.zeroc.IceInternal.TcpTransceiver.read(TcpTransceiver.java:63)
at com.zeroc.IceSSL.TransceiverI.read(TransceiverI.java:225)
at com.zeroc.Ice.ConnectionI.read(ConnectionI.java:3000)
at com.zeroc.Ice.ConnectionI.message(ConnectionI.java:995)
at com.zeroc.IceInternal.ThreadPool.run(ThreadPool.java:417)
at com.zeroc.IceInternal.ThreadPool.access$500(ThreadPool.java:12)
at com.zeroc.IceInternal.ThreadPool$EventHandlerThread.run(ThreadPool.java:786)
at java.lang.Thread.run(Thread.java:761)
unexpected exit status: expected: 0, got 1
``` | 1.0 | sporadic Ice/ami test failure with Android emulator - This occurs when running the android tests on macOS.
```
[ running client/server test - 05/31/18 08:37:14 ]
- Config: ssl,ipv6,mx,androidemulator
testing begin/end invocation... ok
testing local exceptions... ok
testing sent callback... ok
testing sent async callback... ok
testing unexpected exceptions... ok
testing batch requests with proxy... ok
testing batch requests with connection... ok
testing batch requests with communicator... ok
testing future operations... ok
testing graceful close connection with wait... null: unknown exception
java.util.concurrent.CompletionException: com.zeroc.Ice.ConnectionLostException
error = 0
at java.util.concurrent.CompletableFuture.reportJoin(CompletableFuture.java:348)
at java.util.concurrent.CompletableFuture.join(CompletableFuture.java:1931)
at test.Ice.ami.AllTests.allTests(AllTests.java:995)
at test.Ice.ami.Client.run(Client.java:17)
at test.Util.Application.main(Application.java:100)
at test.Util.Application.main(Application.java:42)
at com.zeroc.testcontroller.ControllerApp$MainHelperI.run(ControllerApp.java:366)
Caused by: com.zeroc.Ice.ConnectionLostException
error = 0
at com.zeroc.IceInternal.StreamSocket.read(StreamSocket.java:193)
at com.zeroc.IceInternal.StreamSocket.read(StreamSocket.java:154)
at com.zeroc.IceInternal.TcpTransceiver.read(TcpTransceiver.java:63)
at com.zeroc.IceSSL.TransceiverI.read(TransceiverI.java:225)
at com.zeroc.Ice.ConnectionI.read(ConnectionI.java:3000)
at com.zeroc.Ice.ConnectionI.message(ConnectionI.java:995)
at com.zeroc.IceInternal.ThreadPool.run(ThreadPool.java:417)
at com.zeroc.IceInternal.ThreadPool.access$500(ThreadPool.java:12)
at com.zeroc.IceInternal.ThreadPool$EventHandlerThread.run(ThreadPool.java:786)
at java.lang.Thread.run(Thread.java:761)
unexpected exit status: expected: 0, got 1
``` | non_priority | sporadic ice ami test failure with android emulator this occurs when running the android tests on macos config ssl mx androidemulator testing begin end invocation ok testing local exceptions ok testing sent callback ok testing sent async callback ok testing unexpected exceptions ok testing batch requests with proxy ok testing batch requests with connection ok testing batch requests with communicator ok testing future operations ok testing graceful close connection with wait null unknown exception java util concurrent completionexception com zeroc ice connectionlostexception error at java util concurrent completablefuture reportjoin completablefuture java at java util concurrent completablefuture join completablefuture java at test ice ami alltests alltests alltests java at test ice ami client run client java at test util application main application java at test util application main application java at com zeroc testcontroller controllerapp mainhelperi run controllerapp java caused by com zeroc ice connectionlostexception error at com zeroc iceinternal streamsocket read streamsocket java at com zeroc iceinternal streamsocket read streamsocket java at com zeroc iceinternal tcptransceiver read tcptransceiver java at com zeroc icessl transceiveri read transceiveri java at com zeroc ice connectioni read connectioni java at com zeroc ice connectioni message connectioni java at com zeroc iceinternal threadpool run threadpool java at com zeroc iceinternal threadpool access threadpool java at com zeroc iceinternal threadpool eventhandlerthread run threadpool java at java lang thread run thread java unexpected exit status expected got | 0 |
136,296 | 5,279,573,606 | IssuesEvent | 2017-02-07 11:42:14 | Activisme-be/BK-zeverij | https://api.github.com/repos/Activisme-be/BK-zeverij | closed | Add method that disable the admin for github hook. | bug High Priority | Now the admin can publish a ticket multiple times to github. So we need a disable method. That the admin only can once publish a quetion. This also prevent a overflow on the issue tracker. | 1.0 | Add method that disable the admin for github hook. - Now the admin can publish a ticket multiple times to github. So we need a disable method. That the admin only can once publish a quetion. This also prevent a overflow on the issue tracker. | priority | add method that disable the admin for github hook now the admin can publish a ticket multiple times to github so we need a disable method that the admin only can once publish a quetion this also prevent a overflow on the issue tracker | 1 |
173,186 | 6,521,364,173 | IssuesEvent | 2017-08-28 20:16:53 | Aubron/scoreshots-templates | https://api.github.com/repos/Aubron/scoreshots-templates | closed | Conference Standings Template | Priority: Medium Status: Needs Finalization / Preview Image | <!--- Provide client & short description of template (preview, final, stat, etc) in Title -->
<!--- If thread is long, consider including template thumbnail here for ease of identification. -->
## Requested By:
<!--- PARTNER, CLIENT NAME, DESIRED DELIVERY DATE -->
<!--- This tells us which organization requested the template, and who to contact. -->
<!--- If requested internally, specify this. -->
Internal, by Rick Perko. August 4th delivery date.
## Sharing & Intended Use:
<!--- Specify PUBLIC or PRIVATE, along with SPORT and CATEGORY, if you know it. -->
<!--- Always ask client if they're okay with others using the template, to help with billing/priorities. -->
<!--- Also list any additional tags or categories that might apply to this template. -->
Public, presumably Media Relations.
## Template Description:
<!--- Provide details on what the template ought to look like and do. -->
<!--- Include direct correspondence with clients in quotes IFF necessary. -->
Description from Rick:
> It would be dynamic and include wins and losses for both conference play and overall. It will also need bifurcation into two divisions (e.g., east and west). Would be very cool if there was a way to modify by simply bumping a team up or down.
## Dynamic/Programming Considerations:
<!--- List any additional or unusual features that will require development attention. -->
<!--- If there are none, list N/A. -->
- Dropdown, number of games : 5 , 6, 7 and 8 (not necessarily symmetrical)
## Example Images / Additional Resources:
<!--- Attach any example images or napkin drawings submitted by client. -->
<!--- Specify the source of each image, so we know what we are allowed to use during referencing. -->
Here's a screenshot of a website Rick linked as an example for this task.

| 1.0 | Conference Standings Template - <!--- Provide client & short description of template (preview, final, stat, etc) in Title -->
<!--- If thread is long, consider including template thumbnail here for ease of identification. -->
## Requested By:
<!--- PARTNER, CLIENT NAME, DESIRED DELIVERY DATE -->
<!--- This tells us which organization requested the template, and who to contact. -->
<!--- If requested internally, specify this. -->
Internal, by Rick Perko. August 4th delivery date.
## Sharing & Intended Use:
<!--- Specify PUBLIC or PRIVATE, along with SPORT and CATEGORY, if you know it. -->
<!--- Always ask client if they're okay with others using the template, to help with billing/priorities. -->
<!--- Also list any additional tags or categories that might apply to this template. -->
Public, presumably Media Relations.
## Template Description:
<!--- Provide details on what the template ought to look like and do. -->
<!--- Include direct correspondence with clients in quotes IFF necessary. -->
Description from Rick:
> It would be dynamic and include wins and losses for both conference play and overall. It will also need bifurcation into two divisions (e.g., east and west). Would be very cool if there was a way to modify by simply bumping a team up or down.
## Dynamic/Programming Considerations:
<!--- List any additional or unusual features that will require development attention. -->
<!--- If there are none, list N/A. -->
- Dropdown, number of games : 5 , 6, 7 and 8 (not necessarily symmetrical)
## Example Images / Additional Resources:
<!--- Attach any example images or napkin drawings submitted by client. -->
<!--- Specify the source of each image, so we know what we are allowed to use during referencing. -->
Here's a screenshot of a website Rick linked as an example for this task.

| priority | conference standings template requested by internal by rick perko august delivery date sharing intended use public presumably media relations template description description from rick it would be dynamic and include wins and losses for both conference play and overall it will also need bifurcation into two divisions e g east and west would be very cool if there was a way to modify by simply bumping a team up or down dynamic programming considerations dropdown number of games and not necessarily symmetrical example images additional resources here s a screenshot of a website rick linked as an example for this task | 1 |
433,532 | 12,506,386,785 | IssuesEvent | 2020-06-02 12:29:59 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.corriere.it - site is not usable | browser-chrome priority-normal type-no-css | <!-- @browser: Chrome 83.0.4103 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; ) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/83.0.4103.61 Safari/537.36 -->
<!-- @reported_with: -->
**URL**: https://www.corriere.it/
**Browser / Version**: Chrome 83.0.4103
**Operating System**: Windows 10
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
the site does not load correctly
<details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/5/de1252fb-f1ef-45d7-ab62-f99a43d5dafd.jpeg'></details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.corriere.it - site is not usable - <!-- @browser: Chrome 83.0.4103 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; ) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/83.0.4103.61 Safari/537.36 -->
<!-- @reported_with: -->
**URL**: https://www.corriere.it/
**Browser / Version**: Chrome 83.0.4103
**Operating System**: Windows 10
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
the site does not load correctly
<details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/5/de1252fb-f1ef-45d7-ab62-f99a43d5dafd.jpeg'></details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | site is not usable url browser version chrome operating system windows tested another browser no problem type site is not usable description page not loading correctly steps to reproduce the site does not load correctly view the screenshot img alt screenshot src browser configuration none from with ❤️ | 1 |
500,189 | 14,492,274,386 | IssuesEvent | 2020-12-11 06:38:13 | OneSignal/react-native-onesignal | https://api.github.com/repos/OneSignal/react-native-onesignal | closed | App Crash while setting external user | Android Bug: High Priority Need More Information | <!--
1. IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION WE MAY CLOSE YOUR ISSUE WITHOUT INVESTIGATION
1. SEARCH EXISTING ISSUES FOR AN ANSWER: https://goo.gl/pspQNY
1. See our Common Issues documentation: https://goo.gl/BDcfZZ
1. See our contributing guidelines: https://goo.gl/h19DnX
-->
**Description:**
onesignal.setexternaluserid got 2 arguments, expected 3
**Environment**
```
"react": "16.13.1",
"react-native": "0.63.4",
"react-native-onesignal": "^3.9.3",
```
**Code**
```
OneSignal.setLogLevel(6, 0);
OneSignal.init(config.ONESIGNAL_APP_ID,
{
kOSSettingsKeyAutoPrompt: false,
kOSSettingsKeyInAppLaunchURL: false,
kOSSettingsKeyInFocusDisplayOption: 2
});
OneSignal.inFocusDisplaying(2);
OneSignal.promptForPushNotificationsWithUserResponse(myiOSPromptCallback);
let User = {
"email":"client@example.com",
"id":4,
"name":"Bilal Hussain Shah",
"status":true,
} // This is User.
OneSignal.setExternalUserId(User.id.toString(), (results) => {
console.log('Results of setting external user id');
console.log(results);
});
```

| 1.0 | App Crash while setting external user - <!--
1. IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION WE MAY CLOSE YOUR ISSUE WITHOUT INVESTIGATION
1. SEARCH EXISTING ISSUES FOR AN ANSWER: https://goo.gl/pspQNY
1. See our Common Issues documentation: https://goo.gl/BDcfZZ
1. See our contributing guidelines: https://goo.gl/h19DnX
-->
**Description:**
onesignal.setexternaluserid got 2 arguments, expected 3
**Environment**
```
"react": "16.13.1",
"react-native": "0.63.4",
"react-native-onesignal": "^3.9.3",
```
**Code**
```
OneSignal.setLogLevel(6, 0);
OneSignal.init(config.ONESIGNAL_APP_ID,
{
kOSSettingsKeyAutoPrompt: false,
kOSSettingsKeyInAppLaunchURL: false,
kOSSettingsKeyInFocusDisplayOption: 2
});
OneSignal.inFocusDisplaying(2);
OneSignal.promptForPushNotificationsWithUserResponse(myiOSPromptCallback);
let User = {
"email":"client@example.com",
"id":4,
"name":"Bilal Hussain Shah",
"status":true,
} // This is User.
OneSignal.setExternalUserId(User.id.toString(), (results) => {
console.log('Results of setting external user id');
console.log(results);
});
```

| priority | app crash while setting external user if you don t fill out the following information we may close your issue without investigation search existing issues for an answer see our common issues documentation see our contributing guidelines description onesignal setexternaluserid got arguments expected environment react react native react native onesignal code onesignal setloglevel onesignal init config onesignal app id kossettingskeyautoprompt false kossettingskeyinapplaunchurl false kossettingskeyinfocusdisplayoption onesignal infocusdisplaying onesignal promptforpushnotificationswithuserresponse myiospromptcallback let user email client example com id name bilal hussain shah status true this is user onesignal setexternaluserid user id tostring results console log results of setting external user id console log results | 1 |
18,749 | 11,047,156,129 | IssuesEvent | 2019-12-09 18:21:17 | cityofaustin/atd-knack-signs-markings | https://api.github.com/repos/cityofaustin/atd-knack-signs-markings | opened | Feature: Change GIS QA to after Final Review | Need: 1-Must Have Project: SMB 311 Module Service: Apps Type: Feature Workgroup: SMB | **Page:**
- [ ] Markings Service Request page
- [ ] Service Request Details page
- [x] Work Order Page
- [ ] Work Order/Job Details page
- [ ] Job Details page
### I want to be able to close a Markings WO without having the GIS QA status prevent me from closing the WO.
Removing the conditional rules on the `work order status` field of the `work order markings` object table.
| 1.0 | Feature: Change GIS QA to after Final Review - **Page:**
- [ ] Markings Service Request page
- [ ] Service Request Details page
- [x] Work Order Page
- [ ] Work Order/Job Details page
- [ ] Job Details page
### I want to be able to close a Markings WO without having the GIS QA status prevent me from closing the WO.
Removing the conditional rules on the `work order status` field of the `work order markings` object table.
| non_priority | feature change gis qa to after final review page markings service request page service request details page work order page work order job details page job details page i want to be able to close a markings wo without having the gis qa status prevent me from closing the wo removing the conditional rules on the work order status field of the work order markings object table | 0 |
681,464 | 23,312,032,268 | IssuesEvent | 2022-08-08 09:07:31 | openmsupply/mobile | https://api.github.com/repos/openmsupply/mobile | closed | kiribati Vax: Editing vaccination event, clicking save buttons multiple times quickly is crashing apps and which is corrupting the record | Priority: high Effort: small Bug: development Solution: Agreed | ## Describe the bug
While editing vaccination events and saving application crashes and while reopening the apps and seeing the history for the same patients it shows blank.
### To reproduce
Steps to reproduce the behaviour:
1. Goto vaccination.
2. Select any patient and click on history.
3. Select the event that you are allowed to edit otherwise it will show the alert which is nice 👍
4. Now after editing try to click save multiple times which will crash the application.
5. Now reopen the application and select same patient and check same history.
6. See error
### Expected behaviour
A clear and concise description of what you expected to happen.
### Agreed Solution
- This happens when you click save multiple times. So we need to disable save button after 1st click and not enable it until we get a response back from the server or database.
### Version and device info
- App version: 8.4.2.2022727
- Tablet model:Samsung
- OS version:
### Additional context
Add any other context about the problem here.
| 1.0 | kiribati Vax: Editing vaccination event, clicking save buttons multiple times quickly is crashing apps and which is corrupting the record - ## Describe the bug
While editing vaccination events and saving application crashes and while reopening the apps and seeing the history for the same patients it shows blank.
### To reproduce
Steps to reproduce the behaviour:
1. Goto vaccination.
2. Select any patient and click on history.
3. Select the event that you are allowed to edit otherwise it will show the alert which is nice 👍
4. Now after editing try to click save multiple times which will crash the application.
5. Now reopen the application and select same patient and check same history.
6. See error
### Expected behaviour
A clear and concise description of what you expected to happen.
### Agreed Solution
- This happens when you click save multiple times. So we need to disable save button after 1st click and not enable it until we get a response back from the server or database.
### Version and device info
- App version: 8.4.2.2022727
- Tablet model:Samsung
- OS version:
### Additional context
Add any other context about the problem here.
| priority | kiribati vax editing vaccination event clicking save buttons multiple times quickly is crashing apps and which is corrupting the record describe the bug while editing vaccination events and saving application crashes and while reopening the apps and seeing the history for the same patients it shows blank to reproduce steps to reproduce the behaviour goto vaccination select any patient and click on history select the event that you are allowed to edit otherwise it will show the alert which is nice 👍 now after editing try to click save multiple times which will crash the application now reopen the application and select same patient and check same history see error expected behaviour a clear and concise description of what you expected to happen agreed solution this happens when you click save multiple times so we need to disable save button after click and not enable it until we get a response back from the server or database version and device info app version tablet model samsung os version additional context add any other context about the problem here | 1 |
246,114 | 18,836,086,420 | IssuesEvent | 2021-11-11 01:10:26 | csells/go_router | https://api.github.com/repos/csells/go_router | closed | Update the docs with a link to this pair programming video | documentation | video here: https://twitter.com/mhadaily/status/1455580280924876802
MD link format for "hosted" YT video:
```text
[](https://www.youtube.com/watch?v=YOUTUBE_VIDEO_ID_HERE)
``` | 1.0 | Update the docs with a link to this pair programming video - video here: https://twitter.com/mhadaily/status/1455580280924876802
MD link format for "hosted" YT video:
```text
[](https://www.youtube.com/watch?v=YOUTUBE_VIDEO_ID_HERE)
``` | non_priority | update the docs with a link to this pair programming video video here md link format for hosted yt video text | 0 |
147,216 | 13,203,368,139 | IssuesEvent | 2020-08-14 14:03:16 | oracle/weblogic-kubernetes-operator | https://api.github.com/repos/oracle/weblogic-kubernetes-operator | closed | Connection problem with WLST script. | customer reported documentation polish v1.0-must | Hello,
I'm not able to connect to the Admin using a wlst script. On the kubernetes side it is correctly configured :
```kubectl
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
domain1-admin-server NodePort 10.21.64.58 <none> 7001:30701/TCP 3h
domain1-admin-server-extchannel-t3channel NodePort 10.21.172.82 <none> 30012:30012/TCP 3h
```
I tried from my pc and I have this :
```python
wls:/offline> connect('weblogic','welcome1','t3://PUBLIC_IP:30012')
Connecting to t3://PUBLIC_IP:30012 with userid weblogic ...
Traceback (innermost last):
File "<console>", line 1, in ?
File "<iostream>", line 19, in connect
File "<iostream>", line 553, in raiseWLSTException
WLSTException: Error occurred while performing connect : Error getting the initial context. There is no server running at t3://PUBLIC_IP:30012 : Timed out while attempting to establish connection to :t3://PUBLIC_IP:30012
Use dumpStack() to view the full stacktrace :
```
or
```python
wls:/offline> connect('weblogic','welcome1','http://PUBLIC_IP:30701')
Connecting to http://PUBLIC_IP:30701 with userid weblogic ...
Traceback (innermost last):
File "<console>", line 1, in ?
File "<iostream>", line 19, in connect
File "<iostream>", line 553, in raiseWLSTException
WLSTException: Error occurred while performing connect : Error getting the initial context. There is no server running at http://PUBLIC_IP:30701 : http://PUBLIC_IP:30701: [RJVM:000575]Destination PUBLIC_IP, 30701 unreachable.; nested exception is:
java.io.IOException: Could not connect to http://PUBLIC_IP:30701; [RJVM:000576]No available router to destination.; nested exception is:
java.rmi.ConnectException: [RJVM:000576]No available router to destination.
Use dumpStack() to view the full stacktrace :
```
Tcpdump on my pc and on the OCI servers seems fine (there are packets exchange and also the telnet is working). The curl on the http port works as well:
```curl --user weblogic:welcome1 -H X-Requested-By:MyClient -H Accept:application/json -X GET http://130.61.63.87:30701/management/weblogic```
So I tried also from the server in the kubernetes cluster and i faced the same issue.
The only way that worked was to go to the server running the Admin Container and try to connect to the pod IP :
```python
wls:/offline> connect('weblogic','welcome1','t3://10.99.66.5:7001')
Connecting to t3://10.99.66.5:7001 with userid weblogic ...
Successfully connected to Admin Server "admin-server" that belongs to domain "base_domain".
Warning: An insecure protocol was used to connect to the server.
To ensure on-the-wire security, the SSL port or Admin port should be used instead.
wls:/base_domain/serverConfig/>
```
Any cloue on what is wrong ? | 1.0 | Connection problem with WLST script. - Hello,
I'm not able to connect to the Admin using a wlst script. On the kubernetes side it is correctly configured :
```kubectl
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
domain1-admin-server NodePort 10.21.64.58 <none> 7001:30701/TCP 3h
domain1-admin-server-extchannel-t3channel NodePort 10.21.172.82 <none> 30012:30012/TCP 3h
```
I tried from my pc and I have this :
```python
wls:/offline> connect('weblogic','welcome1','t3://PUBLIC_IP:30012')
Connecting to t3://PUBLIC_IP:30012 with userid weblogic ...
Traceback (innermost last):
File "<console>", line 1, in ?
File "<iostream>", line 19, in connect
File "<iostream>", line 553, in raiseWLSTException
WLSTException: Error occurred while performing connect : Error getting the initial context. There is no server running at t3://PUBLIC_IP:30012 : Timed out while attempting to establish connection to :t3://PUBLIC_IP:30012
Use dumpStack() to view the full stacktrace :
```
or
```python
wls:/offline> connect('weblogic','welcome1','http://PUBLIC_IP:30701')
Connecting to http://PUBLIC_IP:30701 with userid weblogic ...
Traceback (innermost last):
File "<console>", line 1, in ?
File "<iostream>", line 19, in connect
File "<iostream>", line 553, in raiseWLSTException
WLSTException: Error occurred while performing connect : Error getting the initial context. There is no server running at http://PUBLIC_IP:30701 : http://PUBLIC_IP:30701: [RJVM:000575]Destination PUBLIC_IP, 30701 unreachable.; nested exception is:
java.io.IOException: Could not connect to http://PUBLIC_IP:30701; [RJVM:000576]No available router to destination.; nested exception is:
java.rmi.ConnectException: [RJVM:000576]No available router to destination.
Use dumpStack() to view the full stacktrace :
```
Tcpdump on my pc and on the OCI servers seems fine (there are packets exchange and also the telnet is working). The curl on the http port works as well:
```curl --user weblogic:welcome1 -H X-Requested-By:MyClient -H Accept:application/json -X GET http://130.61.63.87:30701/management/weblogic```
So I tried also from the server in the kubernetes cluster and i faced the same issue.
The only way that worked was to go to the server running the Admin Container and try to connect to the pod IP :
```python
wls:/offline> connect('weblogic','welcome1','t3://10.99.66.5:7001')
Connecting to t3://10.99.66.5:7001 with userid weblogic ...
Successfully connected to Admin Server "admin-server" that belongs to domain "base_domain".
Warning: An insecure protocol was used to connect to the server.
To ensure on-the-wire security, the SSL port or Admin port should be used instead.
wls:/base_domain/serverConfig/>
```
Any cloue on what is wrong ? | non_priority | connection problem with wlst script hello i m not able to connect to the admin using a wlst script on the kubernetes side it is correctly configured kubectl name type cluster ip external ip port s age admin server nodeport tcp admin server extchannel nodeport tcp i tried from my pc and i have this python wls offline connect weblogic public ip connecting to public ip with userid weblogic traceback innermost last file line in file line in connect file line in raisewlstexception wlstexception error occurred while performing connect error getting the initial context there is no server running at public ip timed out while attempting to establish connection to public ip use dumpstack to view the full stacktrace or python wls offline connect weblogic connecting to with userid weblogic traceback innermost last file line in file line in connect file line in raisewlstexception wlstexception error occurred while performing connect error getting the initial context there is no server running at destination public ip unreachable nested exception is java io ioexception could not connect to no available router to destination nested exception is java rmi connectexception no available router to destination use dumpstack to view the full stacktrace tcpdump on my pc and on the oci servers seems fine there are packets exchange and also the telnet is working the curl on the http port works as well curl user weblogic h x requested by myclient h accept application json x get so i tried also from the server in the kubernetes cluster and i faced the same issue the only way that worked was to go to the server running the admin container and try to connect to the pod ip python wls offline connect weblogic connecting to with userid weblogic successfully connected to admin server admin server that belongs to domain base domain warning an insecure protocol was used to connect to the server to ensure on the wire security the ssl port or admin port should be used instead wls base domain serverconfig any cloue on what is wrong | 0 |
149,923 | 19,596,281,909 | IssuesEvent | 2022-01-05 18:15:13 | lukebroganws/bag-of-holding | https://api.github.com/repos/lukebroganws/bag-of-holding | opened | CVE-2018-20676 (Medium) detected in bootstrap-3.3.2.min.js | security vulnerability | ## CVE-2018-20676 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.2.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.2/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.2/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /bower_components/select2-bootstrap-css/docs/3.4.4.html</p>
<p>Path to vulnerable library: /bower_components/select2-bootstrap-css/docs/js/bootstrap.min.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.2.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/lukebroganws/bag-of-holding/commit/fd67f5ea4245a25511650d4bf41a6c699a799144">fd67f5ea4245a25511650d4bf41a6c699a799144</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676>CVE-2018-20676</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: bootstrap - 3.4.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"twitter-bootstrap","packageVersion":"3.3.2","packageFilePaths":["/bower_components/select2-bootstrap-css/docs/3.4.4.html"],"isTransitiveDependency":false,"dependencyTree":"twitter-bootstrap:3.3.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bootstrap - 3.4.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-20676","vulnerabilityDetails":"In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | True | CVE-2018-20676 (Medium) detected in bootstrap-3.3.2.min.js - ## CVE-2018-20676 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.2.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.2/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.2/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /bower_components/select2-bootstrap-css/docs/3.4.4.html</p>
<p>Path to vulnerable library: /bower_components/select2-bootstrap-css/docs/js/bootstrap.min.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.2.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/lukebroganws/bag-of-holding/commit/fd67f5ea4245a25511650d4bf41a6c699a799144">fd67f5ea4245a25511650d4bf41a6c699a799144</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676>CVE-2018-20676</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: bootstrap - 3.4.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"twitter-bootstrap","packageVersion":"3.3.2","packageFilePaths":["/bower_components/select2-bootstrap-css/docs/3.4.4.html"],"isTransitiveDependency":false,"dependencyTree":"twitter-bootstrap:3.3.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"bootstrap - 3.4.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-20676","vulnerabilityDetails":"In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | non_priority | cve medium detected in bootstrap min js cve medium severity vulnerability vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file bower components bootstrap css docs html path to vulnerable library bower components bootstrap css docs js bootstrap min js dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before xss is possible in the tooltip data viewport attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree twitter bootstrap isminimumfixversionavailable true minimumfixversion bootstrap isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails in bootstrap before xss is possible in the tooltip data viewport attribute vulnerabilityurl | 0 |
781,540 | 27,441,337,216 | IssuesEvent | 2023-03-02 11:13:09 | GoogleCloudPlatform/cloud-sql-jdbc-socket-factory | https://api.github.com/repos/GoogleCloudPlatform/cloud-sql-jdbc-socket-factory | closed | working with mysql socket and iam auth at the same time | type: question priority: p1 | ### Bug Description
hi i'm using mysql-socket-factory-connector-j-8 to connect a cloud run container to Mysql cloud sql instance.
a lot of options are available to do it, but the one i need to make it work is :
- cloud run has Cloud SQL connections enabled (this creates some kind of a cloud sql proxy side car container that shares a socket in /cloudsql/INSTANCE_ID that can be used from your cloud run container [link to documentation](https://cloud.google.com/sql/docs/mysql/connect-run?_ga=2.40665698.-1100797918.1668437632&_gac=1.185697115.1676974444.CjwKCAiA9NGfBhBvEiwAq5vSy0PaHBu7zPG4FX_efXZ0Yalgkj8mzPI9FJhMBWFrNXN4SxX4a1J2nxoC03kQAvD_BwE&hl=fr)
- using enableIamAuth=true to connect using Service account
each one of those set ups works alone but together i got this error : `SSL connection required for plugin "mysql_clear_password". Check if 'sslMode' is enabled.`
### Example code (or command)
```Java
- working jdbc string with cloud run sql connection feature eanbled and using socket:
jdbc:mysql:///<databaseName>?cloudSqlInstance=<project>:<region>:<dbname>&socketFactory=com.google.cloud.sql.mysql.SocketFactory&=true&ipTypes=PRIVATE&unixSocketPath=/cloudsql/<project>:<region>:<dbname>
- also working jdbc string without socket but with iamAuth enabled :
jdbc:mysql:///<databaseName>?cloudSqlInstance=<project>:<region>:<dbname>&socketFactory=com.google.cloud.sql.mysql.SocketFactory&ipTypes=PRIVATE&enableIamAuth=true
- not working jdbc string when both enableIamAuth and unixSocketPath are used :
jdbc:mysql:///<databaseName>?cloudSqlInstance=<project>:<region>:<dbname>&socketFactory=com.google.cloud.sql.mysql.SocketFactory&&enableIamAuth=true&ipTypes=PRIVATE&unixSocketPath=/cloudsql/<project>:<region>:<dbname>
```
### Stacktrace
```bash
2023-03-01 17:18:13 2023-03-01 17:18:13,993 INFO [com.goo.clo.sql.cor.CoreSocketFactory] (agroal-11) Connecting to Cloud SQL instance [mystic-aileron-190912:europe-west9:test-shi] via unix socket at /cloudsql/mystic-aileron-190912:europe-west9:test-shi.
2023-03-01 17:18:14 2023-03-01 17:18:14,005 WARN [io.agr.pool] (agroal-11) Datasource '<default>': SSL connection required for plugin "mysql_clear_password". Check if 'sslMode' is enabled.
2023-03-01 17:18:14 2023-03-01 17:18:14,005 WARN [org.hib.eng.jdb.spi.SqlExceptionHelper] (executor-thread-0) SQL Error: 0, SQLState: S1000
2023-03-01 17:18:14 2023-03-01 17:18:14,005 ERROR [org.hib.eng.jdb.spi.SqlExceptionHelper] (executor-thread-0) SSL connection required for plugin "mysql_clear_password". Check if 'sslMode' is enabled.
2023-03-01 17:18:14 2023-03-01 17:18:14,006 ERROR [io.qua.ver.htt.run.QuarkusErrorHandler] (executor-thread-0) HTTP Request to /v1/users failed, error id: fcf176b4-bab3-4a53-8410-bf0c85fc1e49-2: javax.persistence.PersistenceException: org.hibernate.exception.GenericJDBCException: Unable to acquire JDBC Connection
2023-03-01 17:18:14 at org.hibernate.internal.ExceptionConverterImpl.convert(ExceptionConverterImpl.java:154)
2023-03-01 17:18:14 at org.hibernate.query.internal.AbstractProducedQuery.list(AbstractProducedQuery.java:1626)
2023-03-01 17:18:14 at org.hibernate.query.Query.getResultList(Query.java:165)
2023-03-01 17:18:14 at io.quarkus.hibernate.orm.panache.common.runtime.CommonPanacheQueryImpl.list(CommonPanacheQueryImpl.java:269)
2023-03-01 17:18:14 at io.quarkus.hibernate.orm.panache.runtime.PanacheQueryImpl.list(PanacheQueryImpl.java:149)
2023-03-01 17:18:14 at io.quarkus.hibernate.orm.panache.runtime.JpaOperations.list(JpaOperations.java:24)
2023-03-01 17:18:14 at io.quarkus.hibernate.orm.panache.runtime.JpaOperations.list(JpaOperations.java:10)
2023-03-01 17:18:14 at io.quarkus.hibernate.orm.panache.common.runtime.AbstractJpaOperations.listAll(AbstractJpaOperations.java:305)
2023-03-01 17:18:14 at org.acme.repository.UserRepository.listAll(UserRepository.java)
2023-03-01 17:18:14 at org.acme.repository.UserRepository_ClientProxy.listAll(Unknown Source)
2023-03-01 17:18:14 at org.acme.service.impl.DefaultUserService.getAllUsers(DefaultUserService.java:30)
2023-03-01 17:18:14 at org.acme.service.impl.DefaultUserService_ClientProxy.getAllUsers(Unknown Source)
2023-03-01 17:18:14 at org.acme.controller.UserController.getUsers(UserController.java:40)
2023-03-01 17:18:14 at org.acme.controller.UserController_ClientProxy.getUsers(Unknown Source)
2023-03-01 17:18:14 at org.acme.controller.UserController$quarkusrestinvoker$getUsers_4b9221263c4fb1e6e88448a563d3364eb951551a.invoke(Unknown Source)
2023-03-01 17:18:14 at org.jboss.resteasy.reactive.server.handlers.InvocationHandler.handle(InvocationHandler.java:29)
2023-03-01 17:18:14 at io.quarkus.resteasy.reactive.server.runtime.QuarkusResteasyReactiveRequestContext.invokeHandler(QuarkusResteasyReactiveRequestContext.java:114)
2023-03-01 17:18:14 at org.jboss.resteasy.reactive.common.core.AbstractResteasyReactiveContext.run(AbstractResteasyReactiveContext.java:145)
2023-03-01 17:18:14 at io.quarkus.vertx.core.runtime.VertxCoreRecorder$14.runWith(VertxCoreRecorder.java:576)
2023-03-01 17:18:14 at org.jboss.threads.EnhancedQueueExecutor$Task.run(EnhancedQueueExecutor.java:2449)
2023-03-01 17:18:14 at org.jboss.threads.EnhancedQueueExecutor$ThreadBody.run(EnhancedQueueExecutor.java:1478)
2023-03-01 17:18:14 at org.jboss.threads.DelegatingRunnable.run(DelegatingRunnable.java:29)
2023-03-01 17:18:14 at org.jboss.threads.ThreadLocalResettingRunnable.run(ThreadLocalResettingRunnable.java:29)
2023-03-01 17:18:14 at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
2023-03-01 17:18:14 at java.base/java.lang.Thread.run(Thread.java:829)
2023-03-01 17:18:14 Caused by: org.hibernate.exception.GenericJDBCException: Unable to acquire JDBC Connection
2023-03-01 17:18:14 at org.hibernate.exception.internal.StandardSQLExceptionConverter.convert(StandardSQLExceptionConverter.java:42)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.spi.SqlExceptionHelper.convert(SqlExceptionHelper.java:113)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.spi.SqlExceptionHelper.convert(SqlExceptionHelper.java:99)
2023-03-01 17:18:14 at org.hibernate.resource.jdbc.internal.LogicalConnectionManagedImpl.acquireConnectionIfNeeded(LogicalConnectionManagedImpl.java:111)
2023-03-01 17:18:14 at org.hibernate.resource.jdbc.internal.LogicalConnectionManagedImpl.getPhysicalConnection(LogicalConnectionManagedImpl.java:138)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.internal.StatementPreparerImpl.connection(StatementPreparerImpl.java:50)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.internal.StatementPreparerImpl$5.doPrepare(StatementPreparerImpl.java:149)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.internal.StatementPreparerImpl$StatementPreparationTemplate.prepareStatement(StatementPreparerImpl.java:176)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.internal.StatementPreparerImpl.prepareQueryStatement(StatementPreparerImpl.java:151)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.prepareQueryStatement(Loader.java:2122)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.executeQueryStatement(Loader.java:2059)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.executeQueryStatement(Loader.java:2037)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.doQuery(Loader.java:956)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.doQueryAndInitializeNonLazyCollections(Loader.java:357)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.doList(Loader.java:2868)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.doList(Loader.java:2850)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.listIgnoreQueryCache(Loader.java:2682)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.list(Loader.java:2677)
2023-03-01 17:18:14 at org.hibernate.loader.hql.QueryLoader.list(QueryLoader.java:540)
2023-03-01 17:18:14 at org.hibernate.hql.internal.ast.QueryTranslatorImpl.list(QueryTranslatorImpl.java:400)
2023-03-01 17:18:14 at org.hibernate.engine.query.spi.HQLQueryPlan.performList(HQLQueryPlan.java:218)
2023-03-01 17:18:14 at org.hibernate.internal.SessionImpl.list(SessionImpl.java:1459)
2023-03-01 17:18:14 at org.hibernate.query.internal.AbstractProducedQuery.doList(AbstractProducedQuery.java:1649)
2023-03-01 17:18:14 at org.hibernate.query.internal.AbstractProducedQuery.list(AbstractProducedQuery.java:1617)
2023-03-01 17:18:14 ... 23 more
2023-03-01 17:18:14 Caused by: java.sql.SQLException: SSL connection required for plugin "mysql_clear_password". Check if 'sslMode' is enabled.
2023-03-01 17:18:14 at com.mysql.cj.jdbc.exceptions.SQLError.createSQLException(SQLError.java:129)
2023-03-01 17:18:14 at com.mysql.cj.jdbc.exceptions.SQLExceptionsMapping.translateException(SQLExceptionsMapping.java:122)
2023-03-01 17:18:14 at com.mysql.cj.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:828)
2023-03-01 17:18:14 at com.mysql.cj.jdbc.ConnectionImpl.<init>(ConnectionImpl.java:448)
2023-03-01 17:18:14 at com.mysql.cj.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:241)
2023-03-01 17:18:14 at com.mysql.cj.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:198)
2023-03-01 17:18:14 at io.agroal.pool.ConnectionFactory.createConnection(ConnectionFactory.java:226)
2023-03-01 17:18:14 at io.agroal.pool.ConnectionPool$CreateConnectionTask.call(ConnectionPool.java:535)
2023-03-01 17:18:14 at io.agroal.pool.ConnectionPool$CreateConnectionTask.call(ConnectionPool.java:516)
2023-03-01 17:18:14 at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
2023-03-01 17:18:14 at io.agroal.pool.util.PriorityScheduledExecutor.beforeExecute(PriorityScheduledExecutor.java:75)
2023-03-01 17:18:14 at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1126)
2023-03-01 17:18:14 at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
2023-03-01 17:18:14 ... 1 more
2023-03-01 17:18:14 2023-03-01 17:18:14,007 ERROR [org.jbo.res.rea.com.cor.AbstractResteasyReactiveContext] (executor-thread-0) Request failed: javax.persistence.PersistenceException: org.hibernate.exception.GenericJDBCException: Unable to acquire JDBC Connection
2023-03-01 17:18:14 at org.hibernate.internal.ExceptionConverterImpl.convert(ExceptionConverterImpl.java:154)
2023-03-01 17:18:14 at org.hibernate.query.internal.AbstractProducedQuery.list(AbstractProducedQuery.java:1626)
2023-03-01 17:18:14 at org.hibernate.query.Query.getResultList(Query.java:165)
2023-03-01 17:18:14 at io.quarkus.hibernate.orm.panache.common.runtime.CommonPanacheQueryImpl.list(CommonPanacheQueryImpl.java:269)
2023-03-01 17:18:14 at io.quarkus.hibernate.orm.panache.runtime.PanacheQueryImpl.list(PanacheQueryImpl.java:149)
2023-03-01 17:18:14 at io.quarkus.hibernate.orm.panache.runtime.JpaOperations.list(JpaOperations.java:24)
2023-03-01 17:18:14 at io.quarkus.hibernate.orm.panache.runtime.JpaOperations.list(JpaOperations.java:10)
2023-03-01 17:18:14 at io.quarkus.hibernate.orm.panache.common.runtime.AbstractJpaOperations.listAll(AbstractJpaOperations.java:305)
2023-03-01 17:18:14 at org.acme.repository.UserRepository.listAll(UserRepository.java)
2023-03-01 17:18:14 at org.acme.repository.UserRepository_ClientProxy.listAll(Unknown Source)
2023-03-01 17:18:14 at org.acme.service.impl.DefaultUserService.getAllUsers(DefaultUserService.java:30)
2023-03-01 17:18:14 at org.acme.service.impl.DefaultUserService_ClientProxy.getAllUsers(Unknown Source)
2023-03-01 17:18:14 at org.acme.controller.UserController.getUsers(UserController.java:40)
2023-03-01 17:18:14 at org.acme.controller.UserController_ClientProxy.getUsers(Unknown Source)
2023-03-01 17:18:14 at org.acme.controller.UserController$quarkusrestinvoker$getUsers_4b9221263c4fb1e6e88448a563d3364eb951551a.invoke(Unknown Source)
2023-03-01 17:18:14 at org.jboss.resteasy.reactive.server.handlers.InvocationHandler.handle(InvocationHandler.java:29)
2023-03-01 17:18:14 at io.quarkus.resteasy.reactive.server.runtime.QuarkusResteasyReactiveRequestContext.invokeHandler(QuarkusResteasyReactiveRequestContext.java:114)
2023-03-01 17:18:14 at org.jboss.resteasy.reactive.common.core.AbstractResteasyReactiveContext.run(AbstractResteasyReactiveContext.java:145)
2023-03-01 17:18:14 at io.quarkus.vertx.core.runtime.VertxCoreRecorder$14.runWith(VertxCoreRecorder.java:576)
2023-03-01 17:18:14 at org.jboss.threads.EnhancedQueueExecutor$Task.run(EnhancedQueueExecutor.java:2449)
2023-03-01 17:18:14 at org.jboss.threads.EnhancedQueueExecutor$ThreadBody.run(EnhancedQueueExecutor.java:1478)
2023-03-01 17:18:14 at org.jboss.threads.DelegatingRunnable.run(DelegatingRunnable.java:29)
2023-03-01 17:18:14 at org.jboss.threads.ThreadLocalResettingRunnable.run(ThreadLocalResettingRunnable.java:29)
2023-03-01 17:18:14 at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
2023-03-01 17:18:14 at java.base/java.lang.Thread.run(Thread.java:829)
2023-03-01 17:18:14 Caused by: org.hibernate.exception.GenericJDBCException: Unable to acquire JDBC Connection
2023-03-01 17:18:14 at org.hibernate.exception.internal.StandardSQLExceptionConverter.convert(StandardSQLExceptionConverter.java:42)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.spi.SqlExceptionHelper.convert(SqlExceptionHelper.java:113)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.spi.SqlExceptionHelper.convert(SqlExceptionHelper.java:99)
2023-03-01 17:18:14 at org.hibernate.resource.jdbc.internal.LogicalConnectionManagedImpl.acquireConnectionIfNeeded(LogicalConnectionManagedImpl.java:111)
2023-03-01 17:18:14 at org.hibernate.resource.jdbc.internal.LogicalConnectionManagedImpl.getPhysicalConnection(LogicalConnectionManagedImpl.java:138)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.internal.StatementPreparerImpl.connection(StatementPreparerImpl.java:50)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.internal.StatementPreparerImpl$5.doPrepare(StatementPreparerImpl.java:149)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.internal.StatementPreparerImpl$StatementPreparationTemplate.prepareStatement(StatementPreparerImpl.java:176)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.internal.StatementPreparerImpl.prepareQueryStatement(StatementPreparerImpl.java:151)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.prepareQueryStatement(Loader.java:2122)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.executeQueryStatement(Loader.java:2059)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.executeQueryStatement(Loader.java:2037)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.doQuery(Loader.java:956)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.doQueryAndInitializeNonLazyCollections(Loader.java:357)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.doList(Loader.java:2868)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.doList(Loader.java:2850)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.listIgnoreQueryCache(Loader.java:2682)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.list(Loader.java:2677)
2023-03-01 17:18:14 at org.hibernate.loader.hql.QueryLoader.list(QueryLoader.java:540)
2023-03-01 17:18:14 at org.hibernate.hql.internal.ast.QueryTranslatorImpl.list(QueryTranslatorImpl.java:400)
2023-03-01 17:18:14 at org.hibernate.engine.query.spi.HQLQueryPlan.performList(HQLQueryPlan.java:218)
2023-03-01 17:18:14 at org.hibernate.internal.SessionImpl.list(SessionImpl.java:1459)
2023-03-01 17:18:14 at org.hibernate.query.internal.AbstractProducedQuery.doList(AbstractProducedQuery.java:1649)
2023-03-01 17:18:14 at org.hibernate.query.internal.AbstractProducedQuery.list(AbstractProducedQuery.java:1617)
2023-03-01 17:18:14 ... 23 more
2023-03-01 17:18:14 Caused by: java.sql.SQLException: SSL connection required for plugin "mysql_clear_password". Check if 'sslMode' is enabled.
2023-03-01 17:18:14 at com.mysql.cj.jdbc.exceptions.SQLError.createSQLException(SQLError.java:129)
2023-03-01 17:18:14 at com.mysql.cj.jdbc.exceptions.SQLExceptionsMapping.translateException(SQLExceptionsMapping.java:122)
2023-03-01 17:18:14 at com.mysql.cj.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:828)
2023-03-01 17:18:14 at com.mysql.cj.jdbc.ConnectionImpl.<init>(ConnectionImpl.java:448)
2023-03-01 17:18:14 at com.mysql.cj.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:241)
2023-03-01 17:18:14 at com.mysql.cj.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:198)
2023-03-01 17:18:14 at io.agroal.pool.ConnectionFactory.createConnection(ConnectionFactory.java:226)
2023-03-01 17:18:14 at io.agroal.pool.ConnectionPool$CreateConnectionTask.call(ConnectionPool.java:535)
2023-03-01 17:18:14 at io.agroal.pool.ConnectionPool$CreateConnectionTask.call(ConnectionPool.java:516)
2023-03-01 17:18:14 at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
2023-03-01 17:18:14 at io.agroal.pool.util.PriorityScheduledExecutor.beforeExecute(PriorityScheduledExecutor.java:75)
2023-03-01 17:18:14 at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1126)
2023-03-01 17:18:14 at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
2023-03-01 17:18:14 ... 1 more
```
### Steps to reproduce?
1. enable sql connection feature in cloud run
2. enable IAM-auth in cloud sql and create SA and grant acces to your DB
3. set your jdbc string with unixSocketPath + enableIamAuth
### Environment
1. OS type and version: registry.access.redhat.com/ubi8/openjdk-11:1.14
2. Java SDK version: registry.access.redhat.com/ubi8/openjdk-11:1.14
3. Cloud SQL Java Socket Factory version: mysql-socket-factory-connector-j-8 1.10.0
### Additional Details
tested with sslMode=REQUIRED and sslMode=DISABLED both doesn't work | 1.0 | working with mysql socket and iam auth at the same time - ### Bug Description
hi i'm using mysql-socket-factory-connector-j-8 to connect a cloud run container to Mysql cloud sql instance.
a lot of options are available to do it, but the one i need to make it work is :
- cloud run has Cloud SQL connections enabled (this creates some kind of a cloud sql proxy side car container that shares a socket in /cloudsql/INSTANCE_ID that can be used from your cloud run container [link to documentation](https://cloud.google.com/sql/docs/mysql/connect-run?_ga=2.40665698.-1100797918.1668437632&_gac=1.185697115.1676974444.CjwKCAiA9NGfBhBvEiwAq5vSy0PaHBu7zPG4FX_efXZ0Yalgkj8mzPI9FJhMBWFrNXN4SxX4a1J2nxoC03kQAvD_BwE&hl=fr)
- using enableIamAuth=true to connect using Service account
each one of those set ups works alone but together i got this error : `SSL connection required for plugin "mysql_clear_password". Check if 'sslMode' is enabled.`
### Example code (or command)
```Java
- working jdbc string with cloud run sql connection feature eanbled and using socket:
jdbc:mysql:///<databaseName>?cloudSqlInstance=<project>:<region>:<dbname>&socketFactory=com.google.cloud.sql.mysql.SocketFactory&=true&ipTypes=PRIVATE&unixSocketPath=/cloudsql/<project>:<region>:<dbname>
- also working jdbc string without socket but with iamAuth enabled :
jdbc:mysql:///<databaseName>?cloudSqlInstance=<project>:<region>:<dbname>&socketFactory=com.google.cloud.sql.mysql.SocketFactory&ipTypes=PRIVATE&enableIamAuth=true
- not working jdbc string when both enableIamAuth and unixSocketPath are used :
jdbc:mysql:///<databaseName>?cloudSqlInstance=<project>:<region>:<dbname>&socketFactory=com.google.cloud.sql.mysql.SocketFactory&&enableIamAuth=true&ipTypes=PRIVATE&unixSocketPath=/cloudsql/<project>:<region>:<dbname>
```
### Stacktrace
```bash
2023-03-01 17:18:13 2023-03-01 17:18:13,993 INFO [com.goo.clo.sql.cor.CoreSocketFactory] (agroal-11) Connecting to Cloud SQL instance [mystic-aileron-190912:europe-west9:test-shi] via unix socket at /cloudsql/mystic-aileron-190912:europe-west9:test-shi.
2023-03-01 17:18:14 2023-03-01 17:18:14,005 WARN [io.agr.pool] (agroal-11) Datasource '<default>': SSL connection required for plugin "mysql_clear_password". Check if 'sslMode' is enabled.
2023-03-01 17:18:14 2023-03-01 17:18:14,005 WARN [org.hib.eng.jdb.spi.SqlExceptionHelper] (executor-thread-0) SQL Error: 0, SQLState: S1000
2023-03-01 17:18:14 2023-03-01 17:18:14,005 ERROR [org.hib.eng.jdb.spi.SqlExceptionHelper] (executor-thread-0) SSL connection required for plugin "mysql_clear_password". Check if 'sslMode' is enabled.
2023-03-01 17:18:14 2023-03-01 17:18:14,006 ERROR [io.qua.ver.htt.run.QuarkusErrorHandler] (executor-thread-0) HTTP Request to /v1/users failed, error id: fcf176b4-bab3-4a53-8410-bf0c85fc1e49-2: javax.persistence.PersistenceException: org.hibernate.exception.GenericJDBCException: Unable to acquire JDBC Connection
2023-03-01 17:18:14 at org.hibernate.internal.ExceptionConverterImpl.convert(ExceptionConverterImpl.java:154)
2023-03-01 17:18:14 at org.hibernate.query.internal.AbstractProducedQuery.list(AbstractProducedQuery.java:1626)
2023-03-01 17:18:14 at org.hibernate.query.Query.getResultList(Query.java:165)
2023-03-01 17:18:14 at io.quarkus.hibernate.orm.panache.common.runtime.CommonPanacheQueryImpl.list(CommonPanacheQueryImpl.java:269)
2023-03-01 17:18:14 at io.quarkus.hibernate.orm.panache.runtime.PanacheQueryImpl.list(PanacheQueryImpl.java:149)
2023-03-01 17:18:14 at io.quarkus.hibernate.orm.panache.runtime.JpaOperations.list(JpaOperations.java:24)
2023-03-01 17:18:14 at io.quarkus.hibernate.orm.panache.runtime.JpaOperations.list(JpaOperations.java:10)
2023-03-01 17:18:14 at io.quarkus.hibernate.orm.panache.common.runtime.AbstractJpaOperations.listAll(AbstractJpaOperations.java:305)
2023-03-01 17:18:14 at org.acme.repository.UserRepository.listAll(UserRepository.java)
2023-03-01 17:18:14 at org.acme.repository.UserRepository_ClientProxy.listAll(Unknown Source)
2023-03-01 17:18:14 at org.acme.service.impl.DefaultUserService.getAllUsers(DefaultUserService.java:30)
2023-03-01 17:18:14 at org.acme.service.impl.DefaultUserService_ClientProxy.getAllUsers(Unknown Source)
2023-03-01 17:18:14 at org.acme.controller.UserController.getUsers(UserController.java:40)
2023-03-01 17:18:14 at org.acme.controller.UserController_ClientProxy.getUsers(Unknown Source)
2023-03-01 17:18:14 at org.acme.controller.UserController$quarkusrestinvoker$getUsers_4b9221263c4fb1e6e88448a563d3364eb951551a.invoke(Unknown Source)
2023-03-01 17:18:14 at org.jboss.resteasy.reactive.server.handlers.InvocationHandler.handle(InvocationHandler.java:29)
2023-03-01 17:18:14 at io.quarkus.resteasy.reactive.server.runtime.QuarkusResteasyReactiveRequestContext.invokeHandler(QuarkusResteasyReactiveRequestContext.java:114)
2023-03-01 17:18:14 at org.jboss.resteasy.reactive.common.core.AbstractResteasyReactiveContext.run(AbstractResteasyReactiveContext.java:145)
2023-03-01 17:18:14 at io.quarkus.vertx.core.runtime.VertxCoreRecorder$14.runWith(VertxCoreRecorder.java:576)
2023-03-01 17:18:14 at org.jboss.threads.EnhancedQueueExecutor$Task.run(EnhancedQueueExecutor.java:2449)
2023-03-01 17:18:14 at org.jboss.threads.EnhancedQueueExecutor$ThreadBody.run(EnhancedQueueExecutor.java:1478)
2023-03-01 17:18:14 at org.jboss.threads.DelegatingRunnable.run(DelegatingRunnable.java:29)
2023-03-01 17:18:14 at org.jboss.threads.ThreadLocalResettingRunnable.run(ThreadLocalResettingRunnable.java:29)
2023-03-01 17:18:14 at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
2023-03-01 17:18:14 at java.base/java.lang.Thread.run(Thread.java:829)
2023-03-01 17:18:14 Caused by: org.hibernate.exception.GenericJDBCException: Unable to acquire JDBC Connection
2023-03-01 17:18:14 at org.hibernate.exception.internal.StandardSQLExceptionConverter.convert(StandardSQLExceptionConverter.java:42)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.spi.SqlExceptionHelper.convert(SqlExceptionHelper.java:113)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.spi.SqlExceptionHelper.convert(SqlExceptionHelper.java:99)
2023-03-01 17:18:14 at org.hibernate.resource.jdbc.internal.LogicalConnectionManagedImpl.acquireConnectionIfNeeded(LogicalConnectionManagedImpl.java:111)
2023-03-01 17:18:14 at org.hibernate.resource.jdbc.internal.LogicalConnectionManagedImpl.getPhysicalConnection(LogicalConnectionManagedImpl.java:138)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.internal.StatementPreparerImpl.connection(StatementPreparerImpl.java:50)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.internal.StatementPreparerImpl$5.doPrepare(StatementPreparerImpl.java:149)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.internal.StatementPreparerImpl$StatementPreparationTemplate.prepareStatement(StatementPreparerImpl.java:176)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.internal.StatementPreparerImpl.prepareQueryStatement(StatementPreparerImpl.java:151)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.prepareQueryStatement(Loader.java:2122)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.executeQueryStatement(Loader.java:2059)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.executeQueryStatement(Loader.java:2037)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.doQuery(Loader.java:956)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.doQueryAndInitializeNonLazyCollections(Loader.java:357)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.doList(Loader.java:2868)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.doList(Loader.java:2850)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.listIgnoreQueryCache(Loader.java:2682)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.list(Loader.java:2677)
2023-03-01 17:18:14 at org.hibernate.loader.hql.QueryLoader.list(QueryLoader.java:540)
2023-03-01 17:18:14 at org.hibernate.hql.internal.ast.QueryTranslatorImpl.list(QueryTranslatorImpl.java:400)
2023-03-01 17:18:14 at org.hibernate.engine.query.spi.HQLQueryPlan.performList(HQLQueryPlan.java:218)
2023-03-01 17:18:14 at org.hibernate.internal.SessionImpl.list(SessionImpl.java:1459)
2023-03-01 17:18:14 at org.hibernate.query.internal.AbstractProducedQuery.doList(AbstractProducedQuery.java:1649)
2023-03-01 17:18:14 at org.hibernate.query.internal.AbstractProducedQuery.list(AbstractProducedQuery.java:1617)
2023-03-01 17:18:14 ... 23 more
2023-03-01 17:18:14 Caused by: java.sql.SQLException: SSL connection required for plugin "mysql_clear_password". Check if 'sslMode' is enabled.
2023-03-01 17:18:14 at com.mysql.cj.jdbc.exceptions.SQLError.createSQLException(SQLError.java:129)
2023-03-01 17:18:14 at com.mysql.cj.jdbc.exceptions.SQLExceptionsMapping.translateException(SQLExceptionsMapping.java:122)
2023-03-01 17:18:14 at com.mysql.cj.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:828)
2023-03-01 17:18:14 at com.mysql.cj.jdbc.ConnectionImpl.<init>(ConnectionImpl.java:448)
2023-03-01 17:18:14 at com.mysql.cj.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:241)
2023-03-01 17:18:14 at com.mysql.cj.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:198)
2023-03-01 17:18:14 at io.agroal.pool.ConnectionFactory.createConnection(ConnectionFactory.java:226)
2023-03-01 17:18:14 at io.agroal.pool.ConnectionPool$CreateConnectionTask.call(ConnectionPool.java:535)
2023-03-01 17:18:14 at io.agroal.pool.ConnectionPool$CreateConnectionTask.call(ConnectionPool.java:516)
2023-03-01 17:18:14 at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
2023-03-01 17:18:14 at io.agroal.pool.util.PriorityScheduledExecutor.beforeExecute(PriorityScheduledExecutor.java:75)
2023-03-01 17:18:14 at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1126)
2023-03-01 17:18:14 at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
2023-03-01 17:18:14 ... 1 more
2023-03-01 17:18:14 2023-03-01 17:18:14,007 ERROR [org.jbo.res.rea.com.cor.AbstractResteasyReactiveContext] (executor-thread-0) Request failed: javax.persistence.PersistenceException: org.hibernate.exception.GenericJDBCException: Unable to acquire JDBC Connection
2023-03-01 17:18:14 at org.hibernate.internal.ExceptionConverterImpl.convert(ExceptionConverterImpl.java:154)
2023-03-01 17:18:14 at org.hibernate.query.internal.AbstractProducedQuery.list(AbstractProducedQuery.java:1626)
2023-03-01 17:18:14 at org.hibernate.query.Query.getResultList(Query.java:165)
2023-03-01 17:18:14 at io.quarkus.hibernate.orm.panache.common.runtime.CommonPanacheQueryImpl.list(CommonPanacheQueryImpl.java:269)
2023-03-01 17:18:14 at io.quarkus.hibernate.orm.panache.runtime.PanacheQueryImpl.list(PanacheQueryImpl.java:149)
2023-03-01 17:18:14 at io.quarkus.hibernate.orm.panache.runtime.JpaOperations.list(JpaOperations.java:24)
2023-03-01 17:18:14 at io.quarkus.hibernate.orm.panache.runtime.JpaOperations.list(JpaOperations.java:10)
2023-03-01 17:18:14 at io.quarkus.hibernate.orm.panache.common.runtime.AbstractJpaOperations.listAll(AbstractJpaOperations.java:305)
2023-03-01 17:18:14 at org.acme.repository.UserRepository.listAll(UserRepository.java)
2023-03-01 17:18:14 at org.acme.repository.UserRepository_ClientProxy.listAll(Unknown Source)
2023-03-01 17:18:14 at org.acme.service.impl.DefaultUserService.getAllUsers(DefaultUserService.java:30)
2023-03-01 17:18:14 at org.acme.service.impl.DefaultUserService_ClientProxy.getAllUsers(Unknown Source)
2023-03-01 17:18:14 at org.acme.controller.UserController.getUsers(UserController.java:40)
2023-03-01 17:18:14 at org.acme.controller.UserController_ClientProxy.getUsers(Unknown Source)
2023-03-01 17:18:14 at org.acme.controller.UserController$quarkusrestinvoker$getUsers_4b9221263c4fb1e6e88448a563d3364eb951551a.invoke(Unknown Source)
2023-03-01 17:18:14 at org.jboss.resteasy.reactive.server.handlers.InvocationHandler.handle(InvocationHandler.java:29)
2023-03-01 17:18:14 at io.quarkus.resteasy.reactive.server.runtime.QuarkusResteasyReactiveRequestContext.invokeHandler(QuarkusResteasyReactiveRequestContext.java:114)
2023-03-01 17:18:14 at org.jboss.resteasy.reactive.common.core.AbstractResteasyReactiveContext.run(AbstractResteasyReactiveContext.java:145)
2023-03-01 17:18:14 at io.quarkus.vertx.core.runtime.VertxCoreRecorder$14.runWith(VertxCoreRecorder.java:576)
2023-03-01 17:18:14 at org.jboss.threads.EnhancedQueueExecutor$Task.run(EnhancedQueueExecutor.java:2449)
2023-03-01 17:18:14 at org.jboss.threads.EnhancedQueueExecutor$ThreadBody.run(EnhancedQueueExecutor.java:1478)
2023-03-01 17:18:14 at org.jboss.threads.DelegatingRunnable.run(DelegatingRunnable.java:29)
2023-03-01 17:18:14 at org.jboss.threads.ThreadLocalResettingRunnable.run(ThreadLocalResettingRunnable.java:29)
2023-03-01 17:18:14 at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
2023-03-01 17:18:14 at java.base/java.lang.Thread.run(Thread.java:829)
2023-03-01 17:18:14 Caused by: org.hibernate.exception.GenericJDBCException: Unable to acquire JDBC Connection
2023-03-01 17:18:14 at org.hibernate.exception.internal.StandardSQLExceptionConverter.convert(StandardSQLExceptionConverter.java:42)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.spi.SqlExceptionHelper.convert(SqlExceptionHelper.java:113)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.spi.SqlExceptionHelper.convert(SqlExceptionHelper.java:99)
2023-03-01 17:18:14 at org.hibernate.resource.jdbc.internal.LogicalConnectionManagedImpl.acquireConnectionIfNeeded(LogicalConnectionManagedImpl.java:111)
2023-03-01 17:18:14 at org.hibernate.resource.jdbc.internal.LogicalConnectionManagedImpl.getPhysicalConnection(LogicalConnectionManagedImpl.java:138)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.internal.StatementPreparerImpl.connection(StatementPreparerImpl.java:50)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.internal.StatementPreparerImpl$5.doPrepare(StatementPreparerImpl.java:149)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.internal.StatementPreparerImpl$StatementPreparationTemplate.prepareStatement(StatementPreparerImpl.java:176)
2023-03-01 17:18:14 at org.hibernate.engine.jdbc.internal.StatementPreparerImpl.prepareQueryStatement(StatementPreparerImpl.java:151)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.prepareQueryStatement(Loader.java:2122)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.executeQueryStatement(Loader.java:2059)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.executeQueryStatement(Loader.java:2037)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.doQuery(Loader.java:956)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.doQueryAndInitializeNonLazyCollections(Loader.java:357)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.doList(Loader.java:2868)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.doList(Loader.java:2850)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.listIgnoreQueryCache(Loader.java:2682)
2023-03-01 17:18:14 at org.hibernate.loader.Loader.list(Loader.java:2677)
2023-03-01 17:18:14 at org.hibernate.loader.hql.QueryLoader.list(QueryLoader.java:540)
2023-03-01 17:18:14 at org.hibernate.hql.internal.ast.QueryTranslatorImpl.list(QueryTranslatorImpl.java:400)
2023-03-01 17:18:14 at org.hibernate.engine.query.spi.HQLQueryPlan.performList(HQLQueryPlan.java:218)
2023-03-01 17:18:14 at org.hibernate.internal.SessionImpl.list(SessionImpl.java:1459)
2023-03-01 17:18:14 at org.hibernate.query.internal.AbstractProducedQuery.doList(AbstractProducedQuery.java:1649)
2023-03-01 17:18:14 at org.hibernate.query.internal.AbstractProducedQuery.list(AbstractProducedQuery.java:1617)
2023-03-01 17:18:14 ... 23 more
2023-03-01 17:18:14 Caused by: java.sql.SQLException: SSL connection required for plugin "mysql_clear_password". Check if 'sslMode' is enabled.
2023-03-01 17:18:14 at com.mysql.cj.jdbc.exceptions.SQLError.createSQLException(SQLError.java:129)
2023-03-01 17:18:14 at com.mysql.cj.jdbc.exceptions.SQLExceptionsMapping.translateException(SQLExceptionsMapping.java:122)
2023-03-01 17:18:14 at com.mysql.cj.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:828)
2023-03-01 17:18:14 at com.mysql.cj.jdbc.ConnectionImpl.<init>(ConnectionImpl.java:448)
2023-03-01 17:18:14 at com.mysql.cj.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:241)
2023-03-01 17:18:14 at com.mysql.cj.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:198)
2023-03-01 17:18:14 at io.agroal.pool.ConnectionFactory.createConnection(ConnectionFactory.java:226)
2023-03-01 17:18:14 at io.agroal.pool.ConnectionPool$CreateConnectionTask.call(ConnectionPool.java:535)
2023-03-01 17:18:14 at io.agroal.pool.ConnectionPool$CreateConnectionTask.call(ConnectionPool.java:516)
2023-03-01 17:18:14 at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
2023-03-01 17:18:14 at io.agroal.pool.util.PriorityScheduledExecutor.beforeExecute(PriorityScheduledExecutor.java:75)
2023-03-01 17:18:14 at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1126)
2023-03-01 17:18:14 at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
2023-03-01 17:18:14 ... 1 more
```
### Steps to reproduce?
1. enable sql connection feature in cloud run
2. enable IAM-auth in cloud sql and create SA and grant acces to your DB
3. set your jdbc string with unixSocketPath + enableIamAuth
### Environment
1. OS type and version: registry.access.redhat.com/ubi8/openjdk-11:1.14
2. Java SDK version: registry.access.redhat.com/ubi8/openjdk-11:1.14
3. Cloud SQL Java Socket Factory version: mysql-socket-factory-connector-j-8 1.10.0
### Additional Details
tested with sslMode=REQUIRED and sslMode=DISABLED both doesn't work | priority | working with mysql socket and iam auth at the same time bug description hi i m using mysql socket factory connector j to connect a cloud run container to mysql cloud sql instance a lot of options are available to do it but the one i need to make it work is cloud run has cloud sql connections enabled this creates some kind of a cloud sql proxy side car container that shares a socket in cloudsql instance id that can be used from your cloud run container using enableiamauth true to connect using service account each one of those set ups works alone but together i got this error ssl connection required for plugin mysql clear password check if sslmode is enabled example code or command java working jdbc string with cloud run sql connection feature eanbled and using socket jdbc mysql cloudsqlinstance socketfactory com google cloud sql mysql socketfactory true iptypes private unixsocketpath cloudsql also working jdbc string without socket but with iamauth enabled jdbc mysql cloudsqlinstance socketfactory com google cloud sql mysql socketfactory iptypes private enableiamauth true not working jdbc string when both enableiamauth and unixsocketpath are used jdbc mysql cloudsqlinstance socketfactory com google cloud sql mysql socketfactory enableiamauth true iptypes private unixsocketpath cloudsql stacktrace bash info agroal connecting to cloud sql instance via unix socket at cloudsql mystic aileron europe test shi warn agroal datasource ssl connection required for plugin mysql clear password check if sslmode is enabled warn executor thread sql error sqlstate error executor thread ssl connection required for plugin mysql clear password check if sslmode is enabled error executor thread http request to users failed error id javax persistence persistenceexception org hibernate exception genericjdbcexception unable to acquire jdbc connection at org hibernate internal exceptionconverterimpl convert exceptionconverterimpl java at org hibernate query internal abstractproducedquery list abstractproducedquery java at org hibernate query query getresultlist query java at io quarkus hibernate orm panache common runtime commonpanachequeryimpl list commonpanachequeryimpl java at io quarkus hibernate orm panache runtime panachequeryimpl list panachequeryimpl java at io quarkus hibernate orm panache runtime jpaoperations list jpaoperations java at io quarkus hibernate orm panache runtime jpaoperations list jpaoperations java at io quarkus hibernate orm panache common runtime abstractjpaoperations listall abstractjpaoperations java at org acme repository userrepository listall userrepository java at org acme repository userrepository clientproxy listall unknown source at org acme service impl defaultuserservice getallusers defaultuserservice java at org acme service impl defaultuserservice clientproxy getallusers unknown source at org acme controller usercontroller getusers usercontroller java at org acme controller usercontroller clientproxy getusers unknown source at org acme controller usercontroller quarkusrestinvoker getusers invoke unknown source at org jboss resteasy reactive server handlers invocationhandler handle invocationhandler java at io quarkus resteasy reactive server runtime quarkusresteasyreactiverequestcontext invokehandler quarkusresteasyreactiverequestcontext java at org jboss resteasy reactive common core abstractresteasyreactivecontext run abstractresteasyreactivecontext java at io quarkus vertx core runtime vertxcorerecorder runwith vertxcorerecorder java at org jboss threads enhancedqueueexecutor task run enhancedqueueexecutor java at org jboss threads enhancedqueueexecutor threadbody run enhancedqueueexecutor java at org jboss threads delegatingrunnable run delegatingrunnable java at org jboss threads threadlocalresettingrunnable run threadlocalresettingrunnable java at io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java base java lang thread run thread java caused by org hibernate exception genericjdbcexception unable to acquire jdbc connection at org hibernate exception internal standardsqlexceptionconverter convert standardsqlexceptionconverter java at org hibernate engine jdbc spi sqlexceptionhelper convert sqlexceptionhelper java at org hibernate engine jdbc spi sqlexceptionhelper convert sqlexceptionhelper java at org hibernate resource jdbc internal logicalconnectionmanagedimpl acquireconnectionifneeded logicalconnectionmanagedimpl java at org hibernate resource jdbc internal logicalconnectionmanagedimpl getphysicalconnection logicalconnectionmanagedimpl java at org hibernate engine jdbc internal statementpreparerimpl connection statementpreparerimpl java at org hibernate engine jdbc internal statementpreparerimpl doprepare statementpreparerimpl java at org hibernate engine jdbc internal statementpreparerimpl statementpreparationtemplate preparestatement statementpreparerimpl java at org hibernate engine jdbc internal statementpreparerimpl preparequerystatement statementpreparerimpl java at org hibernate loader loader preparequerystatement loader java at org hibernate loader loader executequerystatement loader java at org hibernate loader loader executequerystatement loader java at org hibernate loader loader doquery loader java at org hibernate loader loader doqueryandinitializenonlazycollections loader java at org hibernate loader loader dolist loader java at org hibernate loader loader dolist loader java at org hibernate loader loader listignorequerycache loader java at org hibernate loader loader list loader java at org hibernate loader hql queryloader list queryloader java at org hibernate hql internal ast querytranslatorimpl list querytranslatorimpl java at org hibernate engine query spi hqlqueryplan performlist hqlqueryplan java at org hibernate internal sessionimpl list sessionimpl java at org hibernate query internal abstractproducedquery dolist abstractproducedquery java at org hibernate query internal abstractproducedquery list abstractproducedquery java more caused by java sql sqlexception ssl connection required for plugin mysql clear password check if sslmode is enabled at com mysql cj jdbc exceptions sqlerror createsqlexception sqlerror java at com mysql cj jdbc exceptions sqlexceptionsmapping translateexception sqlexceptionsmapping java at com mysql cj jdbc connectionimpl createnewio connectionimpl java at com mysql cj jdbc connectionimpl connectionimpl java at com mysql cj jdbc connectionimpl getinstance connectionimpl java at com mysql cj jdbc nonregisteringdriver connect nonregisteringdriver java at io agroal pool connectionfactory createconnection connectionfactory java at io agroal pool connectionpool createconnectiontask call connectionpool java at io agroal pool connectionpool createconnectiontask call connectionpool java at java base java util concurrent futuretask run futuretask java at io agroal pool util priorityscheduledexecutor beforeexecute priorityscheduledexecutor java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java more error executor thread request failed javax persistence persistenceexception org hibernate exception genericjdbcexception unable to acquire jdbc connection at org hibernate internal exceptionconverterimpl convert exceptionconverterimpl java at org hibernate query internal abstractproducedquery list abstractproducedquery java at org hibernate query query getresultlist query java at io quarkus hibernate orm panache common runtime commonpanachequeryimpl list commonpanachequeryimpl java at io quarkus hibernate orm panache runtime panachequeryimpl list panachequeryimpl java at io quarkus hibernate orm panache runtime jpaoperations list jpaoperations java at io quarkus hibernate orm panache runtime jpaoperations list jpaoperations java at io quarkus hibernate orm panache common runtime abstractjpaoperations listall abstractjpaoperations java at org acme repository userrepository listall userrepository java at org acme repository userrepository clientproxy listall unknown source at org acme service impl defaultuserservice getallusers defaultuserservice java at org acme service impl defaultuserservice clientproxy getallusers unknown source at org acme controller usercontroller getusers usercontroller java at org acme controller usercontroller clientproxy getusers unknown source at org acme controller usercontroller quarkusrestinvoker getusers invoke unknown source at org jboss resteasy reactive server handlers invocationhandler handle invocationhandler java at io quarkus resteasy reactive server runtime quarkusresteasyreactiverequestcontext invokehandler quarkusresteasyreactiverequestcontext java at org jboss resteasy reactive common core abstractresteasyreactivecontext run abstractresteasyreactivecontext java at io quarkus vertx core runtime vertxcorerecorder runwith vertxcorerecorder java at org jboss threads enhancedqueueexecutor task run enhancedqueueexecutor java at org jboss threads enhancedqueueexecutor threadbody run enhancedqueueexecutor java at org jboss threads delegatingrunnable run delegatingrunnable java at org jboss threads threadlocalresettingrunnable run threadlocalresettingrunnable java at io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java base java lang thread run thread java caused by org hibernate exception genericjdbcexception unable to acquire jdbc connection at org hibernate exception internal standardsqlexceptionconverter convert standardsqlexceptionconverter java at org hibernate engine jdbc spi sqlexceptionhelper convert sqlexceptionhelper java at org hibernate engine jdbc spi sqlexceptionhelper convert sqlexceptionhelper java at org hibernate resource jdbc internal logicalconnectionmanagedimpl acquireconnectionifneeded logicalconnectionmanagedimpl java at org hibernate resource jdbc internal logicalconnectionmanagedimpl getphysicalconnection logicalconnectionmanagedimpl java at org hibernate engine jdbc internal statementpreparerimpl connection statementpreparerimpl java at org hibernate engine jdbc internal statementpreparerimpl doprepare statementpreparerimpl java at org hibernate engine jdbc internal statementpreparerimpl statementpreparationtemplate preparestatement statementpreparerimpl java at org hibernate engine jdbc internal statementpreparerimpl preparequerystatement statementpreparerimpl java at org hibernate loader loader preparequerystatement loader java at org hibernate loader loader executequerystatement loader java at org hibernate loader loader executequerystatement loader java at org hibernate loader loader doquery loader java at org hibernate loader loader doqueryandinitializenonlazycollections loader java at org hibernate loader loader dolist loader java at org hibernate loader loader dolist loader java at org hibernate loader loader listignorequerycache loader java at org hibernate loader loader list loader java at org hibernate loader hql queryloader list queryloader java at org hibernate hql internal ast querytranslatorimpl list querytranslatorimpl java at org hibernate engine query spi hqlqueryplan performlist hqlqueryplan java at org hibernate internal sessionimpl list sessionimpl java at org hibernate query internal abstractproducedquery dolist abstractproducedquery java at org hibernate query internal abstractproducedquery list abstractproducedquery java more caused by java sql sqlexception ssl connection required for plugin mysql clear password check if sslmode is enabled at com mysql cj jdbc exceptions sqlerror createsqlexception sqlerror java at com mysql cj jdbc exceptions sqlexceptionsmapping translateexception sqlexceptionsmapping java at com mysql cj jdbc connectionimpl createnewio connectionimpl java at com mysql cj jdbc connectionimpl connectionimpl java at com mysql cj jdbc connectionimpl getinstance connectionimpl java at com mysql cj jdbc nonregisteringdriver connect nonregisteringdriver java at io agroal pool connectionfactory createconnection connectionfactory java at io agroal pool connectionpool createconnectiontask call connectionpool java at io agroal pool connectionpool createconnectiontask call connectionpool java at java base java util concurrent futuretask run futuretask java at io agroal pool util priorityscheduledexecutor beforeexecute priorityscheduledexecutor java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java more steps to reproduce enable sql connection feature in cloud run enable iam auth in cloud sql and create sa and grant acces to your db set your jdbc string with unixsocketpath enableiamauth environment os type and version registry access redhat com openjdk java sdk version registry access redhat com openjdk cloud sql java socket factory version mysql socket factory connector j additional details tested with sslmode required and sslmode disabled both doesn t work | 1 |
578,669 | 17,149,852,818 | IssuesEvent | 2021-07-13 19:00:07 | apcountryman/picolibrary | https://api.github.com/repos/apcountryman/picolibrary | closed | Add blocking, free running ADC concept | priority-normal status-awaiting_review type-feature | Add blocking, free running ADC concept (`::picolibrary::ADC::Blocking_Free_Running_Converter_Concept`) and associated mock (`::picolibrary::Testing::Unit::ADC::Mock_Blocking_Free_Running_Converter`).
Blocking, free running ADCs will be expected to supply the following type aliases:
- `Sample`: ADC sample
Blocking free running ADCs will be expected to support the following operations:
- Default construction
- Move construction
- Destruction
- Move assignment
- `auto initialize() noexcept -> ::picolibrary::Result<::picolibrary::Void, ::picolibrary::Error_Code>` or `auto initialize() noexcept -> ::picolibrary::Result<::picolibrary::Void, ::picolibrary::Void>`: initialize the ADC's hardware
- `auto start_converter() noexcept -> ::picolibrary::Result<::picolibrary::Void, ::picolibrary::Error_Code>` or `auto start_converter() noexcept -> ::picolibrary::Result<::picolibrary::Void, ::picolibrary::Void>`: start the ADC
- `auto stop_converter() noexcept -> ::picolibrary::Result<::picolibrary::Void, ::picolibrary::Error_Code>` or `auto stop_converter() noexcept -> ::picolibrary::Result<::picolibrary::Void, ::picolibrary::Void>`: stop the ADC
- `auto sample() noexcept -> ::picolibrary::Result<Sample, ::picolibrary::Error_Code>` or `auto sample() noexcept -> ::picolibrary::Result<Sample, ::picolibrary::Void>`: get a sample | 1.0 | Add blocking, free running ADC concept - Add blocking, free running ADC concept (`::picolibrary::ADC::Blocking_Free_Running_Converter_Concept`) and associated mock (`::picolibrary::Testing::Unit::ADC::Mock_Blocking_Free_Running_Converter`).
Blocking, free running ADCs will be expected to supply the following type aliases:
- `Sample`: ADC sample
Blocking free running ADCs will be expected to support the following operations:
- Default construction
- Move construction
- Destruction
- Move assignment
- `auto initialize() noexcept -> ::picolibrary::Result<::picolibrary::Void, ::picolibrary::Error_Code>` or `auto initialize() noexcept -> ::picolibrary::Result<::picolibrary::Void, ::picolibrary::Void>`: initialize the ADC's hardware
- `auto start_converter() noexcept -> ::picolibrary::Result<::picolibrary::Void, ::picolibrary::Error_Code>` or `auto start_converter() noexcept -> ::picolibrary::Result<::picolibrary::Void, ::picolibrary::Void>`: start the ADC
- `auto stop_converter() noexcept -> ::picolibrary::Result<::picolibrary::Void, ::picolibrary::Error_Code>` or `auto stop_converter() noexcept -> ::picolibrary::Result<::picolibrary::Void, ::picolibrary::Void>`: stop the ADC
- `auto sample() noexcept -> ::picolibrary::Result<Sample, ::picolibrary::Error_Code>` or `auto sample() noexcept -> ::picolibrary::Result<Sample, ::picolibrary::Void>`: get a sample | priority | add blocking free running adc concept add blocking free running adc concept picolibrary adc blocking free running converter concept and associated mock picolibrary testing unit adc mock blocking free running converter blocking free running adcs will be expected to supply the following type aliases sample adc sample blocking free running adcs will be expected to support the following operations default construction move construction destruction move assignment auto initialize noexcept picolibrary result or auto initialize noexcept picolibrary result initialize the adc s hardware auto start converter noexcept picolibrary result or auto start converter noexcept picolibrary result start the adc auto stop converter noexcept picolibrary result or auto stop converter noexcept picolibrary result stop the adc auto sample noexcept picolibrary result or auto sample noexcept picolibrary result get a sample | 1 |
68,375 | 13,124,906,102 | IssuesEvent | 2020-08-06 05:16:28 | NickDMansfield/workinworkout | https://api.github.com/repos/NickDMansfield/workinworkout | opened | Do something with the User.Salt property | Code | Either remove it or implement it onto the token generation | 1.0 | Do something with the User.Salt property - Either remove it or implement it onto the token generation | non_priority | do something with the user salt property either remove it or implement it onto the token generation | 0 |
7,566 | 18,268,507,427 | IssuesEvent | 2021-10-04 11:18:17 | RasaHQ/rasa | https://api.github.com/repos/RasaHQ/rasa | closed | implement `ConfigValidator` | type:enhancement :sparkles: area:rasa-oss :ferris_wheel: priority:normal effort:enable-squad/2 feature:rasa-3.0/architecture | **Overview of the Solution**:
We need to implement the `ConfigValidator `. This component receives the current training data as `TrainingDataImporter` and validates it against the current train graph schema (accessible via `ExecutionContext`). Implements the logic from [here](https://github.com/RasaHQ/rasa/blob/e0a6a4a0829f510fd33657c918192ebe3f5ba06e/rasa/nlu/components.py#L245).
[This figure](https://www.notion.so/rasa/Rasa-Open-Source-3-0-Architecture-Implementation-Proposal-51ab90b05c41435ca98189a101676a1e#24ef56343a554fe88748651864fe9e77) shows the component's position in the graph.
**Update**:
- this should also check [this](https://github.com/RasaHQ/rasa/blob/9fdbca3a1e7b408fa2999d8e394520185c7a0fe0/rasa/core/policies/ensemble.py#L118)
- and [this](https://github.com/RasaHQ/rasa/blob/9fdbca3a1e7b408fa2999d8e394520185c7a0fe0/rasa/core/policies/ensemble.py#L247)
- consider checking all graph components which have a `validate_against_domain` to validate them against the domain (see [here](https://github.com/RasaHQ/rasa/pull/9576/files#r705073921))
**Input to the component**
- `TrainingDataImporter`
**Output:**
* `none`
* raises exception in case config doesn't match training data
**Blockers**
* https://github.com/RasaHQ/rasa/pull/9184
**Definition of Done**:
- [ ] Component is implemented and unit tested
| 1.0 | implement `ConfigValidator` - **Overview of the Solution**:
We need to implement the `ConfigValidator `. This component receives the current training data as `TrainingDataImporter` and validates it against the current train graph schema (accessible via `ExecutionContext`). Implements the logic from [here](https://github.com/RasaHQ/rasa/blob/e0a6a4a0829f510fd33657c918192ebe3f5ba06e/rasa/nlu/components.py#L245).
[This figure](https://www.notion.so/rasa/Rasa-Open-Source-3-0-Architecture-Implementation-Proposal-51ab90b05c41435ca98189a101676a1e#24ef56343a554fe88748651864fe9e77) shows the component's position in the graph.
**Update**:
- this should also check [this](https://github.com/RasaHQ/rasa/blob/9fdbca3a1e7b408fa2999d8e394520185c7a0fe0/rasa/core/policies/ensemble.py#L118)
- and [this](https://github.com/RasaHQ/rasa/blob/9fdbca3a1e7b408fa2999d8e394520185c7a0fe0/rasa/core/policies/ensemble.py#L247)
- consider checking all graph components which have a `validate_against_domain` to validate them against the domain (see [here](https://github.com/RasaHQ/rasa/pull/9576/files#r705073921))
**Input to the component**
- `TrainingDataImporter`
**Output:**
* `none`
* raises exception in case config doesn't match training data
**Blockers**
* https://github.com/RasaHQ/rasa/pull/9184
**Definition of Done**:
- [ ] Component is implemented and unit tested
| non_priority | implement configvalidator overview of the solution we need to implement the configvalidator this component receives the current training data as trainingdataimporter and validates it against the current train graph schema accessible via executioncontext implements the logic from shows the component s position in the graph update this should also check and consider checking all graph components which have a validate against domain to validate them against the domain see input to the component trainingdataimporter output none raises exception in case config doesn t match training data blockers definition of done component is implemented and unit tested | 0 |
125,579 | 10,347,092,071 | IssuesEvent | 2019-09-04 16:35:02 | devssa/onde-codar-em-salvador | https://api.github.com/repos/devssa/onde-codar-em-salvador | opened | ANALISTA DE TESTES JR (QA) | BDD CENÁRIOS DE TESTE METODOLOGIAS ÁGEIS SALVADOR TDD TESTE TESTE FUNCIONAL | <!--
==================================================
POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS!
Use: "Desenvolvedor Front-end" ao invés de
"Front-End Developer" \o/
Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]`
==================================================
-->
## ANALISTA DE TESTES JR (QA)
## Local
- Salvador
## Benefícios:
- VT + VR + AM + AO
- Jornada: Seg a Sex das 08:00 às 17:30hs
## Requisitos
**Obrigatórios:**
- Elaborar cenário de Testes;
- Executar os testes funcionais;
- Registrar defeitos encontrados;
- Garantir o cumprimento dos prazos e qualidade dos trabalhos realizados;
- Experiência comprovada na área;
- Escolaridade: Ensino superior completo ou em andamento em Análise e Desenvolvimento de Sistemas, Ciência da Computação ou em áreas correlatas.
- Muito interesse em aprender e se desenvolver;
**Desejáveis:**
- Testes automatizados, conceitos de BDD, TDD e Metodologia Ágil.
## Como se candidatar
Email: vagastisd@gmail.com
| 3.0 | ANALISTA DE TESTES JR (QA) - <!--
==================================================
POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS!
Use: "Desenvolvedor Front-end" ao invés de
"Front-End Developer" \o/
Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]`
==================================================
-->
## ANALISTA DE TESTES JR (QA)
## Local
- Salvador
## Benefícios:
- VT + VR + AM + AO
- Jornada: Seg a Sex das 08:00 às 17:30hs
## Requisitos
**Obrigatórios:**
- Elaborar cenário de Testes;
- Executar os testes funcionais;
- Registrar defeitos encontrados;
- Garantir o cumprimento dos prazos e qualidade dos trabalhos realizados;
- Experiência comprovada na área;
- Escolaridade: Ensino superior completo ou em andamento em Análise e Desenvolvimento de Sistemas, Ciência da Computação ou em áreas correlatas.
- Muito interesse em aprender e se desenvolver;
**Desejáveis:**
- Testes automatizados, conceitos de BDD, TDD e Metodologia Ágil.
## Como se candidatar
Email: vagastisd@gmail.com
| non_priority | analista de testes jr qa por favor só poste se a vaga for para salvador e cidades vizinhas use desenvolvedor front end ao invés de front end developer o exemplo desenvolvedor front end na analista de testes jr qa local salvador benefícios vt vr am ao jornada seg a sex das às requisitos obrigatórios elaborar cenário de testes executar os testes funcionais registrar defeitos encontrados garantir o cumprimento dos prazos e qualidade dos trabalhos realizados experiência comprovada na área escolaridade ensino superior completo ou em andamento em análise e desenvolvimento de sistemas ciência da computação ou em áreas correlatas muito interesse em aprender e se desenvolver desejáveis testes automatizados conceitos de bdd tdd e metodologia ágil como se candidatar email vagastisd gmail com | 0 |
484,880 | 13,958,541,171 | IssuesEvent | 2020-10-24 12:28:42 | AY2021S1-CS2103T-W16-3/tp | https://api.github.com/repos/AY2021S1-CS2103T-W16-3/tp | closed | Standardize parse exception precedence across commands | priority.medium :2nd_place_medal: type.bug :bug: | The `parseCommand` method of `FinanceTrackerParser` can throw a `ParseException` for the following main reasons:
1. The command word is unknown.
2. The user is entering a command on a tab that is not applicable (e.g. `add` on Overview or Analytics tab)
3. The user enters a valid command word, but the arguments cannot be parsed (either a malformed argument or data not conforming to validation rules).
For the `add` command, exceptions for reason 2 take precedence over reason 3 as the parsing is done after checking the current tab. However, this is not consistent with the behavior of other commands such as `edit`, where parsing is done before checking the current tab, resulting in exceptions under reason 3 being thrown before reason 2.
We should standardize this behavior. In particular, it should follow the `add` command behavior as it does not make sense to parse the arguments to a command that is not even applicable to the current tab in the first place. This may result in the parser application being duplicated across different `switch` cases, but consistency is more important. | 1.0 | Standardize parse exception precedence across commands - The `parseCommand` method of `FinanceTrackerParser` can throw a `ParseException` for the following main reasons:
1. The command word is unknown.
2. The user is entering a command on a tab that is not applicable (e.g. `add` on Overview or Analytics tab)
3. The user enters a valid command word, but the arguments cannot be parsed (either a malformed argument or data not conforming to validation rules).
For the `add` command, exceptions for reason 2 take precedence over reason 3 as the parsing is done after checking the current tab. However, this is not consistent with the behavior of other commands such as `edit`, where parsing is done before checking the current tab, resulting in exceptions under reason 3 being thrown before reason 2.
We should standardize this behavior. In particular, it should follow the `add` command behavior as it does not make sense to parse the arguments to a command that is not even applicable to the current tab in the first place. This may result in the parser application being duplicated across different `switch` cases, but consistency is more important. | priority | standardize parse exception precedence across commands the parsecommand method of financetrackerparser can throw a parseexception for the following main reasons the command word is unknown the user is entering a command on a tab that is not applicable e g add on overview or analytics tab the user enters a valid command word but the arguments cannot be parsed either a malformed argument or data not conforming to validation rules for the add command exceptions for reason take precedence over reason as the parsing is done after checking the current tab however this is not consistent with the behavior of other commands such as edit where parsing is done before checking the current tab resulting in exceptions under reason being thrown before reason we should standardize this behavior in particular it should follow the add command behavior as it does not make sense to parse the arguments to a command that is not even applicable to the current tab in the first place this may result in the parser application being duplicated across different switch cases but consistency is more important | 1 |
21,852 | 14,899,722,708 | IssuesEvent | 2021-01-21 14:39:31 | Automattic/jetpack | https://api.github.com/repos/Automattic/jetpack | closed | Action Required: Fix Renovate Configuration | [Pri] High [Type] Bug [Type] Infrastructure | There is an error with this repository's Renovate configuration that needs to be fixed. As a precaution, Renovate will stop PRs until it is resolved.
Error type: undefined. Note: this is a *nested* preset so please contact the preset author if you are unable to fix it yourself.
| 1.0 | Action Required: Fix Renovate Configuration - There is an error with this repository's Renovate configuration that needs to be fixed. As a precaution, Renovate will stop PRs until it is resolved.
Error type: undefined. Note: this is a *nested* preset so please contact the preset author if you are unable to fix it yourself.
| non_priority | action required fix renovate configuration there is an error with this repository s renovate configuration that needs to be fixed as a precaution renovate will stop prs until it is resolved error type undefined note this is a nested preset so please contact the preset author if you are unable to fix it yourself | 0 |
54,074 | 11,184,896,506 | IssuesEvent | 2019-12-31 20:54:33 | home-assistant/home-assistant | https://api.github.com/repos/home-assistant/home-assistant | closed | alarmdecoder pyserial requirement breaks insteonplm | integration: alarmdecoder integration: insteon | <!-- READ THIS FIRST:
- If you need additional help with this template please refer to https://www.home-assistant.io/help/reporting_issues/
- Make sure you are running the latest version of Home Assistant before reporting an issue: https://github.com/home-assistant/home-assistant/releases
- Frontend issues should be submitted to the home-assistant-polymer repository: https://github.com/home-assistant/home-assistant-polymer/issues
- iOS issues should be submitted to the home-assistant-iOS repository: https://github.com/home-assistant/home-assistant-iOS/issues
- Do not report issues for integrations if you are using a custom integration: files in <config-dir>/custom_components
- This is for bugs only. Feature and enhancement requests should go in our community forum: https://community.home-assistant.io/c/feature-requests
- Provide as many details as possible. Paste logs, configuration sample and code into the backticks. Do not delete any text from this template!
-->
**Home Assistant release with the issue:**
<!--
- Frontend -> Developer tools -> Info
- Or use this command: hass --version
-->
0.102.3
**Last working Home Assistant release (if known):**
n/a
**Operating environment (Hass.io/Docker/Windows/etc.):**
Debian 10 virtualenv
**Integration:**
https://www.home-assistant.io/integrations/insteon/
https://www.home-assistant.io/integrations/alarmdecoder/
**Description of problem:**
The insteonplm integration requires pyserial==3.2.0, but the alarmdecoder integration requires 2.7. 2.7 is incompatible with the socket:// method used to talk to an insteon hub version 1 that is used in pyserial-asyncio. (This is validated by trying to run the test in that library directly). It does however work in 3.0 and above. However, when you try to install, alarmdecoder complains that you are going to break it.
Turns out, this is not true. alarmdecoder works just fine with 3.2.0, and now my insteon works too.
Please bump the requirement for alarmdecoder for pyserial to 3.2.0 to fix this incompatibility, so both will work together.
```
I generated this while testing different versions:
insteonplm 0.16.5 has requirement pyserial==3.2.0, but you'll have pyserial 3.1.1 which is incompatible.
alarmdecoder 1.13.2 has requirement pyserial==2.7, but you'll have pyserial 3.1.1 which is incompatible.
```
**Problem-relevant `configuration.yaml` entries and (fill out even if it seems unimportant):**
```yaml
insteon:
host: xxx
hub_version: 1
```
**Traceback (if applicable):**
```
2019-12-11 19:27:35 INFO (MainThread) [insteonplm.plm] Connection established to PLM
2019-12-11 19:27:35 ERROR (MainThread) [homeassistant.core] Error doing job: Exception in callback SerialTransport._ensure_reader()
Traceback (most recent call last):
File "/usr/lib/python3.7/asyncio/events.py", line 88, in _run
self._context.run(self._callback, *self._args)
File "/opt/hass/lib/python3.7/site-packages/serial_asyncio/__init__.py", line 304, in _ensure_reader
self._loop.add_reader(self._serial.fileno(), self._read_ready)
io.UnsupportedOperation: fileno
```
**Additional information:**
| 1.0 | alarmdecoder pyserial requirement breaks insteonplm - <!-- READ THIS FIRST:
- If you need additional help with this template please refer to https://www.home-assistant.io/help/reporting_issues/
- Make sure you are running the latest version of Home Assistant before reporting an issue: https://github.com/home-assistant/home-assistant/releases
- Frontend issues should be submitted to the home-assistant-polymer repository: https://github.com/home-assistant/home-assistant-polymer/issues
- iOS issues should be submitted to the home-assistant-iOS repository: https://github.com/home-assistant/home-assistant-iOS/issues
- Do not report issues for integrations if you are using a custom integration: files in <config-dir>/custom_components
- This is for bugs only. Feature and enhancement requests should go in our community forum: https://community.home-assistant.io/c/feature-requests
- Provide as many details as possible. Paste logs, configuration sample and code into the backticks. Do not delete any text from this template!
-->
**Home Assistant release with the issue:**
<!--
- Frontend -> Developer tools -> Info
- Or use this command: hass --version
-->
0.102.3
**Last working Home Assistant release (if known):**
n/a
**Operating environment (Hass.io/Docker/Windows/etc.):**
Debian 10 virtualenv
**Integration:**
https://www.home-assistant.io/integrations/insteon/
https://www.home-assistant.io/integrations/alarmdecoder/
**Description of problem:**
The insteonplm integration requires pyserial==3.2.0, but the alarmdecoder integration requires 2.7. 2.7 is incompatible with the socket:// method used to talk to an insteon hub version 1 that is used in pyserial-asyncio. (This is validated by trying to run the test in that library directly). It does however work in 3.0 and above. However, when you try to install, alarmdecoder complains that you are going to break it.
Turns out, this is not true. alarmdecoder works just fine with 3.2.0, and now my insteon works too.
Please bump the requirement for alarmdecoder for pyserial to 3.2.0 to fix this incompatibility, so both will work together.
```
I generated this while testing different versions:
insteonplm 0.16.5 has requirement pyserial==3.2.0, but you'll have pyserial 3.1.1 which is incompatible.
alarmdecoder 1.13.2 has requirement pyserial==2.7, but you'll have pyserial 3.1.1 which is incompatible.
```
**Problem-relevant `configuration.yaml` entries and (fill out even if it seems unimportant):**
```yaml
insteon:
host: xxx
hub_version: 1
```
**Traceback (if applicable):**
```
2019-12-11 19:27:35 INFO (MainThread) [insteonplm.plm] Connection established to PLM
2019-12-11 19:27:35 ERROR (MainThread) [homeassistant.core] Error doing job: Exception in callback SerialTransport._ensure_reader()
Traceback (most recent call last):
File "/usr/lib/python3.7/asyncio/events.py", line 88, in _run
self._context.run(self._callback, *self._args)
File "/opt/hass/lib/python3.7/site-packages/serial_asyncio/__init__.py", line 304, in _ensure_reader
self._loop.add_reader(self._serial.fileno(), self._read_ready)
io.UnsupportedOperation: fileno
```
**Additional information:**
| non_priority | alarmdecoder pyserial requirement breaks insteonplm read this first if you need additional help with this template please refer to make sure you are running the latest version of home assistant before reporting an issue frontend issues should be submitted to the home assistant polymer repository ios issues should be submitted to the home assistant ios repository do not report issues for integrations if you are using a custom integration files in custom components this is for bugs only feature and enhancement requests should go in our community forum provide as many details as possible paste logs configuration sample and code into the backticks do not delete any text from this template home assistant release with the issue frontend developer tools info or use this command hass version last working home assistant release if known n a operating environment hass io docker windows etc debian virtualenv integration description of problem the insteonplm integration requires pyserial but the alarmdecoder integration requires is incompatible with the socket method used to talk to an insteon hub version that is used in pyserial asyncio this is validated by trying to run the test in that library directly it does however work in and above however when you try to install alarmdecoder complains that you are going to break it turns out this is not true alarmdecoder works just fine with and now my insteon works too please bump the requirement for alarmdecoder for pyserial to to fix this incompatibility so both will work together i generated this while testing different versions insteonplm has requirement pyserial but you ll have pyserial which is incompatible alarmdecoder has requirement pyserial but you ll have pyserial which is incompatible problem relevant configuration yaml entries and fill out even if it seems unimportant yaml insteon host xxx hub version traceback if applicable info mainthread connection established to plm error mainthread error doing job exception in callback serialtransport ensure reader traceback most recent call last file usr lib asyncio events py line in run self context run self callback self args file opt hass lib site packages serial asyncio init py line in ensure reader self loop add reader self serial fileno self read ready io unsupportedoperation fileno additional information | 0 |
182,588 | 6,671,360,882 | IssuesEvent | 2017-10-04 06:51:02 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.dailymotion.com - video or audio doesn't play | browser-firefox priority-important status-needstriage | <!-- @browser: Firefox 57.0 -->
<!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:57.0) Gecko/20100101 Firefox/57.0 -->
<!-- @reported_with: web -->
**URL**: http://www.dailymotion.com/video/x62roq9
**Browser / Version**: Firefox 57.0
**Operating System**: Linux
**Tested Another Browser**: Unknown
**Problem type**: Video or audio doesn't play
**Description**: video does not play
**Steps to Reproduce**:
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.dailymotion.com - video or audio doesn't play - <!-- @browser: Firefox 57.0 -->
<!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:57.0) Gecko/20100101 Firefox/57.0 -->
<!-- @reported_with: web -->
**URL**: http://www.dailymotion.com/video/x62roq9
**Browser / Version**: Firefox 57.0
**Operating System**: Linux
**Tested Another Browser**: Unknown
**Problem type**: Video or audio doesn't play
**Description**: video does not play
**Steps to Reproduce**:
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | video or audio doesn t play url browser version firefox operating system linux tested another browser unknown problem type video or audio doesn t play description video does not play steps to reproduce from with ❤️ | 1 |
476,504 | 13,745,803,495 | IssuesEvent | 2020-10-06 03:56:47 | KSP-SpaceDock/SpaceDock | https://api.github.com/repos/KSP-SpaceDock/SpaceDock | closed | Request: Allow links to download always latest version | Priority: Low Type: Backend Type: Feature | The "Download" button would direct to "latest" version - which changes dynamically.
Example: Mod A is version A at https://spacedock.info/mod/xxx/A
Download button creates https://spacedock.info/mod/xxx/A/download/A
Then author uploads Version B
Original link becomes outdated
Download button creates https://spacedock.info/mod/xxx/A/download/B
After update:
Example: Mod B is version A at https://spacedock.info/mod/yyy/B
Download button creates https://spacedock.info/mod/yyy/B/downloadlatest
Then author uploads Version B
Original link is still up to date
Download button creates https://spacedock.info/mod/yyy/B/downloadlatest | 1.0 | Request: Allow links to download always latest version - The "Download" button would direct to "latest" version - which changes dynamically.
Example: Mod A is version A at https://spacedock.info/mod/xxx/A
Download button creates https://spacedock.info/mod/xxx/A/download/A
Then author uploads Version B
Original link becomes outdated
Download button creates https://spacedock.info/mod/xxx/A/download/B
After update:
Example: Mod B is version A at https://spacedock.info/mod/yyy/B
Download button creates https://spacedock.info/mod/yyy/B/downloadlatest
Then author uploads Version B
Original link is still up to date
Download button creates https://spacedock.info/mod/yyy/B/downloadlatest | priority | request allow links to download always latest version the download button would direct to latest version which changes dynamically example mod a is version a at download button creates then author uploads version b original link becomes outdated download button creates after update example mod b is version a at download button creates then author uploads version b original link is still up to date download button creates | 1 |
105,926 | 4,249,557,900 | IssuesEvent | 2016-07-08 00:39:58 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | opened | [Feature] Integrate closely with flannel and/or Calico, via network plugins. | area/networking kind/feature priority/P1 team/cluster | Integrate closely with flannel and/or Calico, via network plugins.
| 1.0 | [Feature] Integrate closely with flannel and/or Calico, via network plugins. - Integrate closely with flannel and/or Calico, via network plugins.
| priority | integrate closely with flannel and or calico via network plugins integrate closely with flannel and or calico via network plugins | 1 |
292,835 | 22,037,466,751 | IssuesEvent | 2022-05-28 20:32:33 | EquipoBIT/ProyectoLTI | https://api.github.com/repos/EquipoBIT/ProyectoLTI | closed | Etapa 2: Análisis de Ideas | documentation Epic | Se deberá utilizar el template disponible debajo y realizar la entrega en formato pdf.
Descripción de la etapa
Cada equipo deberá realizar el análisis de las ideas-problemáticas entregadas en la etapa 1, debiéndose indicar un orden de preferencia entre ellas (se sugiere enfocarse en la idea/necesidad utilizada en las actividades de la última presencial).
El siguiente mapa conceptual ejemplifica el proceso que cada equipo deberá seguir a la hora de analizar cada una de ellas:
Ideas
Para cada idea, se deberá completar una ficha (ver template publicada), donde se detalle de forma clara y concisa la siguiente información (que se corresponde al mapa conceptual anterior):
1. ¿Cuál es el problema o la necesidad detectada?
2. ¿Quiénes sufren la necesidad o el problema detectado?
3. ¿Cuál es la propuesta de valor?
4. ¿Cuáles son las actividades que considera claves para que la iniciativa sea exitosa?
5. ¿Cuáles son los principales riesgos o razones por los que el proyecto podría llegar a no tener éxito?
A la hora de elaborar las respuestas, será necesario remitirse al contenido solicitado en las preguntas, lo cual resulta clave para el posterior desarrollo de las etapas subsiguientes.
¿Qué sigue luego de la entrega?
Una vez realizada la entrega, por medio de una revisión conjunta entre la coordinación de la carrera y el tutor a cargo del Anteproyecto se determinará cual de las ideas presentadas se adecua mejor a los objetivos de esta unidad curricular y los conocimientos-habilidades que se pretenden desarrollar.
En base a ello, y tratando de contemplar el orden de preferencia indicado en este entrega, se aprobará una de las ideas presentadas sobre la cual el equipo trabajará a lo largo del anteproyecto y posteriormente el Proyecto Final de Carrera.
Sin perjuicio de lo anterior, se le podrá solicitar al equipo que profundice sobre una de las ideas, condicionando la aprobación de la misma a ello. | 1.0 | Etapa 2: Análisis de Ideas - Se deberá utilizar el template disponible debajo y realizar la entrega en formato pdf.
Descripción de la etapa
Cada equipo deberá realizar el análisis de las ideas-problemáticas entregadas en la etapa 1, debiéndose indicar un orden de preferencia entre ellas (se sugiere enfocarse en la idea/necesidad utilizada en las actividades de la última presencial).
El siguiente mapa conceptual ejemplifica el proceso que cada equipo deberá seguir a la hora de analizar cada una de ellas:
Ideas
Para cada idea, se deberá completar una ficha (ver template publicada), donde se detalle de forma clara y concisa la siguiente información (que se corresponde al mapa conceptual anterior):
1. ¿Cuál es el problema o la necesidad detectada?
2. ¿Quiénes sufren la necesidad o el problema detectado?
3. ¿Cuál es la propuesta de valor?
4. ¿Cuáles son las actividades que considera claves para que la iniciativa sea exitosa?
5. ¿Cuáles son los principales riesgos o razones por los que el proyecto podría llegar a no tener éxito?
A la hora de elaborar las respuestas, será necesario remitirse al contenido solicitado en las preguntas, lo cual resulta clave para el posterior desarrollo de las etapas subsiguientes.
¿Qué sigue luego de la entrega?
Una vez realizada la entrega, por medio de una revisión conjunta entre la coordinación de la carrera y el tutor a cargo del Anteproyecto se determinará cual de las ideas presentadas se adecua mejor a los objetivos de esta unidad curricular y los conocimientos-habilidades que se pretenden desarrollar.
En base a ello, y tratando de contemplar el orden de preferencia indicado en este entrega, se aprobará una de las ideas presentadas sobre la cual el equipo trabajará a lo largo del anteproyecto y posteriormente el Proyecto Final de Carrera.
Sin perjuicio de lo anterior, se le podrá solicitar al equipo que profundice sobre una de las ideas, condicionando la aprobación de la misma a ello. | non_priority | etapa análisis de ideas se deberá utilizar el template disponible debajo y realizar la entrega en formato pdf descripción de la etapa cada equipo deberá realizar el análisis de las ideas problemáticas entregadas en la etapa debiéndose indicar un orden de preferencia entre ellas se sugiere enfocarse en la idea necesidad utilizada en las actividades de la última presencial el siguiente mapa conceptual ejemplifica el proceso que cada equipo deberá seguir a la hora de analizar cada una de ellas ideas para cada idea se deberá completar una ficha ver template publicada donde se detalle de forma clara y concisa la siguiente información que se corresponde al mapa conceptual anterior ¿cuál es el problema o la necesidad detectada ¿quiénes sufren la necesidad o el problema detectado ¿cuál es la propuesta de valor ¿cuáles son las actividades que considera claves para que la iniciativa sea exitosa ¿cuáles son los principales riesgos o razones por los que el proyecto podría llegar a no tener éxito a la hora de elaborar las respuestas será necesario remitirse al contenido solicitado en las preguntas lo cual resulta clave para el posterior desarrollo de las etapas subsiguientes ¿qué sigue luego de la entrega una vez realizada la entrega por medio de una revisión conjunta entre la coordinación de la carrera y el tutor a cargo del anteproyecto se determinará cual de las ideas presentadas se adecua mejor a los objetivos de esta unidad curricular y los conocimientos habilidades que se pretenden desarrollar en base a ello y tratando de contemplar el orden de preferencia indicado en este entrega se aprobará una de las ideas presentadas sobre la cual el equipo trabajará a lo largo del anteproyecto y posteriormente el proyecto final de carrera sin perjuicio de lo anterior se le podrá solicitar al equipo que profundice sobre una de las ideas condicionando la aprobación de la misma a ello | 0 |
233,069 | 7,692,922,898 | IssuesEvent | 2018-05-18 00:18:01 | metabase/metabase | https://api.github.com/repos/metabase/metabase | closed | Oh, Start failed after I upgraded to v0.26.1 from v0.25.2 | Bug Priority/P1 Running Metabase | I toke much time to do my upgrade, but when i start dev, the metabase occured an error so thoroughly.
my environments:
```bash
mac: 10.11.1 (15B42)
metabse: v0.25.2 -> 0.26.1
```
What error happened as below:

Why the foreign key `fk_rawtable_ref_database` does not exist? And I scaned all tables in metabase . The result is I cannot found any constraint or fk named `fk_rawtable_ref_database`. And so I attemped to comment the code in `resources/migrations/000_migrations.yml` which changeSet's id is 64.
<img width="929" alt="2121" src="https://user-images.githubusercontent.com/12776545/31752948-4a9af700-b453-11e7-948b-3df744379743.png">
to ----- as below:
<img width="936" alt="222" src="https://user-images.githubusercontent.com/12776545/31752946-4a54f408-b453-11e7-8ae8-f3326f028afe.png">
And I run the command :
```bash
yarn run dev
```
Successed

**But How did this happen?**
| 1.0 | Oh, Start failed after I upgraded to v0.26.1 from v0.25.2 - I toke much time to do my upgrade, but when i start dev, the metabase occured an error so thoroughly.
my environments:
```bash
mac: 10.11.1 (15B42)
metabse: v0.25.2 -> 0.26.1
```
What error happened as below:

Why the foreign key `fk_rawtable_ref_database` does not exist? And I scaned all tables in metabase . The result is I cannot found any constraint or fk named `fk_rawtable_ref_database`. And so I attemped to comment the code in `resources/migrations/000_migrations.yml` which changeSet's id is 64.
<img width="929" alt="2121" src="https://user-images.githubusercontent.com/12776545/31752948-4a9af700-b453-11e7-948b-3df744379743.png">
to ----- as below:
<img width="936" alt="222" src="https://user-images.githubusercontent.com/12776545/31752946-4a54f408-b453-11e7-8ae8-f3326f028afe.png">
And I run the command :
```bash
yarn run dev
```
Successed

**But How did this happen?**
| priority | oh start failed after i upgraded to from i toke much time to do my upgrade but when i start dev the metabase occured an error so thoroughly my environments bash mac metabse what error happened as below why the foreign key fk rawtable ref database does not exist and i scaned all tables in metabase the result is i cannot found any constraint or fk named fk rawtable ref database and so i attemped to comment the code in resources migrations migrations yml which changeset s id is img width alt src to as below img width alt src and i run the command bash yarn run dev successed but how did this happen | 1 |
58,763 | 14,483,142,047 | IssuesEvent | 2020-12-10 14:49:46 | kubevirt/containerized-data-importer | https://api.github.com/repos/kubevirt/containerized-data-importer | closed | [Flaky CI] DataVolume tests Verify DataVolume should succeed creating import dv with given tar archive url. [rfe_id:1947][crit:high][test_id:2145] | kind/bug triage/build-watcher | /kind bug
/triage build-officer
**What happened**:
FlakeFinder reported a flaky test:
```
DataVolume tests Verify DataVolume should
[rfe_id:1947][crit:high][test_id:2220]fail creating import dv with non tar archive url
/root/go/src/kubevirt.io/containerized-data-importer/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:43
STEP: Building a "dv-func-test" namespace api object
INFO: Created new namespace "cdi-e2e-tests-dv-func-test-2s58d"
STEP: Configuring default FeatureGates ["HonorWaitForFirstConsumer"]
STEP: creating new datavolume dv-non-tar-archive
verifying pvc was created for dv dv-non-tar-archive
STEP: waiting for datavolume to match phase ImportInProgress
STEP: Verifying the DV has the correct conditions and messages for those conditions
INFO: Condition.Reason does not match for type: Ready, reason expected [], reason found: [TransferRunning]
STEP: verifying pvc was created
STEP: Verifying event occurred
LAST SEEN TYPE REASON OBJECT MESSAGE
10s Normal Pending datavolume/dv-non-tar-archive PVC dv-non-tar-archive Pending
7s Normal Bound datavolume/dv-non-tar-archive PVC dv-non-tar-archive Bound
9s Normal CreatingScratchSpace datavolume/dv-non-tar-archive Creating scratch space
9s Normal Claim Pending datavolume/dv-non-tar-archive Claim Pending
3s Normal ImportInProgress datavolume/dv-non-tar-archive Import into dv-non-tar-archive in progress
2s Warning Error datavolume/dv-non-tar-archive Unable to process data: exit status 2
```
**[Prow Job log](https://prow.apps.ovirt.org/view/gcs/kubevirt-prow/pr-logs/pull/kubevirt_containerized-data-importer/1320/pull-containerized-data-importer-e2e-k8s-1.17-ceph/1304433405550661632)**
**What you expected to happen**:
Test should pass
**How to reproduce it (as minimally and precisely as possible)**:
Check logs of prow job
**Anything else we need to know?**:
This is a flaky test which passes on other runs. Check what might cause that DataVolume to alternately fail.
| 1.0 | [Flaky CI] DataVolume tests Verify DataVolume should succeed creating import dv with given tar archive url. [rfe_id:1947][crit:high][test_id:2145] - /kind bug
/triage build-officer
**What happened**:
FlakeFinder reported a flaky test:
```
DataVolume tests Verify DataVolume should
[rfe_id:1947][crit:high][test_id:2220]fail creating import dv with non tar archive url
/root/go/src/kubevirt.io/containerized-data-importer/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:43
STEP: Building a "dv-func-test" namespace api object
INFO: Created new namespace "cdi-e2e-tests-dv-func-test-2s58d"
STEP: Configuring default FeatureGates ["HonorWaitForFirstConsumer"]
STEP: creating new datavolume dv-non-tar-archive
verifying pvc was created for dv dv-non-tar-archive
STEP: waiting for datavolume to match phase ImportInProgress
STEP: Verifying the DV has the correct conditions and messages for those conditions
INFO: Condition.Reason does not match for type: Ready, reason expected [], reason found: [TransferRunning]
STEP: verifying pvc was created
STEP: Verifying event occurred
LAST SEEN TYPE REASON OBJECT MESSAGE
10s Normal Pending datavolume/dv-non-tar-archive PVC dv-non-tar-archive Pending
7s Normal Bound datavolume/dv-non-tar-archive PVC dv-non-tar-archive Bound
9s Normal CreatingScratchSpace datavolume/dv-non-tar-archive Creating scratch space
9s Normal Claim Pending datavolume/dv-non-tar-archive Claim Pending
3s Normal ImportInProgress datavolume/dv-non-tar-archive Import into dv-non-tar-archive in progress
2s Warning Error datavolume/dv-non-tar-archive Unable to process data: exit status 2
```
**[Prow Job log](https://prow.apps.ovirt.org/view/gcs/kubevirt-prow/pr-logs/pull/kubevirt_containerized-data-importer/1320/pull-containerized-data-importer-e2e-k8s-1.17-ceph/1304433405550661632)**
**What you expected to happen**:
Test should pass
**How to reproduce it (as minimally and precisely as possible)**:
Check logs of prow job
**Anything else we need to know?**:
This is a flaky test which passes on other runs. Check what might cause that DataVolume to alternately fail.
| non_priority | datavolume tests verify datavolume should succeed creating import dv with given tar archive url kind bug triage build officer what happened flakefinder reported a flaky test datavolume tests verify datavolume should fail creating import dv with non tar archive url root go src kubevirt io containerized data importer vendor github com onsi ginkgo extensions table table entry go step building a dv func test namespace api object info created new namespace cdi tests dv func test step configuring default featuregates step creating new datavolume dv non tar archive verifying pvc was created for dv dv non tar archive step waiting for datavolume to match phase importinprogress step verifying the dv has the correct conditions and messages for those conditions info condition reason does not match for type ready reason expected reason found step verifying pvc was created step verifying event occurred last seen type reason object message normal pending datavolume dv non tar archive pvc dv non tar archive pending normal bound datavolume dv non tar archive pvc dv non tar archive bound normal creatingscratchspace datavolume dv non tar archive creating scratch space normal claim pending datavolume dv non tar archive claim pending normal importinprogress datavolume dv non tar archive import into dv non tar archive in progress warning error datavolume dv non tar archive unable to process data exit status what you expected to happen test should pass how to reproduce it as minimally and precisely as possible check logs of prow job anything else we need to know this is a flaky test which passes on other runs check what might cause that datavolume to alternately fail | 0 |
56,271 | 3,078,738,028 | IssuesEvent | 2015-08-21 12:29:36 | pavel-pimenov/flylinkdc-r5xx | https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx | opened | Быстрое отключение чата (индивидуальное для каждого хаба, аналогичное отключению списка пользователей) | Close Component-Logic Component-UI enhancement imported Priority-Medium Usability | _From [reaor...@gmail.com](https://code.google.com/u/102418317896447533964/) on August 18, 2011 15:23:34_
Общий чат не всегда нужен. Хотелось бы иметь возможность его отключить одной галочкой (как сейчас можно отключить список пользователей).
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=541_ | 1.0 | Быстрое отключение чата (индивидуальное для каждого хаба, аналогичное отключению списка пользователей) - _From [reaor...@gmail.com](https://code.google.com/u/102418317896447533964/) on August 18, 2011 15:23:34_
Общий чат не всегда нужен. Хотелось бы иметь возможность его отключить одной галочкой (как сейчас можно отключить список пользователей).
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=541_ | priority | быстрое отключение чата индивидуальное для каждого хаба аналогичное отключению списка пользователей from on august общий чат не всегда нужен хотелось бы иметь возможность его отключить одной галочкой как сейчас можно отключить список пользователей original issue | 1 |
7,902 | 7,065,296,246 | IssuesEvent | 2018-01-06 18:17:11 | mschlenstedt/Loxberry | https://api.github.com/repos/mschlenstedt/Loxberry | opened | Bridging des Miniservers am LB für Token Auth | Core LB Modules Security enhancement | Dabei geht es erstmal nicht um Websockets, sondern um REST mit Token Auth für eine verschlüsselte LB->MS Kommunikation.
Alles ist für Benutzer transparent. Plugin-Entwickler müssen kleine Modifikationen vornehmen.
Es sind drei Teile nötig:
# Lighty
- Für jeden angelegten MS gibt es einen (zufälligen) Port mit Website. Der Port wird zum MS in die general.cfg geschrieben.
- Nur lokale Verbindungen sind erlaubt.
- Alle Verbindungen müssen angenommen werden, egal ob anonym oder mit falschen! Credentials (es kommen die MS-Credentials)
- Alles, egal welche URL, muss an ein PHP-Script gehen.
- Die verwendete URL mit Parametern muss das Script bekommen.
# PHP-Script
- PHP deswegen, weil es schneller ist als der Perl-CGI-Interpreter.
- Aus get_miniservers werden die Credentials gelesen.
- Das Script holt einen Token und speichert diesen.
- Das Script schickt den nun verschlüsselten Request 1:1 an den jeweiligen Miniserver
- Ist beim Request Der Token abgelaufen , wird ein neuer bezogen.
- Unterstützt der MS keine Token-Auth, wird ohne weitergemacht.
# Module
- get_miniservers liefert den virtuellen Host und Port zurück, wo Plugins hinverbinden müssen.
Für Plugin-Entwickler ist die Änderung folgende:
- Sie holen die Miniserver-Infos per get_miniserver.
- Für die Verbindung werden einfach Host und Port durch die Variablen ersetzt.
- Die URL selbst bleibt gleich, also praktisch keine Änderung am Plugin und die Verbindung ist dann verschlüsselt. | True | Bridging des Miniservers am LB für Token Auth - Dabei geht es erstmal nicht um Websockets, sondern um REST mit Token Auth für eine verschlüsselte LB->MS Kommunikation.
Alles ist für Benutzer transparent. Plugin-Entwickler müssen kleine Modifikationen vornehmen.
Es sind drei Teile nötig:
# Lighty
- Für jeden angelegten MS gibt es einen (zufälligen) Port mit Website. Der Port wird zum MS in die general.cfg geschrieben.
- Nur lokale Verbindungen sind erlaubt.
- Alle Verbindungen müssen angenommen werden, egal ob anonym oder mit falschen! Credentials (es kommen die MS-Credentials)
- Alles, egal welche URL, muss an ein PHP-Script gehen.
- Die verwendete URL mit Parametern muss das Script bekommen.
# PHP-Script
- PHP deswegen, weil es schneller ist als der Perl-CGI-Interpreter.
- Aus get_miniservers werden die Credentials gelesen.
- Das Script holt einen Token und speichert diesen.
- Das Script schickt den nun verschlüsselten Request 1:1 an den jeweiligen Miniserver
- Ist beim Request Der Token abgelaufen , wird ein neuer bezogen.
- Unterstützt der MS keine Token-Auth, wird ohne weitergemacht.
# Module
- get_miniservers liefert den virtuellen Host und Port zurück, wo Plugins hinverbinden müssen.
Für Plugin-Entwickler ist die Änderung folgende:
- Sie holen die Miniserver-Infos per get_miniserver.
- Für die Verbindung werden einfach Host und Port durch die Variablen ersetzt.
- Die URL selbst bleibt gleich, also praktisch keine Änderung am Plugin und die Verbindung ist dann verschlüsselt. | non_priority | bridging des miniservers am lb für token auth dabei geht es erstmal nicht um websockets sondern um rest mit token auth für eine verschlüsselte lb ms kommunikation alles ist für benutzer transparent plugin entwickler müssen kleine modifikationen vornehmen es sind drei teile nötig lighty für jeden angelegten ms gibt es einen zufälligen port mit website der port wird zum ms in die general cfg geschrieben nur lokale verbindungen sind erlaubt alle verbindungen müssen angenommen werden egal ob anonym oder mit falschen credentials es kommen die ms credentials alles egal welche url muss an ein php script gehen die verwendete url mit parametern muss das script bekommen php script php deswegen weil es schneller ist als der perl cgi interpreter aus get miniservers werden die credentials gelesen das script holt einen token und speichert diesen das script schickt den nun verschlüsselten request an den jeweiligen miniserver ist beim request der token abgelaufen wird ein neuer bezogen unterstützt der ms keine token auth wird ohne weitergemacht module get miniservers liefert den virtuellen host und port zurück wo plugins hinverbinden müssen für plugin entwickler ist die änderung folgende sie holen die miniserver infos per get miniserver für die verbindung werden einfach host und port durch die variablen ersetzt die url selbst bleibt gleich also praktisch keine änderung am plugin und die verbindung ist dann verschlüsselt | 0 |
412,490 | 12,042,934,613 | IssuesEvent | 2020-04-14 11:30:31 | threefoldfoundation/tft-stellar | https://api.github.com/repos/threefoldfoundation/tft-stellar | closed | Better UI for the FreeTFT faucet | priority_major type_feature | Currently it looks like
<img width="559" alt="Screenshot 2020-04-07 at 18 22 17" src="https://user-images.githubusercontent.com/3511318/78716050-968ca180-791e-11ea-8c31-9f2fa6dec050.png">
| 1.0 | Better UI for the FreeTFT faucet - Currently it looks like
<img width="559" alt="Screenshot 2020-04-07 at 18 22 17" src="https://user-images.githubusercontent.com/3511318/78716050-968ca180-791e-11ea-8c31-9f2fa6dec050.png">
| priority | better ui for the freetft faucet currently it looks like img width alt screenshot at src | 1 |
68,188 | 7,088,884,194 | IssuesEvent | 2018-01-11 23:22:35 | Iridescent-CM/technovation-app | https://api.github.com/repos/Iridescent-CM/technovation-app | closed | Identify and link each piece with its corresponding curriculum pages | 4 - Test <= 8 [sprint topic] submissions added during sprint | See #1381
<!---
@huboard:{"order":2.4095278261974814e-31,"milestone_order":7.89525459899401e-45,"custom_state":""}
-->
| 1.0 | Identify and link each piece with its corresponding curriculum pages - See #1381
<!---
@huboard:{"order":2.4095278261974814e-31,"milestone_order":7.89525459899401e-45,"custom_state":""}
-->
| non_priority | identify and link each piece with its corresponding curriculum pages see huboard order milestone order custom state | 0 |
243,329 | 20,378,834,836 | IssuesEvent | 2022-02-21 18:40:11 | task-tim/Chere-Sasha | https://api.github.com/repos/task-tim/Chere-Sasha | closed | 2.1 - Prototype pour tester le contrôle de son à l'aide de Max et Arduino | bug test case priority : high | - [x] **2.1 - 1** Faire le code Arduino
- [x] **2.1 - 2** Faire le code Max
- [x] **2.1 - 3** Connecter les 2 ensemble pour voir si le tout marche ensemble
SUCCÈS | 1.0 | 2.1 - Prototype pour tester le contrôle de son à l'aide de Max et Arduino - - [x] **2.1 - 1** Faire le code Arduino
- [x] **2.1 - 2** Faire le code Max
- [x] **2.1 - 3** Connecter les 2 ensemble pour voir si le tout marche ensemble
SUCCÈS | non_priority | prototype pour tester le contrôle de son à l aide de max et arduino faire le code arduino faire le code max connecter les ensemble pour voir si le tout marche ensemble succès | 0 |
413,085 | 27,943,832,827 | IssuesEvent | 2023-03-24 00:13:17 | Angelus123/abidors-nodejs-ch | https://api.github.com/repos/Angelus123/abidors-nodejs-ch | opened | Authentication and JSON patching | documentation | #### Public Endpoints
- Login
Request body should contain an arbitrary username/password pair
Treat it as a mock authentication service and accept any username/password.
Return a signed Json Web Token(JWT, https://jwt.io/) which can be used to validate future requests.
#### Protected Endpoints
The following two endpoints should be protected. The JWT obtained in the “Login” endpoint must be attached to each request header. If the JWT is missing or invalid, these endpoints should reject the request with the appropriate message to the user.
- Apply Json Patch
Request body should contain a JSON object and a JSON patch object (https://jsonpatch.com/).
Apply the json patch to the json object, and return the resulting json object. | 1.0 | Authentication and JSON patching - #### Public Endpoints
- Login
Request body should contain an arbitrary username/password pair
Treat it as a mock authentication service and accept any username/password.
Return a signed Json Web Token(JWT, https://jwt.io/) which can be used to validate future requests.
#### Protected Endpoints
The following two endpoints should be protected. The JWT obtained in the “Login” endpoint must be attached to each request header. If the JWT is missing or invalid, these endpoints should reject the request with the appropriate message to the user.
- Apply Json Patch
Request body should contain a JSON object and a JSON patch object (https://jsonpatch.com/).
Apply the json patch to the json object, and return the resulting json object. | non_priority | authentication and json patching public endpoints login request body should contain an arbitrary username password pair treat it as a mock authentication service and accept any username password return a signed json web token jwt which can be used to validate future requests protected endpoints the following two endpoints should be protected the jwt obtained in the “login” endpoint must be attached to each request header if the jwt is missing or invalid these endpoints should reject the request with the appropriate message to the user apply json patch request body should contain a json object and a json patch object apply the json patch to the json object and return the resulting json object | 0 |
87,004 | 3,735,988,747 | IssuesEvent | 2016-03-08 14:29:53 | PlanHubMe/PlanHub | https://api.github.com/repos/PlanHubMe/PlanHub | opened | HW view should handle invalid subjects properly | Feature: HW View medium priority | Somehow, some events in the database got "disconnected" from their subject. (they point to a subject number that doesn't exist) I'm not sure why this happens, but one side effect of this is that HW view stops loading. It errors out and stays "Loading..." forever, which isn't so good. We should probably display a message at the top or something warning the user and asking them to please contact support. | 1.0 | HW view should handle invalid subjects properly - Somehow, some events in the database got "disconnected" from their subject. (they point to a subject number that doesn't exist) I'm not sure why this happens, but one side effect of this is that HW view stops loading. It errors out and stays "Loading..." forever, which isn't so good. We should probably display a message at the top or something warning the user and asking them to please contact support. | priority | hw view should handle invalid subjects properly somehow some events in the database got disconnected from their subject they point to a subject number that doesn t exist i m not sure why this happens but one side effect of this is that hw view stops loading it errors out and stays loading forever which isn t so good we should probably display a message at the top or something warning the user and asking them to please contact support | 1 |
260,031 | 8,203,288,210 | IssuesEvent | 2018-09-02 19:34:59 | caelum100/Luminite | https://api.github.com/repos/caelum100/Luminite | closed | `combine_models` function doesn't offset indices | category: graphics category: models difficulty: easy priority: high type: bug | The `combine_ models` function doesn't offset indices when combining multiple models, resulting in weird geometry. | 1.0 | `combine_models` function doesn't offset indices - The `combine_ models` function doesn't offset indices when combining multiple models, resulting in weird geometry. | priority | combine models function doesn t offset indices the combine models function doesn t offset indices when combining multiple models resulting in weird geometry | 1 |
297,368 | 9,167,630,767 | IssuesEvent | 2019-03-02 15:25:45 | project-koku/koku-ui | https://api.github.com/repos/project-koku/koku-ui | opened | Some 'Add a Source' modal copy/paste text strings contain '#' | bug priority - medium | **Describe the bug**
Some 'Add a Source' modal copy/paste text strings contain '#'
**To Reproduce**
Steps to reproduce the behavior:
1. Go to 'Overview'
2. Click on 'Add a Source'
3. Navigate through OCP configuration
4. Notice some copy/paste strings start with '#'
5. Pasting and running a string starting with '#' will fail
**Expected behavior**
Strings should not contain '#'s
**Screenshots**


**Desktop (please complete the following information):**
- OS: Fedora
- Browser Chrome
- Version 66
**Additional context**
Testing Against CI insightsbeta
| 1.0 | Some 'Add a Source' modal copy/paste text strings contain '#' - **Describe the bug**
Some 'Add a Source' modal copy/paste text strings contain '#'
**To Reproduce**
Steps to reproduce the behavior:
1. Go to 'Overview'
2. Click on 'Add a Source'
3. Navigate through OCP configuration
4. Notice some copy/paste strings start with '#'
5. Pasting and running a string starting with '#' will fail
**Expected behavior**
Strings should not contain '#'s
**Screenshots**


**Desktop (please complete the following information):**
- OS: Fedora
- Browser Chrome
- Version 66
**Additional context**
Testing Against CI insightsbeta
| priority | some add a source modal copy paste text strings contain describe the bug some add a source modal copy paste text strings contain to reproduce steps to reproduce the behavior go to overview click on add a source navigate through ocp configuration notice some copy paste strings start with pasting and running a string starting with will fail expected behavior strings should not contain s screenshots desktop please complete the following information os fedora browser chrome version additional context testing against ci insightsbeta | 1 |
752,849 | 26,328,937,740 | IssuesEvent | 2023-01-10 09:19:21 | matrixorigin/matrixone | https://api.github.com/repos/matrixorigin/matrixone | opened | [Feature Request]: Support correlated subquery with multi levels | priority/p0 kind/feature source/on-demand | ### Is there an existing issue for the same feature request?
- [X] I have checked the existing issues.
### Is your feature request related to a problem?
_No response_
### Describe the feature you'd like
Currently, MO supports correlated subquery with only 1 level. To accomplish this feature, have to support correlated subquery with more than 1 level
### Describe implementation you've considered
_No response_
### Documentation, Adoption, Use Case, Migration Strategy
_No response_
### Additional information
_No response_ | 1.0 | [Feature Request]: Support correlated subquery with multi levels - ### Is there an existing issue for the same feature request?
- [X] I have checked the existing issues.
### Is your feature request related to a problem?
_No response_
### Describe the feature you'd like
Currently, MO supports correlated subquery with only 1 level. To accomplish this feature, have to support correlated subquery with more than 1 level
### Describe implementation you've considered
_No response_
### Documentation, Adoption, Use Case, Migration Strategy
_No response_
### Additional information
_No response_ | priority | support correlated subquery with multi levels is there an existing issue for the same feature request i have checked the existing issues is your feature request related to a problem no response describe the feature you d like currently mo supports correlated subquery with only level to accomplish this feature have to support correlated subquery with more than level describe implementation you ve considered no response documentation adoption use case migration strategy no response additional information no response | 1 |
60,668 | 3,131,399,679 | IssuesEvent | 2015-09-09 14:34:37 | OCHA-DAP/cod-migration | https://api.github.com/repos/OCHA-DAP/cod-migration | closed | Gospawn failing on node 31900 | priority-high | Failed on two different runs. Execution stops.
Terminal output:
```
Start processing dataset with Node ID = 31900
Government
Creating new dataset: sierra-leone-admin-level-1-boundaries
usage: gospawn.py test (short|full)
```
Log output:
```
2015-08-31T17:55:27+0300 | dataset | START | True | 31900 | None | https://www.humanitarianresponse.info/en/api/v1.0/datasets/31900 | None | None | None | None | None
2015-08-31T17:55:28+0300 | dataset | get | True | 31900 | https://www.humanitarianresponse.info/en/operations/west-and-central-africa/dataset/sierra-leone-admin-level-1-boundaries | https://www.humanitarianresponse.info/en/api/v1.0/datasets/31900 | None | None | None | None | None
2015-08-31T17:55:28+0300 | dataset | format | True | 31900 | https://www.humanitarianresponse.info/en/operations/west-and-central-africa/dataset/sierra-leone-admin-level-1-boundaries | https://www.humanitarianresponse.info/en/api/v1.0/datasets/31900 | None | Sierra Leone - Admin Level 1 Boundaries | None | None | None
2015-08-31T17:55:29+0300 | dataset | create | False | 31900 | https://www.humanitarianresponse.info/en/operations/west-and-central-africa/dataset/sierra-leone-admin-level-1-boundaries | https://www.humanitarianresponse.info/en/api/v1.0/datasets/31900 | None | Sierra Leone - Admin Level 1 Boundaries | None | https://test-data.hdx.rwlabs.org/dataset/sierra-leone-admin-level-1-boundaries | https://test-data.hdx.rwlabs.org/api/3/action/package_show?id=sierra-leone-admin-level-1-boundaries
``` | 1.0 | Gospawn failing on node 31900 - Failed on two different runs. Execution stops.
Terminal output:
```
Start processing dataset with Node ID = 31900
Government
Creating new dataset: sierra-leone-admin-level-1-boundaries
usage: gospawn.py test (short|full)
```
Log output:
```
2015-08-31T17:55:27+0300 | dataset | START | True | 31900 | None | https://www.humanitarianresponse.info/en/api/v1.0/datasets/31900 | None | None | None | None | None
2015-08-31T17:55:28+0300 | dataset | get | True | 31900 | https://www.humanitarianresponse.info/en/operations/west-and-central-africa/dataset/sierra-leone-admin-level-1-boundaries | https://www.humanitarianresponse.info/en/api/v1.0/datasets/31900 | None | None | None | None | None
2015-08-31T17:55:28+0300 | dataset | format | True | 31900 | https://www.humanitarianresponse.info/en/operations/west-and-central-africa/dataset/sierra-leone-admin-level-1-boundaries | https://www.humanitarianresponse.info/en/api/v1.0/datasets/31900 | None | Sierra Leone - Admin Level 1 Boundaries | None | None | None
2015-08-31T17:55:29+0300 | dataset | create | False | 31900 | https://www.humanitarianresponse.info/en/operations/west-and-central-africa/dataset/sierra-leone-admin-level-1-boundaries | https://www.humanitarianresponse.info/en/api/v1.0/datasets/31900 | None | Sierra Leone - Admin Level 1 Boundaries | None | https://test-data.hdx.rwlabs.org/dataset/sierra-leone-admin-level-1-boundaries | https://test-data.hdx.rwlabs.org/api/3/action/package_show?id=sierra-leone-admin-level-1-boundaries
``` | priority | gospawn failing on node failed on two different runs execution stops terminal output start processing dataset with node id government creating new dataset sierra leone admin level boundaries usage gospawn py test short full log output dataset start true none none none none none none dataset get true none none none none none dataset format true none sierra leone admin level boundaries none none none dataset create false none sierra leone admin level boundaries none | 1 |
194,762 | 15,438,903,089 | IssuesEvent | 2021-03-07 22:07:23 | civictechindex/CTI-website-frontend | https://api.github.com/repos/civictechindex/CTI-website-frontend | closed | Autofill for search functions across site | documentation feature: contributors feature: faq feature: search feature: tag generator role: UI/UX role: front end role: project management | ### Overview
We need to document autofill in this issue and share techniques so we can keep it consistent throughout CTI.
### Action Items
- [ ] Add all notes about autofill
- [ ] Provide links to related issues
- [ ] Document where autofill used
### Resources/Instructions
suggested https://react-autosuggest.js.org/ custom render as way to do the autofill
| 1.0 | Autofill for search functions across site - ### Overview
We need to document autofill in this issue and share techniques so we can keep it consistent throughout CTI.
### Action Items
- [ ] Add all notes about autofill
- [ ] Provide links to related issues
- [ ] Document where autofill used
### Resources/Instructions
suggested https://react-autosuggest.js.org/ custom render as way to do the autofill
| non_priority | autofill for search functions across site overview we need to document autofill in this issue and share techniques so we can keep it consistent throughout cti action items add all notes about autofill provide links to related issues document where autofill used resources instructions suggested custom render as way to do the autofill | 0 |
24,178 | 5,037,085,096 | IssuesEvent | 2016-12-17 13:02:24 | dbpedia-spotlight/dbpedia-spotlight | https://api.github.com/repos/dbpedia-spotlight/dbpedia-spotlight | closed | Document how to set up spotters for local use | documentation | Extend the documentation in the Spotting page on the wiki to include how to configure and test each spotter (send to Apache Stanbol's documentation as well)
| 1.0 | Document how to set up spotters for local use - Extend the documentation in the Spotting page on the wiki to include how to configure and test each spotter (send to Apache Stanbol's documentation as well)
| non_priority | document how to set up spotters for local use extend the documentation in the spotting page on the wiki to include how to configure and test each spotter send to apache stanbol s documentation as well | 0 |
323,496 | 27,731,199,632 | IssuesEvent | 2023-03-15 08:10:56 | inverse-inc/packetfence | https://api.github.com/repos/inverse-inc/packetfence | closed | tests(cli_login): RADIUS test service doesn't start | tests | **Describe the problem**
`start_radius_test_service` test can failed on EL8 or Debian 11
**Mandatory informations**
* Branch: 11.2 or 12.0
* Scenario: cli_login(el8|deb11)
* Pipelines URL:
* https://gitlab.com/inverse-inc/packetfence/-/jobs/3926221110
* https://gitlab.com/inverse-inc/packetfence/-/jobs/3936631931
**Symptoms**
How the problem can be identified if it appears again ?

| 1.0 | tests(cli_login): RADIUS test service doesn't start - **Describe the problem**
`start_radius_test_service` test can failed on EL8 or Debian 11
**Mandatory informations**
* Branch: 11.2 or 12.0
* Scenario: cli_login(el8|deb11)
* Pipelines URL:
* https://gitlab.com/inverse-inc/packetfence/-/jobs/3926221110
* https://gitlab.com/inverse-inc/packetfence/-/jobs/3936631931
**Symptoms**
How the problem can be identified if it appears again ?

| non_priority | tests cli login radius test service doesn t start describe the problem start radius test service test can failed on or debian mandatory informations branch or scenario cli login pipelines url symptoms how the problem can be identified if it appears again | 0 |
23,987 | 16,742,112,817 | IssuesEvent | 2021-06-11 11:08:47 | oppia/oppia-android | https://api.github.com/repos/oppia/oppia-android | opened | New Protos for the Platform Parameter Project | Priority: Essential Type: Task Where: Infrastructure | Add new protos for the platform parameter project. Namely
- PlatformParameter
- PlatformParameterWithName
- PlatformParameterDatabase
- PlatformParameterMap
Target PR date: 7 June 2021
Target completion date: 11 June 2021 | 1.0 | New Protos for the Platform Parameter Project - Add new protos for the platform parameter project. Namely
- PlatformParameter
- PlatformParameterWithName
- PlatformParameterDatabase
- PlatformParameterMap
Target PR date: 7 June 2021
Target completion date: 11 June 2021 | non_priority | new protos for the platform parameter project add new protos for the platform parameter project namely platformparameter platformparameterwithname platformparameterdatabase platformparametermap target pr date june target completion date june | 0 |
72,177 | 31,237,051,740 | IssuesEvent | 2023-08-20 12:10:14 | jhipster/generator-jhipster | https://api.github.com/repos/jhipster/generator-jhipster | closed | Switch to Consul as default service discovery mechanism | theme: microservice theme: JDL theme: kubernetes $$ bug-bounty $$ $500 v8 | Since eureka is kind of unmaintained and JHipster registry is deprecated and JHCC is unmaintained, it makes sense to make Consul the default service discovery mechanism in JHipster.
- [x] Change default in generator
- [x] Change default in JDL
- [x] update docs at hipster.tech
- [x] Test docker compose and kubernetes setup
- [x] Update conul kubernetes setup to latest version
- [x] Update JHipster Online: https://github.com/jhipster/jhipster-online -> https://start.jhipster.tech/generate-application
- [x] Update Sample JDL: https://github.com/jhipster/jdl-samples
- [x] Update Sample Apps: https://github.com/jhipster/jhipster-sample-app-gateway and https://github.com/jhipster/jhipster-sample-app-microservice
- [x] Check and update JHipster Base: https://github.com/jhipster/jhipster-base
- [x] Probably IDE too: https://github.com/jhipster/jhipster-ide
| 1.0 | Switch to Consul as default service discovery mechanism - Since eureka is kind of unmaintained and JHipster registry is deprecated and JHCC is unmaintained, it makes sense to make Consul the default service discovery mechanism in JHipster.
- [x] Change default in generator
- [x] Change default in JDL
- [x] update docs at hipster.tech
- [x] Test docker compose and kubernetes setup
- [x] Update conul kubernetes setup to latest version
- [x] Update JHipster Online: https://github.com/jhipster/jhipster-online -> https://start.jhipster.tech/generate-application
- [x] Update Sample JDL: https://github.com/jhipster/jdl-samples
- [x] Update Sample Apps: https://github.com/jhipster/jhipster-sample-app-gateway and https://github.com/jhipster/jhipster-sample-app-microservice
- [x] Check and update JHipster Base: https://github.com/jhipster/jhipster-base
- [x] Probably IDE too: https://github.com/jhipster/jhipster-ide
| non_priority | switch to consul as default service discovery mechanism since eureka is kind of unmaintained and jhipster registry is deprecated and jhcc is unmaintained it makes sense to make consul the default service discovery mechanism in jhipster change default in generator change default in jdl update docs at hipster tech test docker compose and kubernetes setup update conul kubernetes setup to latest version update jhipster online update sample jdl update sample apps and check and update jhipster base probably ide too | 0 |
709,054 | 24,366,410,319 | IssuesEvent | 2022-10-03 15:29:30 | o3de/o3de | https://api.github.com/repos/o3de/o3de | closed | NetworkPlayerSpawnerComponent Allows Users to Select Non-Network Prefabs | feature/networking kind/bug sig/network triage/accepted priority/critical WF6 RTE | **Describe the bug**
NetworkPlayerSpawnerComponent allows users to select non-network prefabs, Net-players will always need a net-binding so we should enforce only prefabs which contain a .network product.
**Steps to reproduce**
Steps to reproduce the behavior:
1. Add an entity with NetworkPlayerSpawnerComponent
2. Select a non-network prefab for the SpawnableAsset field
**Expected behavior**
Editor should give a warning, or block the user from selecting non-net prefabs
**Actual behavior**
Editor gladly accepts an non-net prefab
**Found in Branch**
development 5/10/2022 | 1.0 | NetworkPlayerSpawnerComponent Allows Users to Select Non-Network Prefabs - **Describe the bug**
NetworkPlayerSpawnerComponent allows users to select non-network prefabs, Net-players will always need a net-binding so we should enforce only prefabs which contain a .network product.
**Steps to reproduce**
Steps to reproduce the behavior:
1. Add an entity with NetworkPlayerSpawnerComponent
2. Select a non-network prefab for the SpawnableAsset field
**Expected behavior**
Editor should give a warning, or block the user from selecting non-net prefabs
**Actual behavior**
Editor gladly accepts an non-net prefab
**Found in Branch**
development 5/10/2022 | priority | networkplayerspawnercomponent allows users to select non network prefabs describe the bug networkplayerspawnercomponent allows users to select non network prefabs net players will always need a net binding so we should enforce only prefabs which contain a network product steps to reproduce steps to reproduce the behavior add an entity with networkplayerspawnercomponent select a non network prefab for the spawnableasset field expected behavior editor should give a warning or block the user from selecting non net prefabs actual behavior editor gladly accepts an non net prefab found in branch development | 1 |
709,323 | 24,373,824,802 | IssuesEvent | 2022-10-03 22:00:02 | fecgov/openFEC | https://api.github.com/repos/fecgov/openFEC | closed | Dev API CircleCI builds failing | Needs refinement High priority | ### Summary
Dev API CircleCI builds are failing on the `celery-beat` deployment stage. Error messages include:
`ERR Failed to compress build artifacts: exit status 2`
`ERR Failed to stage build: staging failed`
### Completion criteria
- [ ] Failing Dev API CircleCi builds are fixed | 1.0 | Dev API CircleCI builds failing - ### Summary
Dev API CircleCI builds are failing on the `celery-beat` deployment stage. Error messages include:
`ERR Failed to compress build artifacts: exit status 2`
`ERR Failed to stage build: staging failed`
### Completion criteria
- [ ] Failing Dev API CircleCi builds are fixed | priority | dev api circleci builds failing summary dev api circleci builds are failing on the celery beat deployment stage error messages include err failed to compress build artifacts exit status err failed to stage build staging failed completion criteria failing dev api circleci builds are fixed | 1 |
35,509 | 7,756,333,957 | IssuesEvent | 2018-05-31 13:19:37 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | opened | Do not use InternalVisitListener for ORACLE12C dialect | C: Functionality P: Medium T: Defect | There is an `InternalVisitListener` that collects some column aliases from the `SELECT` clause and makes them available to the `ORDER BY` clause in the same scope, in case `OFFSET` pagination needs to be emulated (see #2080). This is currently being done for:
- DB2
- Oracle
- SQL Data Warehouse
- SQL Server 2008
- Sybase
- Teradata
`VisitListeners` incur some non-trivial overhead on SQL generation and should be avoided if possible. In Oracle 12c, the `InternalVisitListener` is probably not necessary. Perhaps, it can be even replaced by something entirely different anyway. | 1.0 | Do not use InternalVisitListener for ORACLE12C dialect - There is an `InternalVisitListener` that collects some column aliases from the `SELECT` clause and makes them available to the `ORDER BY` clause in the same scope, in case `OFFSET` pagination needs to be emulated (see #2080). This is currently being done for:
- DB2
- Oracle
- SQL Data Warehouse
- SQL Server 2008
- Sybase
- Teradata
`VisitListeners` incur some non-trivial overhead on SQL generation and should be avoided if possible. In Oracle 12c, the `InternalVisitListener` is probably not necessary. Perhaps, it can be even replaced by something entirely different anyway. | non_priority | do not use internalvisitlistener for dialect there is an internalvisitlistener that collects some column aliases from the select clause and makes them available to the order by clause in the same scope in case offset pagination needs to be emulated see this is currently being done for oracle sql data warehouse sql server sybase teradata visitlisteners incur some non trivial overhead on sql generation and should be avoided if possible in oracle the internalvisitlistener is probably not necessary perhaps it can be even replaced by something entirely different anyway | 0 |
260,605 | 8,212,395,011 | IssuesEvent | 2018-09-04 16:11:42 | eclipse-ee4j/jsp-api | https://api.github.com/repos/eclipse-ee4j/jsp-api | closed | Tag file problem (Work in tomcat , dont work in jetty) | Priority: Major Type: Bug invalid | hi everyone,
i'am using a tag file library that works properly in tomcat 6.0.29 but doesn't
work in jetty jetty-distribution-7.1.6.v20100715.zip
the files are below to recreate the problem!!
body.tag
[code]
<%@ attribute name="jsCode" %>
<jsp:doBody var="html"/>
<html>
<head>
//This variable is empty in jetty and not empty in tomcat
<script>
$
{jsCode}
</script>
</head>
<body>
$
{html}
</body>
</html>
[/code]
setParentProperty.tag
[code]
<%@ attribute name="tag" required="true"
type="javax.servlet.jsp.tagext.SimpleTagSupport" %>
<%@ attribute name="property" required="true" type="java.lang.String" %>
<%@ attribute name="value" required="true" type="java.lang.String" %>
<%@ tag
import="
org.apache.commons.beanutils.BeanUtils,
javax.servlet.jsp.tagext.SimpleTagSupport,
javax.servlet.jsp.tagext.SimpleTag" %>
<%
SimpleTagSupport targetTag =
(SimpleTagSupport)findAncestorWithClass(tag,SimpleTag.class);
String parentPropertyValue = (String)BeanUtils.getProperty(targetTag,property);
if (parentPropertyValue!=null)
{ value = parentPropertyValue + value; }
BeanUtils.setProperty(targetTag,property,value);
//Both servers prints write this out
out.println("<br/> property => "<ins>property</ins>" =
"+(String)BeanUtils.getProperty(targetTag,property));
%>
[/code]
jsCode.tag
[code]
<jsp:doBody var="code"></jsp:doBody>
<util:setParentProperty tag="<%=this%>" property="jsCode"
value="$
"></util:setParentProperty>
[/code]
index.jsp
[code]
<html:body>
<html:jsCode>
alert('hi');
</html:jsCode>
</html:body>
[/code]
Is it a bug ? or in jetty way there are more things to do??
Sorry about my poor english!!
#### Environment
Operating System: All
Platform: All
#### Affected Versions
[current] | 1.0 | Tag file problem (Work in tomcat , dont work in jetty) - hi everyone,
i'am using a tag file library that works properly in tomcat 6.0.29 but doesn't
work in jetty jetty-distribution-7.1.6.v20100715.zip
the files are below to recreate the problem!!
body.tag
[code]
<%@ attribute name="jsCode" %>
<jsp:doBody var="html"/>
<html>
<head>
//This variable is empty in jetty and not empty in tomcat
<script>
$
{jsCode}
</script>
</head>
<body>
$
{html}
</body>
</html>
[/code]
setParentProperty.tag
[code]
<%@ attribute name="tag" required="true"
type="javax.servlet.jsp.tagext.SimpleTagSupport" %>
<%@ attribute name="property" required="true" type="java.lang.String" %>
<%@ attribute name="value" required="true" type="java.lang.String" %>
<%@ tag
import="
org.apache.commons.beanutils.BeanUtils,
javax.servlet.jsp.tagext.SimpleTagSupport,
javax.servlet.jsp.tagext.SimpleTag" %>
<%
SimpleTagSupport targetTag =
(SimpleTagSupport)findAncestorWithClass(tag,SimpleTag.class);
String parentPropertyValue = (String)BeanUtils.getProperty(targetTag,property);
if (parentPropertyValue!=null)
{ value = parentPropertyValue + value; }
BeanUtils.setProperty(targetTag,property,value);
//Both servers prints write this out
out.println("<br/> property => "<ins>property</ins>" =
"+(String)BeanUtils.getProperty(targetTag,property));
%>
[/code]
jsCode.tag
[code]
<jsp:doBody var="code"></jsp:doBody>
<util:setParentProperty tag="<%=this%>" property="jsCode"
value="$
"></util:setParentProperty>
[/code]
index.jsp
[code]
<html:body>
<html:jsCode>
alert('hi');
</html:jsCode>
</html:body>
[/code]
Is it a bug ? or in jetty way there are more things to do??
Sorry about my poor english!!
#### Environment
Operating System: All
Platform: All
#### Affected Versions
[current] | priority | tag file problem work in tomcat dont work in jetty hi everyone i am using a tag file library that works properly in tomcat but doesn t work in jetty jetty distribution zip the files are below to recreate the problem body tag this variable is empty in jetty and not empty in tomcat jscode html setparentproperty tag attribute name tag required true type javax servlet jsp tagext simpletagsupport tag import org apache commons beanutils beanutils javax servlet jsp tagext simpletagsupport javax servlet jsp tagext simpletag simpletagsupport targettag simpletagsupport findancestorwithclass tag simpletag class string parentpropertyvalue string beanutils getproperty targettag property if parentpropertyvalue null value parentpropertyvalue value beanutils setproperty targettag property value both servers prints write this out out println property property string beanutils getproperty targettag property jscode tag property jscode value index jsp alert hi is it a bug or in jetty way there are more things to do sorry about my poor english environment operating system all platform all affected versions | 1 |
801 | 14,915,484,101 | IssuesEvent | 2021-01-22 16:47:30 | Automattic/wp-calypso | https://api.github.com/repos/Automattic/wp-calypso | closed | Nav Unification: Subscription / People settings missing for Jetpack sites | Jetpack Nav Unification Nav-Unification [Beta] People Management Subscriptions [Type] Bug | ### Steps to reproduce the behavior
1. Start from a Jetpack site connected to WordPress.com, running either Jetpack `master` or the current stable.
2. In wp-admin, go to Jetpack > Settings > Discussion
3. Click on "View your Email Followers"
4. You'll land in Calypso, in `https://wordpress.com/people/email-followers/yoursite.com`. That page does not have a corresponding menu item in the sidebar, though.
| 1.0 | Nav Unification: Subscription / People settings missing for Jetpack sites - ### Steps to reproduce the behavior
1. Start from a Jetpack site connected to WordPress.com, running either Jetpack `master` or the current stable.
2. In wp-admin, go to Jetpack > Settings > Discussion
3. Click on "View your Email Followers"
4. You'll land in Calypso, in `https://wordpress.com/people/email-followers/yoursite.com`. That page does not have a corresponding menu item in the sidebar, though.
| non_priority | nav unification subscription people settings missing for jetpack sites steps to reproduce the behavior start from a jetpack site connected to wordpress com running either jetpack master or the current stable in wp admin go to jetpack settings discussion click on view your email followers you ll land in calypso in that page does not have a corresponding menu item in the sidebar though | 0 |
285,939 | 31,155,791,331 | IssuesEvent | 2023-08-16 13:03:31 | nidhi7598/linux-4.1.15_CVE-2018-5873 | https://api.github.com/repos/nidhi7598/linux-4.1.15_CVE-2018-5873 | opened | CVE-2015-8575 (Medium) detected in linuxlinux-4.1.52 | Mend: dependency security vulnerability | ## CVE-2015-8575 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.1.52</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/sco.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/sco.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The sco_sock_bind function in net/bluetooth/sco.c in the Linux kernel before 4.3.4 does not verify an address length, which allows local users to obtain sensitive information from kernel memory and bypass the KASLR protection mechanism via a crafted application.
<p>Publish Date: 2016-02-08
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-8575>CVE-2015-8575</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-8575">https://nvd.nist.gov/vuln/detail/CVE-2015-8575</a></p>
<p>Release Date: 2016-02-08</p>
<p>Fix Resolution: 4.3.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2015-8575 (Medium) detected in linuxlinux-4.1.52 - ## CVE-2015-8575 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.1.52</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/sco.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/sco.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The sco_sock_bind function in net/bluetooth/sco.c in the Linux kernel before 4.3.4 does not verify an address length, which allows local users to obtain sensitive information from kernel memory and bypass the KASLR protection mechanism via a crafted application.
<p>Publish Date: 2016-02-08
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-8575>CVE-2015-8575</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-8575">https://nvd.nist.gov/vuln/detail/CVE-2015-8575</a></p>
<p>Release Date: 2016-02-08</p>
<p>Fix Resolution: 4.3.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in base branch master vulnerable source files net bluetooth sco c net bluetooth sco c vulnerability details the sco sock bind function in net bluetooth sco c in the linux kernel before does not verify an address length which allows local users to obtain sensitive information from kernel memory and bypass the kaslr protection mechanism via a crafted application publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
56,146 | 23,708,331,016 | IssuesEvent | 2022-08-30 05:00:22 | hashicorp/terraform-provider-azurerm | https://api.github.com/repos/hashicorp/terraform-provider-azurerm | closed | Make it possible to run tasks rights away with azurerm_container_registry_task | question service/container-registry | ### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
### Description
When using azurerm_container_registry_task I would like these to be supported:
A) Point to a Dockerfile, for example in a git repo, and run the build right away.
B) Run a build right away, where I'm pointing to a Dockerfile in my Terraform project directory, together with my tf files.
### New or Affected Resource(s)
* `azurerm_container_registry_task `
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
resource "azurerm_container_registry_task" "example" {
name = "Example"
container_registry_id = azurerm_container_registry.example.id
platform {
os = "Linux"
architecture = "amd64"
}
docker_step {
dockerfile_path = "Dockerfile"
context_path = "https://dev.azure.com/<org>/<project>/_git/<repo>"
context_access_token = "<PAT>"
image_names = ["image:latest"]
# Something like this could be added
# Terraform task should not finish before the build is finished
trigger_onetime_run = true
}
# And also be able to add the source_trigger as well for future updates
source_trigger {
name = "newcommits"
events = ["commit"]
repository_url = "https://dev.azure.com/<org>/<project>/_git/<repo>"
source_type = "VisualStudioTeamService"
authentication {
token = "<PAT>"
token_type = "PAT"
}
}
}
``` | 1.0 | Make it possible to run tasks rights away with azurerm_container_registry_task - ### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
### Description
When using azurerm_container_registry_task I would like these to be supported:
A) Point to a Dockerfile, for example in a git repo, and run the build right away.
B) Run a build right away, where I'm pointing to a Dockerfile in my Terraform project directory, together with my tf files.
### New or Affected Resource(s)
* `azurerm_container_registry_task `
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
resource "azurerm_container_registry_task" "example" {
name = "Example"
container_registry_id = azurerm_container_registry.example.id
platform {
os = "Linux"
architecture = "amd64"
}
docker_step {
dockerfile_path = "Dockerfile"
context_path = "https://dev.azure.com/<org>/<project>/_git/<repo>"
context_access_token = "<PAT>"
image_names = ["image:latest"]
# Something like this could be added
# Terraform task should not finish before the build is finished
trigger_onetime_run = true
}
# And also be able to add the source_trigger as well for future updates
source_trigger {
name = "newcommits"
events = ["commit"]
repository_url = "https://dev.azure.com/<org>/<project>/_git/<repo>"
source_type = "VisualStudioTeamService"
authentication {
token = "<PAT>"
token_type = "PAT"
}
}
}
``` | non_priority | make it possible to run tasks rights away with azurerm container registry task community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description when using azurerm container registry task i would like these to be supported a point to a dockerfile for example in a git repo and run the build right away b run a build right away where i m pointing to a dockerfile in my terraform project directory together with my tf files new or affected resource s azurerm container registry task potential terraform configuration hcl resource azurerm container registry task example name example container registry id azurerm container registry example id platform os linux architecture docker step dockerfile path dockerfile context path context access token image names something like this could be added terraform task should not finish before the build is finished trigger onetime run true and also be able to add the source trigger as well for future updates source trigger name newcommits events repository url source type visualstudioteamservice authentication token token type pat | 0 |
409,602 | 11,965,347,154 | IssuesEvent | 2020-04-05 23:01:04 | Kaktushose/levelbot2 | https://api.github.com/repos/Kaktushose/levelbot2 | opened | Rechtschreibfehler in Fehlermeldung | bug high priority member | **Beschreibe den Bug**
In der Fehlermeldung (siehe Anhang) steht `ausgschlossen` anstatt `ausgeschlossen`.
**Schritte zum Reproduzieren**
!info @Bot
**Screenshots**

**Dein Discordname + Discriminator**
z.B. Kaktushose#0436
| 1.0 | Rechtschreibfehler in Fehlermeldung - **Beschreibe den Bug**
In der Fehlermeldung (siehe Anhang) steht `ausgschlossen` anstatt `ausgeschlossen`.
**Schritte zum Reproduzieren**
!info @Bot
**Screenshots**

**Dein Discordname + Discriminator**
z.B. Kaktushose#0436
| priority | rechtschreibfehler in fehlermeldung beschreibe den bug in der fehlermeldung siehe anhang steht ausgschlossen anstatt ausgeschlossen schritte zum reproduzieren info bot screenshots dein discordname discriminator z b kaktushose | 1 |
793,334 | 27,990,998,484 | IssuesEvent | 2023-03-27 03:43:16 | lowRISC/opentitan | https://api.github.com/repos/lowRISC/opentitan | opened | [ottf] add stack trace logging feature to aid in manufacturing test failures | Priority:P1 Type:Enhancement Type:Task SW:OTTF Manufacturing | ### Description
With #17495, we are refactoring the testutils to return `status_t`s to enable gracefully unwinding the stack when test errors are encountered in a test. This allows us to print a test result in JSON format, in a loop over the OTTF console so the ATE host can pick up the error. However, if the same function is called in many places throughout a test, it may be hard to determine the exact failure point.
Maintaining a lightweight, fixed-size, stack trace to print to the console on a test failure would be simple way to address this issue.
- [ ] investigate how this feature can be added to the OTTF console library
- [ ] implement mechanism in OTTF console library
| 1.0 | [ottf] add stack trace logging feature to aid in manufacturing test failures - ### Description
With #17495, we are refactoring the testutils to return `status_t`s to enable gracefully unwinding the stack when test errors are encountered in a test. This allows us to print a test result in JSON format, in a loop over the OTTF console so the ATE host can pick up the error. However, if the same function is called in many places throughout a test, it may be hard to determine the exact failure point.
Maintaining a lightweight, fixed-size, stack trace to print to the console on a test failure would be simple way to address this issue.
- [ ] investigate how this feature can be added to the OTTF console library
- [ ] implement mechanism in OTTF console library
| priority | add stack trace logging feature to aid in manufacturing test failures description with we are refactoring the testutils to return status t s to enable gracefully unwinding the stack when test errors are encountered in a test this allows us to print a test result in json format in a loop over the ottf console so the ate host can pick up the error however if the same function is called in many places throughout a test it may be hard to determine the exact failure point maintaining a lightweight fixed size stack trace to print to the console on a test failure would be simple way to address this issue investigate how this feature can be added to the ottf console library implement mechanism in ottf console library | 1 |
448,947 | 12,960,671,084 | IssuesEvent | 2020-07-20 14:40:16 | siteorigin/so-widgets-bundle | https://api.github.com/repos/siteorigin/so-widgets-bundle | closed | Hero: Buttons aren't being correctly rendered | bug priority-2 | The issue isn't present in the `master` branch. In `develop` the Hero's buttons aren't being correctly outputted.

P2 because this issue isn't present in the release version. We'll need to resolve this before the next release next week. | 1.0 | Hero: Buttons aren't being correctly rendered - The issue isn't present in the `master` branch. In `develop` the Hero's buttons aren't being correctly outputted.

P2 because this issue isn't present in the release version. We'll need to resolve this before the next release next week. | priority | hero buttons aren t being correctly rendered the issue isn t present in the master branch in develop the hero s buttons aren t being correctly outputted because this issue isn t present in the release version we ll need to resolve this before the next release next week | 1 |
162,783 | 13,901,064,561 | IssuesEvent | 2020-10-20 02:01:02 | maxitg/SetReplace | https://api.github.com/repos/maxitg/SetReplace | opened | Document passing evolution objects to RulePlot | analysis documentation english | ### The problem
We need to mention in the documentation that it is now possible to pass evolution objects directly to `RulePlot`, e.g.,
<img width="491" alt="image" src="https://user-images.githubusercontent.com/1479325/96530909-11c7fc00-124e-11eb-91e1-65f6a159268e.png"> | 1.0 | Document passing evolution objects to RulePlot - ### The problem
We need to mention in the documentation that it is now possible to pass evolution objects directly to `RulePlot`, e.g.,
<img width="491" alt="image" src="https://user-images.githubusercontent.com/1479325/96530909-11c7fc00-124e-11eb-91e1-65f6a159268e.png"> | non_priority | document passing evolution objects to ruleplot the problem we need to mention in the documentation that it is now possible to pass evolution objects directly to ruleplot e g img width alt image src | 0 |
41,323 | 6,900,133,550 | IssuesEvent | 2017-11-24 16:46:33 | Azolo/websockex | https://api.github.com/repos/Azolo/websockex | closed | Handling {:remote, :closed} ? | documentation | Hey, I've been using websockex in a project and it's working great (thanks for writing it). I do have one problem though.
I initiate a Supervisor with a simple_one_for_one that is responsible for initializing a websockex. It calls websockex with `start` (not `start_link`), but whenever the socket in which I'm listening is closed websockex will raise an exception, that in production gets logged as this:
CRASH REPORT==== 11-Nov-2017::12:47:29 ===
crasher:
initial call: Elixir.WebSockex:init/5
pid: <0.1340.0>
registered_name: []
exception exit: {remote,closed}
in function 'Elixir.WebSockex':terminate/4 (lib/websockex.ex, line 870)
ancestors: ['Elixir.Channeler.Monitor','Elixir.Channeler.Supervisor',
<0.1284.0>]
message_queue_len: 0
messages: []
links: []
dictionary: []
trap_exit: false
status: running
heap_size: 28690
stack_size: 27
reductions: 33897241
neighbours:
I have a terminate callback and I can see on the reason {:remote, :closed}, but would like to know how I can handle the :closed event so that it doesn't technically "crash"? I'm fine with writing the explanation to the docs and PR'ing if you would like, along with a section making notice of both `start` and `start_link`. Thanks | 1.0 | Handling {:remote, :closed} ? - Hey, I've been using websockex in a project and it's working great (thanks for writing it). I do have one problem though.
I initiate a Supervisor with a simple_one_for_one that is responsible for initializing a websockex. It calls websockex with `start` (not `start_link`), but whenever the socket in which I'm listening is closed websockex will raise an exception, that in production gets logged as this:
CRASH REPORT==== 11-Nov-2017::12:47:29 ===
crasher:
initial call: Elixir.WebSockex:init/5
pid: <0.1340.0>
registered_name: []
exception exit: {remote,closed}
in function 'Elixir.WebSockex':terminate/4 (lib/websockex.ex, line 870)
ancestors: ['Elixir.Channeler.Monitor','Elixir.Channeler.Supervisor',
<0.1284.0>]
message_queue_len: 0
messages: []
links: []
dictionary: []
trap_exit: false
status: running
heap_size: 28690
stack_size: 27
reductions: 33897241
neighbours:
I have a terminate callback and I can see on the reason {:remote, :closed}, but would like to know how I can handle the :closed event so that it doesn't technically "crash"? I'm fine with writing the explanation to the docs and PR'ing if you would like, along with a section making notice of both `start` and `start_link`. Thanks | non_priority | handling remote closed hey i ve been using websockex in a project and it s working great thanks for writing it i do have one problem though i initiate a supervisor with a simple one for one that is responsible for initializing a websockex it calls websockex with start not start link but whenever the socket in which i m listening is closed websockex will raise an exception that in production gets logged as this crash report nov crasher initial call elixir websockex init pid registered name exception exit remote closed in function elixir websockex terminate lib websockex ex line ancestors elixir channeler monitor elixir channeler supervisor message queue len messages links dictionary trap exit false status running heap size stack size reductions neighbours i have a terminate callback and i can see on the reason remote closed but would like to know how i can handle the closed event so that it doesn t technically crash i m fine with writing the explanation to the docs and pr ing if you would like along with a section making notice of both start and start link thanks | 0 |
500,441 | 14,499,523,729 | IssuesEvent | 2020-12-11 16:49:34 | ppc64le-cloud/pvsadm | https://api.github.com/repos/ppc64le-cloud/pvsadm | closed | pvsadm image import is not picking the IBMCLOUD_API_KEY from environment | kind/bug priority/critical-urgent | ```
🍏 Friday December 11 2020 07:28:32 PM 🍏
╭─github.com/ocp-power-automation/full-flow ⍉
╰─▶ export IBMCLOUD_API_KEY=******************
🍏 Friday December 11 2020 07:28:44 PM 🍏
╭─github.com/ocp-power-automation/full-flow
╰─▶ ./pvsadm image import --pvs-instance-name ocp-cicd-tokyo-04 --bucket my-test-bucket-10 --bucket-region us-south --object CentOS-83-11122020.ova.gz --pvs-image-name CentOS-83-11122020
Error: The ApiKey property is required but was not specified.
Usage:
pvsadm image import [flags]
Flags:
-n, --pvs-instance-name string PowerVS Instance name.
-i, --pvs-instance-id string PowerVS Instance ID.
-b, --bucket string Cloud Object Storage bucket name.
-r, --bucket-region string Cloud Object Storage bucket location.
-o, --object string Cloud Object Storage object name.
--accesskey string Cloud Object Storage HMAC access key.
--secretkey string Cloud Object Storage HMAC secret key.
--pvs-image-name string Name to PowerVS imported image.
--ostype string Image OS Type, accepted values are[aix, ibmi, redhat, sles]. (default "redhat")
--pvs-storagetype string PowerVS Storage type, accepted values are [tier1, tier3]. (default "tier3")
--pvs-service-cred string Service Credential name to be auto generated. (default "pvsadm-service-cred")
-h, --help help for import
Global Flags:
--add_dir_header If true, adds the file directory to the header of the log messages
--alsologtostderr log to standard error as well as files
-k, --api-key string IBMCLOUD API Key(env name: IBMCLOUD_API_KEY)
--audit-file string Audit logs for the tool (default "pvsadm.log")
--debug Enable PowerVS debug option(ATTENTION: dev only option, may print sensitive data from APIs)
--log_backtrace_at traceLocation when logging hits line file:N, emit a stack trace (default :0)
--log_dir string If non-empty, write log files in this directory
--log_file string If non-empty, use this log file
--log_file_max_size uint Defines the maximum size a log file can grow to. Unit is megabytes. If the value is 0, the maximum file size is unlimited. (default 1800)
--logtostderr log to standard error instead of files (default true)
--one_output If true, only write logs to their native severity level (vs also writing to each lower severity level
--skip_headers If true, avoid header prefixes in the log messages
--skip_log_headers If true, avoid headers when opening log files
--stderrthreshold severity logs at or above this threshold go to stderr (default 2)
-v, --v Level number for the log level verbosity
--vmodule moduleSpec comma-separated list of pattern=N settings for file-filtered logging
E1211 19:28:50.723929 50173 root.go:68] The ApiKey property is required but was not specified.
🍏 Friday December 11 2020 07:28:50 PM 🍏
╭─github.com/ocp-power-automation/full-flow ⍉
╰─▶
🍏 Friday December 11 2020 07:28:57 PM 🍏
╭─github.com/ocp-power-automation/full-flow ⍉
╰─▶ ./pvsadm version
I1211 19:29:00.830006 50244 root.go:29] Using an API key from IBMCLOUD_API_KEY environment variable
Version: v0.1-alpha.10-4-g57c47c5-dirty, GoVersion: go1.15.5
🍏 Friday December 11 2020 07:29:00 PM 🍏
╭─github.com/ocp-power-automation/full-flow
╰─▶
``` | 1.0 | pvsadm image import is not picking the IBMCLOUD_API_KEY from environment - ```
🍏 Friday December 11 2020 07:28:32 PM 🍏
╭─github.com/ocp-power-automation/full-flow ⍉
╰─▶ export IBMCLOUD_API_KEY=******************
🍏 Friday December 11 2020 07:28:44 PM 🍏
╭─github.com/ocp-power-automation/full-flow
╰─▶ ./pvsadm image import --pvs-instance-name ocp-cicd-tokyo-04 --bucket my-test-bucket-10 --bucket-region us-south --object CentOS-83-11122020.ova.gz --pvs-image-name CentOS-83-11122020
Error: The ApiKey property is required but was not specified.
Usage:
pvsadm image import [flags]
Flags:
-n, --pvs-instance-name string PowerVS Instance name.
-i, --pvs-instance-id string PowerVS Instance ID.
-b, --bucket string Cloud Object Storage bucket name.
-r, --bucket-region string Cloud Object Storage bucket location.
-o, --object string Cloud Object Storage object name.
--accesskey string Cloud Object Storage HMAC access key.
--secretkey string Cloud Object Storage HMAC secret key.
--pvs-image-name string Name to PowerVS imported image.
--ostype string Image OS Type, accepted values are[aix, ibmi, redhat, sles]. (default "redhat")
--pvs-storagetype string PowerVS Storage type, accepted values are [tier1, tier3]. (default "tier3")
--pvs-service-cred string Service Credential name to be auto generated. (default "pvsadm-service-cred")
-h, --help help for import
Global Flags:
--add_dir_header If true, adds the file directory to the header of the log messages
--alsologtostderr log to standard error as well as files
-k, --api-key string IBMCLOUD API Key(env name: IBMCLOUD_API_KEY)
--audit-file string Audit logs for the tool (default "pvsadm.log")
--debug Enable PowerVS debug option(ATTENTION: dev only option, may print sensitive data from APIs)
--log_backtrace_at traceLocation when logging hits line file:N, emit a stack trace (default :0)
--log_dir string If non-empty, write log files in this directory
--log_file string If non-empty, use this log file
--log_file_max_size uint Defines the maximum size a log file can grow to. Unit is megabytes. If the value is 0, the maximum file size is unlimited. (default 1800)
--logtostderr log to standard error instead of files (default true)
--one_output If true, only write logs to their native severity level (vs also writing to each lower severity level
--skip_headers If true, avoid header prefixes in the log messages
--skip_log_headers If true, avoid headers when opening log files
--stderrthreshold severity logs at or above this threshold go to stderr (default 2)
-v, --v Level number for the log level verbosity
--vmodule moduleSpec comma-separated list of pattern=N settings for file-filtered logging
E1211 19:28:50.723929 50173 root.go:68] The ApiKey property is required but was not specified.
🍏 Friday December 11 2020 07:28:50 PM 🍏
╭─github.com/ocp-power-automation/full-flow ⍉
╰─▶
🍏 Friday December 11 2020 07:28:57 PM 🍏
╭─github.com/ocp-power-automation/full-flow ⍉
╰─▶ ./pvsadm version
I1211 19:29:00.830006 50244 root.go:29] Using an API key from IBMCLOUD_API_KEY environment variable
Version: v0.1-alpha.10-4-g57c47c5-dirty, GoVersion: go1.15.5
🍏 Friday December 11 2020 07:29:00 PM 🍏
╭─github.com/ocp-power-automation/full-flow
╰─▶
``` | priority | pvsadm image import is not picking the ibmcloud api key from environment 🍏 friday december pm 🍏 ╭─github com ocp power automation full flow ⍉ ╰─▶ export ibmcloud api key 🍏 friday december pm 🍏 ╭─github com ocp power automation full flow ╰─▶ pvsadm image import pvs instance name ocp cicd tokyo bucket my test bucket bucket region us south object centos ova gz pvs image name centos error the apikey property is required but was not specified usage pvsadm image import flags n pvs instance name string powervs instance name i pvs instance id string powervs instance id b bucket string cloud object storage bucket name r bucket region string cloud object storage bucket location o object string cloud object storage object name accesskey string cloud object storage hmac access key secretkey string cloud object storage hmac secret key pvs image name string name to powervs imported image ostype string image os type accepted values are default redhat pvs storagetype string powervs storage type accepted values are default pvs service cred string service credential name to be auto generated default pvsadm service cred h help help for import global flags add dir header if true adds the file directory to the header of the log messages alsologtostderr log to standard error as well as files k api key string ibmcloud api key env name ibmcloud api key audit file string audit logs for the tool default pvsadm log debug enable powervs debug option attention dev only option may print sensitive data from apis log backtrace at tracelocation when logging hits line file n emit a stack trace default log dir string if non empty write log files in this directory log file string if non empty use this log file log file max size uint defines the maximum size a log file can grow to unit is megabytes if the value is the maximum file size is unlimited default logtostderr log to standard error instead of files default true one output if true only write logs to their native severity level vs also writing to each lower severity level skip headers if true avoid header prefixes in the log messages skip log headers if true avoid headers when opening log files stderrthreshold severity logs at or above this threshold go to stderr default v v level number for the log level verbosity vmodule modulespec comma separated list of pattern n settings for file filtered logging root go the apikey property is required but was not specified 🍏 friday december pm 🍏 ╭─github com ocp power automation full flow ⍉ ╰─▶ 🍏 friday december pm 🍏 ╭─github com ocp power automation full flow ⍉ ╰─▶ pvsadm version root go using an api key from ibmcloud api key environment variable version alpha dirty goversion 🍏 friday december pm 🍏 ╭─github com ocp power automation full flow ╰─▶ | 1 |
699,108 | 24,004,548,657 | IssuesEvent | 2022-09-14 13:56:40 | ramp4-pcar4/ramp4-pcar4 | https://api.github.com/repos/ramp4-pcar4/ramp4-pcar4 | closed | Class Breaks Symbol Lookup Bug | effort: small flavour: bug priority: must type: corrective | To replicate:
1. Open the latest "fancy legend" main sample page.
2. Open the grid for `Releases and Disposals by Mining Facilities` layer.
3. Notice one symbol does not draw, and error appears in console.
If sample is missing the layer, can add via legend using [this service url](https://maps-cartes.ec.gc.ca/arcgis/rest/services/StoryRAMP/410b88da_0ed1_4749_903f_5e76c24e2e5f/MapServer/2).
It looks like this layer is symbolizing on the `Total` field, but [Feature 3](https://maps-cartes.ec.gc.ca/arcgis/rest/services/StoryRAMP/410b88da_0ed1_4749_903f_5e76c24e2e5f/MapServer/2/3?f=pjson) has a value that is exactly the top-end of the first class break bucket (the one named `< 16,500`).
My guess is we have a `<` or `>` in the code that needs to be `<=` or `>=`, or vice versa. Or we have a floating point accuracy problem lurking. But it appears the "which bucket does the value fall into" code is failing to land this feature in the first or second bucket. | 1.0 | Class Breaks Symbol Lookup Bug - To replicate:
1. Open the latest "fancy legend" main sample page.
2. Open the grid for `Releases and Disposals by Mining Facilities` layer.
3. Notice one symbol does not draw, and error appears in console.
If sample is missing the layer, can add via legend using [this service url](https://maps-cartes.ec.gc.ca/arcgis/rest/services/StoryRAMP/410b88da_0ed1_4749_903f_5e76c24e2e5f/MapServer/2).
It looks like this layer is symbolizing on the `Total` field, but [Feature 3](https://maps-cartes.ec.gc.ca/arcgis/rest/services/StoryRAMP/410b88da_0ed1_4749_903f_5e76c24e2e5f/MapServer/2/3?f=pjson) has a value that is exactly the top-end of the first class break bucket (the one named `< 16,500`).
My guess is we have a `<` or `>` in the code that needs to be `<=` or `>=`, or vice versa. Or we have a floating point accuracy problem lurking. But it appears the "which bucket does the value fall into" code is failing to land this feature in the first or second bucket. | priority | class breaks symbol lookup bug to replicate open the latest fancy legend main sample page open the grid for releases and disposals by mining facilities layer notice one symbol does not draw and error appears in console if sample is missing the layer can add via legend using it looks like this layer is symbolizing on the total field but has a value that is exactly the top end of the first class break bucket the one named my guess is we have a in the code that needs to be or vice versa or we have a floating point accuracy problem lurking but it appears the which bucket does the value fall into code is failing to land this feature in the first or second bucket | 1 |
366,361 | 10,819,735,665 | IssuesEvent | 2019-11-08 14:59:24 | minio/minio | https://api.github.com/repos/minio/minio | closed | Using MooseFS as the storage directory, MinIO cannot run. | community priority: low triage | We use the Client provided by the MooseFS cluster as the MinIO data storage directory (/ofs), but MinIO cannot run. Run: minio server /ofs, the hard disk will be unresponsive.When I looked at a lot of information, I found that there is not much information about this issue. If the MinIO team can do this, I hope that MinIO developers can help us solve this problem together, thank you.
## Expected Behavior
MinIO can run normally on a MooseFS cluster.
## Current Behavior
We have a MinIO server and also serve as the Client role for the MooseFS cluster.
I created the /ofs directory in the root directory to be used as the MinIO storage directory and mounted moosefs-client.
When I run:
```shell
minio server /ofs
```
minio does not respond, and the hard drive enters a suspended state. When I force kill minio, an error log appears.

## Steps to Reproduce (for bugs)
- **Not using MooseFS**

- **Use MooseFS**

- **Test the most error log multiple times**

## Your Environment
```shell
[root@MinIO_C01_A-1 /]# minio version
Version: 2019-10-12T01:39:57Z
Release-Tag: RELEASE.2019-10-12T01-39-57Z
Commit-ID: bd106408462ecef70debf51f1e6179de950c5812
```
```shell
[root@MinIO_C01_A-1 /]# cat /etc/os-release
NAME="Amazon Linux"
VERSION="2"
ID="amzn"
ID_LIKE="centos rhel fedora"
VERSION_ID="2"
PRETTY_NAME="Amazon Linux 2"
ANSI_COLOR="0;33"
CPE_NAME="cpe:2.3:o:amazon:amazon_linux:2"
HOME_URL="https://amazonlinux.com/"
```
```shell
MooseFS: CGI version: 3.0.105 ; python: 2.7
``` | 1.0 | Using MooseFS as the storage directory, MinIO cannot run. - We use the Client provided by the MooseFS cluster as the MinIO data storage directory (/ofs), but MinIO cannot run. Run: minio server /ofs, the hard disk will be unresponsive.When I looked at a lot of information, I found that there is not much information about this issue. If the MinIO team can do this, I hope that MinIO developers can help us solve this problem together, thank you.
## Expected Behavior
MinIO can run normally on a MooseFS cluster.
## Current Behavior
We have a MinIO server and also serve as the Client role for the MooseFS cluster.
I created the /ofs directory in the root directory to be used as the MinIO storage directory and mounted moosefs-client.
When I run:
```shell
minio server /ofs
```
minio does not respond, and the hard drive enters a suspended state. When I force kill minio, an error log appears.

## Steps to Reproduce (for bugs)
- **Not using MooseFS**

- **Use MooseFS**

- **Test the most error log multiple times**

## Your Environment
```shell
[root@MinIO_C01_A-1 /]# minio version
Version: 2019-10-12T01:39:57Z
Release-Tag: RELEASE.2019-10-12T01-39-57Z
Commit-ID: bd106408462ecef70debf51f1e6179de950c5812
```
```shell
[root@MinIO_C01_A-1 /]# cat /etc/os-release
NAME="Amazon Linux"
VERSION="2"
ID="amzn"
ID_LIKE="centos rhel fedora"
VERSION_ID="2"
PRETTY_NAME="Amazon Linux 2"
ANSI_COLOR="0;33"
CPE_NAME="cpe:2.3:o:amazon:amazon_linux:2"
HOME_URL="https://amazonlinux.com/"
```
```shell
MooseFS: CGI version: 3.0.105 ; python: 2.7
``` | priority | using moosefs as the storage directory minio cannot run we use the client provided by the moosefs cluster as the minio data storage directory ofs but minio cannot run run minio server ofs the hard disk will be unresponsive when i looked at a lot of information i found that there is not much information about this issue if the minio team can do this i hope that minio developers can help us solve this problem together thank you expected behavior minio can run normally on a moosefs cluster current behavior we have a minio server and also serve as the client role for the moosefs cluster i created the ofs directory in the root directory to be used as the minio storage directory and mounted moosefs client when i run shell minio server ofs minio does not respond and the hard drive enters a suspended state when i force kill minio an error log appears steps to reproduce for bugs not using moosefs use moosefs test the most error log multiple times your environment shell minio version version release tag release commit id shell cat etc os release name amazon linux version id amzn id like centos rhel fedora version id pretty name amazon linux ansi color cpe name cpe o amazon amazon linux home url shell moosefs cgi version python | 1 |
372,577 | 11,017,103,720 | IssuesEvent | 2019-12-05 07:32:29 | pombase/canto | https://api.github.com/repos/pombase/canto | closed | new conditions do not always show in red | bug display medium priority |
Did somebody else mention this. They
seem to GO red eventually, but I'm not sure why they don't appear red immediately?
 | 1.0 | new conditions do not always show in red -
Did somebody else mention this. They
seem to GO red eventually, but I'm not sure why they don't appear red immediately?
 | priority | new conditions do not always show in red did somebody else mention this they seem to go red eventually but i m not sure why they don t appear red immediately | 1 |
814,731 | 30,519,623,055 | IssuesEvent | 2023-07-19 07:07:39 | ut-issl/c2a-aobc | https://api.github.com/repos/ut-issl/c2a-aobc | closed | formatter CIを導入する | :rocket: priority::high :fish: patch :wrench: tools | ## 詳細
GitLab時代のものをGitHub用に修正して設定する必要がある。
## close条件
FormatterがGitHub Actionsに追加されたら。
## 備考
NA
## 注意
NA | 1.0 | formatter CIを導入する - ## 詳細
GitLab時代のものをGitHub用に修正して設定する必要がある。
## close条件
FormatterがGitHub Actionsに追加されたら。
## 備考
NA
## 注意
NA | priority | formatter ciを導入する 詳細 gitlab時代のものをgithub用に修正して設定する必要がある。 close条件 formatterがgithub actionsに追加されたら。 備考 na 注意 na | 1 |
19,286 | 25,466,249,303 | IssuesEvent | 2022-11-25 04:53:40 | GoogleCloudPlatform/fda-mystudies | https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies | closed | [GCI] [PM] UI issue in the sign in screen when entered invalid credentials | Bug P0 Participant manager Process: Fixed Process: Tested QA Process: Tested dev | UI issue is observed in the sign in screen when entered invalid credentials.

| 3.0 | [GCI] [PM] UI issue in the sign in screen when entered invalid credentials - UI issue is observed in the sign in screen when entered invalid credentials.

| non_priority | ui issue in the sign in screen when entered invalid credentials ui issue is observed in the sign in screen when entered invalid credentials | 0 |
685,340 | 23,453,802,621 | IssuesEvent | 2022-08-16 06:58:33 | Greenstand/treetracker-query-api | https://api.github.com/repos/Greenstand/treetracker-query-api | opened | The GET /trees/{id} should return wallet_id and wallet_name | Node.js postgresql priority | The GET /trees/{id} endpoint should return wallet_id and wallet_name fields, by joining `wallet.token` on `wallet.token.capture_id` = `trees.uuid` and join `wallet.token` and `wallet.wallet` on `wallet.token.wallet_id` = `wallet.wallet.id` , so we can get the id and the name of the wallet
---
Some hints:
- Please read our readme for more information/guide/tutorial.
- Here is [an engineering book](https://greenstand.gitbook.io/engineering/) in Greenstand.
- To know more about our organization, visit our [website](https://greenstand.org).
- If you want to join the slack community (some resources need the community member's permission), please leave your email address.
| 1.0 | The GET /trees/{id} should return wallet_id and wallet_name - The GET /trees/{id} endpoint should return wallet_id and wallet_name fields, by joining `wallet.token` on `wallet.token.capture_id` = `trees.uuid` and join `wallet.token` and `wallet.wallet` on `wallet.token.wallet_id` = `wallet.wallet.id` , so we can get the id and the name of the wallet
---
Some hints:
- Please read our readme for more information/guide/tutorial.
- Here is [an engineering book](https://greenstand.gitbook.io/engineering/) in Greenstand.
- To know more about our organization, visit our [website](https://greenstand.org).
- If you want to join the slack community (some resources need the community member's permission), please leave your email address.
| priority | the get trees id should return wallet id and wallet name the get trees id endpoint should return wallet id and wallet name fields by joining wallet token on wallet token capture id trees uuid and join wallet token and wallet wallet on wallet token wallet id wallet wallet id so we can get the id and the name of the wallet some hints please read our readme for more information guide tutorial here is in greenstand to know more about our organization visit our if you want to join the slack community some resources need the community member s permission please leave your email address | 1 |
163,419 | 20,363,748,184 | IssuesEvent | 2022-02-21 01:23:06 | howlr-me/howlr-front | https://api.github.com/repos/howlr-me/howlr-front | opened | CVE-2021-3664 (Medium) detected in url-parse-1.4.7.tgz | security vulnerability | ## CVE-2021-3664 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p>
<p>Path to dependency file: /howlr-front/package.json</p>
<p>Path to vulnerable library: /node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.0.1.tgz (Root Library)
- react-dev-utils-9.0.1.tgz
- sockjs-client-1.3.0.tgz
- :x: **url-parse-1.4.7.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
url-parse is vulnerable to URL Redirection to Untrusted Site
<p>Publish Date: 2021-07-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3664>CVE-2021-3664</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-3664">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-3664</a></p>
<p>Release Date: 2021-07-26</p>
<p>Fix Resolution (url-parse): 1.5.2</p>
<p>Direct dependency fix Resolution (react-scripts): 3.1.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-3664 (Medium) detected in url-parse-1.4.7.tgz - ## CVE-2021-3664 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p>
<p>Path to dependency file: /howlr-front/package.json</p>
<p>Path to vulnerable library: /node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.0.1.tgz (Root Library)
- react-dev-utils-9.0.1.tgz
- sockjs-client-1.3.0.tgz
- :x: **url-parse-1.4.7.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
url-parse is vulnerable to URL Redirection to Untrusted Site
<p>Publish Date: 2021-07-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3664>CVE-2021-3664</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-3664">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-3664</a></p>
<p>Release Date: 2021-07-26</p>
<p>Fix Resolution (url-parse): 1.5.2</p>
<p>Direct dependency fix Resolution (react-scripts): 3.1.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in url parse tgz cve medium severity vulnerability vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href path to dependency file howlr front package json path to vulnerable library node modules url parse package json dependency hierarchy react scripts tgz root library react dev utils tgz sockjs client tgz x url parse tgz vulnerable library vulnerability details url parse is vulnerable to url redirection to untrusted site publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution url parse direct dependency fix resolution react scripts step up your open source security game with whitesource | 0 |
289,462 | 24,991,407,627 | IssuesEvent | 2022-11-02 19:07:19 | streamnative/pulsar | https://api.github.com/repos/streamnative/pulsar | opened | ISSUE-18300: Flaky-test: MetadataStoreStatsTest#testBatchMetadataStoreMetrics | component/test flaky-tests | Original Issue: apache/pulsar#18300
---
### Search before asking
- [X] I searched in the [issues](https://github.com/apache/pulsar/issues) and found nothing similar.
### Example failure
https://github.com/labuladong/pulsar/actions/runs/3376057575/jobs/5603501145#step:10:1208
### Exception stacktrace
<!-- optionally provide the full stacktrace -->
<details>
<summary>Full exception stacktrace</summary>
<pre><code>
Error: Tests run: 7, Failures: 2, Errors: 0, Skipped: 5, Time elapsed: 28.876 s <<< FAILURE! - in org.apache.pulsar.broker.stats.MetadataStoreStatsTest
Error: testBatchMetadataStoreMetrics(org.apache.pulsar.broker.stats.MetadataStoreStatsTest) Time elapsed: 10.577 s <<< FAILURE!
java.lang.AssertionError: expected object to not be null
at org.testng.Assert.fail(Assert.java:99)
at org.testng.Assert.assertNotNull(Assert.java:942)
at org.testng.Assert.assertNotNull(Assert.java:926)
at org.apache.pulsar.broker.stats.MetadataStoreStatsTest.testBatchMetadataStoreMetrics(MetadataStoreStatsTest.java:185)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:568)
at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:132)
at org.testng.internal.InvokeMethodRunnable.runOne(InvokeMethodRunnable.java:45)
at org.testng.internal.InvokeMethodRunnable.call(InvokeMethodRunnable.java:73)
at org.testng.internal.InvokeMethodRunnable.call(InvokeMethodRunnable.java:11)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)
at java.base/java.lang.Thread.run(Thread.java:833)
</code></pre>
</details>
### Are you willing to submit a PR?
- [ ] I'm willing to submit a PR! | 2.0 | ISSUE-18300: Flaky-test: MetadataStoreStatsTest#testBatchMetadataStoreMetrics - Original Issue: apache/pulsar#18300
---
### Search before asking
- [X] I searched in the [issues](https://github.com/apache/pulsar/issues) and found nothing similar.
### Example failure
https://github.com/labuladong/pulsar/actions/runs/3376057575/jobs/5603501145#step:10:1208
### Exception stacktrace
<!-- optionally provide the full stacktrace -->
<details>
<summary>Full exception stacktrace</summary>
<pre><code>
Error: Tests run: 7, Failures: 2, Errors: 0, Skipped: 5, Time elapsed: 28.876 s <<< FAILURE! - in org.apache.pulsar.broker.stats.MetadataStoreStatsTest
Error: testBatchMetadataStoreMetrics(org.apache.pulsar.broker.stats.MetadataStoreStatsTest) Time elapsed: 10.577 s <<< FAILURE!
java.lang.AssertionError: expected object to not be null
at org.testng.Assert.fail(Assert.java:99)
at org.testng.Assert.assertNotNull(Assert.java:942)
at org.testng.Assert.assertNotNull(Assert.java:926)
at org.apache.pulsar.broker.stats.MetadataStoreStatsTest.testBatchMetadataStoreMetrics(MetadataStoreStatsTest.java:185)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:568)
at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:132)
at org.testng.internal.InvokeMethodRunnable.runOne(InvokeMethodRunnable.java:45)
at org.testng.internal.InvokeMethodRunnable.call(InvokeMethodRunnable.java:73)
at org.testng.internal.InvokeMethodRunnable.call(InvokeMethodRunnable.java:11)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)
at java.base/java.lang.Thread.run(Thread.java:833)
</code></pre>
</details>
### Are you willing to submit a PR?
- [ ] I'm willing to submit a PR! | non_priority | issue flaky test metadatastorestatstest testbatchmetadatastoremetrics original issue apache pulsar search before asking i searched in the and found nothing similar example failure exception stacktrace full exception stacktrace error tests run failures errors skipped time elapsed s failure in org apache pulsar broker stats metadatastorestatstest error testbatchmetadatastoremetrics org apache pulsar broker stats metadatastorestatstest time elapsed s failure java lang assertionerror expected object to not be null at org testng assert fail assert java at org testng assert assertnotnull assert java at org testng assert assertnotnull assert java at org apache pulsar broker stats metadatastorestatstest testbatchmetadatastoremetrics metadatastorestatstest java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org testng internal methodinvocationhelper invokemethod methodinvocationhelper java at org testng internal invokemethodrunnable runone invokemethodrunnable java at org testng internal invokemethodrunnable call invokemethodrunnable java at org testng internal invokemethodrunnable call invokemethodrunnable java at java base java util concurrent futuretask run futuretask java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java base java lang thread run thread java are you willing to submit a pr i m willing to submit a pr | 0 |
650,348 | 21,388,270,873 | IssuesEvent | 2022-04-21 02:46:45 | forgojs/forgo | https://api.github.com/repos/forgojs/forgo | closed | More tests are needed | low-priority | The current test cases are mostly integration tests and are incomplete.
We need
- (a) more tests
- (b) a path to unit testing individual functions
(a) is more important as of now.
| 1.0 | More tests are needed - The current test cases are mostly integration tests and are incomplete.
We need
- (a) more tests
- (b) a path to unit testing individual functions
(a) is more important as of now.
| priority | more tests are needed the current test cases are mostly integration tests and are incomplete we need a more tests b a path to unit testing individual functions a is more important as of now | 1 |
16,513 | 6,219,567,320 | IssuesEvent | 2017-07-09 14:55:28 | travis-ci/packer-templates | https://api.github.com/repos/travis-ci/packer-templates | closed | Update git to 2.x | build-env-update/2017Q1 enhancement ubuntu/precise ubuntu/trusty | With https://github.com/travis-ci/dpl/pull/516, we can now use `git` 2.10.x to deploy to Heroku via `git`. I do not think there is any other reason to stay on `git` 1.9.x now.
- [x] sudo-enabled Trusty
- [x] <del>sudo-enabled Precise</del>
- [ ] container-based Trusty
- [x] <del>container-based Precise</del>
Related to https://github.com/travis-ci/travis-ci/issues/6328 | 1.0 | Update git to 2.x - With https://github.com/travis-ci/dpl/pull/516, we can now use `git` 2.10.x to deploy to Heroku via `git`. I do not think there is any other reason to stay on `git` 1.9.x now.
- [x] sudo-enabled Trusty
- [x] <del>sudo-enabled Precise</del>
- [ ] container-based Trusty
- [x] <del>container-based Precise</del>
Related to https://github.com/travis-ci/travis-ci/issues/6328 | non_priority | update git to x with we can now use git x to deploy to heroku via git i do not think there is any other reason to stay on git x now sudo enabled trusty sudo enabled precise container based trusty container based precise related to | 0 |
14,663 | 11,044,933,025 | IssuesEvent | 2019-12-09 14:15:17 | jhu-sheridan-libraries/jhuda-general-issues | https://api.github.com/repos/jhu-sheridan-libraries/jhuda-general-issues | closed | investigate and propose a specific messaging technology | infrastructure | the technologies mentioned include: ActiveMQ, Kafka, Microsoft Azure Messaging Service
Please make a recommendation based on known restrictions/conditions of JHU and Data Archive specific infrastructure.
part of ticket #13 | 1.0 | investigate and propose a specific messaging technology - the technologies mentioned include: ActiveMQ, Kafka, Microsoft Azure Messaging Service
Please make a recommendation based on known restrictions/conditions of JHU and Data Archive specific infrastructure.
part of ticket #13 | non_priority | investigate and propose a specific messaging technology the technologies mentioned include activemq kafka microsoft azure messaging service please make a recommendation based on known restrictions conditions of jhu and data archive specific infrastructure part of ticket | 0 |
715,328 | 24,594,483,140 | IssuesEvent | 2022-10-14 07:04:01 | objectiv/objectiv-analytics | https://api.github.com/repos/objectiv/objectiv-analytics | closed | Fix BQ tests.functional.bach.test_df_sort_values.test_sort_values_expression | Priority: 3. Medium Area: Bach Component: BigQuery | * Make code changes if needed
* Make test use `engine` instead of `pg_engine` | 1.0 | Fix BQ tests.functional.bach.test_df_sort_values.test_sort_values_expression - * Make code changes if needed
* Make test use `engine` instead of `pg_engine` | priority | fix bq tests functional bach test df sort values test sort values expression make code changes if needed make test use engine instead of pg engine | 1 |
829,234 | 31,860,516,511 | IssuesEvent | 2023-09-15 10:32:34 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.wowhead.com - Article content does not load with "Trackers and Scripts to Block- Content" option enabled | priority-normal browser-focus-geckoview engine-gecko android 13 | <!-- @browser: Firefox Mobile 117.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 13; Mobile; rv:109.0) Gecko/117.0 Firefox/117.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/127078 -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://www.wowhead.com/news/patch-10-2-ptr-development-notes-sept-12-more-class-tuning-335043
**Browser / Version**: Firefox Mobile 117.0
**Operating System**: Android 13
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Missing items
**Steps to Reproduce**:
Article content does not load on Firefox focus
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2023/9/0044649d-a025-4020-8c4a-44650b06afd1.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>buildID: 20230912013654</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2023/9/8eeed0a1-eade-4ccb-b2f6-af3d5911d03d)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.wowhead.com - Article content does not load with "Trackers and Scripts to Block- Content" option enabled - <!-- @browser: Firefox Mobile 117.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 13; Mobile; rv:109.0) Gecko/117.0 Firefox/117.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/127078 -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://www.wowhead.com/news/patch-10-2-ptr-development-notes-sept-12-more-class-tuning-335043
**Browser / Version**: Firefox Mobile 117.0
**Operating System**: Android 13
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Missing items
**Steps to Reproduce**:
Article content does not load on Firefox focus
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2023/9/0044649d-a025-4020-8c4a-44650b06afd1.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>buildID: 20230912013654</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2023/9/8eeed0a1-eade-4ccb-b2f6-af3d5911d03d)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | article content does not load with trackers and scripts to block content option enabled url browser version firefox mobile operating system android tested another browser yes chrome problem type site is not usable description missing items steps to reproduce article content does not load on firefox focus view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false buildid channel release hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 1 |
163,929 | 20,364,282,023 | IssuesEvent | 2022-02-21 02:28:51 | delaval-htps/ProjetDevJava | https://api.github.com/repos/delaval-htps/ProjetDevJava | opened | CVE-2022-0639 (Medium) detected in url-parse-1.4.7.tgz | security vulnerability | ## CVE-2022-0639 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p>
<p>Path to dependency file: /AppAngular/package.json</p>
<p>Path to vulnerable library: /AppAngular/node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.1001.7.tgz (Root Library)
- webpack-dev-server-3.11.0.tgz
- sockjs-client-1.4.0.tgz
- :x: **url-parse-1.4.7.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.7.
<p>Publish Date: 2022-02-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0639>CVE-2022-0639</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0639">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0639</a></p>
<p>Release Date: 2022-02-17</p>
<p>Fix Resolution: url-parse - 1.5.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-0639 (Medium) detected in url-parse-1.4.7.tgz - ## CVE-2022-0639 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p>
<p>Path to dependency file: /AppAngular/package.json</p>
<p>Path to vulnerable library: /AppAngular/node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.1001.7.tgz (Root Library)
- webpack-dev-server-3.11.0.tgz
- sockjs-client-1.4.0.tgz
- :x: **url-parse-1.4.7.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.7.
<p>Publish Date: 2022-02-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0639>CVE-2022-0639</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0639">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0639</a></p>
<p>Release Date: 2022-02-17</p>
<p>Fix Resolution: url-parse - 1.5.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in url parse tgz cve medium severity vulnerability vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href path to dependency file appangular package json path to vulnerable library appangular node modules url parse package json dependency hierarchy build angular tgz root library webpack dev server tgz sockjs client tgz x url parse tgz vulnerable library found in base branch master vulnerability details authorization bypass through user controlled key in npm url parse prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution url parse step up your open source security game with whitesource | 0 |
99,313 | 30,331,242,625 | IssuesEvent | 2023-07-11 06:32:23 | PaddlePaddle/Paddle | https://api.github.com/repos/PaddlePaddle/Paddle | closed | 在Mac M1 pro平台使用Debian:stable (stable-20220622) 编译arm版Paddlepaddle成功,但是ctest未能完全通过。 | status/following-up type/build | ### 问题描述 Issue Description
ctest日志:[LastTest.log](https://github.com/PaddlePaddle/Paddle/files/9034266/LastTest.log)
- Paddlepaddle源码版本: `git checkout v2.3.1`
- 使用的CMake命令: `cmake .. -DPY_VERSION=3.9 -DWITH_TESTING=ON -DON_INFER=ON -DWITH_ARM=ON -DCMAKE_BUILD_TYPE=Release`
- Docker Engine: `v20.10.16`
- 容器内核: `Linux 9093ee9c043a 5.10.104-linuxkit #1 SMP PREEMPT Thu Mar 17 17:05:54 UTC 2022 aarch64 GNU/Linux`
### 版本&环境信息 Version & Environment Information
****************************************
Paddle version: 2.3.1
Paddle With CUDA: False
OS: Debian GNU/Linux 11
Python version: 3.9.2
CUDA version: None
cuDNN version: None.None.None
Nvidia driver version: None
****************************************
python3 -m pip list
Package Version
------------------ ---------
astor 0.8.1
attrs 21.4.0
Babel 2.10.3
bce-python-sdk 0.8.73
certifi 2022.6.15
charset-normalizer 2.1.0
click 8.1.3
cloudpickle 2.1.0
cycler 0.11.0
decorator 5.1.1
distro 1.7.0
exceptiongroup 1.0.0rc8
Flask 2.1.2
Flask-Babel 2.0.0
fonttools 4.33.3
future 0.18.2
gym 0.24.1
gym-notices 0.0.7
hypothesis 6.48.2
idna 3.3
importlib-metadata 4.12.0
itsdangerous 2.1.2
Jinja2 3.1.2
kiwisolver 1.4.3
MarkupSafe 2.1.1
matplotlib 3.5.2
numpy 1.23.0
onnx 1.12.0
opencv-python 4.6.0.66
opt-einsum 3.3.0
packaging 21.3
paddle-bfloat 0.1.7
paddle2onnx 0.9.9
paddlepaddle 2.3.1
pandas 1.4.3
Pillow 9.2.0
pip 20.3.4
protobuf 3.20.0
pycryptodome 3.15.0
pyparsing 3.0.9
python-dateutil 2.8.2
pytz 2022.1
PyYAML 6.0
requests 2.28.1
scipy 1.8.1
setuptools 52.0.0
six 1.16.0
sortedcontainers 2.4.0
typing-extensions 4.3.0
urllib3 1.26.9
visualdl 2.3.0
Werkzeug 2.1.2
wheel 0.34.2
zipp 3.8.0
**************************************** | 1.0 | 在Mac M1 pro平台使用Debian:stable (stable-20220622) 编译arm版Paddlepaddle成功,但是ctest未能完全通过。 - ### 问题描述 Issue Description
ctest日志:[LastTest.log](https://github.com/PaddlePaddle/Paddle/files/9034266/LastTest.log)
- Paddlepaddle源码版本: `git checkout v2.3.1`
- 使用的CMake命令: `cmake .. -DPY_VERSION=3.9 -DWITH_TESTING=ON -DON_INFER=ON -DWITH_ARM=ON -DCMAKE_BUILD_TYPE=Release`
- Docker Engine: `v20.10.16`
- 容器内核: `Linux 9093ee9c043a 5.10.104-linuxkit #1 SMP PREEMPT Thu Mar 17 17:05:54 UTC 2022 aarch64 GNU/Linux`
### 版本&环境信息 Version & Environment Information
****************************************
Paddle version: 2.3.1
Paddle With CUDA: False
OS: Debian GNU/Linux 11
Python version: 3.9.2
CUDA version: None
cuDNN version: None.None.None
Nvidia driver version: None
****************************************
python3 -m pip list
Package Version
------------------ ---------
astor 0.8.1
attrs 21.4.0
Babel 2.10.3
bce-python-sdk 0.8.73
certifi 2022.6.15
charset-normalizer 2.1.0
click 8.1.3
cloudpickle 2.1.0
cycler 0.11.0
decorator 5.1.1
distro 1.7.0
exceptiongroup 1.0.0rc8
Flask 2.1.2
Flask-Babel 2.0.0
fonttools 4.33.3
future 0.18.2
gym 0.24.1
gym-notices 0.0.7
hypothesis 6.48.2
idna 3.3
importlib-metadata 4.12.0
itsdangerous 2.1.2
Jinja2 3.1.2
kiwisolver 1.4.3
MarkupSafe 2.1.1
matplotlib 3.5.2
numpy 1.23.0
onnx 1.12.0
opencv-python 4.6.0.66
opt-einsum 3.3.0
packaging 21.3
paddle-bfloat 0.1.7
paddle2onnx 0.9.9
paddlepaddle 2.3.1
pandas 1.4.3
Pillow 9.2.0
pip 20.3.4
protobuf 3.20.0
pycryptodome 3.15.0
pyparsing 3.0.9
python-dateutil 2.8.2
pytz 2022.1
PyYAML 6.0
requests 2.28.1
scipy 1.8.1
setuptools 52.0.0
six 1.16.0
sortedcontainers 2.4.0
typing-extensions 4.3.0
urllib3 1.26.9
visualdl 2.3.0
Werkzeug 2.1.2
wheel 0.34.2
zipp 3.8.0
**************************************** | non_priority | 在mac pro平台使用debian stable stable 编译arm版paddlepaddle成功,但是ctest未能完全通过。 问题描述 issue description ctest日志: paddlepaddle源码版本 git checkout 使用的cmake命令 cmake dpy version dwith testing on don infer on dwith arm on dcmake build type release docker engine 容器内核 linux linuxkit smp preempt thu mar utc gnu linux 版本 环境信息 version environment information paddle version paddle with cuda false os debian gnu linux python version cuda version none cudnn version none none none nvidia driver version none m pip list package version astor attrs babel bce python sdk certifi charset normalizer click cloudpickle cycler decorator distro exceptiongroup flask flask babel fonttools future gym gym notices hypothesis idna importlib metadata itsdangerous kiwisolver markupsafe matplotlib numpy onnx opencv python opt einsum packaging paddle bfloat paddlepaddle pandas pillow pip protobuf pycryptodome pyparsing python dateutil pytz pyyaml requests scipy setuptools six sortedcontainers typing extensions visualdl werkzeug wheel zipp | 0 |
309,176 | 23,286,651,733 | IssuesEvent | 2022-08-05 17:15:43 | wanted-group-4/wanted-preonboarding-tripbtoz | https://api.github.com/repos/wanted-group-4/wanted-preonboarding-tripbtoz | opened | docs: Readme 작성을 위한 screenshots | documentation | ## ISSUE
- Group: `client`
- Type: `feature`
- Detail: Readme
-
<img width="563" alt="cal_mobile_dot" src="https://user-images.githubusercontent.com/87647245/183127717-06b07f53-e95a-41f1-9ef8-2913c8962bab.png">
<img width="563" alt="cal_mobile_init" src="https://user-images.githubusercontent.com/87647245/183127738-c4b29e17-0e57-434f-93bc-a860cb4bd3d0.png">
<img width="563" alt="cal_mobile_long" src="https://user-images.githubusercontent.com/87647245/183127745-7ade8dd8-caca-4dca-8e7f-376a48b50745.png">
<img width="563" alt="cal_mobile_max" src="https://user-images.githubusercontent.com/87647245/183127749-840aaeda-9920-4b2d-927d-d142dd679833.png">
<img width="1312" alt="cal_web_ing" src="https://user-images.githubusercontent.com/87647245/183127754-85e8e370-a5c3-42f2-9bd6-56f456447583.png">
<img width="1312" alt="cal_web" src="https://user-images.githubusercontent.com/87647245/183127757-d0a3ac4a-300e-4d8d-a4b0-8620bc3542e7.png">
<img width="385" alt="count_mobile" src="https://user-images.githubusercontent.com/87647245/183127761-46d6310a-638e-4827-98e5-d15dc936f344.png">
<img width="1312" alt="count_web" src="https://user-images.githubusercontent.com/87647245/183127765-8472d9e7-a98d-4308-bfc0-ad7f0448c9f5.png">
<img width="791" alt="figma" src="https://user-images.githubusercontent.com/87647245/183127767-568f3996-657c-4c5c-ad84-89f6dd0b4589.png">
<img width="387" alt="main_mobile_last" src="https://user-images.githubusercontent.com/87647245/183127769-ab79e443-d5c8-4179-a58c-f30095b93131.png">
<img width="386" alt="main_mobile" src="https://user-images.githubusercontent.com/87647245/183127772-6aaeba2f-6cb2-4da5-acde-3f641323828e.png">
<img width="1312" alt="main" src="https://user-images.githubusercontent.com/87647245/183127777-07738a97-5ced-4001-8665-61955eefafe6.png">
<img width="1312" alt="search_01" src="https://user-images.githubusercontent.com/87647245/183127778-9c2fb1de-f6c4-44f7-9543-95adfbb99c60.png">
<img width="1312" alt="search_02" src="https://user-images.githubusercontent.com/87647245/183127780-380a8b64-b795-4cf4-a47d-1311abef89ec.png">
<img width="385" alt="search_transition" src="https://user-images.githubusercontent.com/87647245/183127783-3a0496e3-37fa-4613-b8ce-42e765583156.png">
<img width="386" alt="error" src="https://user-images.githubusercontent.com/87647245/183127958-65fa1161-17fb-4908-a586-639add3e8bd6.png">
| 1.0 | docs: Readme 작성을 위한 screenshots - ## ISSUE
- Group: `client`
- Type: `feature`
- Detail: Readme
-
<img width="563" alt="cal_mobile_dot" src="https://user-images.githubusercontent.com/87647245/183127717-06b07f53-e95a-41f1-9ef8-2913c8962bab.png">
<img width="563" alt="cal_mobile_init" src="https://user-images.githubusercontent.com/87647245/183127738-c4b29e17-0e57-434f-93bc-a860cb4bd3d0.png">
<img width="563" alt="cal_mobile_long" src="https://user-images.githubusercontent.com/87647245/183127745-7ade8dd8-caca-4dca-8e7f-376a48b50745.png">
<img width="563" alt="cal_mobile_max" src="https://user-images.githubusercontent.com/87647245/183127749-840aaeda-9920-4b2d-927d-d142dd679833.png">
<img width="1312" alt="cal_web_ing" src="https://user-images.githubusercontent.com/87647245/183127754-85e8e370-a5c3-42f2-9bd6-56f456447583.png">
<img width="1312" alt="cal_web" src="https://user-images.githubusercontent.com/87647245/183127757-d0a3ac4a-300e-4d8d-a4b0-8620bc3542e7.png">
<img width="385" alt="count_mobile" src="https://user-images.githubusercontent.com/87647245/183127761-46d6310a-638e-4827-98e5-d15dc936f344.png">
<img width="1312" alt="count_web" src="https://user-images.githubusercontent.com/87647245/183127765-8472d9e7-a98d-4308-bfc0-ad7f0448c9f5.png">
<img width="791" alt="figma" src="https://user-images.githubusercontent.com/87647245/183127767-568f3996-657c-4c5c-ad84-89f6dd0b4589.png">
<img width="387" alt="main_mobile_last" src="https://user-images.githubusercontent.com/87647245/183127769-ab79e443-d5c8-4179-a58c-f30095b93131.png">
<img width="386" alt="main_mobile" src="https://user-images.githubusercontent.com/87647245/183127772-6aaeba2f-6cb2-4da5-acde-3f641323828e.png">
<img width="1312" alt="main" src="https://user-images.githubusercontent.com/87647245/183127777-07738a97-5ced-4001-8665-61955eefafe6.png">
<img width="1312" alt="search_01" src="https://user-images.githubusercontent.com/87647245/183127778-9c2fb1de-f6c4-44f7-9543-95adfbb99c60.png">
<img width="1312" alt="search_02" src="https://user-images.githubusercontent.com/87647245/183127780-380a8b64-b795-4cf4-a47d-1311abef89ec.png">
<img width="385" alt="search_transition" src="https://user-images.githubusercontent.com/87647245/183127783-3a0496e3-37fa-4613-b8ce-42e765583156.png">
<img width="386" alt="error" src="https://user-images.githubusercontent.com/87647245/183127958-65fa1161-17fb-4908-a586-639add3e8bd6.png">
| non_priority | docs readme 작성을 위한 screenshots issue group client type feature detail readme img width alt cal mobile dot src img width alt cal mobile init src img width alt cal mobile long src img width alt cal mobile max src img width alt cal web ing src img width alt cal web src img width alt count mobile src img width alt count web src img width alt figma src img width alt main mobile last src img width alt main mobile src img width alt main src img width alt search src img width alt search src img width alt search transition src img width alt error src | 0 |
692,368 | 23,731,604,097 | IssuesEvent | 2022-08-31 02:35:40 | kubernetes/ingress-nginx | https://api.github.com/repos/kubernetes/ingress-nginx | closed | Am I able to ssl tls terminate mqtt protocol or is it limited to http/s only? If not, how? | kind/feature needs-triage needs-priority | I've installed ingress in this way via Helm.
```
helm upgrade --install ingress-nginx ingress-nginx/ingress-nginx --create-namespace --namespace $NAMESPACE --set controller.service.annotations.service\.beta\.kubernetes\.io/azure-load-balancer-health-probe-request-path=/healthz,controller.service.annotations.service\.beta\.kubernetes\.io/azure-dns-label-name=$DNS_LABEL --set controller.service.loadBalancerIP=$STATIC_IP --set tcp.18083=$NAMESPACE/emqx-ee:18083,tcp.1883=$NAMESPACE/emqx-ee:1883,tcp.8883=$NAMESPACE/emqx-ee:8883
```
What I am wondering is, can I rewrite, and thus terminate the incoming ssl mqtts:// traffic to tcp mqtt://.
The way I've setup the Ingress is in this way from a letsencrypt RA cert listed below. From here should I just be able to rewrite a backend service with port 1883? Would that work? Or will it not work because it's not http/s? If it won't work is there a way to handle this as a work around. A lot of the documentation is from long ago and not sure if it is relevant.
```
apiVersion: networking.k8s.io/v1
kind: Ingress
metadata:
name: emqx-ee-ingress
annotations:
nginx.ingress.kubernetes.io/rewrite-target: /$2
nginx.ingress.kubernetes.io/use-regex: "true"
cert-manager.io/cluster-issuer: letsencrypt
spec:
ingressClassName: nginx
tls:
- hosts:
- <fdqn-part>.cloudapp.azure.com
secretName: tls-secret
rules:
- host: <fdqn-part>.cloudapp.azure.com
http:
paths:
- path: /(.*)
pathType: Prefix
backend:
service:
name: emqx-ee
port:
number: 18083
```
references:
https://github.com/kubernetes/ingress-nginx/issues/636
https://github.com/nginxinc/kubernetes-ingress/issues/831
| 1.0 | Am I able to ssl tls terminate mqtt protocol or is it limited to http/s only? If not, how? - I've installed ingress in this way via Helm.
```
helm upgrade --install ingress-nginx ingress-nginx/ingress-nginx --create-namespace --namespace $NAMESPACE --set controller.service.annotations.service\.beta\.kubernetes\.io/azure-load-balancer-health-probe-request-path=/healthz,controller.service.annotations.service\.beta\.kubernetes\.io/azure-dns-label-name=$DNS_LABEL --set controller.service.loadBalancerIP=$STATIC_IP --set tcp.18083=$NAMESPACE/emqx-ee:18083,tcp.1883=$NAMESPACE/emqx-ee:1883,tcp.8883=$NAMESPACE/emqx-ee:8883
```
What I am wondering is, can I rewrite, and thus terminate the incoming ssl mqtts:// traffic to tcp mqtt://.
The way I've setup the Ingress is in this way from a letsencrypt RA cert listed below. From here should I just be able to rewrite a backend service with port 1883? Would that work? Or will it not work because it's not http/s? If it won't work is there a way to handle this as a work around. A lot of the documentation is from long ago and not sure if it is relevant.
```
apiVersion: networking.k8s.io/v1
kind: Ingress
metadata:
name: emqx-ee-ingress
annotations:
nginx.ingress.kubernetes.io/rewrite-target: /$2
nginx.ingress.kubernetes.io/use-regex: "true"
cert-manager.io/cluster-issuer: letsencrypt
spec:
ingressClassName: nginx
tls:
- hosts:
- <fdqn-part>.cloudapp.azure.com
secretName: tls-secret
rules:
- host: <fdqn-part>.cloudapp.azure.com
http:
paths:
- path: /(.*)
pathType: Prefix
backend:
service:
name: emqx-ee
port:
number: 18083
```
references:
https://github.com/kubernetes/ingress-nginx/issues/636
https://github.com/nginxinc/kubernetes-ingress/issues/831
| priority | am i able to ssl tls terminate mqtt protocol or is it limited to http s only if not how i ve installed ingress in this way via helm helm upgrade install ingress nginx ingress nginx ingress nginx create namespace namespace namespace set controller service annotations service beta kubernetes io azure load balancer health probe request path healthz controller service annotations service beta kubernetes io azure dns label name dns label set controller service loadbalancerip static ip set tcp namespace emqx ee tcp namespace emqx ee tcp namespace emqx ee what i am wondering is can i rewrite and thus terminate the incoming ssl mqtts traffic to tcp mqtt the way i ve setup the ingress is in this way from a letsencrypt ra cert listed below from here should i just be able to rewrite a backend service with port would that work or will it not work because it s not http s if it won t work is there a way to handle this as a work around a lot of the documentation is from long ago and not sure if it is relevant apiversion networking io kind ingress metadata name emqx ee ingress annotations nginx ingress kubernetes io rewrite target nginx ingress kubernetes io use regex true cert manager io cluster issuer letsencrypt spec ingressclassname nginx tls hosts cloudapp azure com secretname tls secret rules host cloudapp azure com http paths path pathtype prefix backend service name emqx ee port number references | 1 |
69,240 | 14,980,486,304 | IssuesEvent | 2021-01-28 13:42:48 | ConnectionMaster/create-probot-app | https://api.github.com/repos/ConnectionMaster/create-probot-app | opened | WS-2019-0307 (Medium) detected in tagstooTagstoo_2.0.1_sourcecode, nodev12.16.1 | security vulnerability | ## WS-2019-0307 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tagstooTagstoo_2.0.1_sourcecode</b>, <b>nodev12.16.1</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In 'mem' before v4.0.0 there is a Denial of Service (DoS) vulnerability as a result of a failure in removal old values from the cache.
<p>Publish Date: 2018-08-27
<p>URL: <a href=https://github.com/sindresorhus/mem/commit/da4e4398cb27b602de3bd55f746efa9b4a31702b>WS-2019-0307</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1084">https://www.npmjs.com/advisories/1084</a></p>
<p>Release Date: 2019-12-01</p>
<p>Fix Resolution: mem - 4.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2019-0307 (Medium) detected in tagstooTagstoo_2.0.1_sourcecode, nodev12.16.1 - ## WS-2019-0307 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tagstooTagstoo_2.0.1_sourcecode</b>, <b>nodev12.16.1</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In 'mem' before v4.0.0 there is a Denial of Service (DoS) vulnerability as a result of a failure in removal old values from the cache.
<p>Publish Date: 2018-08-27
<p>URL: <a href=https://github.com/sindresorhus/mem/commit/da4e4398cb27b602de3bd55f746efa9b4a31702b>WS-2019-0307</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1084">https://www.npmjs.com/advisories/1084</a></p>
<p>Release Date: 2019-12-01</p>
<p>Fix Resolution: mem - 4.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | ws medium detected in tagstootagstoo sourcecode ws medium severity vulnerability vulnerable libraries tagstootagstoo sourcecode vulnerability details in mem before there is a denial of service dos vulnerability as a result of a failure in removal old values from the cache publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution mem step up your open source security game with whitesource | 0 |
94,173 | 27,135,597,507 | IssuesEvent | 2023-02-16 13:00:41 | expo/expo | https://api.github.com/repos/expo/expo | closed | Build error related to expo-dev-menu when jsEngine=jsc | Issue accepted Development Builds | ### Summary
I upgraded to Expo 48, which has Hermes enables by default and my build was successful. However, Hermes is leading to my iOS build being really big which is an issue because I'm building an App Clip which has a size limit of 15 MB. So I switched back to jsc (using the jsEngine param in app.json), which leads to the following build errors in EAS Build:
```
❌ (../../node_modules/expo-dev-menu/vendored/react-native-reanimated/ios/native/NativeProxy.mm:27:10)
25 | #else
26 | #if RNVERSION >= 71
> 27 | #include <jsc/JSCRuntime.h>
| ^ 'jsc/JSCRuntime.h' file not found
28 | #else
29 | #include <jsi/JSCRuntime.h>
30 | #endif // RNVersion
```
and
```
▸ CompileC /Users/expo/Library/Developer/Xcode/DerivedData/Feld-ejrefokefgxjrfcfsvxvjoonuons/Build/Intermediates.noindex/ArchiveIntermediates/Feld/IntermediateBuildFilesPath/Pods.build/Release-iphoneos/expo-dev-menu.build/Objects-normal/arm64/NativeProxy.o /Users/expo/workingdir/build/node_modules/expo-dev-menu/vendored/react-native-reanimated/ios/native/NativeProxy.mm normal arm64 objective-c++ com.apple.compilers.llvm.clang.1_0.compiler (in target 'expo-dev-menu' from project 'Pods')
```
### Managed or bare workflow?
managed
### What platform(s) does this occur on?
iOS
### Package versions
"expo": "48.0.1",
"expo-dev-client": "~2.1.1",
"react-native-reanimated": "~2.14.4",
### Environment
expo-env-info 1.0.5 environment info:
System:
OS: macOS 13.2.1
Shell: 5.8.1 - /bin/zsh
Binaries:
Node: 18.12.1 - ~/Library/pnpm/node
Yarn: 1.22.19 - ~/Developer/GitHub/feld/node_modules/.bin/yarn
npm: 8.19.2 - ~/Library/pnpm/npm
Watchman: 2023.02.13.00 - /opt/homebrew/bin/watchman
Managers:
CocoaPods: 1.11.3 - /opt/homebrew/bin/pod
SDKs:
iOS SDK:
Platforms: DriverKit 22.2, iOS 16.2, macOS 13.1, tvOS 16.1, watchOS 9.1
IDEs:
Xcode: 14.2/14C18 - /usr/bin/xcodebuild
Expo Workflow: managed
### Reproducible demo
- Create new Expo app
- Add expo-dev-client
- Add jsEngine="jsc" in app.json
- Build app
### Stacktrace (if a crash is involved)
_No response_ | 1.0 | Build error related to expo-dev-menu when jsEngine=jsc - ### Summary
I upgraded to Expo 48, which has Hermes enables by default and my build was successful. However, Hermes is leading to my iOS build being really big which is an issue because I'm building an App Clip which has a size limit of 15 MB. So I switched back to jsc (using the jsEngine param in app.json), which leads to the following build errors in EAS Build:
```
❌ (../../node_modules/expo-dev-menu/vendored/react-native-reanimated/ios/native/NativeProxy.mm:27:10)
25 | #else
26 | #if RNVERSION >= 71
> 27 | #include <jsc/JSCRuntime.h>
| ^ 'jsc/JSCRuntime.h' file not found
28 | #else
29 | #include <jsi/JSCRuntime.h>
30 | #endif // RNVersion
```
and
```
▸ CompileC /Users/expo/Library/Developer/Xcode/DerivedData/Feld-ejrefokefgxjrfcfsvxvjoonuons/Build/Intermediates.noindex/ArchiveIntermediates/Feld/IntermediateBuildFilesPath/Pods.build/Release-iphoneos/expo-dev-menu.build/Objects-normal/arm64/NativeProxy.o /Users/expo/workingdir/build/node_modules/expo-dev-menu/vendored/react-native-reanimated/ios/native/NativeProxy.mm normal arm64 objective-c++ com.apple.compilers.llvm.clang.1_0.compiler (in target 'expo-dev-menu' from project 'Pods')
```
### Managed or bare workflow?
managed
### What platform(s) does this occur on?
iOS
### Package versions
"expo": "48.0.1",
"expo-dev-client": "~2.1.1",
"react-native-reanimated": "~2.14.4",
### Environment
expo-env-info 1.0.5 environment info:
System:
OS: macOS 13.2.1
Shell: 5.8.1 - /bin/zsh
Binaries:
Node: 18.12.1 - ~/Library/pnpm/node
Yarn: 1.22.19 - ~/Developer/GitHub/feld/node_modules/.bin/yarn
npm: 8.19.2 - ~/Library/pnpm/npm
Watchman: 2023.02.13.00 - /opt/homebrew/bin/watchman
Managers:
CocoaPods: 1.11.3 - /opt/homebrew/bin/pod
SDKs:
iOS SDK:
Platforms: DriverKit 22.2, iOS 16.2, macOS 13.1, tvOS 16.1, watchOS 9.1
IDEs:
Xcode: 14.2/14C18 - /usr/bin/xcodebuild
Expo Workflow: managed
### Reproducible demo
- Create new Expo app
- Add expo-dev-client
- Add jsEngine="jsc" in app.json
- Build app
### Stacktrace (if a crash is involved)
_No response_ | non_priority | build error related to expo dev menu when jsengine jsc summary i upgraded to expo which has hermes enables by default and my build was successful however hermes is leading to my ios build being really big which is an issue because i m building an app clip which has a size limit of mb so i switched back to jsc using the jsengine param in app json which leads to the following build errors in eas build ❌ node modules expo dev menu vendored react native reanimated ios native nativeproxy mm else if rnversion include jsc jscruntime h file not found else include endif rnversion and ▸ compilec users expo library developer xcode deriveddata feld ejrefokefgxjrfcfsvxvjoonuons build intermediates noindex archiveintermediates feld intermediatebuildfilespath pods build release iphoneos expo dev menu build objects normal nativeproxy o users expo workingdir build node modules expo dev menu vendored react native reanimated ios native nativeproxy mm normal objective c com apple compilers llvm clang compiler in target expo dev menu from project pods managed or bare workflow managed what platform s does this occur on ios package versions expo expo dev client react native reanimated environment expo env info environment info system os macos shell bin zsh binaries node library pnpm node yarn developer github feld node modules bin yarn npm library pnpm npm watchman opt homebrew bin watchman managers cocoapods opt homebrew bin pod sdks ios sdk platforms driverkit ios macos tvos watchos ides xcode usr bin xcodebuild expo workflow managed reproducible demo create new expo app add expo dev client add jsengine jsc in app json build app stacktrace if a crash is involved no response | 0 |
77,218 | 3,506,273,361 | IssuesEvent | 2016-01-08 05:11:43 | OregonCore/OregonCore | https://api.github.com/repos/OregonCore/OregonCore | closed | frequent crashes (BB #281) | migrated Priority: Medium Type: Bug | This issue was migrated from bitbucket.
**Original Reporter:**
**Original Date:** 02.09.2010 16:56:31 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** resolved
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/281
<hr>
I'm on your newest rev (682) and my server crashes every 5-10min, sometimes even earlier. I guess it's because of joining arenas' queues, but i'm not sure. Please check it. | 1.0 | frequent crashes (BB #281) - This issue was migrated from bitbucket.
**Original Reporter:**
**Original Date:** 02.09.2010 16:56:31 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** resolved
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/281
<hr>
I'm on your newest rev (682) and my server crashes every 5-10min, sometimes even earlier. I guess it's because of joining arenas' queues, but i'm not sure. Please check it. | priority | frequent crashes bb this issue was migrated from bitbucket original reporter original date gmt original priority major original type bug original state resolved direct link i m on your newest rev and my server crashes every sometimes even earlier i guess it s because of joining arenas queues but i m not sure please check it | 1 |
613,463 | 19,091,147,720 | IssuesEvent | 2021-11-29 12:15:01 | qgis/QGIS | https://api.github.com/repos/qgis/QGIS | closed | [node tool] nodes on the locked item remains highlighted after toggle editing | High Priority Bug Digitizing | Author Name: **Saber Razmjooei** (@saberraz)
Original Redmine Issue: [21532](https://issues.qgis.org/issues/21532)
Affected QGIS version: 3.7(master)
Redmine category:digitising
---
After toggle editing (exit edit mode), the highlighted nodes of the locked item remains highlighted.
---
- [node_tool_artifact.gif](https://issues.qgis.org/attachments/download/14559/node_tool_artifact.gif) (Saber Razmjooei) | 1.0 | [node tool] nodes on the locked item remains highlighted after toggle editing - Author Name: **Saber Razmjooei** (@saberraz)
Original Redmine Issue: [21532](https://issues.qgis.org/issues/21532)
Affected QGIS version: 3.7(master)
Redmine category:digitising
---
After toggle editing (exit edit mode), the highlighted nodes of the locked item remains highlighted.
---
- [node_tool_artifact.gif](https://issues.qgis.org/attachments/download/14559/node_tool_artifact.gif) (Saber Razmjooei) | priority | nodes on the locked item remains highlighted after toggle editing author name saber razmjooei saberraz original redmine issue affected qgis version master redmine category digitising after toggle editing exit edit mode the highlighted nodes of the locked item remains highlighted saber razmjooei | 1 |
737,217 | 25,506,369,558 | IssuesEvent | 2022-11-28 09:45:41 | bash-lsp/bash-language-server | https://api.github.com/repos/bash-lsp/bash-language-server | closed | Unable to disable shellcheck with `SHELLCHECK_PATH=""` | bug priority ⭐️ | #555 seems to have broken the functionality allowing shellcheck to be disabled by setting the environment variable `SHELLCHECK_PATH` to the empty string. See [here](https://github.com/bash-lsp/bash-language-server/blob/938bf337ec6afb7b909167165e03759a65a751d4/server/src/config.ts#L5) and #432.
Specifically, I think this change was the cause:
```diff
diff --git a/server/src/linter.ts b/server/src/linter.ts
index 936341a3..6cb7087c 100644
--- a/server/src/linter.ts
+++ b/server/src/linter.ts
@@ -1,6 +1,9 @@
import { spawn } from 'child_process'
import * as LSP from 'vscode-languageserver'
+import * as config from './config'
+import { execShellScript } from './util/sh'
+
function formatMessage(comment: ShellcheckComment): string {
return (comment.code ? `SC${comment.code}: ` : '') + comment.message
}
@@ -10,8 +13,23 @@ type LinterOptions = {
cwd?: string
}
-export default class Linter {
- private executablePath: string | null
+export async function getLinterExecutablePath(): Promise<string | null> {
+ const pathFromConfig = config.getShellcheckPath()
+
+ if (pathFromConfig) {
+ return pathFromConfig
+ }
+
+ try {
+ const path = (await execShellScript('which shellcheck')).trim()
+ return path === '' ? null : path
+ } catch (e) {
+ return null
+ }
+}
+
+export class Linter {
+ public executablePath: string | null
private cwd: string
_canLint: boolean
``` | 1.0 | Unable to disable shellcheck with `SHELLCHECK_PATH=""` - #555 seems to have broken the functionality allowing shellcheck to be disabled by setting the environment variable `SHELLCHECK_PATH` to the empty string. See [here](https://github.com/bash-lsp/bash-language-server/blob/938bf337ec6afb7b909167165e03759a65a751d4/server/src/config.ts#L5) and #432.
Specifically, I think this change was the cause:
```diff
diff --git a/server/src/linter.ts b/server/src/linter.ts
index 936341a3..6cb7087c 100644
--- a/server/src/linter.ts
+++ b/server/src/linter.ts
@@ -1,6 +1,9 @@
import { spawn } from 'child_process'
import * as LSP from 'vscode-languageserver'
+import * as config from './config'
+import { execShellScript } from './util/sh'
+
function formatMessage(comment: ShellcheckComment): string {
return (comment.code ? `SC${comment.code}: ` : '') + comment.message
}
@@ -10,8 +13,23 @@ type LinterOptions = {
cwd?: string
}
-export default class Linter {
- private executablePath: string | null
+export async function getLinterExecutablePath(): Promise<string | null> {
+ const pathFromConfig = config.getShellcheckPath()
+
+ if (pathFromConfig) {
+ return pathFromConfig
+ }
+
+ try {
+ const path = (await execShellScript('which shellcheck')).trim()
+ return path === '' ? null : path
+ } catch (e) {
+ return null
+ }
+}
+
+export class Linter {
+ public executablePath: string | null
private cwd: string
_canLint: boolean
``` | priority | unable to disable shellcheck with shellcheck path seems to have broken the functionality allowing shellcheck to be disabled by setting the environment variable shellcheck path to the empty string see and specifically i think this change was the cause diff diff git a server src linter ts b server src linter ts index a server src linter ts b server src linter ts import spawn from child process import as lsp from vscode languageserver import as config from config import execshellscript from util sh function formatmessage comment shellcheckcomment string return comment code sc comment code comment message type linteroptions cwd string export default class linter private executablepath string null export async function getlinterexecutablepath promise const pathfromconfig config getshellcheckpath if pathfromconfig return pathfromconfig try const path await execshellscript which shellcheck trim return path null path catch e return null export class linter public executablepath string null private cwd string canlint boolean | 1 |
800,042 | 28,323,514,529 | IssuesEvent | 2023-04-11 04:39:30 | sygmaprotocol/explorer-ui | https://api.github.com/repos/sygmaprotocol/explorer-ui | opened | create component `number of records showed` | Priority: P3 | <!--- Provide a general summary of the issue in the Title above -->
We are going to provide with the option for the user to decide the amounts of records to display
## Implementation details
<!-- Enter description of implementation that may help dev team -->
- Create dropdown list to select number of records to show. Numbers: 10, 50, 100
- on click table re-renders with the current number selected
- create functional test
## Testing details
<!-- Enter description of special test-cases-->
- if we select another number rather than the default, table shows that number of records
## Acceptance Criteria
<!-- Enter the conditions of satisfaction here. That is, the conditions that will satisfy the user/persona that the goal/benefit/value has been achieved -->
| 1.0 | create component `number of records showed` - <!--- Provide a general summary of the issue in the Title above -->
We are going to provide with the option for the user to decide the amounts of records to display
## Implementation details
<!-- Enter description of implementation that may help dev team -->
- Create dropdown list to select number of records to show. Numbers: 10, 50, 100
- on click table re-renders with the current number selected
- create functional test
## Testing details
<!-- Enter description of special test-cases-->
- if we select another number rather than the default, table shows that number of records
## Acceptance Criteria
<!-- Enter the conditions of satisfaction here. That is, the conditions that will satisfy the user/persona that the goal/benefit/value has been achieved -->
| priority | create component number of records showed we are going to provide with the option for the user to decide the amounts of records to display implementation details create dropdown list to select number of records to show numbers on click table re renders with the current number selected create functional test testing details if we select another number rather than the default table shows that number of records acceptance criteria | 1 |
343,832 | 10,337,258,032 | IssuesEvent | 2019-09-03 14:32:17 | level73/membernet | https://api.github.com/repos/level73/membernet | closed | hierarchy in nes/cbi contacts on landing page | Platform: Membernet Priority: Medium Type: Development | could we please have the list of facilitators be organised alphabetically for NES contacts? ie, NES Albania, NES Bolivia, Colombia etc.?

Same for CBI
| 1.0 | hierarchy in nes/cbi contacts on landing page - could we please have the list of facilitators be organised alphabetically for NES contacts? ie, NES Albania, NES Bolivia, Colombia etc.?

Same for CBI
| priority | hierarchy in nes cbi contacts on landing page could we please have the list of facilitators be organised alphabetically for nes contacts ie nes albania nes bolivia colombia etc same for cbi | 1 |
300,865 | 9,213,075,419 | IssuesEvent | 2019-03-10 08:27:09 | CS2103-AY1819S2-W15-2/main | https://api.github.com/repos/CS2103-AY1819S2-W15-2/main | closed | As a user I want to be able to set a budget | priority.High type.Epic type.Story | so I can better manage my expenses and see track how much more i can spend for the remaining part of the day/week/month/year | 1.0 | As a user I want to be able to set a budget - so I can better manage my expenses and see track how much more i can spend for the remaining part of the day/week/month/year | priority | as a user i want to be able to set a budget so i can better manage my expenses and see track how much more i can spend for the remaining part of the day week month year | 1 |
555,542 | 16,457,453,454 | IssuesEvent | 2021-05-21 14:21:25 | GoogleCloudPlatform/ruby-docs-samples | https://api.github.com/repos/GoogleCloudPlatform/ruby-docs-samples | closed | rspec: Google Cloud Spanner API samples list backups with various filters failed | api: spanner flakybot: issue priority: p2 samples type: bug | Note: #755 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 0f1c000e62f9f5a457f16b52784f887c1ba03741
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/dd8ec378-661a-458e-9948-b7977cb10cdf), [Sponge](http://sponge2/dd8ec378-661a-458e-9948-b7977cb10cdf)
status: failed
<details><summary>Test output</summary><br><pre>Failure/Error:
expect(captured_output).to include(
"All backups\n#{backup.backup_id}"
)
expected "All backups\ntest_bu_a38f75b97487de4b\ntest_bu_8b0f42761f7ae284\nAll backups with backup name contai...b0f42761f7ae284\nAll backups with pagination:\ntest_bu_a38f75b97487de4b\ntest_bu_8b0f42761f7ae284\n" to include "All backups\ntest_bu_8b0f42761f7ae284"
Diff:
@@ -1,17 +1,33 @@
-All backups\ntest_bu_8b0f42761f7ae284
+All backups
+test_bu_a38f75b97487de4b
+test_bu_8b0f42761f7ae284
+All backups with backup name containing "test_bu_8b0f42761f7ae284":
+test_bu_8b0f42761f7ae284
+All backups for databases with a name containing "test_db_8b0f42761f7ae284":
+test_bu_8b0f42761f7ae284
+All backups that expire before a timestamp:
+test_bu_a38f75b97487de4b
+test_bu_8b0f42761f7ae284
+All backups with a size greater than 500 bytes:
+test_bu_8b0f42761f7ae284
+All backups that were created after a timestamp that are also ready:
+test_bu_8b0f42761f7ae284
+All backups with pagination:
+test_bu_a38f75b97487de4b
+test_bu_8b0f42761f7ae284
./spec/spanner_samples_spec.rb:1680:in `block (2 levels) in <top (required)>'</pre></details> | 1.0 | rspec: Google Cloud Spanner API samples list backups with various filters failed - Note: #755 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 0f1c000e62f9f5a457f16b52784f887c1ba03741
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/dd8ec378-661a-458e-9948-b7977cb10cdf), [Sponge](http://sponge2/dd8ec378-661a-458e-9948-b7977cb10cdf)
status: failed
<details><summary>Test output</summary><br><pre>Failure/Error:
expect(captured_output).to include(
"All backups\n#{backup.backup_id}"
)
expected "All backups\ntest_bu_a38f75b97487de4b\ntest_bu_8b0f42761f7ae284\nAll backups with backup name contai...b0f42761f7ae284\nAll backups with pagination:\ntest_bu_a38f75b97487de4b\ntest_bu_8b0f42761f7ae284\n" to include "All backups\ntest_bu_8b0f42761f7ae284"
Diff:
@@ -1,17 +1,33 @@
-All backups\ntest_bu_8b0f42761f7ae284
+All backups
+test_bu_a38f75b97487de4b
+test_bu_8b0f42761f7ae284
+All backups with backup name containing "test_bu_8b0f42761f7ae284":
+test_bu_8b0f42761f7ae284
+All backups for databases with a name containing "test_db_8b0f42761f7ae284":
+test_bu_8b0f42761f7ae284
+All backups that expire before a timestamp:
+test_bu_a38f75b97487de4b
+test_bu_8b0f42761f7ae284
+All backups with a size greater than 500 bytes:
+test_bu_8b0f42761f7ae284
+All backups that were created after a timestamp that are also ready:
+test_bu_8b0f42761f7ae284
+All backups with pagination:
+test_bu_a38f75b97487de4b
+test_bu_8b0f42761f7ae284
./spec/spanner_samples_spec.rb:1680:in `block (2 levels) in <top (required)>'</pre></details> | priority | rspec google cloud spanner api samples list backups with various filters failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output failure error expect captured output to include all backups n backup backup id expected all backups ntest bu ntest bu nall backups with backup name contai nall backups with pagination ntest bu ntest bu n to include all backups ntest bu diff all backups ntest bu all backups test bu test bu all backups with backup name containing test bu test bu all backups for databases with a name containing test db test bu all backups that expire before a timestamp test bu test bu all backups with a size greater than bytes test bu all backups that were created after a timestamp that are also ready test bu all backups with pagination test bu test bu spec spanner samples spec rb in block levels in | 1 |
144,844 | 19,308,483,191 | IssuesEvent | 2021-12-13 14:03:23 | jgeraigery/tc-ta3-api-bindings-java | https://api.github.com/repos/jgeraigery/tc-ta3-api-bindings-java | opened | CVE-2019-17571 (High) detected in log4j-1.2.17.jar | security vulnerability | ## CVE-2019-17571 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.17.jar</b></p></summary>
<p>Apache Log4j 1.2</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.17.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/tc-ta3-api-bindings-java/commit/2c226dda8de79bfd0db3e278181aebf361463b78">2c226dda8de79bfd0db3e278181aebf361463b78</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Included in Log4j 1.2 is a SocketServer class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data. This affects Log4j versions up to 1.2 up to 1.2.17.
<p>Publish Date: 2019-12-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17571>CVE-2019-17571</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"log4j","packageName":"log4j","packageVersion":"1.2.17","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"log4j:log4j:1.2.17","isMinimumFixVersionAvailable":false,"isBinary":true}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-17571","vulnerabilityDetails":"Included in Log4j 1.2 is a SocketServer class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data. This affects Log4j versions up to 1.2 up to 1.2.17.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17571","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2019-17571 (High) detected in log4j-1.2.17.jar - ## CVE-2019-17571 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.17.jar</b></p></summary>
<p>Apache Log4j 1.2</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.17.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/tc-ta3-api-bindings-java/commit/2c226dda8de79bfd0db3e278181aebf361463b78">2c226dda8de79bfd0db3e278181aebf361463b78</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Included in Log4j 1.2 is a SocketServer class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data. This affects Log4j versions up to 1.2 up to 1.2.17.
<p>Publish Date: 2019-12-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17571>CVE-2019-17571</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"log4j","packageName":"log4j","packageVersion":"1.2.17","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"log4j:log4j:1.2.17","isMinimumFixVersionAvailable":false,"isBinary":true}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-17571","vulnerabilityDetails":"Included in Log4j 1.2 is a SocketServer class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data. This affects Log4j versions up to 1.2 up to 1.2.17.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17571","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in jar cve high severity vulnerability vulnerable library jar apache dependency hierarchy x jar vulnerable library found in head commit a href found in base branch master vulnerability details included in is a socketserver class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data this affects versions up to up to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree isminimumfixversionavailable false isbinary true basebranches vulnerabilityidentifier cve vulnerabilitydetails included in is a socketserver class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data this affects versions up to up to vulnerabilityurl | 0 |
656,748 | 21,774,322,161 | IssuesEvent | 2022-05-13 12:22:54 | barbacbd/nautical | https://api.github.com/repos/barbacbd/nautical | opened | use haversine module | enhancement good first issue Low Priority | Remove the haversine functions from the location/util.py. We can use the haversine module. | 1.0 | use haversine module - Remove the haversine functions from the location/util.py. We can use the haversine module. | priority | use haversine module remove the haversine functions from the location util py we can use the haversine module | 1 |
10,295 | 4,031,576,015 | IssuesEvent | 2016-05-18 17:39:06 | teotidev/guide | https://api.github.com/repos/teotidev/guide | closed | Project information needs more data | code work | - File name
- Location
- Size
- Date created/modified
- .caustic files import... maybe a list with any metadata that was tagged on | 1.0 | Project information needs more data - - File name
- Location
- Size
- Date created/modified
- .caustic files import... maybe a list with any metadata that was tagged on | non_priority | project information needs more data file name location size date created modified caustic files import maybe a list with any metadata that was tagged on | 0 |
436,767 | 30,568,914,047 | IssuesEvent | 2023-07-20 20:11:56 | Alqor-UG/qlued | https://api.github.com/repos/Alqor-UG/qlued | closed | Versionized documentation | documentation | **Is your feature request related to a problem? Please describe.**
We currently only have one documentation running. We need one for different versions.
**Describe the solution you'd like**
It would seem that `mike` is one solution. However, its documentation is really not easy to understand.
**Describe alternatives you've considered**
Going with a single version. But this really leads to plenty of problems for different users.
**Additional context**
We recently introduced versions. So each version should have a proper documentation.
| 1.0 | Versionized documentation - **Is your feature request related to a problem? Please describe.**
We currently only have one documentation running. We need one for different versions.
**Describe the solution you'd like**
It would seem that `mike` is one solution. However, its documentation is really not easy to understand.
**Describe alternatives you've considered**
Going with a single version. But this really leads to plenty of problems for different users.
**Additional context**
We recently introduced versions. So each version should have a proper documentation.
| non_priority | versionized documentation is your feature request related to a problem please describe we currently only have one documentation running we need one for different versions describe the solution you d like it would seem that mike is one solution however its documentation is really not easy to understand describe alternatives you ve considered going with a single version but this really leads to plenty of problems for different users additional context we recently introduced versions so each version should have a proper documentation | 0 |
161,887 | 12,598,360,992 | IssuesEvent | 2020-06-11 02:42:02 | elastic/kibana | https://api.github.com/repos/elastic/kibana | opened | [test-failed]: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/visualize/reporting·ts - Visualize Visualize Print PDF button downloaded PDF has OK status | failed-test test-cloud | **Version: 7.8.0**
**Class: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/visualize/reporting·ts**
**Stack Trace:**
Error: expected 400 to equal 200
at Assertion.assert (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/linux-immutable/ci/cloud/common/build/kibana/packages/kbn-expect/expect.js:100:11)
at Assertion.equal (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/linux-immutable/ci/cloud/common/build/kibana/packages/kbn-expect/expect.js:227:8)
at Context.<anonymous> (test/functional/apps/visualize/reporting.ts:65:35)
at process._tickCallback (internal/process/next_tick.js:68:7)
_Platform: cloud_
_Build Num: 324_ | 2.0 | [test-failed]: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/visualize/reporting·ts - Visualize Visualize Print PDF button downloaded PDF has OK status - **Version: 7.8.0**
**Class: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/visualize/reporting·ts**
**Stack Trace:**
Error: expected 400 to equal 200
at Assertion.assert (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/linux-immutable/ci/cloud/common/build/kibana/packages/kbn-expect/expect.js:100:11)
at Assertion.equal (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp2/TASK/saas_run_kibana_tests/node/linux-immutable/ci/cloud/common/build/kibana/packages/kbn-expect/expect.js:227:8)
at Context.<anonymous> (test/functional/apps/visualize/reporting.ts:65:35)
at process._tickCallback (internal/process/next_tick.js:68:7)
_Platform: cloud_
_Build Num: 324_ | non_priority | chrome x pack ui functional x pack test functional apps visualize reporting·ts visualize visualize print pdf button downloaded pdf has ok status version class chrome x pack ui functional x pack test functional apps visualize reporting·ts stack trace error expected to equal at assertion assert var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node linux immutable ci cloud common build kibana packages kbn expect expect js at assertion equal var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node linux immutable ci cloud common build kibana packages kbn expect expect js at context test functional apps visualize reporting ts at process tickcallback internal process next tick js platform cloud build num | 0 |
16,216 | 10,451,666,323 | IssuesEvent | 2019-09-19 13:18:43 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | Please elaborate on diagnostic settings not being swapped as it pertains to the app's config | Pri1 app-service/svc cxp doc-enhancement triaged | The App Service logs blade appears to use the app settings in a way that's inconsistent with the documentation. For example, in the App Service logs when you make changes (eg. retention days or use of blob storage) Azure will create, and/or edit, settings like WEBSITE_HTTPLOGGING_RETENTION_DAYS, DIAGNOSTICS_AZUREBLOBRETENTIONDAYS, and others. While these are diagnostic settings and are not swapped, they do not appear visibly as "slot settings" which can lead to confusion.
In general, if the platform is able to modify app settings to hold certain key/value pairs it would be nice if they followed the same rules as the end users :-)
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: f6e09089-1ae2-8943-5ce2-9d48f458c81f
* Version Independent ID: ba780cba-f604-b0a4-a81a-23c7d2384762
* Content: [Set up staging environments for web apps in Azure App Service](https://docs.microsoft.com/en-us/azure/app-service/deploy-staging-slots)
* Content Source: [articles/app-service/deploy-staging-slots.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/deploy-staging-slots.md)
* Service: **app-service**
* GitHub Login: @cephalin
* Microsoft Alias: **cephalin** | 1.0 | Please elaborate on diagnostic settings not being swapped as it pertains to the app's config - The App Service logs blade appears to use the app settings in a way that's inconsistent with the documentation. For example, in the App Service logs when you make changes (eg. retention days or use of blob storage) Azure will create, and/or edit, settings like WEBSITE_HTTPLOGGING_RETENTION_DAYS, DIAGNOSTICS_AZUREBLOBRETENTIONDAYS, and others. While these are diagnostic settings and are not swapped, they do not appear visibly as "slot settings" which can lead to confusion.
In general, if the platform is able to modify app settings to hold certain key/value pairs it would be nice if they followed the same rules as the end users :-)
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: f6e09089-1ae2-8943-5ce2-9d48f458c81f
* Version Independent ID: ba780cba-f604-b0a4-a81a-23c7d2384762
* Content: [Set up staging environments for web apps in Azure App Service](https://docs.microsoft.com/en-us/azure/app-service/deploy-staging-slots)
* Content Source: [articles/app-service/deploy-staging-slots.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/deploy-staging-slots.md)
* Service: **app-service**
* GitHub Login: @cephalin
* Microsoft Alias: **cephalin** | non_priority | please elaborate on diagnostic settings not being swapped as it pertains to the app s config the app service logs blade appears to use the app settings in a way that s inconsistent with the documentation for example in the app service logs when you make changes eg retention days or use of blob storage azure will create and or edit settings like website httplogging retention days diagnostics azureblobretentiondays and others while these are diagnostic settings and are not swapped they do not appear visibly as slot settings which can lead to confusion in general if the platform is able to modify app settings to hold certain key value pairs it would be nice if they followed the same rules as the end users document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service app service github login cephalin microsoft alias cephalin | 0 |
700,602 | 24,066,357,680 | IssuesEvent | 2022-09-17 14:48:16 | GaloisInc/cryptol | https://api.github.com/repos/GaloisInc/cryptol | closed | Panic - `not a word value` `ccatV left` and `ccatV right` | bug priority | ```
nightly: Pulling from galoisinc/cryptol
Digest: sha256:87b94650e66333dadcca914f77ebfdad158ec920b960cf19f714e4181b57eb10
Status: Image is up to date for ghcr.io/galoisinc/cryptol:nightly
ghcr.io/galoisinc/cryptol:nightly
┏━╸┏━┓╻ ╻┏━┓╺┳╸┏━┓╻
┃ ┣┳┛┗┳┛┣━┛ ┃ ┃ ┃┃
┗━╸╹┗╸ ╹ ╹ ╹ ┗━┛┗━╸
version 2.13.0.99
https://cryptol.net :? for help
Loading module Cryptol
Cryptol> 0
Showing a specific instance of polymorphic result:
* Using 'Integer' for the type of '(expression)'
0
Cryptol> [0..0]
Showing a specific instance of polymorphic result:
* Using 'Integer' for type argument 'a' of 'Cryptol::fromTo'
[0]
Cryptol> 0 # [0..0]
Showing a specific instance of polymorphic result:
* Using '0' for type argument 'front' of '(Cryptol::#)'
cryptol: You have encountered a bug in Cryptol's implementation.
*** Please create an issue at https://github.com/GaloisInc/cryptol/issues
%< ---------------------------------------------------
Revision: UNKNOWN
Branch: UNKNOWN
Location: [Eval] fromWordVal
Message: not a word value
ccatV right
seq:1
CallStack (from HasCallStack):
panic, called at src/Cryptol/Utils/Panic.hs:21:9 in cryptol-2.13.0.99-inplace:Cryptol.Utils.Panic
panic, called at src/Cryptol/Backend/Monad.hs:410:17 in cryptol-2.13.0.99-inplace:Cryptol.Backend.Monad
evalPanic, called at src/Cryptol/Eval/Value.hs:377:23 in cryptol-2.13.0.99-inplace:Cryptol.Eval.Value
%< ---------------------------------------------------
...restarting cryptol...
Cryptol> [0..0] # 0
Showing a specific instance of polymorphic result:
* Using '0' for type argument 'back' of '(Cryptol::#)'
cryptol: You have encountered a bug in Cryptol's implementation.
*** Please create an issue at https://github.com/GaloisInc/cryptol/issues
%< ---------------------------------------------------
Revision: UNKNOWN
Branch: UNKNOWN
Location: [Eval] fromWordVal
Message: not a word value
ccatV left
seq:1
CallStack (from HasCallStack):
panic, called at src/Cryptol/Utils/Panic.hs:21:9 in cryptol-2.13.0.99-inplace:Cryptol.Utils.Panic
panic, called at src/Cryptol/Backend/Monad.hs:410:17 in cryptol-2.13.0.99-inplace:Cryptol.Backend.Monad
evalPanic, called at src/Cryptol/Eval/Value.hs:377:23 in cryptol-2.13.0.99-inplace:Cryptol.Eval.Value
%< ---------------------------------------------------
``` | 1.0 | Panic - `not a word value` `ccatV left` and `ccatV right` - ```
nightly: Pulling from galoisinc/cryptol
Digest: sha256:87b94650e66333dadcca914f77ebfdad158ec920b960cf19f714e4181b57eb10
Status: Image is up to date for ghcr.io/galoisinc/cryptol:nightly
ghcr.io/galoisinc/cryptol:nightly
┏━╸┏━┓╻ ╻┏━┓╺┳╸┏━┓╻
┃ ┣┳┛┗┳┛┣━┛ ┃ ┃ ┃┃
┗━╸╹┗╸ ╹ ╹ ╹ ┗━┛┗━╸
version 2.13.0.99
https://cryptol.net :? for help
Loading module Cryptol
Cryptol> 0
Showing a specific instance of polymorphic result:
* Using 'Integer' for the type of '(expression)'
0
Cryptol> [0..0]
Showing a specific instance of polymorphic result:
* Using 'Integer' for type argument 'a' of 'Cryptol::fromTo'
[0]
Cryptol> 0 # [0..0]
Showing a specific instance of polymorphic result:
* Using '0' for type argument 'front' of '(Cryptol::#)'
cryptol: You have encountered a bug in Cryptol's implementation.
*** Please create an issue at https://github.com/GaloisInc/cryptol/issues
%< ---------------------------------------------------
Revision: UNKNOWN
Branch: UNKNOWN
Location: [Eval] fromWordVal
Message: not a word value
ccatV right
seq:1
CallStack (from HasCallStack):
panic, called at src/Cryptol/Utils/Panic.hs:21:9 in cryptol-2.13.0.99-inplace:Cryptol.Utils.Panic
panic, called at src/Cryptol/Backend/Monad.hs:410:17 in cryptol-2.13.0.99-inplace:Cryptol.Backend.Monad
evalPanic, called at src/Cryptol/Eval/Value.hs:377:23 in cryptol-2.13.0.99-inplace:Cryptol.Eval.Value
%< ---------------------------------------------------
...restarting cryptol...
Cryptol> [0..0] # 0
Showing a specific instance of polymorphic result:
* Using '0' for type argument 'back' of '(Cryptol::#)'
cryptol: You have encountered a bug in Cryptol's implementation.
*** Please create an issue at https://github.com/GaloisInc/cryptol/issues
%< ---------------------------------------------------
Revision: UNKNOWN
Branch: UNKNOWN
Location: [Eval] fromWordVal
Message: not a word value
ccatV left
seq:1
CallStack (from HasCallStack):
panic, called at src/Cryptol/Utils/Panic.hs:21:9 in cryptol-2.13.0.99-inplace:Cryptol.Utils.Panic
panic, called at src/Cryptol/Backend/Monad.hs:410:17 in cryptol-2.13.0.99-inplace:Cryptol.Backend.Monad
evalPanic, called at src/Cryptol/Eval/Value.hs:377:23 in cryptol-2.13.0.99-inplace:Cryptol.Eval.Value
%< ---------------------------------------------------
``` | priority | panic not a word value ccatv left and ccatv right nightly pulling from galoisinc cryptol digest status image is up to date for ghcr io galoisinc cryptol nightly ghcr io galoisinc cryptol nightly ┏━╸┏━┓╻ ╻┏━┓╺┳╸┏━┓╻ ┃ ┣┳┛┗┳┛┣━┛ ┃ ┃ ┃┃ ┗━╸╹┗╸ ╹ ╹ ╹ ┗━┛┗━╸ version for help loading module cryptol cryptol showing a specific instance of polymorphic result using integer for the type of expression cryptol showing a specific instance of polymorphic result using integer for type argument a of cryptol fromto cryptol showing a specific instance of polymorphic result using for type argument front of cryptol cryptol you have encountered a bug in cryptol s implementation please create an issue at revision unknown branch unknown location fromwordval message not a word value ccatv right seq callstack from hascallstack panic called at src cryptol utils panic hs in cryptol inplace cryptol utils panic panic called at src cryptol backend monad hs in cryptol inplace cryptol backend monad evalpanic called at src cryptol eval value hs in cryptol inplace cryptol eval value restarting cryptol cryptol showing a specific instance of polymorphic result using for type argument back of cryptol cryptol you have encountered a bug in cryptol s implementation please create an issue at revision unknown branch unknown location fromwordval message not a word value ccatv left seq callstack from hascallstack panic called at src cryptol utils panic hs in cryptol inplace cryptol utils panic panic called at src cryptol backend monad hs in cryptol inplace cryptol backend monad evalpanic called at src cryptol eval value hs in cryptol inplace cryptol eval value | 1 |
147,909 | 23,292,396,995 | IssuesEvent | 2022-08-06 03:20:09 | diego-ss/Dodgeball-Masters | https://api.github.com/repos/diego-ss/Dodgeball-Masters | closed | Personagens - Prefabs e Animações | unity design script | - [x] Adicionar novos personagens do Mixamo e adequar animações
- [x] Tela de seleção de personagem | 1.0 | Personagens - Prefabs e Animações - - [x] Adicionar novos personagens do Mixamo e adequar animações
- [x] Tela de seleção de personagem | non_priority | personagens prefabs e animações adicionar novos personagens do mixamo e adequar animações tela de seleção de personagem | 0 |
38,988 | 6,716,648,923 | IssuesEvent | 2017-10-14 11:17:59 | XNightHawk/SE2 | https://api.github.com/repos/XNightHawk/SE2 | closed | Creazione pagina wiki Speech outline | documentation in progress | Come sviluppatore, capire cosa verrà presentato mi aiuta a farmi una migliore idea del prodotto che devo realizzare e di come dovrà funzionare nel suo complesso.
How to demo: Aprendo la wiki del progetto e cliccando su Speech outline comparirà una pagina contenete tutte le informazioni riguardanti la presentazione. | 1.0 | Creazione pagina wiki Speech outline - Come sviluppatore, capire cosa verrà presentato mi aiuta a farmi una migliore idea del prodotto che devo realizzare e di come dovrà funzionare nel suo complesso.
How to demo: Aprendo la wiki del progetto e cliccando su Speech outline comparirà una pagina contenete tutte le informazioni riguardanti la presentazione. | non_priority | creazione pagina wiki speech outline come sviluppatore capire cosa verrà presentato mi aiuta a farmi una migliore idea del prodotto che devo realizzare e di come dovrà funzionare nel suo complesso how to demo aprendo la wiki del progetto e cliccando su speech outline comparirà una pagina contenete tutte le informazioni riguardanti la presentazione | 0 |
67,025 | 27,708,408,792 | IssuesEvent | 2023-03-14 12:48:15 | kubeshop/testkube | https://api.github.com/repos/kubeshop/testkube | closed | UI: Only show chart axes (lines) and labels after 2 executions | service:dashboard 🎡 | The chart axis (lines) and labels are very confusing when we only have 1 execution on the test.
<img width="1280" alt="image" src="https://user-images.githubusercontent.com/132332/221956275-2e7e40ea-e71d-414c-8ac5-3646c2a81561.png">
We'd like to not show them in that case and only display the lines and axis labels after 2 or more executions
| 1.0 | UI: Only show chart axes (lines) and labels after 2 executions - The chart axis (lines) and labels are very confusing when we only have 1 execution on the test.
<img width="1280" alt="image" src="https://user-images.githubusercontent.com/132332/221956275-2e7e40ea-e71d-414c-8ac5-3646c2a81561.png">
We'd like to not show them in that case and only display the lines and axis labels after 2 or more executions
| non_priority | ui only show chart axes lines and labels after executions the chart axis lines and labels are very confusing when we only have execution on the test img width alt image src we d like to not show them in that case and only display the lines and axis labels after or more executions | 0 |
331,831 | 10,077,315,414 | IssuesEvent | 2019-07-24 18:26:19 | grpc/grpc | https://api.github.com/repos/grpc/grpc | opened | [Aio] Channel Argument Implementation | kind/enhancement lang/Python priority/P2 | The `grpc.aio` module underlying is using `asyncio` custom IO manager for both client and server. So, the socket is no longer managed by C-Core. One option that might cause trouble came to my mind is `SO_REUSEPORT`.
We needs to go through available channel arguments in `grpc_types.h`, and figure out which option needs special treatment.
Stretch: since we are implementing channel arguments, probably it is a good chance to promote some common arguments to channel method.
@gnossen @pfreixes | 1.0 | [Aio] Channel Argument Implementation - The `grpc.aio` module underlying is using `asyncio` custom IO manager for both client and server. So, the socket is no longer managed by C-Core. One option that might cause trouble came to my mind is `SO_REUSEPORT`.
We needs to go through available channel arguments in `grpc_types.h`, and figure out which option needs special treatment.
Stretch: since we are implementing channel arguments, probably it is a good chance to promote some common arguments to channel method.
@gnossen @pfreixes | priority | channel argument implementation the grpc aio module underlying is using asyncio custom io manager for both client and server so the socket is no longer managed by c core one option that might cause trouble came to my mind is so reuseport we needs to go through available channel arguments in grpc types h and figure out which option needs special treatment stretch since we are implementing channel arguments probably it is a good chance to promote some common arguments to channel method gnossen pfreixes | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.