Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
757
| labels
stringlengths 4
664
| body
stringlengths 3
261k
| index
stringclasses 10
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
232k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
96,792
| 12,157,448,750
|
IssuesEvent
|
2020-04-25 22:03:19
|
ethereum/solidity
|
https://api.github.com/repos/ethereum/solidity
|
opened
|
Allow constants in interfaces
|
language design :rage4:
|
(Moved this from #3411 which, now confusingly, refers to a different feature.)
Since we allow accessing constants through contract names (#1290) and enums in interfaces (#4087), it may make sense to reconsider this feature, at least for some healthy discussion.
I am still not sure if it is a good idea or not (it definitely can be useful from a users' perspective, but could be dangerous?)
The case I ran into is the following (https://github.com/ralexstokes/deposit-contract-verifying-proxy/pull/2):
```
interface DepositContract {
uint constant PUBLIC_KEY_LENGTH = 48;
uint constant SIGNATURE_LENGTH = 96;
function deposit(
bytes calldata publicKey,
bytes32 withdrawalCredentials,
bytes calldata signature,
bytes32 depositDataRoot
) external payable;
}
```
In this example the interface uses basic types (`bytes`), but they have a limit enforced (the two constant above) so it would be useful from the caller to have `require` statements:
```
contract DepositProxy {
function verifyAndDeposit(...) public {
require(publicKey.length == DepositContract.PUBLIC_KEY_LENGTH, "incorrectly sized public key");
require(signature.length == DepositContract.SIGNATURE_LENGTH, "incorrectly sized signature");
// ...
DepositContract(0x...).deposit(...);
}
}
```
|
1.0
|
Allow constants in interfaces - (Moved this from #3411 which, now confusingly, refers to a different feature.)
Since we allow accessing constants through contract names (#1290) and enums in interfaces (#4087), it may make sense to reconsider this feature, at least for some healthy discussion.
I am still not sure if it is a good idea or not (it definitely can be useful from a users' perspective, but could be dangerous?)
The case I ran into is the following (https://github.com/ralexstokes/deposit-contract-verifying-proxy/pull/2):
```
interface DepositContract {
uint constant PUBLIC_KEY_LENGTH = 48;
uint constant SIGNATURE_LENGTH = 96;
function deposit(
bytes calldata publicKey,
bytes32 withdrawalCredentials,
bytes calldata signature,
bytes32 depositDataRoot
) external payable;
}
```
In this example the interface uses basic types (`bytes`), but they have a limit enforced (the two constant above) so it would be useful from the caller to have `require` statements:
```
contract DepositProxy {
function verifyAndDeposit(...) public {
require(publicKey.length == DepositContract.PUBLIC_KEY_LENGTH, "incorrectly sized public key");
require(signature.length == DepositContract.SIGNATURE_LENGTH, "incorrectly sized signature");
// ...
DepositContract(0x...).deposit(...);
}
}
```
|
non_defect
|
allow constants in interfaces moved this from which now confusingly refers to a different feature since we allow accessing constants through contract names and enums in interfaces it may make sense to reconsider this feature at least for some healthy discussion i am still not sure if it is a good idea or not it definitely can be useful from a users perspective but could be dangerous the case i ran into is the following interface depositcontract uint constant public key length uint constant signature length function deposit bytes calldata publickey withdrawalcredentials bytes calldata signature depositdataroot external payable in this example the interface uses basic types bytes but they have a limit enforced the two constant above so it would be useful from the caller to have require statements contract depositproxy function verifyanddeposit public require publickey length depositcontract public key length incorrectly sized public key require signature length depositcontract signature length incorrectly sized signature depositcontract deposit
| 0
|
7,107
| 2,610,327,391
|
IssuesEvent
|
2015-02-26 19:45:30
|
chrsmith/republic-at-war
|
https://api.github.com/repos/chrsmith/republic-at-war
|
closed
|
Typo
|
auto-migrated Priority-Low Type-Defect
|
```
who is an should be between "Deltasquad" and "explosive" on Scorch's description
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 9 Jun 2011 at 12:21
|
1.0
|
Typo - ```
who is an should be between "Deltasquad" and "explosive" on Scorch's description
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 9 Jun 2011 at 12:21
|
defect
|
typo who is an should be between deltasquad and explosive on scorch s description original issue reported on code google com by gmail com on jun at
| 1
|
49,132
| 13,185,249,346
|
IssuesEvent
|
2020-08-12 21:01:14
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
opened
|
[sim-services] sanity checker - tau samples (Trac #805)
|
Incomplete Migration Migrated from Trac combo simulation defect
|
<details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/805
, reported by olivas and owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-03-18T21:14:03",
"description": "https://icecube-spno.slack.com/files/dschultz/F02V2RF5A/v04-01-07_sanity_checker_error.txt\n\nThis is likely due to the fact that the taus are InIce but outside the active volume. The stochastic aren't tracked in this case. So this is the fix :\n\n1) Make PROPOSAL set InActiveVolume appropriately.\n2) Write a test for the tree checker that reproduces this failure.\n3) Fix it.\n\n",
"reporter": "olivas",
"cc": "",
"resolution": "fixed",
"_ts": "1458335643235016",
"component": "combo simulation",
"summary": "[sim-services] sanity checker - tau samples",
"priority": "blocker",
"keywords": "",
"time": "2014-11-12T01:31:33",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
[sim-services] sanity checker - tau samples (Trac #805) - <details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/805
, reported by olivas and owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-03-18T21:14:03",
"description": "https://icecube-spno.slack.com/files/dschultz/F02V2RF5A/v04-01-07_sanity_checker_error.txt\n\nThis is likely due to the fact that the taus are InIce but outside the active volume. The stochastic aren't tracked in this case. So this is the fix :\n\n1) Make PROPOSAL set InActiveVolume appropriately.\n2) Write a test for the tree checker that reproduces this failure.\n3) Fix it.\n\n",
"reporter": "olivas",
"cc": "",
"resolution": "fixed",
"_ts": "1458335643235016",
"component": "combo simulation",
"summary": "[sim-services] sanity checker - tau samples",
"priority": "blocker",
"keywords": "",
"time": "2014-11-12T01:31:33",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
|
defect
|
sanity checker tau samples trac migrated from reported by olivas and owned by olivas json status closed changetime description is likely due to the fact that the taus are inice but outside the active volume the stochastic aren t tracked in this case so this is the fix n make proposal set inactivevolume appropriately write a test for the tree checker that reproduces this failure fix it n n reporter olivas cc resolution fixed ts component combo simulation summary sanity checker tau samples priority blocker keywords time milestone owner olivas type defect
| 1
|
626,768
| 19,842,560,418
|
IssuesEvent
|
2022-01-21 00:02:33
|
SahilSawantUSA/Scouterdeck-Project
|
https://api.github.com/repos/SahilSawantUSA/Scouterdeck-Project
|
closed
|
Add electron base
|
Priority: High Type: Feature Hardware: RPI
|
Add the base for an electron app which will serve as the main application for the Scouter Deck.
|
1.0
|
Add electron base - Add the base for an electron app which will serve as the main application for the Scouter Deck.
|
non_defect
|
add electron base add the base for an electron app which will serve as the main application for the scouter deck
| 0
|
451,462
| 13,036,287,920
|
IssuesEvent
|
2020-07-28 11:58:50
|
strapi/strapi
|
https://api.github.com/repos/strapi/strapi
|
closed
|
Model using component fails with error since update
|
priority: medium source: plugin:content-manager status: confirmed
|
**Describe the bug**
Any model referencing a component as an attribute fails to open admin collection type screen.
**Steps to reproduce the behavior**
1. Clicking on any collection type that uses a component within Strapi admin results in follow error:
```
Something went wrong.
Details
TypeError: S.includes is not a function
in w
in w
in withRouter(undefined)
in u
in q
in q
in z
in t
in t
in Suspense
in m
in t
in t
in Suspense
in Unknown
in P
in z
in n
in div
in m
in m
in t
in t
in div
in styled.div
in div
in div
in styled.div
in Vi
in le
in n
in z
in injectIntl(Connect(n))
in t
in d
in t
in div
in styled.div
in div
in styled.div
in Unknown
in l
in C
in z
in t
in t
in IntlProvider
in n
in z
in $c
in l
```
**Expected behavior**
The collection type admin screen opens.
**Code snippets**
Model collection type:
```
{
"kind": "collectionType",
"collectionName": "actors",
"info": {
"name": "Actor"
},
"options": {
"increments": true,
"timestamps": false
},
"attributes": {
"fullName": {
"type": "string",
"required": true,
"minLength": 2,
"maxLength": 30
},
"picture": {
"model": "file",
"via": "related",
"allowedTypes": ["images"],
"plugin": "upload",
"required": true
},
"bio": {
"type": "text",
"required": true,
"minLength": 25,
"maxLength": 1000
},
"social": {
"type": "component",
"repeatable": false,
"component": "social.social",
"required": false
},
"website": {
"type": "string",
"maxLength": 253
},
"characters": {
"via": "actor",
"collection": "character"
}
}
}
```
Component type:
```
{
"collectionName": "components_social_socials",
"info": {
"name": "Social",
"icon": "at"
},
"options": {
"timestamps": false
},
"attributes": {
"facebook": {
"type": "string",
"maxLength": 50
},
"instagram": {
"type": "string",
"maxLength": 30
},
"tikTok": {
"type": "string",
"maxLength": 24
},
"twitter": {
"type": "string",
"maxLength": 15
},
"youTube": {
"type": "string",
"maxLength": 32
}
}
}
```
**System**
- Node.js version: 14.4.0
- NPM version: 6.14.5
- Strapi version: 3.1.1
- Database: Mongo
- Operating system: Alpine Linux (Docker Node Official Image)
**Additional context**
This was working on 3.0.1 but since the upgrade to 3.1.1 with no changes the above error occurs.
|
1.0
|
Model using component fails with error since update - **Describe the bug**
Any model referencing a component as an attribute fails to open admin collection type screen.
**Steps to reproduce the behavior**
1. Clicking on any collection type that uses a component within Strapi admin results in follow error:
```
Something went wrong.
Details
TypeError: S.includes is not a function
in w
in w
in withRouter(undefined)
in u
in q
in q
in z
in t
in t
in Suspense
in m
in t
in t
in Suspense
in Unknown
in P
in z
in n
in div
in m
in m
in t
in t
in div
in styled.div
in div
in div
in styled.div
in Vi
in le
in n
in z
in injectIntl(Connect(n))
in t
in d
in t
in div
in styled.div
in div
in styled.div
in Unknown
in l
in C
in z
in t
in t
in IntlProvider
in n
in z
in $c
in l
```
**Expected behavior**
The collection type admin screen opens.
**Code snippets**
Model collection type:
```
{
"kind": "collectionType",
"collectionName": "actors",
"info": {
"name": "Actor"
},
"options": {
"increments": true,
"timestamps": false
},
"attributes": {
"fullName": {
"type": "string",
"required": true,
"minLength": 2,
"maxLength": 30
},
"picture": {
"model": "file",
"via": "related",
"allowedTypes": ["images"],
"plugin": "upload",
"required": true
},
"bio": {
"type": "text",
"required": true,
"minLength": 25,
"maxLength": 1000
},
"social": {
"type": "component",
"repeatable": false,
"component": "social.social",
"required": false
},
"website": {
"type": "string",
"maxLength": 253
},
"characters": {
"via": "actor",
"collection": "character"
}
}
}
```
Component type:
```
{
"collectionName": "components_social_socials",
"info": {
"name": "Social",
"icon": "at"
},
"options": {
"timestamps": false
},
"attributes": {
"facebook": {
"type": "string",
"maxLength": 50
},
"instagram": {
"type": "string",
"maxLength": 30
},
"tikTok": {
"type": "string",
"maxLength": 24
},
"twitter": {
"type": "string",
"maxLength": 15
},
"youTube": {
"type": "string",
"maxLength": 32
}
}
}
```
**System**
- Node.js version: 14.4.0
- NPM version: 6.14.5
- Strapi version: 3.1.1
- Database: Mongo
- Operating system: Alpine Linux (Docker Node Official Image)
**Additional context**
This was working on 3.0.1 but since the upgrade to 3.1.1 with no changes the above error occurs.
|
non_defect
|
model using component fails with error since update describe the bug any model referencing a component as an attribute fails to open admin collection type screen steps to reproduce the behavior clicking on any collection type that uses a component within strapi admin results in follow error something went wrong details typeerror s includes is not a function in w in w in withrouter undefined in u in q in q in z in t in t in suspense in m in t in t in suspense in unknown in p in z in n in div in m in m in t in t in div in styled div in div in div in styled div in vi in le in n in z in injectintl connect n in t in d in t in div in styled div in div in styled div in unknown in l in c in z in t in t in intlprovider in n in z in c in l expected behavior the collection type admin screen opens code snippets model collection type kind collectiontype collectionname actors info name actor options increments true timestamps false attributes fullname type string required true minlength maxlength picture model file via related allowedtypes plugin upload required true bio type text required true minlength maxlength social type component repeatable false component social social required false website type string maxlength characters via actor collection character component type collectionname components social socials info name social icon at options timestamps false attributes facebook type string maxlength instagram type string maxlength tiktok type string maxlength twitter type string maxlength youtube type string maxlength system node js version npm version strapi version database mongo operating system alpine linux docker node official image additional context this was working on but since the upgrade to with no changes the above error occurs
| 0
|
78,723
| 27,734,516,200
|
IssuesEvent
|
2023-03-15 10:18:38
|
SeleniumHQ/selenium
|
https://api.github.com/repos/SeleniumHQ/selenium
|
closed
|
[🐛 Bug]: send_keys() is messing up indentation
|
I-defect I-issue-template
|
### What happened?


When using "ctrl+c" and "ctrl+v" to copy and paste the code to the textarea, its' indentation doesn't change. But when we use send_keys() to enter the text in, its' indentation will be in a mess.
### How can we reproduce the issue?
```shell
We just wanna enter the code in a common textarea.
```
### Relevant log output
```shell
.
```
### Operating System
macOS
### Selenium version
Python
### What are the browser(s) and version(s) where you see this issue?
Chrome 110
### What are the browser driver(s) and version(s) where you see this issue?
ChromeDriver
### Are you using Selenium Grid?
4.8.2
|
1.0
|
[🐛 Bug]: send_keys() is messing up indentation - ### What happened?


When using "ctrl+c" and "ctrl+v" to copy and paste the code to the textarea, its' indentation doesn't change. But when we use send_keys() to enter the text in, its' indentation will be in a mess.
### How can we reproduce the issue?
```shell
We just wanna enter the code in a common textarea.
```
### Relevant log output
```shell
.
```
### Operating System
macOS
### Selenium version
Python
### What are the browser(s) and version(s) where you see this issue?
Chrome 110
### What are the browser driver(s) and version(s) where you see this issue?
ChromeDriver
### Are you using Selenium Grid?
4.8.2
|
defect
|
send keys is messing up indentation what happened when using ctrl c and ctrl v to copy and paste the code to the textarea its indentation doesn t change but when we use send keys to enter the text in its indentation will be in a mess how can we reproduce the issue shell we just wanna enter the code in a common textarea relevant log output shell operating system macos selenium version python what are the browser s and version s where you see this issue chrome what are the browser driver s and version s where you see this issue chromedriver are you using selenium grid
| 1
|
59,339
| 17,023,091,059
|
IssuesEvent
|
2021-07-03 00:20:32
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
interactive map edit queries
|
Component: admin Priority: major Resolution: fixed Type: defect
|
**[Submitted to the original trac issue database at 12.36pm, Friday, 16th December 2005]**
From the editor, or another interface, ability to query the edit history of points/segments.
|
1.0
|
interactive map edit queries - **[Submitted to the original trac issue database at 12.36pm, Friday, 16th December 2005]**
From the editor, or another interface, ability to query the edit history of points/segments.
|
defect
|
interactive map edit queries from the editor or another interface ability to query the edit history of points segments
| 1
|
74,962
| 25,457,811,317
|
IssuesEvent
|
2022-11-24 15:35:52
|
SAP/fundamental-styles
|
https://api.github.com/repos/SAP/fundamental-styles
|
closed
|
Input field truncation bug
|
Bug Defect Hunting version 1.0
|
discussed with @manu-kr and the issue is coming from Input component, not related to File Uploader
It's coming from Fundamental-styles
#### Is this a bug, enhancement, or feature request?
bug
#### Briefly describe your proposal.
- [ ] File Uploader with truncation loses truncation when a file is selected:
Initially:
<img width="535" alt="Screen Shot 2021-06-17 at 5 11 48 PM" src="https://user-images.githubusercontent.com/39598672/122472994-4633b980-cf8f-11eb-8ef9-f4a8cffc62f9.png">
After you've clicked to upload a file:
<img width="636" alt="Screen Shot 2021-06-17 at 5 11 57 PM" src="https://user-images.githubusercontent.com/39598672/122473013-4d5ac780-cf8f-11eb-815a-918ea6020efb.png">
|
1.0
|
Input field truncation bug - discussed with @manu-kr and the issue is coming from Input component, not related to File Uploader
It's coming from Fundamental-styles
#### Is this a bug, enhancement, or feature request?
bug
#### Briefly describe your proposal.
- [ ] File Uploader with truncation loses truncation when a file is selected:
Initially:
<img width="535" alt="Screen Shot 2021-06-17 at 5 11 48 PM" src="https://user-images.githubusercontent.com/39598672/122472994-4633b980-cf8f-11eb-8ef9-f4a8cffc62f9.png">
After you've clicked to upload a file:
<img width="636" alt="Screen Shot 2021-06-17 at 5 11 57 PM" src="https://user-images.githubusercontent.com/39598672/122473013-4d5ac780-cf8f-11eb-815a-918ea6020efb.png">
|
defect
|
input field truncation bug discussed with manu kr and the issue is coming from input component not related to file uploader it s coming from fundamental styles is this a bug enhancement or feature request bug briefly describe your proposal file uploader with truncation loses truncation when a file is selected initially img width alt screen shot at pm src after you ve clicked to upload a file img width alt screen shot at pm src
| 1
|
810,336
| 30,237,292,845
|
IssuesEvent
|
2023-07-06 11:08:54
|
slynch8/10x
|
https://api.github.com/repos/slynch8/10x
|
closed
|
AddBuildFinishedFunction always has result as true
|
bug Priority 2 trivial current done
|
Result passed to the function given to AddBuildFinishedFunction always seems to be true even tho there're build errors & warnings. Also it would be nice to have GetErrorCount & GetWarningCount.
|
1.0
|
AddBuildFinishedFunction always has result as true - Result passed to the function given to AddBuildFinishedFunction always seems to be true even tho there're build errors & warnings. Also it would be nice to have GetErrorCount & GetWarningCount.
|
non_defect
|
addbuildfinishedfunction always has result as true result passed to the function given to addbuildfinishedfunction always seems to be true even tho there re build errors warnings also it would be nice to have geterrorcount getwarningcount
| 0
|
403,046
| 11,834,993,639
|
IssuesEvent
|
2020-03-23 09:53:38
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
m.youtube.com - desktop site instead of mobile site
|
browser-firefox-mobile engine-gecko priority-critical
|
<!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 6.0; Mobile; rv:68.0) Gecko/20100101 Firefox/68.0 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/50527 -->
**URL**: https://m.youtube.com/watch?v=-f0IF0NYmVI
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android 6.0
**Tested Another Browser**: Yes
**Problem type**: Desktop site instead of mobile site
**Description**: 8
**Steps to Reproduce**:
https://m.youtube.com/watch?v=-f0IF0NYmVI
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200406090101</li><li>channel: alpha</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/3/c6f04e00-cc48-4c0a-8641-a1e74b22395f)
Submitted in the name of `@peivado`
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
m.youtube.com - desktop site instead of mobile site - <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 6.0; Mobile; rv:68.0) Gecko/20100101 Firefox/68.0 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/50527 -->
**URL**: https://m.youtube.com/watch?v=-f0IF0NYmVI
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android 6.0
**Tested Another Browser**: Yes
**Problem type**: Desktop site instead of mobile site
**Description**: 8
**Steps to Reproduce**:
https://m.youtube.com/watch?v=-f0IF0NYmVI
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200406090101</li><li>channel: alpha</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/3/c6f04e00-cc48-4c0a-8641-a1e74b22395f)
Submitted in the name of `@peivado`
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_defect
|
m youtube com desktop site instead of mobile site url browser version firefox mobile operating system android tested another browser yes problem type desktop site instead of mobile site description steps to reproduce browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel alpha hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false submitted in the name of peivado from with ❤️
| 0
|
49,079
| 13,185,220,561
|
IssuesEvent
|
2020-08-12 20:57:55
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
opened
|
Update offline_trunk/metaproject/quickstart.html (Trac #701)
|
Incomplete Migration Migrated from Trac defect documentation
|
<details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/701
, reported by blaufuss and owned by blaufuss</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-02-11T19:37:48",
"description": "These are old and crufty. \n\nProbably best to leave specifics out, bur rather make it a collection of internal\nlinks to other rst docs. Limit updates to a single place.\n\n",
"reporter": "blaufuss",
"cc": "",
"resolution": "wontfix",
"_ts": "1423683468641230",
"component": "documentation",
"summary": "Update offline_trunk/metaproject/quickstart.html",
"priority": "normal",
"keywords": "",
"time": "2012-11-06T02:09:11",
"milestone": "",
"owner": "blaufuss",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
Update offline_trunk/metaproject/quickstart.html (Trac #701) - <details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/701
, reported by blaufuss and owned by blaufuss</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-02-11T19:37:48",
"description": "These are old and crufty. \n\nProbably best to leave specifics out, bur rather make it a collection of internal\nlinks to other rst docs. Limit updates to a single place.\n\n",
"reporter": "blaufuss",
"cc": "",
"resolution": "wontfix",
"_ts": "1423683468641230",
"component": "documentation",
"summary": "Update offline_trunk/metaproject/quickstart.html",
"priority": "normal",
"keywords": "",
"time": "2012-11-06T02:09:11",
"milestone": "",
"owner": "blaufuss",
"type": "defect"
}
```
</p>
</details>
|
defect
|
update offline trunk metaproject quickstart html trac migrated from reported by blaufuss and owned by blaufuss json status closed changetime description these are old and crufty n nprobably best to leave specifics out bur rather make it a collection of internal nlinks to other rst docs limit updates to a single place n n reporter blaufuss cc resolution wontfix ts component documentation summary update offline trunk metaproject quickstart html priority normal keywords time milestone owner blaufuss type defect
| 1
|
14,953
| 2,832,206,518
|
IssuesEvent
|
2015-05-25 05:26:40
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
closed
|
Footer moves down with Horizontal Scrolling and Selection
|
defect
|
In Internet Explorer 9 (IE9), with a DataTable if we enable Row Selection (single or multiple) and we enable scrollable with a scrollWidth small enough to cause horizontal scrollbar, the footer of the DataTable moves down the page whenever we mouseover the data rows. The more you move the mouse on mouseover, the more it moves down. Mouse movement can be either side-to-side or up-and-down. Clicking on an item in the list will reset the footer back to normal.
This is reproducible by taking the PrimeFaces ShowCase application and modifying either ui/datatableRowSelectionSingle.xhtml or ui/datatableRowSelectionMultiple.xhtml by adding the following attributes to the <p:dataTable> tag:
scrollable="true" scrollWidth="250"
Note that this appears to only occur in IE9. IE8 (or IE8 compatibility mode) and IE10 do not exhibit this behavior.
|
1.0
|
Footer moves down with Horizontal Scrolling and Selection - In Internet Explorer 9 (IE9), with a DataTable if we enable Row Selection (single or multiple) and we enable scrollable with a scrollWidth small enough to cause horizontal scrollbar, the footer of the DataTable moves down the page whenever we mouseover the data rows. The more you move the mouse on mouseover, the more it moves down. Mouse movement can be either side-to-side or up-and-down. Clicking on an item in the list will reset the footer back to normal.
This is reproducible by taking the PrimeFaces ShowCase application and modifying either ui/datatableRowSelectionSingle.xhtml or ui/datatableRowSelectionMultiple.xhtml by adding the following attributes to the <p:dataTable> tag:
scrollable="true" scrollWidth="250"
Note that this appears to only occur in IE9. IE8 (or IE8 compatibility mode) and IE10 do not exhibit this behavior.
|
defect
|
footer moves down with horizontal scrolling and selection in internet explorer with a datatable if we enable row selection single or multiple and we enable scrollable with a scrollwidth small enough to cause horizontal scrollbar the footer of the datatable moves down the page whenever we mouseover the data rows the more you move the mouse on mouseover the more it moves down mouse movement can be either side to side or up and down clicking on an item in the list will reset the footer back to normal this is reproducible by taking the primefaces showcase application and modifying either ui datatablerowselectionsingle xhtml or ui datatablerowselectionmultiple xhtml by adding the following attributes to the tag scrollable true scrollwidth note that this appears to only occur in or compatibility mode and do not exhibit this behavior
| 1
|
333,171
| 24,365,774,611
|
IssuesEvent
|
2022-10-03 15:03:35
|
Punpun1643/alpha6
|
https://api.github.com/repos/Punpun1643/alpha6
|
opened
|
Bug0 - trying all sort of markdown stuff
|
type.DocumentationBug severity.Medium
|
---
__Advertisement :)__
- __[pica](https://nodeca.github.io/pica/demo/)__ - high quality and fast image
resize in browser.
- __[babelfish](https://github.com/nodeca/babelfish/)__ - developer friendly
i18n with plurals support and easy syntax.
You will like those projects!
---
# h1 Heading 8-)
## h2 Heading
### h3 Heading
#### h4 Heading
##### h5 Heading
###### h6 Heading
## Horizontal Rules
___
---
***
## Typographic replacements
Enable typographer option to see result.
(c) (C) (r) (R) (tm) (TM) (p) (P) +-
test.. test... test..... test?..... test!....
!!!!!! ???? ,, -- ---
"Smartypants, double quotes" and 'single quotes'
## Emphasis
**This is bold text**
__This is bold text__
*This is italic text*
_This is italic text_
~~Strikethrough~~
## Blockquotes
> Blockquotes can also be nested...
>> ...by using additional greater-than signs right next to each other...
> > > ...or with spaces between arrows.
## Lists
Unordered
+ Create a list by starting a line with `+`, `-`, or `*`
+ Sub-lists are made by indenting 2 spaces:
- Marker character change forces new list start:
* Ac tristique libero volutpat at
+ Facilisis in pretium nisl aliquet
- Nulla volutpat aliquam velit
+ Very easy!
Ordered
1. Lorem ipsum dolor sit amet
2. Consectetur adipiscing elit
3. Integer molestie lorem at massa
1. You can use sequential numbers...
1. ...or keep all the numbers as `1.`
Start numbering with offset:
57. foo
1. bar
## Code
Inline `code`
Indented code
// Some comments
line 1 of code
line 2 of code
line 3 of code
Block code "fences"
```
Sample text here...
```
Syntax highlighting
``` js
var foo = function (bar) {
return bar++;
};
console.log(foo(5));
```
## Tables
| Option | Description |
| ------ | ----------- |
| data | path to data files to supply the data that will be passed into templates. |
| engine | engine to be used for processing templates. Handlebars is the default. |
| ext | extension to be used for dest files. |
Right aligned columns
| Option | Description |
| ------:| -----------:|
| data | path to data files to supply the data that will be passed into templates. |
| engine | engine to be used for processing templates. Handlebars is the default. |
| ext | extension to be used for dest files. |
## Links
[link text](http://dev.nodeca.com)
[link with title](http://nodeca.github.io/pica/demo/ "title text!")
Autoconverted link https://github.com/nodeca/pica (enable linkify to see)
## Images


Like links, Images also have a footnote style syntax
![Alt text][id]
With a reference later in the document defining the URL location:
[id]: https://octodex.github.com/images/dojocat.jpg "The Dojocat"
## Plugins
The killer feature of `markdown-it` is very effective support of
[syntax plugins](https://www.npmjs.org/browse/keyword/markdown-it-plugin).
### [Emojies](https://github.com/markdown-it/markdown-it-emoji)
> Classic markup: :wink: :crush: :cry: :tear: :laughing: :yum:
>
> Shortcuts (emoticons): :-) :-( 8-) ;)
see [how to change output](https://github.com/markdown-it/markdown-it-emoji#change-output) with twemoji.
### [Subscript](https://github.com/markdown-it/markdown-it-sub) / [Superscript](https://github.com/markdown-it/markdown-it-sup)
- 19^th^
- H~2~O
### [\<ins>](https://github.com/markdown-it/markdown-it-ins)
++Inserted text++
### [\<mark>](https://github.com/markdown-it/markdown-it-mark)
==Marked text==
### [Footnotes](https://github.com/markdown-it/markdown-it-footnote)
Footnote 1 link[^first].
Footnote 2 link[^second].
Inline footnote^[Text of inline footnote] definition.
Duplicated footnote reference[^second].
[^first]: Footnote **can have markup**
and multiple paragraphs.
[^second]: Footnote text.
### [Definition lists](https://github.com/markdown-it/markdown-it-deflist)
Term 1
: Definition 1
with lazy continuation.
Term 2 with *inline markup*
: Definition 2
{ some code, part of Definition 2 }
Third paragraph of definition 2.
_Compact style:_
Term 1
~ Definition 1
Term 2
~ Definition 2a
~ Definition 2b
### [Abbreviations](https://github.com/markdown-it/markdown-it-abbr)
This is HTML abbreviation example.
It converts "HTML", but keep intact partial entries like "xxxHTMLyyy" and so on.
*[HTML]: Hyper Text Markup Language
### [Custom containers](https://github.com/markdown-it/markdown-it-container)
::: warning
*here be dragons*
:::
<!--session: 1664809250731-e4bde0ca-c0c8-49fd-a50b-c8dacb2781db-->
<!--Version: Web v3.4.3-->
|
1.0
|
Bug0 - trying all sort of markdown stuff - ---
__Advertisement :)__
- __[pica](https://nodeca.github.io/pica/demo/)__ - high quality and fast image
resize in browser.
- __[babelfish](https://github.com/nodeca/babelfish/)__ - developer friendly
i18n with plurals support and easy syntax.
You will like those projects!
---
# h1 Heading 8-)
## h2 Heading
### h3 Heading
#### h4 Heading
##### h5 Heading
###### h6 Heading
## Horizontal Rules
___
---
***
## Typographic replacements
Enable typographer option to see result.
(c) (C) (r) (R) (tm) (TM) (p) (P) +-
test.. test... test..... test?..... test!....
!!!!!! ???? ,, -- ---
"Smartypants, double quotes" and 'single quotes'
## Emphasis
**This is bold text**
__This is bold text__
*This is italic text*
_This is italic text_
~~Strikethrough~~
## Blockquotes
> Blockquotes can also be nested...
>> ...by using additional greater-than signs right next to each other...
> > > ...or with spaces between arrows.
## Lists
Unordered
+ Create a list by starting a line with `+`, `-`, or `*`
+ Sub-lists are made by indenting 2 spaces:
- Marker character change forces new list start:
* Ac tristique libero volutpat at
+ Facilisis in pretium nisl aliquet
- Nulla volutpat aliquam velit
+ Very easy!
Ordered
1. Lorem ipsum dolor sit amet
2. Consectetur adipiscing elit
3. Integer molestie lorem at massa
1. You can use sequential numbers...
1. ...or keep all the numbers as `1.`
Start numbering with offset:
57. foo
1. bar
## Code
Inline `code`
Indented code
// Some comments
line 1 of code
line 2 of code
line 3 of code
Block code "fences"
```
Sample text here...
```
Syntax highlighting
``` js
var foo = function (bar) {
return bar++;
};
console.log(foo(5));
```
## Tables
| Option | Description |
| ------ | ----------- |
| data | path to data files to supply the data that will be passed into templates. |
| engine | engine to be used for processing templates. Handlebars is the default. |
| ext | extension to be used for dest files. |
Right aligned columns
| Option | Description |
| ------:| -----------:|
| data | path to data files to supply the data that will be passed into templates. |
| engine | engine to be used for processing templates. Handlebars is the default. |
| ext | extension to be used for dest files. |
## Links
[link text](http://dev.nodeca.com)
[link with title](http://nodeca.github.io/pica/demo/ "title text!")
Autoconverted link https://github.com/nodeca/pica (enable linkify to see)
## Images


Like links, Images also have a footnote style syntax
![Alt text][id]
With a reference later in the document defining the URL location:
[id]: https://octodex.github.com/images/dojocat.jpg "The Dojocat"
## Plugins
The killer feature of `markdown-it` is very effective support of
[syntax plugins](https://www.npmjs.org/browse/keyword/markdown-it-plugin).
### [Emojies](https://github.com/markdown-it/markdown-it-emoji)
> Classic markup: :wink: :crush: :cry: :tear: :laughing: :yum:
>
> Shortcuts (emoticons): :-) :-( 8-) ;)
see [how to change output](https://github.com/markdown-it/markdown-it-emoji#change-output) with twemoji.
### [Subscript](https://github.com/markdown-it/markdown-it-sub) / [Superscript](https://github.com/markdown-it/markdown-it-sup)
- 19^th^
- H~2~O
### [\<ins>](https://github.com/markdown-it/markdown-it-ins)
++Inserted text++
### [\<mark>](https://github.com/markdown-it/markdown-it-mark)
==Marked text==
### [Footnotes](https://github.com/markdown-it/markdown-it-footnote)
Footnote 1 link[^first].
Footnote 2 link[^second].
Inline footnote^[Text of inline footnote] definition.
Duplicated footnote reference[^second].
[^first]: Footnote **can have markup**
and multiple paragraphs.
[^second]: Footnote text.
### [Definition lists](https://github.com/markdown-it/markdown-it-deflist)
Term 1
: Definition 1
with lazy continuation.
Term 2 with *inline markup*
: Definition 2
{ some code, part of Definition 2 }
Third paragraph of definition 2.
_Compact style:_
Term 1
~ Definition 1
Term 2
~ Definition 2a
~ Definition 2b
### [Abbreviations](https://github.com/markdown-it/markdown-it-abbr)
This is HTML abbreviation example.
It converts "HTML", but keep intact partial entries like "xxxHTMLyyy" and so on.
*[HTML]: Hyper Text Markup Language
### [Custom containers](https://github.com/markdown-it/markdown-it-container)
::: warning
*here be dragons*
:::
<!--session: 1664809250731-e4bde0ca-c0c8-49fd-a50b-c8dacb2781db-->
<!--Version: Web v3.4.3-->
|
non_defect
|
trying all sort of markdown stuff advertisement high quality and fast image resize in browser developer friendly with plurals support and easy syntax you will like those projects heading heading heading heading heading heading horizontal rules typographic replacements enable typographer option to see result c c r r tm tm p p test test test test test smartypants double quotes and single quotes emphasis this is bold text this is bold text this is italic text this is italic text strikethrough blockquotes blockquotes can also be nested by using additional greater than signs right next to each other or with spaces between arrows lists unordered create a list by starting a line with or sub lists are made by indenting spaces marker character change forces new list start ac tristique libero volutpat at facilisis in pretium nisl aliquet nulla volutpat aliquam velit very easy ordered lorem ipsum dolor sit amet consectetur adipiscing elit integer molestie lorem at massa you can use sequential numbers or keep all the numbers as start numbering with offset foo bar code inline code indented code some comments line of code line of code line of code block code fences sample text here syntax highlighting js var foo function bar return bar console log foo tables option description data path to data files to supply the data that will be passed into templates engine engine to be used for processing templates handlebars is the default ext extension to be used for dest files right aligned columns option description data path to data files to supply the data that will be passed into templates engine engine to be used for processing templates handlebars is the default ext extension to be used for dest files links title text autoconverted link enable linkify to see images the stormtroopocat like links images also have a footnote style syntax with a reference later in the document defining the url location the dojocat plugins the killer feature of markdown it is very effective support of classic markup wink crush cry tear laughing yum shortcuts emoticons see with twemoji th h o inserted text marked text footnote link footnote link inline footnote definition duplicated footnote reference footnote can have markup and multiple paragraphs footnote text term definition with lazy continuation term with inline markup definition some code part of definition third paragraph of definition compact style term definition term definition definition this is html abbreviation example it converts html but keep intact partial entries like xxxhtmlyyy and so on hyper text markup language warning here be dragons
| 0
|
185,054
| 21,785,058,176
|
IssuesEvent
|
2022-05-14 02:19:36
|
Yash-Handa/GitHub-Org-Geographics
|
https://api.github.com/repos/Yash-Handa/GitHub-Org-Geographics
|
closed
|
CVE-2020-8116 (High) detected in dot-prop-4.2.0.tgz - autoclosed
|
security vulnerability
|
## CVE-2020-8116 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>dot-prop-4.2.0.tgz</b></p></summary>
<p>Get, set, or delete a property from a nested object using a dot path</p>
<p>Library home page: <a href="https://registry.npmjs.org/dot-prop/-/dot-prop-4.2.0.tgz">https://registry.npmjs.org/dot-prop/-/dot-prop-4.2.0.tgz</a></p>
<p>Path to dependency file: /functions/package.json</p>
<p>Path to vulnerable library: /functions/node_modules/dot-prop/package.json</p>
<p>
Dependency Hierarchy:
- firebase-admin-7.0.0.tgz (Root Library)
- storage-2.5.0.tgz
- gcs-resumable-upload-1.1.0.tgz
- configstore-4.0.0.tgz
- :x: **dot-prop-4.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Yash-Handa/GitHub-Org-Geographics/commit/0f5764a8d4fca2f9276cb62d0d976a22d119f8ee">0f5764a8d4fca2f9276cb62d0d976a22d119f8ee</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype pollution vulnerability in dot-prop npm package versions before 4.2.1 and versions 5.x before 5.1.1 allows an attacker to add arbitrary properties to JavaScript language constructs such as objects.
<p>Publish Date: 2020-02-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8116>CVE-2020-8116</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8116">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8116</a></p>
<p>Release Date: 2020-02-04</p>
<p>Fix Resolution: dot-prop - 5.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-8116 (High) detected in dot-prop-4.2.0.tgz - autoclosed - ## CVE-2020-8116 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>dot-prop-4.2.0.tgz</b></p></summary>
<p>Get, set, or delete a property from a nested object using a dot path</p>
<p>Library home page: <a href="https://registry.npmjs.org/dot-prop/-/dot-prop-4.2.0.tgz">https://registry.npmjs.org/dot-prop/-/dot-prop-4.2.0.tgz</a></p>
<p>Path to dependency file: /functions/package.json</p>
<p>Path to vulnerable library: /functions/node_modules/dot-prop/package.json</p>
<p>
Dependency Hierarchy:
- firebase-admin-7.0.0.tgz (Root Library)
- storage-2.5.0.tgz
- gcs-resumable-upload-1.1.0.tgz
- configstore-4.0.0.tgz
- :x: **dot-prop-4.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Yash-Handa/GitHub-Org-Geographics/commit/0f5764a8d4fca2f9276cb62d0d976a22d119f8ee">0f5764a8d4fca2f9276cb62d0d976a22d119f8ee</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype pollution vulnerability in dot-prop npm package versions before 4.2.1 and versions 5.x before 5.1.1 allows an attacker to add arbitrary properties to JavaScript language constructs such as objects.
<p>Publish Date: 2020-02-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8116>CVE-2020-8116</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8116">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8116</a></p>
<p>Release Date: 2020-02-04</p>
<p>Fix Resolution: dot-prop - 5.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in dot prop tgz autoclosed cve high severity vulnerability vulnerable library dot prop tgz get set or delete a property from a nested object using a dot path library home page a href path to dependency file functions package json path to vulnerable library functions node modules dot prop package json dependency hierarchy firebase admin tgz root library storage tgz gcs resumable upload tgz configstore tgz x dot prop tgz vulnerable library found in head commit a href vulnerability details prototype pollution vulnerability in dot prop npm package versions before and versions x before allows an attacker to add arbitrary properties to javascript language constructs such as objects publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution dot prop step up your open source security game with whitesource
| 0
|
157,646
| 6,010,260,995
|
IssuesEvent
|
2017-06-06 12:47:59
|
theam/haskell-do
|
https://api.github.com/repos/theam/haskell-do
|
closed
|
Backend doesn't spawn on first opening
|
Bug High priority
|
It seems like the backend can't spawn when I reboot my laptop and pc on first opening (ubuntu 16.04 and windows 10). The web socket is right the first time I execute the app but it doesn't show in haskell-do.
|
1.0
|
Backend doesn't spawn on first opening - It seems like the backend can't spawn when I reboot my laptop and pc on first opening (ubuntu 16.04 and windows 10). The web socket is right the first time I execute the app but it doesn't show in haskell-do.
|
non_defect
|
backend doesn t spawn on first opening it seems like the backend can t spawn when i reboot my laptop and pc on first opening ubuntu and windows the web socket is right the first time i execute the app but it doesn t show in haskell do
| 0
|
113,485
| 24,424,546,801
|
IssuesEvent
|
2022-10-06 00:42:16
|
alexander-wise/RvLineList
|
https://api.github.com/repos/alexander-wise/RvLineList
|
closed
|
Are you using a floating point variable to store integers?
|
bug code quality
|
https://github.com/alexander-wise/RvLineList/blob/6f57cd6572fb8678bb252fe64019d1e76394317a/src/empirical_line_lists.jl#L72
If I were starting from scratch, I'd use a different data structure here, since (I think) you're storing different information in different columns.
|
1.0
|
Are you using a floating point variable to store integers? - https://github.com/alexander-wise/RvLineList/blob/6f57cd6572fb8678bb252fe64019d1e76394317a/src/empirical_line_lists.jl#L72
If I were starting from scratch, I'd use a different data structure here, since (I think) you're storing different information in different columns.
|
non_defect
|
are you using a floating point variable to store integers if i were starting from scratch i d use a different data structure here since i think you re storing different information in different columns
| 0
|
244,870
| 18,768,653,829
|
IssuesEvent
|
2021-11-06 12:08:28
|
girlscript/winter-of-contributing
|
https://api.github.com/repos/girlscript/winter-of-contributing
|
closed
|
Matrix
|
documentation GWOC21 Assigned C/CPP
|
### Description
What is Matrix
Declaration, working, and traversal
Matrix will be a folder.
This is will be the intro matrix .md file
Further files will be added accordingly!
### Domain
C/CPP
### Type of Contribution
Documentation
### Code of Conduct
- [X] I follow [Contributing Guidelines](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CONTRIBUTING.md) & [Code of conduct](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CODE_OF_CONDUCT.md) of this project.
|
1.0
|
Matrix - ### Description
What is Matrix
Declaration, working, and traversal
Matrix will be a folder.
This is will be the intro matrix .md file
Further files will be added accordingly!
### Domain
C/CPP
### Type of Contribution
Documentation
### Code of Conduct
- [X] I follow [Contributing Guidelines](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CONTRIBUTING.md) & [Code of conduct](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CODE_OF_CONDUCT.md) of this project.
|
non_defect
|
matrix description what is matrix declaration working and traversal matrix will be a folder this is will be the intro matrix md file further files will be added accordingly domain c cpp type of contribution documentation code of conduct i follow of this project
| 0
|
171
| 2,517,527,541
|
IssuesEvent
|
2015-01-16 15:28:18
|
contao/core
|
https://api.github.com/repos/contao/core
|
closed
|
Input.php - Unset for static::$arrCache['postUnsafeRaw'] missing.
|
defect
|
With the new function "postUnsafeRaw" you added a new "static::$arrCache['postUnsafeRaw']".
See #4614. We have the same porblem again. It would be nice if the "setPost" function removes the "static::$arrCache['postUnsafeRaw'][$strKey]", too.
See https://github.com/contao/core/blob/support/3.2/system/modules/core/library/Contao/Input.php#L359
|
1.0
|
Input.php - Unset for static::$arrCache['postUnsafeRaw'] missing. - With the new function "postUnsafeRaw" you added a new "static::$arrCache['postUnsafeRaw']".
See #4614. We have the same porblem again. It would be nice if the "setPost" function removes the "static::$arrCache['postUnsafeRaw'][$strKey]", too.
See https://github.com/contao/core/blob/support/3.2/system/modules/core/library/Contao/Input.php#L359
|
defect
|
input php unset for static arrcache missing with the new function postunsaferaw you added a new static arrcache see we have the same porblem again it would be nice if the setpost function removes the static arrcache too see
| 1
|
25,981
| 4,540,072,141
|
IssuesEvent
|
2016-09-09 13:33:00
|
stunpix/stacklessexamples
|
https://api.github.com/repos/stunpix/stacklessexamples
|
closed
|
receive bug in stacklesssocket30.py
|
auto-migrated Priority-Medium Type-Defect
|
```
The bug is in stacklesssocket30.py.
What steps will reproduce the problem?
1. Try receiving data on a socket using the recv_into() function.
2.
3.
What is the expected output? What do you see instead?
It will fail like this:
File "stacklesssocket.py", line 249, in recv
self.recv_into(b, byteCount, flags)
File "/stacklesssocket.py", line 303, in recv_into
buffer[:] = self.readBytes
TypeError: an integer is required
What version of the product are you using? On what operating system?
I use the latest version of the stacklesssocket30.py module from the trunk.
Please provide any additional information below.
Fix:
The problem is the statement at line 283.
It says:
"self.readBytes = self.recvChannel.receive()"
but should actually say:
"self.readBytes, address = self.recvChannel.receive()"
```
Original issue reported on code.google.com by `sebastia...@gmail.com` on 23 Jun 2011 at 7:09
|
1.0
|
receive bug in stacklesssocket30.py - ```
The bug is in stacklesssocket30.py.
What steps will reproduce the problem?
1. Try receiving data on a socket using the recv_into() function.
2.
3.
What is the expected output? What do you see instead?
It will fail like this:
File "stacklesssocket.py", line 249, in recv
self.recv_into(b, byteCount, flags)
File "/stacklesssocket.py", line 303, in recv_into
buffer[:] = self.readBytes
TypeError: an integer is required
What version of the product are you using? On what operating system?
I use the latest version of the stacklesssocket30.py module from the trunk.
Please provide any additional information below.
Fix:
The problem is the statement at line 283.
It says:
"self.readBytes = self.recvChannel.receive()"
but should actually say:
"self.readBytes, address = self.recvChannel.receive()"
```
Original issue reported on code.google.com by `sebastia...@gmail.com` on 23 Jun 2011 at 7:09
|
defect
|
receive bug in py the bug is in py what steps will reproduce the problem try receiving data on a socket using the recv into function what is the expected output what do you see instead it will fail like this file stacklesssocket py line in recv self recv into b bytecount flags file stacklesssocket py line in recv into buffer self readbytes typeerror an integer is required what version of the product are you using on what operating system i use the latest version of the py module from the trunk please provide any additional information below fix the problem is the statement at line it says self readbytes self recvchannel receive but should actually say self readbytes address self recvchannel receive original issue reported on code google com by sebastia gmail com on jun at
| 1
|
127,123
| 18,010,290,683
|
IssuesEvent
|
2021-09-16 07:48:40
|
maddyCode23/linux-4.1.15
|
https://api.github.com/repos/maddyCode23/linux-4.1.15
|
opened
|
CVE-2017-9075 (High) detected in linux-stable-rtv4.1.33
|
security vulnerability
|
## CVE-2017-9075 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sctp/ipv6.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The sctp_v6_create_accept_sk function in net/sctp/ipv6.c in the Linux kernel through 4.11.1 mishandles inheritance, which allows local users to cause a denial of service or possibly have unspecified other impact via crafted system calls, a related issue to CVE-2017-8890.
<p>Publish Date: 2017-05-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-9075>CVE-2017-9075</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-9075">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-9075</a></p>
<p>Release Date: 2017-05-19</p>
<p>Fix Resolution: v4.12-rc2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2017-9075 (High) detected in linux-stable-rtv4.1.33 - ## CVE-2017-9075 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sctp/ipv6.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The sctp_v6_create_accept_sk function in net/sctp/ipv6.c in the Linux kernel through 4.11.1 mishandles inheritance, which allows local users to cause a denial of service or possibly have unspecified other impact via crafted system calls, a related issue to CVE-2017-8890.
<p>Publish Date: 2017-05-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-9075>CVE-2017-9075</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-9075">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-9075</a></p>
<p>Release Date: 2017-05-19</p>
<p>Fix Resolution: v4.12-rc2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href vulnerable source files net sctp c vulnerability details the sctp create accept sk function in net sctp c in the linux kernel through mishandles inheritance which allows local users to cause a denial of service or possibly have unspecified other impact via crafted system calls a related issue to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
289,424
| 21,782,752,342
|
IssuesEvent
|
2022-05-13 20:55:38
|
lumeland/docker
|
https://api.github.com/repos/lumeland/docker
|
closed
|
Improve Docker Hub page
|
documentation
|
Lume's [Docker Hub](https://hub.docker.com/r/oscarotero/lume/) page is blank currently. The initial work for the Docker images is done, so it's about time to improve the Docker Hub page.
[There](https://hub.docker.com/_/gradle/) [are](https://hub.docker.com/_/eclipse-temurin) [good](https://hub.docker.com/r/denoland/deno) examples to take a look at and probably use them as a guide.
|
1.0
|
Improve Docker Hub page - Lume's [Docker Hub](https://hub.docker.com/r/oscarotero/lume/) page is blank currently. The initial work for the Docker images is done, so it's about time to improve the Docker Hub page.
[There](https://hub.docker.com/_/gradle/) [are](https://hub.docker.com/_/eclipse-temurin) [good](https://hub.docker.com/r/denoland/deno) examples to take a look at and probably use them as a guide.
|
non_defect
|
improve docker hub page lume s page is blank currently the initial work for the docker images is done so it s about time to improve the docker hub page examples to take a look at and probably use them as a guide
| 0
|
330,726
| 24,274,747,089
|
IssuesEvent
|
2022-09-28 13:07:03
|
Qiskit/qiskit.org
|
https://api.github.com/repos/Qiskit/qiskit.org
|
closed
|
Document local development instructions
|
documentation project: marketing
|
We need to document how to work on development tasks in the `q-site` project.
|
1.0
|
Document local development instructions - We need to document how to work on development tasks in the `q-site` project.
|
non_defect
|
document local development instructions we need to document how to work on development tasks in the q site project
| 0
|
5,724
| 8,183,087,581
|
IssuesEvent
|
2018-08-29 07:58:00
|
pingcap/tidb
|
https://api.github.com/repos/pingcap/tidb
|
closed
|
SHOW GRANTS FOR CURRENT_USER() does not work
|
type/compatibility
|
Hi.
I'm execute command(by mysql cli)
`SHOW GRANTS FOR CURRENT_USER();`
get the error
`
ERROR 1105 (HY000): line 1 column 28 near "()" (total length 30)
`
tidb version:
`5.7.1-TiDB-v1.1.0-alpha-459-ga2a48b3`
BTW .
tidb support MySQL Protocol , but not very compatible with MySQL's existing client tools ( `Navicat`, `MySQL-Front`,`mysql workbench` and so on ) . Less convenient for developers to use it.
|
True
|
SHOW GRANTS FOR CURRENT_USER() does not work - Hi.
I'm execute command(by mysql cli)
`SHOW GRANTS FOR CURRENT_USER();`
get the error
`
ERROR 1105 (HY000): line 1 column 28 near "()" (total length 30)
`
tidb version:
`5.7.1-TiDB-v1.1.0-alpha-459-ga2a48b3`
BTW .
tidb support MySQL Protocol , but not very compatible with MySQL's existing client tools ( `Navicat`, `MySQL-Front`,`mysql workbench` and so on ) . Less convenient for developers to use it.
|
non_defect
|
show grants for current user does not work hi i m execute command(by mysql cli) show grants for current user get the error error line column near total length tidb version tidb alpha btw tidb support mysql protocol but not very compatible with mysql s existing client tools navicat , mysql front , mysql workbench and so on less convenient for developers to use it
| 0
|
40,813
| 10,167,638,399
|
IssuesEvent
|
2019-08-07 18:43:18
|
NREL/EnergyPlus
|
https://api.github.com/repos/NREL/EnergyPlus
|
closed
|
Allow evaporative coolers to cycle
|
Defect InProgress MigratedFromUserVoice SeverityHigh
|
Problem: Many residential evaporative cooling technologies (e.g., swamp coolers) are controlled to cycle on and off to meet the load. As stated in the I/O reference, these units cannot cycle, meaning that their pumps run continuously even when the required airflow rate is close to zero.
Solution: Give zone evaporative coolers the capability to cycle like unitary equipment.
Context: See the second paragraph [here](http://bigladdersoftware.com/epx/docs/8-3/input-output-reference/group-zone-forced-air-units.html#zonehvacevaporativecoolerunit)
Migrated from [UserVoice](http://energyplus.uservoice.com/forums/258860/suggestions/9118099) feedback from @nealkruis
|
1.0
|
Allow evaporative coolers to cycle - Problem: Many residential evaporative cooling technologies (e.g., swamp coolers) are controlled to cycle on and off to meet the load. As stated in the I/O reference, these units cannot cycle, meaning that their pumps run continuously even when the required airflow rate is close to zero.
Solution: Give zone evaporative coolers the capability to cycle like unitary equipment.
Context: See the second paragraph [here](http://bigladdersoftware.com/epx/docs/8-3/input-output-reference/group-zone-forced-air-units.html#zonehvacevaporativecoolerunit)
Migrated from [UserVoice](http://energyplus.uservoice.com/forums/258860/suggestions/9118099) feedback from @nealkruis
|
defect
|
allow evaporative coolers to cycle problem many residential evaporative cooling technologies e g swamp coolers are controlled to cycle on and off to meet the load as stated in the i o reference these units cannot cycle meaning that their pumps run continuously even when the required airflow rate is close to zero solution give zone evaporative coolers the capability to cycle like unitary equipment context see the second paragraph migrated from feedback from nealkruis
| 1
|
393,066
| 26,971,414,758
|
IssuesEvent
|
2023-02-09 05:23:49
|
SashenJayathilaka/Library_Management_System
|
https://api.github.com/repos/SashenJayathilaka/Library_Management_System
|
closed
|
Library Management System in java
|
documentation
|
# Library_Management_System

Library Management System in java
# MySql Database





|
1.0
|
Library Management System in java - # Library_Management_System

Library Management System in java
# MySql Database





|
non_defect
|
library management system in java library management system library management system in java mysql database
| 0
|
16,879
| 2,955,895,136
|
IssuesEvent
|
2015-07-08 07:47:25
|
mclinker/mclinker
|
https://api.github.com/repos/mclinker/mclinker
|
closed
|
DebugInfo has wrong info when gdb debugging
|
auto-migrated Milestone-Release2.0 Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. Write a simple helloworld.c and gcc it to a.out (Use -v to dump detail
commands)
3. Replace /usr/bin/ld to use ld.mcld to link
4. Now, use 'gdb ./a.out' and 'b main'
Expected:
Breakpoint 1, main() at test.c:3
Result:
Breakpoint 1, GNU C 4.8.2 -mtune=generic -march=x86-64 -g -O2 -fstack-protector
() at test.c:3
What version of the product are you using? On what operating system?
x86_64 host, linux
I guess it is introduced by previously DebugInfo patchset.
Please check. Thanks!
```
Original issue reported on code.google.com by `wenhan...@gmail.com` on 31 Dec 2014 at 9:43
|
1.0
|
DebugInfo has wrong info when gdb debugging - ```
What steps will reproduce the problem?
1. Write a simple helloworld.c and gcc it to a.out (Use -v to dump detail
commands)
3. Replace /usr/bin/ld to use ld.mcld to link
4. Now, use 'gdb ./a.out' and 'b main'
Expected:
Breakpoint 1, main() at test.c:3
Result:
Breakpoint 1, GNU C 4.8.2 -mtune=generic -march=x86-64 -g -O2 -fstack-protector
() at test.c:3
What version of the product are you using? On what operating system?
x86_64 host, linux
I guess it is introduced by previously DebugInfo patchset.
Please check. Thanks!
```
Original issue reported on code.google.com by `wenhan...@gmail.com` on 31 Dec 2014 at 9:43
|
defect
|
debuginfo has wrong info when gdb debugging what steps will reproduce the problem write a simple helloworld c and gcc it to a out use v to dump detail commands replace usr bin ld to use ld mcld to link now use gdb a out and b main expected breakpoint main at test c result breakpoint gnu c mtune generic march g fstack protector at test c what version of the product are you using on what operating system host linux i guess it is introduced by previously debuginfo patchset please check thanks original issue reported on code google com by wenhan gmail com on dec at
| 1
|
84,957
| 3,682,518,633
|
IssuesEvent
|
2016-02-24 10:03:58
|
MoOx/statinamic
|
https://api.github.com/repos/MoOx/statinamic
|
closed
|
Internal link (from markdown) are normal link and trigger full page load
|
level: high-priority type: enhancement
|
That's a shame isn't it?
Before handling this, we need to work on #11 (cause this will handle markdown as react components instead of html).
|
1.0
|
Internal link (from markdown) are normal link and trigger full page load - That's a shame isn't it?
Before handling this, we need to work on #11 (cause this will handle markdown as react components instead of html).
|
non_defect
|
internal link from markdown are normal link and trigger full page load that s a shame isn t it before handling this we need to work on cause this will handle markdown as react components instead of html
| 0
|
775,033
| 27,216,001,045
|
IssuesEvent
|
2023-02-20 21:55:22
|
GoogleCloudPlatform/cloud-sql-python-connector
|
https://api.github.com/repos/GoogleCloudPlatform/cloud-sql-python-connector
|
closed
|
system.test_asyncpg_connection: test_connection_with_asyncpg failed
|
type: bug priority: p2 flakybot: issue
|
Note: #597 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 79d227784531f61cca70ebbaa8e9c75c438fd4f4
buildURL: https://github.com/GoogleCloudPlatform/cloud-sql-python-connector/actions/runs/4227214516
status: failed
<details><summary>Test output</summary><br><pre>@pytest.fixture(name="conn")
async def setup() -> AsyncGenerator:
# initialize Cloud SQL Python Connector object
connector = await create_async_connector()
> conn: asyncpg.Connection = await connector.connect_async(
os.environ["POSTGRES_CONNECTION_NAME"],
"asyncpg",
user=os.environ["POSTGRES_USER"],
password=os.environ["POSTGRES_PASS"],
db=os.environ["POSTGRES_DB"],
)
tests/system/test_asyncpg_connection.py:31:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
google/cloud/sql/connector/connector.py:261: in connect_async
return await asyncio.wait_for(get_connection(), timeout)
../../../hostedtoolcache/Python/3.8.16/x64/lib/python3.8/asyncio/tasks.py:494: in wait_for
return fut.result()
google/cloud/sql/connector/connector.py:237: in get_connection
instance_data, ip_address = await instance.connect_info(ip_type)
google/cloud/sql/connector/instance.py:445: in connect_info
instance_data = await self._current
google/cloud/sql/connector/instance.py:388: in _refresh_task
refresh_data = await refresh_task
google/cloud/sql/connector/instance.py:325: in _perform_refresh
ephemeral_cert = await ephemeral_task
google/cloud/sql/connector/refresh_utils.py:191: in _get_ephemeral
resp = await client_session.post(
.nox/system-3-8/lib/python3.8/site-packages/aiohttp/client.py:643: in _request
resp.raise_for_status()
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <ClientResponse(https://sqladmin.googleapis.com/sql/v1beta4/projects/cloud-sql-connector-testing/instances/postgres-pr...ype-Options': 'nosniff', 'Alt-Svc': 'h3=":443"; ma=2592000,h3-29=":443"; ma=2592000', 'Transfer-Encoding': 'chunked')>
def raise_for_status(self) -> None:
if not self.ok:
# reason should always be not None for a started response
assert self.reason is not None
self.release()
> raise ClientResponseError(
self.request_info,
self.history,
status=self.status,
message=self.reason,
headers=self.headers,
)
E aiohttp.client_exceptions.ClientResponseError: 429, message='Too Many Requests', url=URL('https://sqladmin.googleapis.com/sql/v1beta4/projects/cloud-sql-connector-testing/instances/postgres-proxy-testing:generateEphemeralCert')
.nox/system-3-8/lib/python3.8/site-packages/aiohttp/client_reqrep.py:1005: ClientResponseError</pre></details>
|
1.0
|
system.test_asyncpg_connection: test_connection_with_asyncpg failed - Note: #597 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 79d227784531f61cca70ebbaa8e9c75c438fd4f4
buildURL: https://github.com/GoogleCloudPlatform/cloud-sql-python-connector/actions/runs/4227214516
status: failed
<details><summary>Test output</summary><br><pre>@pytest.fixture(name="conn")
async def setup() -> AsyncGenerator:
# initialize Cloud SQL Python Connector object
connector = await create_async_connector()
> conn: asyncpg.Connection = await connector.connect_async(
os.environ["POSTGRES_CONNECTION_NAME"],
"asyncpg",
user=os.environ["POSTGRES_USER"],
password=os.environ["POSTGRES_PASS"],
db=os.environ["POSTGRES_DB"],
)
tests/system/test_asyncpg_connection.py:31:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
google/cloud/sql/connector/connector.py:261: in connect_async
return await asyncio.wait_for(get_connection(), timeout)
../../../hostedtoolcache/Python/3.8.16/x64/lib/python3.8/asyncio/tasks.py:494: in wait_for
return fut.result()
google/cloud/sql/connector/connector.py:237: in get_connection
instance_data, ip_address = await instance.connect_info(ip_type)
google/cloud/sql/connector/instance.py:445: in connect_info
instance_data = await self._current
google/cloud/sql/connector/instance.py:388: in _refresh_task
refresh_data = await refresh_task
google/cloud/sql/connector/instance.py:325: in _perform_refresh
ephemeral_cert = await ephemeral_task
google/cloud/sql/connector/refresh_utils.py:191: in _get_ephemeral
resp = await client_session.post(
.nox/system-3-8/lib/python3.8/site-packages/aiohttp/client.py:643: in _request
resp.raise_for_status()
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <ClientResponse(https://sqladmin.googleapis.com/sql/v1beta4/projects/cloud-sql-connector-testing/instances/postgres-pr...ype-Options': 'nosniff', 'Alt-Svc': 'h3=":443"; ma=2592000,h3-29=":443"; ma=2592000', 'Transfer-Encoding': 'chunked')>
def raise_for_status(self) -> None:
if not self.ok:
# reason should always be not None for a started response
assert self.reason is not None
self.release()
> raise ClientResponseError(
self.request_info,
self.history,
status=self.status,
message=self.reason,
headers=self.headers,
)
E aiohttp.client_exceptions.ClientResponseError: 429, message='Too Many Requests', url=URL('https://sqladmin.googleapis.com/sql/v1beta4/projects/cloud-sql-connector-testing/instances/postgres-proxy-testing:generateEphemeralCert')
.nox/system-3-8/lib/python3.8/site-packages/aiohttp/client_reqrep.py:1005: ClientResponseError</pre></details>
|
non_defect
|
system test asyncpg connection test connection with asyncpg failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output pytest fixture name conn async def setup asyncgenerator initialize cloud sql python connector object connector await create async connector conn asyncpg connection await connector connect async os environ asyncpg user os environ password os environ db os environ tests system test asyncpg connection py google cloud sql connector connector py in connect async return await asyncio wait for get connection timeout hostedtoolcache python lib asyncio tasks py in wait for return fut result google cloud sql connector connector py in get connection instance data ip address await instance connect info ip type google cloud sql connector instance py in connect info instance data await self current google cloud sql connector instance py in refresh task refresh data await refresh task google cloud sql connector instance py in perform refresh ephemeral cert await ephemeral task google cloud sql connector refresh utils py in get ephemeral resp await client session post nox system lib site packages aiohttp client py in request resp raise for status self def raise for status self none if not self ok reason should always be not none for a started response assert self reason is not none self release raise clientresponseerror self request info self history status self status message self reason headers self headers e aiohttp client exceptions clientresponseerror message too many requests url url nox system lib site packages aiohttp client reqrep py clientresponseerror
| 0
|
39,995
| 6,793,831,045
|
IssuesEvent
|
2017-11-01 09:29:12
|
commons-app/apps-android-commons
|
https://api.github.com/repos/commons-app/apps-android-commons
|
opened
|
Tidy up Commons:Mobile app page
|
documentation
|
https://commons.wikimedia.org/wiki/Commons:Mobile_app could potentially be due for a bit of a tidy-up, to reflect the current state of the app. Any ideas for how we could structure the information and what else to add?
|
1.0
|
Tidy up Commons:Mobile app page - https://commons.wikimedia.org/wiki/Commons:Mobile_app could potentially be due for a bit of a tidy-up, to reflect the current state of the app. Any ideas for how we could structure the information and what else to add?
|
non_defect
|
tidy up commons mobile app page could potentially be due for a bit of a tidy up to reflect the current state of the app any ideas for how we could structure the information and what else to add
| 0
|
21,717
| 11,351,730,635
|
IssuesEvent
|
2020-01-24 11:57:06
|
kyma-project/kyma
|
https://api.github.com/repos/kyma-project/kyma
|
closed
|
Production grade profile for Console Backend Service
|
area/console area/performance
|
**Description**
Looks like Console Backend Service hits its memory limits on some clusters (see [linked issue](https://github.com/kyma-project/kyma/issues/6667)). This usually happens right after pod starts. There seems to be a memory consumption peak that happens immediately after pod starts
Analyse realistic memory requirements for CBS. Identify and adress potential memory leaks.
**Define a production grade setup for CBS chart that will ensure stability and eliminates the crash loops**
**Reasons**
Ensure reliability and stability of CBS component on production kyma runtimes
https://github.com/kyma-project/kyma/issues/4139
|
True
|
Production grade profile for Console Backend Service - **Description**
Looks like Console Backend Service hits its memory limits on some clusters (see [linked issue](https://github.com/kyma-project/kyma/issues/6667)). This usually happens right after pod starts. There seems to be a memory consumption peak that happens immediately after pod starts
Analyse realistic memory requirements for CBS. Identify and adress potential memory leaks.
**Define a production grade setup for CBS chart that will ensure stability and eliminates the crash loops**
**Reasons**
Ensure reliability and stability of CBS component on production kyma runtimes
https://github.com/kyma-project/kyma/issues/4139
|
non_defect
|
production grade profile for console backend service description looks like console backend service hits its memory limits on some clusters see this usually happens right after pod starts there seems to be a memory consumption peak that happens immediately after pod starts analyse realistic memory requirements for cbs identify and adress potential memory leaks define a production grade setup for cbs chart that will ensure stability and eliminates the crash loops reasons ensure reliability and stability of cbs component on production kyma runtimes
| 0
|
716,779
| 24,648,135,151
|
IssuesEvent
|
2022-10-17 16:19:05
|
wazuh/wazuh-kibana-app
|
https://api.github.com/repos/wazuh/wazuh-kibana-app
|
closed
|
Blank window due to conflicts with the cached assets
|
bug priority/high compatibility cat-4
|
| Wazuh | OpenSearch | Rev | Security |
| ----- | ------- | ---- | -------- |
| 4.3.6 | 1.2.0 | 4xxx | - |
| Browser |
| ------- |
| Chrome, Firefox, Safari, etc|
## Description

This issue has been affecting our plugin for a long time. We need to investigate how to solve this issue in order to refresh the assets cache when necessary, this is, when the version of the cached assets and the actual plugin version do not match, for example, after a plugin update.
For other team members, check [this conversation in Slack](https://wazuh-team.slack.com/archives/C01RN8WTRHS/p1656663010549839).
## Kibana
We researched how the cache versioning works in Kibana, and found out that there was an issue with it, being fixed in Kibana 7.13.0.
- Related issue in Kibana:
- https://github.com/elastic/kibana/issues/94951
- Included in Kibana v7.13.0
For older versions, our plugin notifies the user with a toast, indicating that there might be a problem with the cached assets, and that a clear refresh is required.
For versions newer than 7.13.0, the assets are properly cached using the plugin build number, so no toast is shown.

## OpenSearch Dashboards
As OpenSearch was forked from Kibana at v7.10.2, this fix wasn't included. They have an issue on their repo about this matter, being fixed in OSD v1.3.1.
- Related issue in OpenSearch:
- https://github.com/opensearch-project/OpenSearch-Dashboards/issues/1363
- Included in OSD v1.3.1, and back ported to previous versions.
As Wazuh currently uses the OpenSearch 1.2.0 stack (indexer + dashboard), this issue is present in our OSD - Wazuh Dashboard plugin. A toast must be shown to notify the user about this.



However, the PR that fixed the broken cache buster on OSD do not take into account the plugins, but the platform itself. So, using different Wazuh versions with the same OSD version will result in conflicts with the plugin's cache. The `buildNumber` in the `package.json` from OS defines how the cached assets are used, as seen in this [PR](https://github.com/opensearch-project/OpenSearch-Dashboards/pull/1371/files).
## Proposed solution
We need the @wazuh/cicd team to change the OSD build number on the `package.json` with a unique number for each Wazuh version, using the Wazuh version itself or the App revision number.
### PoC


|
1.0
|
Blank window due to conflicts with the cached assets - | Wazuh | OpenSearch | Rev | Security |
| ----- | ------- | ---- | -------- |
| 4.3.6 | 1.2.0 | 4xxx | - |
| Browser |
| ------- |
| Chrome, Firefox, Safari, etc|
## Description

This issue has been affecting our plugin for a long time. We need to investigate how to solve this issue in order to refresh the assets cache when necessary, this is, when the version of the cached assets and the actual plugin version do not match, for example, after a plugin update.
For other team members, check [this conversation in Slack](https://wazuh-team.slack.com/archives/C01RN8WTRHS/p1656663010549839).
## Kibana
We researched how the cache versioning works in Kibana, and found out that there was an issue with it, being fixed in Kibana 7.13.0.
- Related issue in Kibana:
- https://github.com/elastic/kibana/issues/94951
- Included in Kibana v7.13.0
For older versions, our plugin notifies the user with a toast, indicating that there might be a problem with the cached assets, and that a clear refresh is required.
For versions newer than 7.13.0, the assets are properly cached using the plugin build number, so no toast is shown.

## OpenSearch Dashboards
As OpenSearch was forked from Kibana at v7.10.2, this fix wasn't included. They have an issue on their repo about this matter, being fixed in OSD v1.3.1.
- Related issue in OpenSearch:
- https://github.com/opensearch-project/OpenSearch-Dashboards/issues/1363
- Included in OSD v1.3.1, and back ported to previous versions.
As Wazuh currently uses the OpenSearch 1.2.0 stack (indexer + dashboard), this issue is present in our OSD - Wazuh Dashboard plugin. A toast must be shown to notify the user about this.



However, the PR that fixed the broken cache buster on OSD do not take into account the plugins, but the platform itself. So, using different Wazuh versions with the same OSD version will result in conflicts with the plugin's cache. The `buildNumber` in the `package.json` from OS defines how the cached assets are used, as seen in this [PR](https://github.com/opensearch-project/OpenSearch-Dashboards/pull/1371/files).
## Proposed solution
We need the @wazuh/cicd team to change the OSD build number on the `package.json` with a unique number for each Wazuh version, using the Wazuh version itself or the App revision number.
### PoC


|
non_defect
|
blank window due to conflicts with the cached assets wazuh opensearch rev security browser chrome firefox safari etc description this issue has been affecting our plugin for a long time we need to investigate how to solve this issue in order to refresh the assets cache when necessary this is when the version of the cached assets and the actual plugin version do not match for example after a plugin update for other team members check kibana we researched how the cache versioning works in kibana and found out that there was an issue with it being fixed in kibana related issue in kibana included in kibana for older versions our plugin notifies the user with a toast indicating that there might be a problem with the cached assets and that a clear refresh is required for versions newer than the assets are properly cached using the plugin build number so no toast is shown opensearch dashboards as opensearch was forked from kibana at this fix wasn t included they have an issue on their repo about this matter being fixed in osd related issue in opensearch included in osd and back ported to previous versions as wazuh currently uses the opensearch stack indexer dashboard this issue is present in our osd wazuh dashboard plugin a toast must be shown to notify the user about this however the pr that fixed the broken cache buster on osd do not take into account the plugins but the platform itself so using different wazuh versions with the same osd version will result in conflicts with the plugin s cache the buildnumber in the package json from os defines how the cached assets are used as seen in this proposed solution we need the wazuh cicd team to change the osd build number on the package json with a unique number for each wazuh version using the wazuh version itself or the app revision number poc
| 0
|
583,313
| 17,382,100,822
|
IssuesEvent
|
2021-07-31 23:28:03
|
chuck2147/offseason-2021
|
https://api.github.com/repos/chuck2147/offseason-2021
|
closed
|
Shooter - Run-Time Settings
|
Medium Priority enhancement
|
Need a way to change at run-time the speed of the upper and lower wheels for the shooter. This can be done by setting the speeds in the Network Tables. A UI for the Driver Station needs to be established to change the Network Table settings.
|
1.0
|
Shooter - Run-Time Settings - Need a way to change at run-time the speed of the upper and lower wheels for the shooter. This can be done by setting the speeds in the Network Tables. A UI for the Driver Station needs to be established to change the Network Table settings.
|
non_defect
|
shooter run time settings need a way to change at run time the speed of the upper and lower wheels for the shooter this can be done by setting the speeds in the network tables a ui for the driver station needs to be established to change the network table settings
| 0
|
71,626
| 23,734,777,878
|
IssuesEvent
|
2022-08-31 07:04:10
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Unable to decrypt
|
T-Defect S-Major A-E2EE Z-UISI O-Uncommon
|
### Steps to reproduce
1. I'm getting "Unable to decrypt" errors in a room where I used to have a normal conversation with the sender
### Outcome
#### What did you expect?
I expected messages to be decrypted
### Operating system
Debian Linux
### Browser information
Firefox 103.0
### URL for webapp
chat.element.io
### Application version
Element version: 1.11.3 Olm version: 3.2.12
### Homeserver
_No response_
### Will you send logs?
Yes
|
1.0
|
Unable to decrypt - ### Steps to reproduce
1. I'm getting "Unable to decrypt" errors in a room where I used to have a normal conversation with the sender
### Outcome
#### What did you expect?
I expected messages to be decrypted
### Operating system
Debian Linux
### Browser information
Firefox 103.0
### URL for webapp
chat.element.io
### Application version
Element version: 1.11.3 Olm version: 3.2.12
### Homeserver
_No response_
### Will you send logs?
Yes
|
defect
|
unable to decrypt steps to reproduce i m getting unable to decrypt errors in a room where i used to have a normal conversation with the sender outcome what did you expect i expected messages to be decrypted operating system debian linux browser information firefox url for webapp chat element io application version element version olm version homeserver no response will you send logs yes
| 1
|
29,445
| 5,693,473,654
|
IssuesEvent
|
2017-04-15 01:52:32
|
hugotacito/django-pagination
|
https://api.github.com/repos/hugotacito/django-pagination
|
closed
|
Making it work with Jinja2 (coffin)
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. use Jinja2 as template engine via coffin (http://github.com/dcramer/coffin)
2. overriding pagination/pagination.html with own Jinja2 template
3. pagination templatetags are unknown
TemplateSyntaxError: unknown tag 'autopaginate' File ...
Try to add them:
from coffin.template import Library
from pagination.templatetags import pagination_tags
from jinja2 import contextfunction
#make django pagination work with jinja2
register = Library()
#make paginate a contextfunction in jinja2
paginate = contextfunction(pagination_tags.paginate)
paginate = register.object(paginate)
autopaginate = register.tag('autopaginate', pagination_tags.do_autopaginate)
What is the expected output? What do you see instead?
A paginated view. An exception.
What version of the product are you using? On what operating system?
1.0.5
Ubuntu 9.10
Please provide any additional information below.
None that I can think of.
Cheers for any help!
```
Original issue reported on code.google.com by `sportjun...@gmail.com` on 17 Dec 2009 at 9:55
|
1.0
|
Making it work with Jinja2 (coffin) - ```
What steps will reproduce the problem?
1. use Jinja2 as template engine via coffin (http://github.com/dcramer/coffin)
2. overriding pagination/pagination.html with own Jinja2 template
3. pagination templatetags are unknown
TemplateSyntaxError: unknown tag 'autopaginate' File ...
Try to add them:
from coffin.template import Library
from pagination.templatetags import pagination_tags
from jinja2 import contextfunction
#make django pagination work with jinja2
register = Library()
#make paginate a contextfunction in jinja2
paginate = contextfunction(pagination_tags.paginate)
paginate = register.object(paginate)
autopaginate = register.tag('autopaginate', pagination_tags.do_autopaginate)
What is the expected output? What do you see instead?
A paginated view. An exception.
What version of the product are you using? On what operating system?
1.0.5
Ubuntu 9.10
Please provide any additional information below.
None that I can think of.
Cheers for any help!
```
Original issue reported on code.google.com by `sportjun...@gmail.com` on 17 Dec 2009 at 9:55
|
defect
|
making it work with coffin what steps will reproduce the problem use as template engine via coffin overriding pagination pagination html with own template pagination templatetags are unknown templatesyntaxerror unknown tag autopaginate file try to add them from coffin template import library from pagination templatetags import pagination tags from import contextfunction make django pagination work with register library make paginate a contextfunction in paginate contextfunction pagination tags paginate paginate register object paginate autopaginate register tag autopaginate pagination tags do autopaginate what is the expected output what do you see instead a paginated view an exception what version of the product are you using on what operating system ubuntu please provide any additional information below none that i can think of cheers for any help original issue reported on code google com by sportjun gmail com on dec at
| 1
|
24,431
| 3,979,988,029
|
IssuesEvent
|
2016-05-06 03:54:47
|
extnet/Ext.NET
|
https://api.github.com/repos/extnet/Ext.NET
|
opened
|
Ext.grid.plugin.RowExpander has expanded rows reset when view is refreshed
|
3.x 4.x defect
|
Reported on this forum thread: [Illogical work collapse\expand in grouping gridpanel](http://forums.ext.net/showthread.php?61113)
The grid with an expanded row like this:

When a group is collapsed, the row expanded body is no more:

This discussion in a sencha thread looks much like it: [Extended Grid: renderer problem with store update](https://www.sencha.com/forum/showthread.php?90760). Thus a limitation from the grid panel's design causes this.
There may be a way to auto-reload the expanded row's contents when the view is refreshed.
On the grid above, the following is enough to break rowExpander: `App.GridPanel1.view.refreshView()`.
This gets called by grouping plugin when expanding/collapsing groups, thus the grid breaks.
|
1.0
|
Ext.grid.plugin.RowExpander has expanded rows reset when view is refreshed - Reported on this forum thread: [Illogical work collapse\expand in grouping gridpanel](http://forums.ext.net/showthread.php?61113)
The grid with an expanded row like this:

When a group is collapsed, the row expanded body is no more:

This discussion in a sencha thread looks much like it: [Extended Grid: renderer problem with store update](https://www.sencha.com/forum/showthread.php?90760). Thus a limitation from the grid panel's design causes this.
There may be a way to auto-reload the expanded row's contents when the view is refreshed.
On the grid above, the following is enough to break rowExpander: `App.GridPanel1.view.refreshView()`.
This gets called by grouping plugin when expanding/collapsing groups, thus the grid breaks.
|
defect
|
ext grid plugin rowexpander has expanded rows reset when view is refreshed reported on this forum thread the grid with an expanded row like this when a group is collapsed the row expanded body is no more this discussion in a sencha thread looks much like it thus a limitation from the grid panel s design causes this there may be a way to auto reload the expanded row s contents when the view is refreshed on the grid above the following is enough to break rowexpander app view refreshview this gets called by grouping plugin when expanding collapsing groups thus the grid breaks
| 1
|
499,349
| 14,445,411,840
|
IssuesEvent
|
2020-12-07 22:56:14
|
GoogleCloudPlatform/cloud-code-samples
|
https://api.github.com/repos/GoogleCloudPlatform/cloud-code-samples
|
closed
|
Validation error in VS Code in hello.deployment.yaml when creating NodeJS K8s Hello World template
|
:rotating_light: priority: p1 type: bug
|
**Repro Steps:**
1. Create a new K8s NodeJS Hello World application in VS Code
2. Expand **kubernetes-manifests** and select **hello.deployment.yaml** (may require waiting a little or switching between files for LSP to catch up I guess?)
**Expected:**
**hello.deployment.yaml** contains no schema validation errors
**Actual:**
**hello.deployment.yaml** appears to contain schema errors since lines 20-29 have yellow squiggly underlines. I don't know what's wrong...?
After hovering over line 28, I was eventually able to see this message "One or more containers do not have resource limits - this could starve other processes" which I assume is why there is the yellow underline?

|
1.0
|
Validation error in VS Code in hello.deployment.yaml when creating NodeJS K8s Hello World template - **Repro Steps:**
1. Create a new K8s NodeJS Hello World application in VS Code
2. Expand **kubernetes-manifests** and select **hello.deployment.yaml** (may require waiting a little or switching between files for LSP to catch up I guess?)
**Expected:**
**hello.deployment.yaml** contains no schema validation errors
**Actual:**
**hello.deployment.yaml** appears to contain schema errors since lines 20-29 have yellow squiggly underlines. I don't know what's wrong...?
After hovering over line 28, I was eventually able to see this message "One or more containers do not have resource limits - this could starve other processes" which I assume is why there is the yellow underline?

|
non_defect
|
validation error in vs code in hello deployment yaml when creating nodejs hello world template repro steps create a new nodejs hello world application in vs code expand kubernetes manifests and select hello deployment yaml may require waiting a little or switching between files for lsp to catch up i guess expected hello deployment yaml contains no schema validation errors actual hello deployment yaml appears to contain schema errors since lines have yellow squiggly underlines i don t know what s wrong after hovering over line i was eventually able to see this message one or more containers do not have resource limits this could starve other processes which i assume is why there is the yellow underline
| 0
|
56,245
| 14,993,835,179
|
IssuesEvent
|
2021-01-29 11:57:55
|
radon-h2020/radon-defect-prediction-api
|
https://api.github.com/repos/radon-h2020/radon-defect-prediction-api
|
closed
|
R-T3.4-5: The defect-prediction tool MUST provide a set of rules that identify defect-prone scripts and an interpretation of the final decision
|
Defect prediction IDE MUST WP3
|
ID | R-T3.4-5
-- | --
Section | WP3: Methodology and Quality Assurance Requirements
Type | FUNCTIONAL_SUITABILITY
User Story | As an Operations Engineer/QoS Engineer/Release Manager I want the tool to show me an interpretation of the final decision for the classification of a script as defective.
Requirement | The defect-prediction tool must provide a set of rules that identify defect-prone scripts and an interpretation of the final decision.
Extended Description | The user must have the opportunity to prioritize the actions to solve defects.
Priority | Must have
Affected Tools | DEFECT_PRED_TOOL
Means of Verification | Direct implementation on IDE, feature checklist, case-study
This requirement replaces requirement #4.
The previous requirement stated that the defect prediction tool could provide a defect threat level to architecture elements and predict threat-level defects under certain infrastructure assumptions. This is a typical regression problem that requires to identify the number of bugs in the infrastructure establish a threat-level. It in turns requires an ontology of IaC bugs, which does not exist yet. In addition, the current defect predictor is a classification model based on Decision Tree or Random Forest. Therefore, the requirement has been changed to address its explainability by providing the user with a set of rules that identify defective-prone IaC scripts and the decision path that led to the final prediction. This requirement has also been raised by one of the industrial partners (PRQ) and therefore it’s priority changed from COULD to MUST HAVE.
|
1.0
|
R-T3.4-5: The defect-prediction tool MUST provide a set of rules that identify defect-prone scripts and an interpretation of the final decision - ID | R-T3.4-5
-- | --
Section | WP3: Methodology and Quality Assurance Requirements
Type | FUNCTIONAL_SUITABILITY
User Story | As an Operations Engineer/QoS Engineer/Release Manager I want the tool to show me an interpretation of the final decision for the classification of a script as defective.
Requirement | The defect-prediction tool must provide a set of rules that identify defect-prone scripts and an interpretation of the final decision.
Extended Description | The user must have the opportunity to prioritize the actions to solve defects.
Priority | Must have
Affected Tools | DEFECT_PRED_TOOL
Means of Verification | Direct implementation on IDE, feature checklist, case-study
This requirement replaces requirement #4.
The previous requirement stated that the defect prediction tool could provide a defect threat level to architecture elements and predict threat-level defects under certain infrastructure assumptions. This is a typical regression problem that requires to identify the number of bugs in the infrastructure establish a threat-level. It in turns requires an ontology of IaC bugs, which does not exist yet. In addition, the current defect predictor is a classification model based on Decision Tree or Random Forest. Therefore, the requirement has been changed to address its explainability by providing the user with a set of rules that identify defective-prone IaC scripts and the decision path that led to the final prediction. This requirement has also been raised by one of the industrial partners (PRQ) and therefore it’s priority changed from COULD to MUST HAVE.
|
defect
|
r the defect prediction tool must provide a set of rules that identify defect prone scripts and an interpretation of the final decision id r section methodology and quality assurance requirements type functional suitability user story as an operations engineer qos engineer release manager i want the tool to show me an interpretation of the final decision for the classification of a script as defective requirement the defect prediction tool must provide a set of rules that identify defect prone scripts and an interpretation of the final decision extended description the user must have the opportunity to prioritize the actions to solve defects priority must have affected tools defect pred tool means of verification direct implementation on ide feature checklist case study this requirement replaces requirement the previous requirement stated that the defect prediction tool could provide a defect threat level to architecture elements and predict threat level defects under certain infrastructure assumptions this is a typical regression problem that requires to identify the number of bugs in the infrastructure establish a threat level it in turns requires an ontology of iac bugs which does not exist yet in addition the current defect predictor is a classification model based on decision tree or random forest therefore the requirement has been changed to address its explainability by providing the user with a set of rules that identify defective prone iac scripts and the decision path that led to the final prediction this requirement has also been raised by one of the industrial partners prq and therefore it’s priority changed from could to must have
| 1
|
189,157
| 6,794,662,945
|
IssuesEvent
|
2017-11-01 13:07:43
|
cilium/cilium
|
https://api.github.com/repos/cilium/cilium
|
closed
|
Kafka: access.log to be limited to a single topic
|
area/proxy kind/microtask priority/high project/1.0-gap
|
For efficient logging purposes, we should have access.log logging records per topic. This includes request as well as responses.
|
1.0
|
Kafka: access.log to be limited to a single topic - For efficient logging purposes, we should have access.log logging records per topic. This includes request as well as responses.
|
non_defect
|
kafka access log to be limited to a single topic for efficient logging purposes we should have access log logging records per topic this includes request as well as responses
| 0
|
227,772
| 17,399,186,051
|
IssuesEvent
|
2021-08-02 17:06:40
|
Nyasita/HackBioAssignmentPauling
|
https://api.github.com/repos/Nyasita/HackBioAssignmentPauling
|
closed
|
Personal details/ Biographies
|
documentation
|
Hello everyone,
Please update your biography under contributors on the readme, complete with:
1. Your name
2. What you do/ affiliation
3. Your social media handles/ github profile
4. An image of yourself
|
1.0
|
Personal details/ Biographies - Hello everyone,
Please update your biography under contributors on the readme, complete with:
1. Your name
2. What you do/ affiliation
3. Your social media handles/ github profile
4. An image of yourself
|
non_defect
|
personal details biographies hello everyone please update your biography under contributors on the readme complete with your name what you do affiliation your social media handles github profile an image of yourself
| 0
|
72,232
| 24,006,936,389
|
IssuesEvent
|
2022-09-14 15:27:36
|
NREL/EnergyPlus
|
https://api.github.com/repos/NREL/EnergyPlus
|
closed
|
IDFVersionUpdater not working for V22-1-0 files to V22-2-0-IOFreeze
|
Defect AuxiliaryTool
|
Issue overview
--------------
IDFVersionUpdater not working on Linux for V22-2-0-IOFreeze.
```shell
mkdir tmp
cd tmp
cp /usr/local/EnergyPlus-22-1-0/ExampleFiles/1ZoneUncontrolled.idf
grep Version 1ZoneUncontrolled.idf
> Version,22.1;
../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/IDFVersionUpdater
```

The "New Version" dropdown also only shows 22.1, it does not detect 22.2
All the required files are there though:
```
$ ls ../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/*22*
../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/OutputChanges22-1-0-to-22-2-0.md
../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/OutputChanges9-6-0-to-22-1-0.md
'../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/Report Variables 22-1-0 to 22-2-0.csv'
'../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/Report Variables 9-6-0 to 22-1-0.csv'
../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/Rules22-1-0-to-22-2-0.md
../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/Rules9-6-0-to-22-1-0.md
../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/Transition-V22-1-0-to-V22-2-0
../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/Transition-V9-6-0-to-V22-1-0
../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/V22-1-0-Energy+.idd
../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/V22-2-0-Energy+.idd
```
### Details
Some additional details for this issue (if relevant):
- Platform (Operating system, version): Ubuntu 20.04
- Version of EnergyPlus (if using an intermediate build, include SHA): V22-2-0-IOFreeze
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [ ] Defect file added (list location of defect file here)
- [ ] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
|
1.0
|
IDFVersionUpdater not working for V22-1-0 files to V22-2-0-IOFreeze - Issue overview
--------------
IDFVersionUpdater not working on Linux for V22-2-0-IOFreeze.
```shell
mkdir tmp
cd tmp
cp /usr/local/EnergyPlus-22-1-0/ExampleFiles/1ZoneUncontrolled.idf
grep Version 1ZoneUncontrolled.idf
> Version,22.1;
../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/IDFVersionUpdater
```

The "New Version" dropdown also only shows 22.1, it does not detect 22.2
All the required files are there though:
```
$ ls ../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/*22*
../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/OutputChanges22-1-0-to-22-2-0.md
../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/OutputChanges9-6-0-to-22-1-0.md
'../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/Report Variables 22-1-0 to 22-2-0.csv'
'../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/Report Variables 9-6-0 to 22-1-0.csv'
../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/Rules22-1-0-to-22-2-0.md
../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/Rules9-6-0-to-22-1-0.md
../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/Transition-V22-1-0-to-V22-2-0
../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/Transition-V9-6-0-to-V22-1-0
../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/V22-1-0-Energy+.idd
../EnergyPlus-22.2.0-56f3a890b4-Linux-Ubuntu20.04-x86_64/PreProcess/IDFVersionUpdater/V22-2-0-Energy+.idd
```
### Details
Some additional details for this issue (if relevant):
- Platform (Operating system, version): Ubuntu 20.04
- Version of EnergyPlus (if using an intermediate build, include SHA): V22-2-0-IOFreeze
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [ ] Defect file added (list location of defect file here)
- [ ] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
|
defect
|
idfversionupdater not working for files to iofreeze issue overview idfversionupdater not working on linux for iofreeze shell mkdir tmp cd tmp cp usr local energyplus examplefiles idf grep version idf version energyplus linux preprocess idfversionupdater idfversionupdater the new version dropdown also only shows it does not detect all the required files are there though ls energyplus linux preprocess idfversionupdater energyplus linux preprocess idfversionupdater to md energyplus linux preprocess idfversionupdater to md energyplus linux preprocess idfversionupdater report variables to csv energyplus linux preprocess idfversionupdater report variables to csv energyplus linux preprocess idfversionupdater to md energyplus linux preprocess idfversionupdater to md energyplus linux preprocess idfversionupdater transition to energyplus linux preprocess idfversionupdater transition to energyplus linux preprocess idfversionupdater energy idd energyplus linux preprocess idfversionupdater energy idd details some additional details for this issue if relevant platform operating system version ubuntu version of energyplus if using an intermediate build include sha iofreeze checklist add to this list or remove from it as applicable this is a simple templated set of guidelines defect file added list location of defect file here ticket added to pivotal for defect development team task pull request created the pull request will have additional tasks related to reviewing changes that fix this defect
| 1
|
157,589
| 13,697,315,764
|
IssuesEvent
|
2020-10-01 02:40:24
|
Xcov19/mycovidconnect
|
https://api.github.com/repos/Xcov19/mycovidconnect
|
closed
|
Update the README file
|
Hacktoberfest documentation enhancement first-timers-only good first issue help wanted up-for-grabs
|
Follow this format:
- Project name: Your project’s name is the first thing people will see upon scrolling down to your README, and is included upon creation of your README file.
- Description: A description of your project follows. A good description is clear, short, and to the point. Describe the importance of your project, and what it does.
- Table of Contents: Optionally, include a table of contents in order to allow other people to quickly navigate especially long or detailed READMEs.
- Installation: Installation is the next section in an effective README. Tell other users how to install your project locally. Optionally, include a gif to make the process even more clear for other people.
- Usage: The next section is usage, in which you instruct other people on how to use your project after they’ve installed it. This would also be a good place to include screenshots of your project in action.
Contributing: Link to a file in project root called `CONTRIBUTING.md`
Credits: Include a section for credits in order to highlight and link to the authors of your project.
Document the process of the web service flow in the readme as well.
|
1.0
|
Update the README file - Follow this format:
- Project name: Your project’s name is the first thing people will see upon scrolling down to your README, and is included upon creation of your README file.
- Description: A description of your project follows. A good description is clear, short, and to the point. Describe the importance of your project, and what it does.
- Table of Contents: Optionally, include a table of contents in order to allow other people to quickly navigate especially long or detailed READMEs.
- Installation: Installation is the next section in an effective README. Tell other users how to install your project locally. Optionally, include a gif to make the process even more clear for other people.
- Usage: The next section is usage, in which you instruct other people on how to use your project after they’ve installed it. This would also be a good place to include screenshots of your project in action.
Contributing: Link to a file in project root called `CONTRIBUTING.md`
Credits: Include a section for credits in order to highlight and link to the authors of your project.
Document the process of the web service flow in the readme as well.
|
non_defect
|
update the readme file follow this format project name your project’s name is the first thing people will see upon scrolling down to your readme and is included upon creation of your readme file description a description of your project follows a good description is clear short and to the point describe the importance of your project and what it does table of contents optionally include a table of contents in order to allow other people to quickly navigate especially long or detailed readmes installation installation is the next section in an effective readme tell other users how to install your project locally optionally include a gif to make the process even more clear for other people usage the next section is usage in which you instruct other people on how to use your project after they’ve installed it this would also be a good place to include screenshots of your project in action contributing link to a file in project root called contributing md credits include a section for credits in order to highlight and link to the authors of your project document the process of the web service flow in the readme as well
| 0
|
10,463
| 2,622,165,027
|
IssuesEvent
|
2015-03-04 00:11:56
|
byzhang/graphchi
|
https://api.github.com/repos/byzhang/graphchi
|
opened
|
Sharder crashes is input not in the right format
|
auto-migrated Priority-High Type-Defect
|
```
If file was adjacency, but try to read with edgelist format -> segfault.
```
Original issue reported on code.google.com by `akyrola...@gmail.com` on 28 Jun 2012 at 5:08
|
1.0
|
Sharder crashes is input not in the right format - ```
If file was adjacency, but try to read with edgelist format -> segfault.
```
Original issue reported on code.google.com by `akyrola...@gmail.com` on 28 Jun 2012 at 5:08
|
defect
|
sharder crashes is input not in the right format if file was adjacency but try to read with edgelist format segfault original issue reported on code google com by akyrola gmail com on jun at
| 1
|
342,090
| 24,728,579,223
|
IssuesEvent
|
2022-10-20 15:42:12
|
hypothesis/frontend-shared
|
https://api.github.com/repos/hypothesis/frontend-shared
|
closed
|
Size documentation for `Button` is incomplete
|
documentation pattern library
|
Documentation for `size` prop on `ButtonPage` is incomplete.
|
1.0
|
Size documentation for `Button` is incomplete - Documentation for `size` prop on `ButtonPage` is incomplete.
|
non_defect
|
size documentation for button is incomplete documentation for size prop on buttonpage is incomplete
| 0
|
182,377
| 14,912,867,277
|
IssuesEvent
|
2021-01-22 13:20:27
|
KonstantinEger/Bau-Abrechnungsprogramm
|
https://api.github.com/repos/KonstantinEger/Bau-Abrechnungsprogramm
|
opened
|
Wiki not up to date with main branch
|
documentation
|
**Describe the bug**
The wiki is not up to date with the current and new features of the app.
**Solution**
Add a branch to the wiki, which is up to date and push the new stuff whenever there is a new release.
|
1.0
|
Wiki not up to date with main branch - **Describe the bug**
The wiki is not up to date with the current and new features of the app.
**Solution**
Add a branch to the wiki, which is up to date and push the new stuff whenever there is a new release.
|
non_defect
|
wiki not up to date with main branch describe the bug the wiki is not up to date with the current and new features of the app solution add a branch to the wiki which is up to date and push the new stuff whenever there is a new release
| 0
|
334,352
| 29,831,928,230
|
IssuesEvent
|
2023-06-18 11:35:49
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
opened
|
Fix comparison_ops.test_torch_not_equal
|
PyTorch Frontend Sub Task Failing Test
|
| | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5303169054/jobs/9598603920"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5303169054/jobs/9598603920"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5303169054/jobs/9598603920"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5303169054/jobs/9598603920"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5303169054/jobs/9598603920"><img src=https://img.shields.io/badge/-success-success></a>
|
1.0
|
Fix comparison_ops.test_torch_not_equal - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5303169054/jobs/9598603920"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5303169054/jobs/9598603920"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5303169054/jobs/9598603920"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5303169054/jobs/9598603920"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5303169054/jobs/9598603920"><img src=https://img.shields.io/badge/-success-success></a>
|
non_defect
|
fix comparison ops test torch not equal tensorflow a href src torch a href src numpy a href src jax a href src paddle a href src
| 0
|
209,302
| 16,014,734,261
|
IssuesEvent
|
2021-04-20 14:46:13
|
istio/istio
|
https://api.github.com/repos/istio/istio
|
closed
|
TestTraffic/virtualservice/shifting test flakes in multicluster
|
area/networking kind/test failure
| ERROR: type should be string, got "https://prow.istio.io/view/gs/istio-prow/pr-logs/pull/istio_istio/32206/integ-pilot-multicluster-tests_istio/1382736986078449664\r\n\r\nWe are getting errors about unbalanced load. I verified the client side load balancing is correct - we send exactly 1/3 of traffic to each gw\r\n\r\nI think this is the current state:\r\n\r\n\r\nBasically because Envoy client will do HTTP connection pool, we may have only a single connection to the Gateways. Because its TCP routing, that connection will go to a single location. There is a 50/50 shot of both gateways choosing the same destination, at which point our test fails?\r\n\r\nI think envoy is not always using exactly 1 connection, but even with 2 connections each its a 12.% chance of failure, and with 3 connections each 1%\r\n\r\ncc @stevenctl \r\n"
|
1.0
|
TestTraffic/virtualservice/shifting test flakes in multicluster - https://prow.istio.io/view/gs/istio-prow/pr-logs/pull/istio_istio/32206/integ-pilot-multicluster-tests_istio/1382736986078449664
We are getting errors about unbalanced load. I verified the client side load balancing is correct - we send exactly 1/3 of traffic to each gw
I think this is the current state:

Basically because Envoy client will do HTTP connection pool, we may have only a single connection to the Gateways. Because its TCP routing, that connection will go to a single location. There is a 50/50 shot of both gateways choosing the same destination, at which point our test fails?
I think envoy is not always using exactly 1 connection, but even with 2 connections each its a 12.% chance of failure, and with 3 connections each 1%
cc @stevenctl
|
non_defect
|
testtraffic virtualservice shifting test flakes in multicluster we are getting errors about unbalanced load i verified the client side load balancing is correct we send exactly of traffic to each gw i think this is the current state basically because envoy client will do http connection pool we may have only a single connection to the gateways because its tcp routing that connection will go to a single location there is a shot of both gateways choosing the same destination at which point our test fails i think envoy is not always using exactly connection but even with connections each its a chance of failure and with connections each cc stevenctl
| 0
|
63,189
| 17,421,038,384
|
IssuesEvent
|
2021-08-04 01:23:43
|
Cockatrice/Cockatrice
|
https://api.github.com/repos/Cockatrice/Cockatrice
|
closed
|
Both CI builds for windows are failing
|
CI Defect - Regression High Priority OS - Windows
|
I spotted this after the most recent merge to master happened (https://github.com/Cockatrice/Cockatrice/pull/4398).
64-bit and 32-bit builds are failing in the `Restore or setup vcpkg` step:
https://github.com/Cockatrice/Cockatrice/runs/3157059370
It looks like the issue was already visible in the PR itself: https://github.com/Cockatrice/Cockatrice/actions/runs/1061233135
|
1.0
|
Both CI builds for windows are failing - I spotted this after the most recent merge to master happened (https://github.com/Cockatrice/Cockatrice/pull/4398).
64-bit and 32-bit builds are failing in the `Restore or setup vcpkg` step:
https://github.com/Cockatrice/Cockatrice/runs/3157059370
It looks like the issue was already visible in the PR itself: https://github.com/Cockatrice/Cockatrice/actions/runs/1061233135
|
defect
|
both ci builds for windows are failing i spotted this after the most recent merge to master happened bit and bit builds are failing in the restore or setup vcpkg step it looks like the issue was already visible in the pr itself
| 1
|
75,159
| 25,562,775,172
|
IssuesEvent
|
2022-11-30 12:06:44
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
closed
|
Informix CURRENT_TIMESTAMP emulation doesn't work as DDL DEFAULT expression
|
T: Defect C: Functionality C: DB: Informix P: Medium E: Enterprise Edition
|
This is valid in Informix:
```sql
SELECT CURRENT
```
But this isn't:
```sql
CREATE TABLE x (y DATETIME YEAR TO FRACTION (5) DEFAULT CURRENT);
```
Despite the datetime precision being clear from context, we have to supply it explicitly, as `CURRENT YEAR TO FRACTION (5)`
|
1.0
|
Informix CURRENT_TIMESTAMP emulation doesn't work as DDL DEFAULT expression - This is valid in Informix:
```sql
SELECT CURRENT
```
But this isn't:
```sql
CREATE TABLE x (y DATETIME YEAR TO FRACTION (5) DEFAULT CURRENT);
```
Despite the datetime precision being clear from context, we have to supply it explicitly, as `CURRENT YEAR TO FRACTION (5)`
|
defect
|
informix current timestamp emulation doesn t work as ddl default expression this is valid in informix sql select current but this isn t sql create table x y datetime year to fraction default current despite the datetime precision being clear from context we have to supply it explicitly as current year to fraction
| 1
|
532,827
| 15,571,790,864
|
IssuesEvent
|
2021-03-17 05:46:27
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
sso.rajasthan.gov.in - see bug description
|
browser-fenix engine-gecko ml-needsdiagnosis-false priority-normal status-needsinfo
|
<!-- @browser: Firefox Mobile 85.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 7.1.2; Mobile; rv:85.0) Gecko/85.0 Firefox/85.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/66450 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://sso.rajasthan.gov.in/dashboard
**Browser / Version**: Firefox Mobile 85.0
**Operating System**: Android 7.1.2
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: popup
**Steps to Reproduce**:
ok 6ur4u4ji55k5i5i5i6kiyhgvgggggGfdddfffffgggggghhhyuu
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/1/8a8c735b-ea71-40f9-b70d-e429fc23423e.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20210115192513</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/1/2d85814c-2a4d-4848-a239-59ed2417df42)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
sso.rajasthan.gov.in - see bug description - <!-- @browser: Firefox Mobile 85.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 7.1.2; Mobile; rv:85.0) Gecko/85.0 Firefox/85.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/66450 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://sso.rajasthan.gov.in/dashboard
**Browser / Version**: Firefox Mobile 85.0
**Operating System**: Android 7.1.2
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: popup
**Steps to Reproduce**:
ok 6ur4u4ji55k5i5i5i6kiyhgvgggggGfdddfffffgggggghhhyuu
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/1/8a8c735b-ea71-40f9-b70d-e429fc23423e.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20210115192513</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/1/2d85814c-2a4d-4848-a239-59ed2417df42)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_defect
|
sso rajasthan gov in see bug description url browser version firefox mobile operating system android tested another browser yes chrome problem type something else description popup steps to reproduce ok view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
22,622
| 3,670,924,093
|
IssuesEvent
|
2016-02-22 02:39:50
|
gperftools/gperftools
|
https://api.github.com/repos/gperftools/gperftools
|
closed
|
Function names are not printed when running pprof for CPU profiler
|
Priority-Medium Status-New Type-Defect
|
Originally reported on Google Code with ID 520
```
What steps will reproduce the problem?
1.Compile the code with option -g, o, tcmalloc libraries and -lprofile
2.Run the executable with CPUPROFILE flag set and get profiled file ls.profile
3.Run pprof --callgrind /bin/ls > ls.callgrind
What is the expected output? What do you see instead?
ls.callgrind file doesn't contain the function names. Contains only the address of
the stack.
What version of the product are you using? On what operating system?
Latest gperftools-2.0
Linux version 2.6.18-164.el5PAE (mockbuild@x86-002.build.bos.redhat.com) (gcc version
4.1.2 20080704 (Red Hat 4.1.2-46))
Please provide any additional information below.
Compiling the code with compile flag -g and o. Want to know any other additinal flags
needed to get the function names in the stack trace. Getting only the address?
```
Reported by `immaneni.arun` on 2013-04-19 00:10:55
|
1.0
|
Function names are not printed when running pprof for CPU profiler - Originally reported on Google Code with ID 520
```
What steps will reproduce the problem?
1.Compile the code with option -g, o, tcmalloc libraries and -lprofile
2.Run the executable with CPUPROFILE flag set and get profiled file ls.profile
3.Run pprof --callgrind /bin/ls > ls.callgrind
What is the expected output? What do you see instead?
ls.callgrind file doesn't contain the function names. Contains only the address of
the stack.
What version of the product are you using? On what operating system?
Latest gperftools-2.0
Linux version 2.6.18-164.el5PAE (mockbuild@x86-002.build.bos.redhat.com) (gcc version
4.1.2 20080704 (Red Hat 4.1.2-46))
Please provide any additional information below.
Compiling the code with compile flag -g and o. Want to know any other additinal flags
needed to get the function names in the stack trace. Getting only the address?
```
Reported by `immaneni.arun` on 2013-04-19 00:10:55
|
defect
|
function names are not printed when running pprof for cpu profiler originally reported on google code with id what steps will reproduce the problem compile the code with option g o tcmalloc libraries and lprofile run the executable with cpuprofile flag set and get profiled file ls profile run pprof callgrind bin ls ls callgrind what is the expected output what do you see instead ls callgrind file doesn t contain the function names contains only the address of the stack what version of the product are you using on what operating system latest gperftools linux version mockbuild build bos redhat com gcc version red hat please provide any additional information below compiling the code with compile flag g and o want to know any other additinal flags needed to get the function names in the stack trace getting only the address reported by immaneni arun on
| 1
|
10,320
| 2,622,143,910
|
IssuesEvent
|
2015-03-04 00:03:17
|
byzhang/i7z
|
https://api.github.com/repos/byzhang/i7z
|
closed
|
double code in i7z.c
|
auto-migrated Priority-Medium Type-Defect
|
```
The lines may be deleted from 396 to 414.
double code with 422 to 440
```
Original issue reported on code.google.com by `rmattusc...@gmx.net` on 1 Apr 2010 at 12:29
|
1.0
|
double code in i7z.c - ```
The lines may be deleted from 396 to 414.
double code with 422 to 440
```
Original issue reported on code.google.com by `rmattusc...@gmx.net` on 1 Apr 2010 at 12:29
|
defect
|
double code in c the lines may be deleted from to double code with to original issue reported on code google com by rmattusc gmx net on apr at
| 1
|
64,879
| 18,951,612,846
|
IssuesEvent
|
2021-11-18 15:41:58
|
vector-im/element-android
|
https://api.github.com/repos/vector-im/element-android
|
opened
|
"CALL BACK" button sometimes is not displayed
|
T-Defect
|
### Steps to reproduce
1. Where are you starting? What can you see?
I called someone and looked at different times into the chat.
First time I made a screenshot:

Second time I made a screenshot:

(same messages at different times)
2. What do you click?
I did nothing but calling and scrolling.
It happened on a more recent call, too, but I cannot reproduce it.
### Outcome
#### What did you expect?
I expect to see the same each time
#### What happened instead?
The "call back" button is missing temporarily
### Your phone model
Fairphone 3+
### Operating system version
Android 10
### Application version and app store
Element version 1.3.8 [40103082] (G-b4480)
### Homeserver
matrix.org
### Will you send logs?
Yes
|
1.0
|
"CALL BACK" button sometimes is not displayed - ### Steps to reproduce
1. Where are you starting? What can you see?
I called someone and looked at different times into the chat.
First time I made a screenshot:

Second time I made a screenshot:

(same messages at different times)
2. What do you click?
I did nothing but calling and scrolling.
It happened on a more recent call, too, but I cannot reproduce it.
### Outcome
#### What did you expect?
I expect to see the same each time
#### What happened instead?
The "call back" button is missing temporarily
### Your phone model
Fairphone 3+
### Operating system version
Android 10
### Application version and app store
Element version 1.3.8 [40103082] (G-b4480)
### Homeserver
matrix.org
### Will you send logs?
Yes
|
defect
|
call back button sometimes is not displayed steps to reproduce where are you starting what can you see i called someone and looked at different times into the chat first time i made a screenshot second time i made a screenshot same messages at different times what do you click i did nothing but calling and scrolling it happened on a more recent call too but i cannot reproduce it outcome what did you expect i expect to see the same each time what happened instead the call back button is missing temporarily your phone model fairphone operating system version android application version and app store element version g homeserver matrix org will you send logs yes
| 1
|
34,502
| 7,452,416,444
|
IssuesEvent
|
2018-03-29 08:18:54
|
kerdokullamae/test_koik_issued
|
https://api.github.com/repos/kerdokullamae/test_koik_issued
|
closed
|
Täpse päringu lähteandmetesse ei saa valida Valdkonda ega Isik/Org.
|
C: AIS P: high R: fixed T: defect
|
**Reported by katrin vesterblom on 5 Jun 2013 11:30 UTC**
rahvusarhiiv.tietotest.ee
"Täpsemat otsingut" tehes ei saa valida lähteandmeteks Valdkonda (ei reageeri mingisugustele täheühenditele) ega Isikut/Organisatsiooni (samuti ei reageeri mingitele täheühenditele).
Nende puuduste kohta on mingid varasemad ticketid ka, aga igal juhul on vead alles praeguses versioonis.
Valdkonna valimist lähteandmeteks (ja samuti kirjeldusüksusega seostamiseks) puudutab veel see asi, et Valdkonna nimetus peaks sinna ilmuma koos oma kõrgema tasandi(te) Valdkonna(de) nimetus(t)ega - nt. täheühendile "tee" : Haridus->Ministeeriumid ja keskasutused; Riigikaitse->Ministeeriumid ja keskasutused (mitte ainult "ministeeriumid ja keskasutused", mida tuleks 10rida).
|
1.0
|
Täpse päringu lähteandmetesse ei saa valida Valdkonda ega Isik/Org. - **Reported by katrin vesterblom on 5 Jun 2013 11:30 UTC**
rahvusarhiiv.tietotest.ee
"Täpsemat otsingut" tehes ei saa valida lähteandmeteks Valdkonda (ei reageeri mingisugustele täheühenditele) ega Isikut/Organisatsiooni (samuti ei reageeri mingitele täheühenditele).
Nende puuduste kohta on mingid varasemad ticketid ka, aga igal juhul on vead alles praeguses versioonis.
Valdkonna valimist lähteandmeteks (ja samuti kirjeldusüksusega seostamiseks) puudutab veel see asi, et Valdkonna nimetus peaks sinna ilmuma koos oma kõrgema tasandi(te) Valdkonna(de) nimetus(t)ega - nt. täheühendile "tee" : Haridus->Ministeeriumid ja keskasutused; Riigikaitse->Ministeeriumid ja keskasutused (mitte ainult "ministeeriumid ja keskasutused", mida tuleks 10rida).
|
defect
|
täpse päringu lähteandmetesse ei saa valida valdkonda ega isik org reported by katrin vesterblom on jun utc rahvusarhiiv tietotest ee täpsemat otsingut tehes ei saa valida lähteandmeteks valdkonda ei reageeri mingisugustele täheühenditele ega isikut organisatsiooni samuti ei reageeri mingitele täheühenditele nende puuduste kohta on mingid varasemad ticketid ka aga igal juhul on vead alles praeguses versioonis valdkonna valimist lähteandmeteks ja samuti kirjeldusüksusega seostamiseks puudutab veel see asi et valdkonna nimetus peaks sinna ilmuma koos oma kõrgema tasandi te valdkonna de nimetus t ega nt täheühendile tee haridus ministeeriumid ja keskasutused riigikaitse ministeeriumid ja keskasutused mitte ainult ministeeriumid ja keskasutused mida tuleks
| 1
|
44,792
| 12,391,241,769
|
IssuesEvent
|
2020-05-20 12:09:38
|
ontopia/ontopia
|
https://api.github.com/repos/ontopia/ontopia
|
opened
|
XTMTopicMapReader always validates
|
Component-Engine Defect Newbie Syntax-XTM bug
|
Using `XTMTopicMapReader.setValidation(false)` has no effect
|
1.0
|
XTMTopicMapReader always validates - Using `XTMTopicMapReader.setValidation(false)` has no effect
|
defect
|
xtmtopicmapreader always validates using xtmtopicmapreader setvalidation false has no effect
| 1
|
20,263
| 3,322,171,473
|
IssuesEvent
|
2015-11-09 13:13:05
|
bridgedotnet/Bridge
|
https://api.github.com/repos/bridgedotnet/Bridge
|
closed
|
Awaiter in iterator block of for loop generates wrong state machine js code
|
defect
|
http://forums.bridge.net/forum/community/help/768?p=771#post771
```
for( var nextPage = await msgPage.nextPage(); nextPage != null; nextPage = await nextPage.nextPage())
{
foreach (Element item in await nextPage.removeOfflineParallel())
msgPage.appendItem(item);
}
```
|
1.0
|
Awaiter in iterator block of for loop generates wrong state machine js code - http://forums.bridge.net/forum/community/help/768?p=771#post771
```
for( var nextPage = await msgPage.nextPage(); nextPage != null; nextPage = await nextPage.nextPage())
{
foreach (Element item in await nextPage.removeOfflineParallel())
msgPage.appendItem(item);
}
```
|
defect
|
awaiter in iterator block of for loop generates wrong state machine js code for var nextpage await msgpage nextpage nextpage null nextpage await nextpage nextpage foreach element item in await nextpage removeofflineparallel msgpage appenditem item
| 1
|
13,382
| 3,330,579,177
|
IssuesEvent
|
2015-11-11 11:20:18
|
mantidproject/mantid
|
https://api.github.com/repos/mantidproject/mantid
|
opened
|
OSIRIS FuryAndFuryFitMulti System test failing
|
Component: Direct Inelastic Misc: Bugfix Priority: High Quality: System Tests
|
After the update to VS2015, ISISIndirectInelastic system tests are failing on the OSIRISFuryAndFuryFitMulti test
http://builds.mantidproject.org/job/master_systemtests-win7/lastCompletedBuild/testReport/SystemTests/ISISIndirectInelastic/OSIRISFuryAndFuryFitMulti/
|
1.0
|
OSIRIS FuryAndFuryFitMulti System test failing - After the update to VS2015, ISISIndirectInelastic system tests are failing on the OSIRISFuryAndFuryFitMulti test
http://builds.mantidproject.org/job/master_systemtests-win7/lastCompletedBuild/testReport/SystemTests/ISISIndirectInelastic/OSIRISFuryAndFuryFitMulti/
|
non_defect
|
osiris furyandfuryfitmulti system test failing after the update to isisindirectinelastic system tests are failing on the osirisfuryandfuryfitmulti test
| 0
|
389,988
| 26,842,092,210
|
IssuesEvent
|
2023-02-03 01:57:05
|
ophub/amlogic-s9xxx-armbian
|
https://api.github.com/repos/ophub/amlogic-s9xxx-armbian
|
closed
|
Get eMMC partition info with nice-looking layout with single command / 一条命令简单获取结构清晰的eMMC分区信息
|
documentation essence
|
On the terminal of the target Amlogic device that has Android on its eMMC, type the following command (copy&paste is enough):
在eMMC上尚有安卓的目标晶晨设备上,输入下面的命令(复制粘贴即可):
```
echo "https://7ji.github.io/ampart-web-reporter/?dsnapshot=$(ampart /dev/mmcblk2 --mode dsnapshot 2>/dev/null | head -n 1)&esnapshot=$(ampart /dev/mmcblk2 --mode esnapshot 2>/dev/null | head -n 1)"
```
And you will get a URL like the following on the terminal (the device does not to be online):
你就能在终端上得到像下面这样的URL(设备本身不需要联网)
```
https://7ji.github.io/ampart-web-reporter/?esnapshot=bootloader:0:4194304:0 reserved:37748736:67108864:0 cache:113246208:754974720:2 env:876609536:8388608:0 logo:893386752:33554432:1 recovery:935329792:33554432:1 rsv:977272832:8388608:1 tee:994050048:8388608:1 crypt:1010827264:33554432:1 misc:1052770304:33554432:1 instaboot:1094713344:536870912:1 boot:1639972864:33554432:1 system:1681915904:1073741824:1 params:2764046336:67108864:2 bootfiles:2839543808:754974720:2 data:3602907136:4131389440:4&dsnapshot=logo::33554432:1 recovery::33554432:1 rsv::8388608:1 tee::8388608:1 crypt::33554432:1 misc::33554432:1 instaboot::536870912:1 boot::33554432:1 system::1073741824:1 cache::536870912:2 params::67108864:2 data::-1:4
```
(It's quite long, but self-explanatory, and well below the 2048-character limit of modern browsers)
(很长,不过看了就明白意思,并且远短于现代浏览器的2048个字符的限制)
Copy the URL to your browser on your Windows/Linux PC, and you will get a well-formatted page with two tables:
复制URL到你Windows/Linux电脑的浏览器打开,你就能在网页上看到格式整齐的两张表


|
1.0
|
Get eMMC partition info with nice-looking layout with single command / 一条命令简单获取结构清晰的eMMC分区信息 - On the terminal of the target Amlogic device that has Android on its eMMC, type the following command (copy&paste is enough):
在eMMC上尚有安卓的目标晶晨设备上,输入下面的命令(复制粘贴即可):
```
echo "https://7ji.github.io/ampart-web-reporter/?dsnapshot=$(ampart /dev/mmcblk2 --mode dsnapshot 2>/dev/null | head -n 1)&esnapshot=$(ampart /dev/mmcblk2 --mode esnapshot 2>/dev/null | head -n 1)"
```
And you will get a URL like the following on the terminal (the device does not to be online):
你就能在终端上得到像下面这样的URL(设备本身不需要联网)
```
https://7ji.github.io/ampart-web-reporter/?esnapshot=bootloader:0:4194304:0 reserved:37748736:67108864:0 cache:113246208:754974720:2 env:876609536:8388608:0 logo:893386752:33554432:1 recovery:935329792:33554432:1 rsv:977272832:8388608:1 tee:994050048:8388608:1 crypt:1010827264:33554432:1 misc:1052770304:33554432:1 instaboot:1094713344:536870912:1 boot:1639972864:33554432:1 system:1681915904:1073741824:1 params:2764046336:67108864:2 bootfiles:2839543808:754974720:2 data:3602907136:4131389440:4&dsnapshot=logo::33554432:1 recovery::33554432:1 rsv::8388608:1 tee::8388608:1 crypt::33554432:1 misc::33554432:1 instaboot::536870912:1 boot::33554432:1 system::1073741824:1 cache::536870912:2 params::67108864:2 data::-1:4
```
(It's quite long, but self-explanatory, and well below the 2048-character limit of modern browsers)
(很长,不过看了就明白意思,并且远短于现代浏览器的2048个字符的限制)
Copy the URL to your browser on your Windows/Linux PC, and you will get a well-formatted page with two tables:
复制URL到你Windows/Linux电脑的浏览器打开,你就能在网页上看到格式整齐的两张表


|
non_defect
|
get emmc partition info with nice looking layout with single command 一条命令简单获取结构清晰的emmc分区信息 on the terminal of the target amlogic device that has android on its emmc type the following command copy paste is enough 在emmc上尚有安卓的目标晶晨设备上,输入下面的命令(复制粘贴即可): echo dev mode dsnapshot dev null head n esnapshot ampart dev mode esnapshot dev null head n and you will get a url like the following on the terminal the device does not to be online 你就能在终端上得到像下面这样的url(设备本身不需要联网) reserved cache env logo recovery rsv tee crypt misc instaboot boot system params bootfiles data dsnapshot logo recovery rsv tee crypt misc instaboot boot system cache params data it s quite long but self explanatory and well below the character limit of modern browsers (很长,不过看了就明白意思, ) copy the url to your browser on your windows linux pc and you will get a well formatted page with two tables 复制url到你windows linux电脑的浏览器打开,你就能在网页上看到格式整齐的两张表
| 0
|
8,697
| 2,611,536,386
|
IssuesEvent
|
2015-02-27 06:06:08
|
chrsmith/hedgewars
|
https://api.github.com/repos/chrsmith/hedgewars
|
opened
|
Artifacts/Flashing on the map island on AMD
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. Select a big sized map on Island
2. play for a while (it may take 2 rounds till it happens)
3. wait for some strange flickering appearing for a split second
What version of the product are you using? On what operating system?
0.9.20
Please provide any additional information below.
Seems like it is relate to AMD cards. I even replaced my old HD 7850 with a
HD7950 but the issue happens on both cards and it happens with any driver.
```
Original issue reported on code.google.com by `knuddelw...@hotmail.de` on 21 Feb 2014 at 12:17
|
1.0
|
Artifacts/Flashing on the map island on AMD - ```
What steps will reproduce the problem?
1. Select a big sized map on Island
2. play for a while (it may take 2 rounds till it happens)
3. wait for some strange flickering appearing for a split second
What version of the product are you using? On what operating system?
0.9.20
Please provide any additional information below.
Seems like it is relate to AMD cards. I even replaced my old HD 7850 with a
HD7950 but the issue happens on both cards and it happens with any driver.
```
Original issue reported on code.google.com by `knuddelw...@hotmail.de` on 21 Feb 2014 at 12:17
|
defect
|
artifacts flashing on the map island on amd what steps will reproduce the problem select a big sized map on island play for a while it may take rounds till it happens wait for some strange flickering appearing for a split second what version of the product are you using on what operating system please provide any additional information below seems like it is relate to amd cards i even replaced my old hd with a but the issue happens on both cards and it happens with any driver original issue reported on code google com by knuddelw hotmail de on feb at
| 1
|
65,295
| 6,954,775,533
|
IssuesEvent
|
2017-12-07 03:25:59
|
equella/Equella
|
https://api.github.com/repos/equella/Equella
|
closed
|
This ResultSet is closed.
|
Ready for Testing
|
Did a fresh install on Windows server using JDK 8 152 and PostgreSQL 10.1. When first firing up the application I get a database error saying the result set is closed.
java.sql.SQLTransientConnectionException: HikariPool-1 - Connection is not available, request timed out after 30100ms.
at com.zaxxer.hikari.pool.HikariPool.createTimeoutException(HikariPool.java:602)
at com.zaxxer.hikari.pool.HikariPool.getConnection(HikariPool.java:195)
at com.zaxxer.hikari.pool.HikariPool.getConnection(HikariPool.java:145)
at com.zaxxer.hikari.HikariDataSource.getConnection(HikariDataSource.java:85)
at com.tle.core.hibernate.impl.DynamicDataSource.getConnection(DynamicDataSource.java:85)
at com.tle.core.hibernate.HibernateFactory$DataSourceProvider.getConnection(HibernateFactory.java:152)
at org.hibernate.cfg.SettingsFactory.buildSettings(SettingsFactory.java:113)
at org.hibernate.cfg.Configuration.buildSettingsInternal(Configuration.java:2863)
at org.hibernate.cfg.Configuration.buildSettings(Configuration.java:2859)
at org.hibernate.cfg.Configuration.buildSessionFactory(Configuration.java:1870)
at com.tle.core.hibernate.HibernateFactory.getSessionFactory(HibernateFactory.java:114)
at com.tle.core.migration.impl.HibernateMigrationService.getSessionFactory(HibernateMigrationService.java:83)
at com.tle.core.migration.impl.HibernateMigrationService.checkSchemaForMigrations(HibernateMigrationService.java:103)
at com.tle.core.migration.impl.MigrationGlobalTask.checkSystemSchema(MigrationGlobalTask.java:374)
at com.tle.core.migration.impl.MigrationGlobalTask.init(MigrationGlobalTask.java:122)
at com.tle.core.services.impl.AlwaysRunningTask.call(AlwaysRunningTask.java:38)
at com.tle.core.services.impl.AlwaysRunningTask.call(AlwaysRunningTask.java:26)
at com.tle.core.services.impl.LocalTaskServiceImpl$LocalTask.run(LocalTaskServiceImpl.java:597)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: com.zaxxer.hikari.pool.PoolBase$ConnectionSetupException: org.postgresql.util.PSQLException: This ResultSet is closed.
at com.zaxxer.hikari.pool.PoolBase.setupConnection(PoolBase.java:418)
at com.zaxxer.hikari.pool.PoolBase.newConnection(PoolBase.java:361)
at com.zaxxer.hikari.pool.PoolBase.newPoolEntry(PoolBase.java:199)
at com.zaxxer.hikari.pool.HikariPool.createPoolEntry(HikariPool.java:444)
at com.zaxxer.hikari.pool.HikariPool.access$200(HikariPool.java:71)
at com.zaxxer.hikari.pool.HikariPool$PoolEntryCreator.call(HikariPool.java:631)
at com.zaxxer.hikari.pool.HikariPool$PoolEntryCreator.call(HikariPool.java:617)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at com.tle.core.hibernate.DataSourceServiceImpl$CreateDataSourceFunction$1$1.run(DataSourceServiceImpl.java:136)
Caused by: org.postgresql.util.PSQLException: This ResultSet is closed.
at org.postgresql.jdbc2.AbstractJdbc2ResultSet.checkClosed(AbstractJdbc2ResultSet.java:2852)
at org.postgresql.jdbc2.AbstractJdbc2ResultSet.setFetchSize(AbstractJdbc2ResultSet.java:1875)
at org.postgresql.jdbc4.Jdbc4Statement.createResultSet(Jdbc4Statement.java:37)
at org.postgresql.jdbc2.AbstractJdbc2Statement$StatementResultHandler.handleResultRows(AbstractJdbc2Statement.java:219)
at org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:1816)
at org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:255)
at org.postgresql.jdbc2.AbstractJdbc2Statement.execute(AbstractJdbc2Statement.java:559)
at org.postgresql.jdbc2.AbstractJdbc2Statement.executeWithFlags(AbstractJdbc2Statement.java:403)
at org.postgresql.jdbc2.AbstractJdbc2Connection.execSQLUpdate(AbstractJdbc2Connection.java:376)
at org.postgresql.jdbc2.AbstractJdbc2Connection.getTransactionIsolation(AbstractJdbc2Connection.java:898)
at com.zaxxer.hikari.pool.PoolBase.checkDriverSupport(PoolBase.java:444)
at com.zaxxer.hikari.pool.PoolBase.setupConnection(PoolBase.java:403)
... 10 more
|
1.0
|
This ResultSet is closed. - Did a fresh install on Windows server using JDK 8 152 and PostgreSQL 10.1. When first firing up the application I get a database error saying the result set is closed.
java.sql.SQLTransientConnectionException: HikariPool-1 - Connection is not available, request timed out after 30100ms.
at com.zaxxer.hikari.pool.HikariPool.createTimeoutException(HikariPool.java:602)
at com.zaxxer.hikari.pool.HikariPool.getConnection(HikariPool.java:195)
at com.zaxxer.hikari.pool.HikariPool.getConnection(HikariPool.java:145)
at com.zaxxer.hikari.HikariDataSource.getConnection(HikariDataSource.java:85)
at com.tle.core.hibernate.impl.DynamicDataSource.getConnection(DynamicDataSource.java:85)
at com.tle.core.hibernate.HibernateFactory$DataSourceProvider.getConnection(HibernateFactory.java:152)
at org.hibernate.cfg.SettingsFactory.buildSettings(SettingsFactory.java:113)
at org.hibernate.cfg.Configuration.buildSettingsInternal(Configuration.java:2863)
at org.hibernate.cfg.Configuration.buildSettings(Configuration.java:2859)
at org.hibernate.cfg.Configuration.buildSessionFactory(Configuration.java:1870)
at com.tle.core.hibernate.HibernateFactory.getSessionFactory(HibernateFactory.java:114)
at com.tle.core.migration.impl.HibernateMigrationService.getSessionFactory(HibernateMigrationService.java:83)
at com.tle.core.migration.impl.HibernateMigrationService.checkSchemaForMigrations(HibernateMigrationService.java:103)
at com.tle.core.migration.impl.MigrationGlobalTask.checkSystemSchema(MigrationGlobalTask.java:374)
at com.tle.core.migration.impl.MigrationGlobalTask.init(MigrationGlobalTask.java:122)
at com.tle.core.services.impl.AlwaysRunningTask.call(AlwaysRunningTask.java:38)
at com.tle.core.services.impl.AlwaysRunningTask.call(AlwaysRunningTask.java:26)
at com.tle.core.services.impl.LocalTaskServiceImpl$LocalTask.run(LocalTaskServiceImpl.java:597)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: com.zaxxer.hikari.pool.PoolBase$ConnectionSetupException: org.postgresql.util.PSQLException: This ResultSet is closed.
at com.zaxxer.hikari.pool.PoolBase.setupConnection(PoolBase.java:418)
at com.zaxxer.hikari.pool.PoolBase.newConnection(PoolBase.java:361)
at com.zaxxer.hikari.pool.PoolBase.newPoolEntry(PoolBase.java:199)
at com.zaxxer.hikari.pool.HikariPool.createPoolEntry(HikariPool.java:444)
at com.zaxxer.hikari.pool.HikariPool.access$200(HikariPool.java:71)
at com.zaxxer.hikari.pool.HikariPool$PoolEntryCreator.call(HikariPool.java:631)
at com.zaxxer.hikari.pool.HikariPool$PoolEntryCreator.call(HikariPool.java:617)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at com.tle.core.hibernate.DataSourceServiceImpl$CreateDataSourceFunction$1$1.run(DataSourceServiceImpl.java:136)
Caused by: org.postgresql.util.PSQLException: This ResultSet is closed.
at org.postgresql.jdbc2.AbstractJdbc2ResultSet.checkClosed(AbstractJdbc2ResultSet.java:2852)
at org.postgresql.jdbc2.AbstractJdbc2ResultSet.setFetchSize(AbstractJdbc2ResultSet.java:1875)
at org.postgresql.jdbc4.Jdbc4Statement.createResultSet(Jdbc4Statement.java:37)
at org.postgresql.jdbc2.AbstractJdbc2Statement$StatementResultHandler.handleResultRows(AbstractJdbc2Statement.java:219)
at org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:1816)
at org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:255)
at org.postgresql.jdbc2.AbstractJdbc2Statement.execute(AbstractJdbc2Statement.java:559)
at org.postgresql.jdbc2.AbstractJdbc2Statement.executeWithFlags(AbstractJdbc2Statement.java:403)
at org.postgresql.jdbc2.AbstractJdbc2Connection.execSQLUpdate(AbstractJdbc2Connection.java:376)
at org.postgresql.jdbc2.AbstractJdbc2Connection.getTransactionIsolation(AbstractJdbc2Connection.java:898)
at com.zaxxer.hikari.pool.PoolBase.checkDriverSupport(PoolBase.java:444)
at com.zaxxer.hikari.pool.PoolBase.setupConnection(PoolBase.java:403)
... 10 more
|
non_defect
|
this resultset is closed did a fresh install on windows server using jdk and postgresql when first firing up the application i get a database error saying the result set is closed java sql sqltransientconnectionexception hikaripool connection is not available request timed out after at com zaxxer hikari pool hikaripool createtimeoutexception hikaripool java at com zaxxer hikari pool hikaripool getconnection hikaripool java at com zaxxer hikari pool hikaripool getconnection hikaripool java at com zaxxer hikari hikaridatasource getconnection hikaridatasource java at com tle core hibernate impl dynamicdatasource getconnection dynamicdatasource java at com tle core hibernate hibernatefactory datasourceprovider getconnection hibernatefactory java at org hibernate cfg settingsfactory buildsettings settingsfactory java at org hibernate cfg configuration buildsettingsinternal configuration java at org hibernate cfg configuration buildsettings configuration java at org hibernate cfg configuration buildsessionfactory configuration java at com tle core hibernate hibernatefactory getsessionfactory hibernatefactory java at com tle core migration impl hibernatemigrationservice getsessionfactory hibernatemigrationservice java at com tle core migration impl hibernatemigrationservice checkschemaformigrations hibernatemigrationservice java at com tle core migration impl migrationglobaltask checksystemschema migrationglobaltask java at com tle core migration impl migrationglobaltask init migrationglobaltask java at com tle core services impl alwaysrunningtask call alwaysrunningtask java at com tle core services impl alwaysrunningtask call alwaysrunningtask java at com tle core services impl localtaskserviceimpl localtask run localtaskserviceimpl java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java caused by com zaxxer hikari pool poolbase connectionsetupexception org postgresql util psqlexception this resultset is closed at com zaxxer hikari pool poolbase setupconnection poolbase java at com zaxxer hikari pool poolbase newconnection poolbase java at com zaxxer hikari pool poolbase newpoolentry poolbase java at com zaxxer hikari pool hikaripool createpoolentry hikaripool java at com zaxxer hikari pool hikaripool access hikaripool java at com zaxxer hikari pool hikaripool poolentrycreator call hikaripool java at com zaxxer hikari pool hikaripool poolentrycreator call hikaripool java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at com tle core hibernate datasourceserviceimpl createdatasourcefunction run datasourceserviceimpl java caused by org postgresql util psqlexception this resultset is closed at org postgresql checkclosed java at org postgresql setfetchsize java at org postgresql createresultset java at org postgresql statementresulthandler handleresultrows java at org postgresql core queryexecutorimpl processresults queryexecutorimpl java at org postgresql core queryexecutorimpl execute queryexecutorimpl java at org postgresql execute java at org postgresql executewithflags java at org postgresql execsqlupdate java at org postgresql gettransactionisolation java at com zaxxer hikari pool poolbase checkdriversupport poolbase java at com zaxxer hikari pool poolbase setupconnection poolbase java more
| 0
|
51,492
| 13,207,501,770
|
IssuesEvent
|
2020-08-14 23:21:02
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
opened
|
remove remaining should-be-private headers from modules in core projects (Trac #525)
|
IceTray Incomplete Migration Migrated from Trac defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/525">https://code.icecube.wisc.edu/projects/icecube/ticket/525</a>, reported by troyand owned by blaufuss</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-02-11T19:42:30",
"_ts": "1423683750462763",
"description": "especially the utility modules in icetray\n\n",
"reporter": "troy",
"cc": "",
"resolution": "wontfix",
"time": "2009-01-23T17:06:51",
"component": "IceTray",
"summary": "remove remaining should-be-private headers from modules in core projects",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "blaufuss",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
remove remaining should-be-private headers from modules in core projects (Trac #525) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/525">https://code.icecube.wisc.edu/projects/icecube/ticket/525</a>, reported by troyand owned by blaufuss</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-02-11T19:42:30",
"_ts": "1423683750462763",
"description": "especially the utility modules in icetray\n\n",
"reporter": "troy",
"cc": "",
"resolution": "wontfix",
"time": "2009-01-23T17:06:51",
"component": "IceTray",
"summary": "remove remaining should-be-private headers from modules in core projects",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "blaufuss",
"type": "defect"
}
```
</p>
</details>
|
defect
|
remove remaining should be private headers from modules in core projects trac migrated from json status closed changetime ts description especially the utility modules in icetray n n reporter troy cc resolution wontfix time component icetray summary remove remaining should be private headers from modules in core projects priority normal keywords milestone owner blaufuss type defect
| 1
|
57,359
| 15,731,471,829
|
IssuesEvent
|
2021-03-29 17:06:52
|
danmar/testissues
|
https://api.github.com/repos/danmar/testissues
|
opened
|
False positive with "char a[]" (Trac #377)
|
False positive Incomplete Migration Migrated from Trac defect hyd_danmar
|
Migrated from https://trac.cppcheck.net/ticket/377
```json
{
"status": "closed",
"changetime": "2009-06-10T18:37:37",
"description": "{{{\nchar *f()\n{\n char a[] = new char[10];\n return a;\n}\n}}}\n\n[a.c:4]: (error) Returning pointer to local array variable",
"reporter": "aggro80",
"cc": "",
"resolution": "fixed",
"_ts": "1244659057000000",
"component": "False positive",
"summary": "False positive with \"char a[]\"",
"priority": "",
"keywords": "",
"time": "2009-06-08T07:32:42",
"milestone": "1.34",
"owner": "hyd_danmar",
"type": "defect"
}
```
|
1.0
|
False positive with "char a[]" (Trac #377) - Migrated from https://trac.cppcheck.net/ticket/377
```json
{
"status": "closed",
"changetime": "2009-06-10T18:37:37",
"description": "{{{\nchar *f()\n{\n char a[] = new char[10];\n return a;\n}\n}}}\n\n[a.c:4]: (error) Returning pointer to local array variable",
"reporter": "aggro80",
"cc": "",
"resolution": "fixed",
"_ts": "1244659057000000",
"component": "False positive",
"summary": "False positive with \"char a[]\"",
"priority": "",
"keywords": "",
"time": "2009-06-08T07:32:42",
"milestone": "1.34",
"owner": "hyd_danmar",
"type": "defect"
}
```
|
defect
|
false positive with char a trac migrated from json status closed changetime description nchar f n n char a new char n return a n n n n error returning pointer to local array variable reporter cc resolution fixed ts component false positive summary false positive with char a priority keywords time milestone owner hyd danmar type defect
| 1
|
21,841
| 3,568,027,725
|
IssuesEvent
|
2016-01-26 02:15:50
|
antang/NewCapstoneProject
|
https://api.github.com/repos/antang/NewCapstoneProject
|
closed
|
Defect - Not show status question in phase
|
Defect Medium priority
|
When user set up game then they don't know about questions have shown or not, it is setting questions for phase screen.
|
1.0
|
Defect - Not show status question in phase - When user set up game then they don't know about questions have shown or not, it is setting questions for phase screen.
|
defect
|
defect not show status question in phase when user set up game then they don t know about questions have shown or not it is setting questions for phase screen
| 1
|
81,109
| 30,716,813,287
|
IssuesEvent
|
2023-07-27 13:32:19
|
SeleniumHQ/selenium
|
https://api.github.com/repos/SeleniumHQ/selenium
|
opened
|
[🐛 Bug]:
|
I-defect needs-triaging
|
### What happened?
So I'm running several dockerized python crawlers from my k8s cluster. All of them are working just fine, some with proxy some without and I'm having issues with this one that just doesn't want to work anymore. It did work just fine a month ago and I can't figure out what's causing the issue. I did try with proxy and without for this specific crawler.
### How can we reproduce the issue?
```shell
I don't think it has to do anything with the code since It worked a month ago and nothing was changed in the code.
```
### Relevant log output
```shell
Opening browser: https://<redacted_url>?IGNORE_EXTERNAL_CALENDAR=false&viewmode=list
WebDriverException occured, reopening browser ... Exception: [Message: unknown error: net::ERR_CONNECTION_TIMED_OUT
(Session info: headless chrome=115.0.5790.110)
Stacktrace:
#0 0x564f256ee4e3 <unknown>
#1 0x564f2541dc76 <unknown>
#2 0x564f25415c7f <unknown>
#3 0x564f25407ca2 <unknown>
#4 0x564f25409412 <unknown>
#5 0x564f254080ca <unknown>
#6 0x564f25407168 <unknown>
#7 0x564f25406fa0 <unknown>
#8 0x564f254059bf <unknown>
#9 0x564f25405fed <unknown>
#10 0x564f2541fb06 <unknown>
#11 0x564f254919e5 <unknown>
#12 0x564f25479012 <unknown>
#13 0x564f2549130e <unknown>
#14 0x564f25478de3 <unknown>
#15 0x564f2544e2dd <unknown>
#16 0x564f2544f34e <unknown>
#17 0x564f256ae3e4 <unknown>
#18 0x564f256b23d7 <unknown>
#19 0x564f256bcb20 <unknown>
#20 0x564f256b3023 <unknown>
#21 0x564f256811aa <unknown>
#22 0x564f256d76b8 <unknown>
#23 0x564f256d7847 <unknown>
#24 0x564f256e7243 <unknown>
#25 0x7f1b975fffa3 start_thread
]
```
### Operating System
Docker / k8s
### Selenium version
python 4.9.0
### What are the browser(s) and version(s) where you see this issue?
Chrome 115
### What are the browser driver(s) and version(s) where you see this issue?
ChromeDriver 115.0.5790.110
### Are you using Selenium Grid?
None
|
1.0
|
[🐛 Bug]: - ### What happened?
So I'm running several dockerized python crawlers from my k8s cluster. All of them are working just fine, some with proxy some without and I'm having issues with this one that just doesn't want to work anymore. It did work just fine a month ago and I can't figure out what's causing the issue. I did try with proxy and without for this specific crawler.
### How can we reproduce the issue?
```shell
I don't think it has to do anything with the code since It worked a month ago and nothing was changed in the code.
```
### Relevant log output
```shell
Opening browser: https://<redacted_url>?IGNORE_EXTERNAL_CALENDAR=false&viewmode=list
WebDriverException occured, reopening browser ... Exception: [Message: unknown error: net::ERR_CONNECTION_TIMED_OUT
(Session info: headless chrome=115.0.5790.110)
Stacktrace:
#0 0x564f256ee4e3 <unknown>
#1 0x564f2541dc76 <unknown>
#2 0x564f25415c7f <unknown>
#3 0x564f25407ca2 <unknown>
#4 0x564f25409412 <unknown>
#5 0x564f254080ca <unknown>
#6 0x564f25407168 <unknown>
#7 0x564f25406fa0 <unknown>
#8 0x564f254059bf <unknown>
#9 0x564f25405fed <unknown>
#10 0x564f2541fb06 <unknown>
#11 0x564f254919e5 <unknown>
#12 0x564f25479012 <unknown>
#13 0x564f2549130e <unknown>
#14 0x564f25478de3 <unknown>
#15 0x564f2544e2dd <unknown>
#16 0x564f2544f34e <unknown>
#17 0x564f256ae3e4 <unknown>
#18 0x564f256b23d7 <unknown>
#19 0x564f256bcb20 <unknown>
#20 0x564f256b3023 <unknown>
#21 0x564f256811aa <unknown>
#22 0x564f256d76b8 <unknown>
#23 0x564f256d7847 <unknown>
#24 0x564f256e7243 <unknown>
#25 0x7f1b975fffa3 start_thread
]
```
### Operating System
Docker / k8s
### Selenium version
python 4.9.0
### What are the browser(s) and version(s) where you see this issue?
Chrome 115
### What are the browser driver(s) and version(s) where you see this issue?
ChromeDriver 115.0.5790.110
### Are you using Selenium Grid?
None
|
defect
|
what happened so i m running several dockerized python crawlers from my cluster all of them are working just fine some with proxy some without and i m having issues with this one that just doesn t want to work anymore it did work just fine a month ago and i can t figure out what s causing the issue i did try with proxy and without for this specific crawler how can we reproduce the issue shell i don t think it has to do anything with the code since it worked a month ago and nothing was changed in the code relevant log output shell opening browser webdriverexception occured reopening browser exception message unknown error net err connection timed out session info headless chrome stacktrace start thread operating system docker selenium version python what are the browser s and version s where you see this issue chrome what are the browser driver s and version s where you see this issue chromedriver are you using selenium grid none
| 1
|
94,570
| 11,886,278,093
|
IssuesEvent
|
2020-03-27 21:30:57
|
microsoft/BotFramework-Composer
|
https://api.github.com/repos/microsoft/BotFramework-Composer
|
closed
|
The Asking Questions Sample - a few questions
|
Needs investigation UX Design
|
Here are a few questions Steve (@WashingtonKayaker) and I have about the [AskingQuestions](https://github.com/microsoft/BotFramework-Composer/tree/stable/Composer/packages/server/assets/projects/AskingQuestionsSample/ComposerDialogs) Sample.
### 1. About _Always prompt_
Does **Max turn count** overwrite **Always prompt**? Testing the **Always prompt** functionality in the sample and no difference found in bot's behavior b/w when it is checked or unchecked.

### 2. About _Confirm Options_
The **Confirm Options** section in the **ConfirmInput** prompt is confusing. Does **ConfirmInput** prompt needs **Confirm Options** as **ChoiceInput** does? Testing adding the following options and found selecting the first option will always return "confirmation: True" and the rest all return "confirmation: False".

Testing in the Emulator:

### 3. About _Append choices_
What does **Append choice** do in **ChoiceInput** prompt? What impacts will this option have when checked and unchecked? what other fields it depends on or is related to?

### 4. About _OAuth login_
The **OAuth login** prompt under the **Ask a quesiton** action. Is there any difference b/w the **OAuth login** prompt under the **Ask a quesiton** action and the **OAuth login** action under the **Access external resources** action menu?


|
1.0
|
The Asking Questions Sample - a few questions - Here are a few questions Steve (@WashingtonKayaker) and I have about the [AskingQuestions](https://github.com/microsoft/BotFramework-Composer/tree/stable/Composer/packages/server/assets/projects/AskingQuestionsSample/ComposerDialogs) Sample.
### 1. About _Always prompt_
Does **Max turn count** overwrite **Always prompt**? Testing the **Always prompt** functionality in the sample and no difference found in bot's behavior b/w when it is checked or unchecked.

### 2. About _Confirm Options_
The **Confirm Options** section in the **ConfirmInput** prompt is confusing. Does **ConfirmInput** prompt needs **Confirm Options** as **ChoiceInput** does? Testing adding the following options and found selecting the first option will always return "confirmation: True" and the rest all return "confirmation: False".

Testing in the Emulator:

### 3. About _Append choices_
What does **Append choice** do in **ChoiceInput** prompt? What impacts will this option have when checked and unchecked? what other fields it depends on or is related to?

### 4. About _OAuth login_
The **OAuth login** prompt under the **Ask a quesiton** action. Is there any difference b/w the **OAuth login** prompt under the **Ask a quesiton** action and the **OAuth login** action under the **Access external resources** action menu?


|
non_defect
|
the asking questions sample a few questions here are a few questions steve washingtonkayaker and i have about the sample about always prompt does max turn count overwrite always prompt testing the always prompt functionality in the sample and no difference found in bot s behavior b w when it is checked or unchecked about confirm options the confirm options section in the confirminput prompt is confusing does confirminput prompt needs confirm options as choiceinput does testing adding the following options and found selecting the first option will always return confirmation true and the rest all return confirmation false testing in the emulator about append choices what does append choice do in choiceinput prompt what impacts will this option have when checked and unchecked what other fields it depends on or is related to about oauth login the oauth login prompt under the ask a quesiton action is there any difference b w the oauth login prompt under the ask a quesiton action and the oauth login action under the access external resources action menu
| 0
|
79,985
| 29,810,203,749
|
IssuesEvent
|
2023-06-16 14:31:31
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
closed
|
Ambiguous match found when using aliases with implicit join and joining the same table twice
|
T: Defect C: Functionality P: Medium E: All Editions
|
Using the reproducer from a previous issue (https://github.com/jOOQ/jOOQ/issues/12455), then this query here:
```java
@Test
public void testMultisetMappingIntoJavaRecords() {
String atsType = "foobar";
AmPushRunMatch push_run_match = Tables.AM_PUSH_RUN_MATCH.as("push_run_match");
AmPushRun push_run = push_run_match.amPushRun().as("push_run");
AmRun run = push_run.amRun().as("run");
Am am = run.am().as("am");
Ats sourceAts = am.amSourceAtsUidFkey().as("source_ats");
Ats targetAts = am.amTargetAtsUidFkey().as("target_ats");
println(ctx.select(targetAts.UID, targetAts.ID)
.from(push_run_match)
.where(
sourceAts.TYPE.eq(atsType),
targetAts.TYPE.eq(atsType),
push_run_match.STATUS.eq(AmPushRunMatchStatus.NEW)
)
.groupBy(targetAts.UID)
.orderBy(targetAts.UID)
.limit(10))
.fetch(record -> new AtsRef(atsType, record.get(targetAts.UID), record.get(targetAts.ID)));
}
```
which:
- Uses aliased implicit joins (see https://github.com/jOOQ/jOOQ/issues/12210)
- Joins the same table twice using distinct aliases (`Ats` as `"source_ats"` and `"target_ats"`)
Then, the query is generated correctly:
```sql
select "target_ats"."uid", "target_ats"."id"
from (
"public"."am_push_run_match" as "push_run_match"
join (
"public"."am_push_run" as "push_run"
join (
"public"."am_run" as "run"
join (
"public"."am" as "am"
join "public"."ats" as "target_ats"
on "am"."target_ats_uid" = "target_ats"."uid"
join "public"."ats" as "source_ats"
on "am"."source_ats_uid" = "source_ats"."uid"
)
on "run"."am_uid" = "am"."uid"
)
on "push_run"."am_run_uid" = "run"."uid"
)
on "push_run_match"."am_push_run_uid" = "push_run"."uid"
)
where (
"source_ats"."type" = 'foobar'
and "target_ats"."type" = 'foobar'
and "push_run_match"."status" = 'NEW'
)
group by "target_ats"."uid"
order by "target_ats"."uid"
fetch next 10 rows only
```
But a warning is logged, which hints at there possibly being a hidden bug somewhere, e.g. if the order of join path references is changed, etc:
```
09:24:12,929 INFO [org.jooq.impl.FieldsImpl ] - Ambiguous match found for uid. Both "am"."uid" and "source_ats"."uid" match.
java.sql.SQLWarning: null
at org.jooq.impl.FieldsImpl.field0(FieldsImpl.java:274) [jooq-3.15.3.jar:?]
at org.jooq.impl.FieldsImpl.field(FieldsImpl.java:213) [jooq-3.15.3.jar:?]
at org.jooq.impl.AbstractRow.field(AbstractRow.java:238) [jooq-3.15.3.jar:?]
at org.jooq.impl.FieldsTrait.field(FieldsTrait.java:67) [jooq-3.15.3.jar:?]
at org.jooq.impl.JoinTable.onKey(JoinTable.java:748) [jooq-3.15.3.jar:?]
at org.jooq.impl.JoinTable.onKey(JoinTable.java:732) [jooq-3.15.3.jar:?]
at org.jooq.impl.JoinTable.onKey(JoinTable.java:148) [jooq-3.15.3.jar:?]
at org.jooq.impl.AbstractContext$JoinNode.joinTree(AbstractContext.java:1049) [jooq-3.15.3.jar:?]
at org.jooq.impl.AbstractContext$JoinNode.joinTree(AbstractContext.java:1049) [jooq-3.15.3.jar:?]
at org.jooq.impl.AbstractContext$JoinNode.joinTree(AbstractContext.java:1049) [jooq-3.15.3.jar:?]
at org.jooq.impl.DefaultRenderContext.scopeEnd0(DefaultRenderContext.java:306) [jooq-3.15.3.jar:?]
at org.jooq.impl.AbstractContext.scopeEnd(AbstractContext.java:739) [jooq-3.15.3.jar:?]
at org.jooq.impl.SelectQueryImpl.accept0(SelectQueryImpl.java:1850) [jooq-3.15.3.jar:?]
at org.jooq.impl.SelectQueryImpl.accept(SelectQueryImpl.java:1435) [jooq-3.15.3.jar:?]
at org.jooq.impl.DefaultRenderContext.visit0(DefaultRenderContext.java:720) [jooq-3.15.3.jar:?]
at org.jooq.impl.AbstractContext.visit(AbstractContext.java:296) [jooq-3.15.3.jar:?]
at org.jooq.impl.AbstractQuery.getSQL0(AbstractQuery.java:469) [jooq-3.15.3.jar:?]
at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:287) [jooq-3.15.3.jar:?]
at org.jooq.impl.AbstractResultQuery.fetchLazy(AbstractResultQuery.java:295) [jooq-3.15.3.jar:?]
at org.jooq.impl.AbstractResultQuery.fetchLazyNonAutoClosing(AbstractResultQuery.java:316) [jooq-3.15.3.jar:?]
at org.jooq.impl.SelectImpl.fetchLazyNonAutoClosing(SelectImpl.java:2866) [jooq-3.15.3.jar:?]
at org.jooq.impl.ResultQueryTrait.collect(ResultQueryTrait.java:357) [jooq-3.15.3.jar:?]
at org.jooq.impl.ResultQueryTrait.fetch(ResultQueryTrait.java:1454) [jooq-3.15.3.jar:?]
at org.jooq.example.test.containers.TestContainersTest.testMultisetMappingIntoJavaRecords(TestContainersTest.java:75) [test-classes/:?]
```
There doesn't seem to be anything wrong with the query usage, just yet another issue where aliased implicit join paths may cause some trouble.
The workaround is to remove all the aliases
|
1.0
|
Ambiguous match found when using aliases with implicit join and joining the same table twice - Using the reproducer from a previous issue (https://github.com/jOOQ/jOOQ/issues/12455), then this query here:
```java
@Test
public void testMultisetMappingIntoJavaRecords() {
String atsType = "foobar";
AmPushRunMatch push_run_match = Tables.AM_PUSH_RUN_MATCH.as("push_run_match");
AmPushRun push_run = push_run_match.amPushRun().as("push_run");
AmRun run = push_run.amRun().as("run");
Am am = run.am().as("am");
Ats sourceAts = am.amSourceAtsUidFkey().as("source_ats");
Ats targetAts = am.amTargetAtsUidFkey().as("target_ats");
println(ctx.select(targetAts.UID, targetAts.ID)
.from(push_run_match)
.where(
sourceAts.TYPE.eq(atsType),
targetAts.TYPE.eq(atsType),
push_run_match.STATUS.eq(AmPushRunMatchStatus.NEW)
)
.groupBy(targetAts.UID)
.orderBy(targetAts.UID)
.limit(10))
.fetch(record -> new AtsRef(atsType, record.get(targetAts.UID), record.get(targetAts.ID)));
}
```
which:
- Uses aliased implicit joins (see https://github.com/jOOQ/jOOQ/issues/12210)
- Joins the same table twice using distinct aliases (`Ats` as `"source_ats"` and `"target_ats"`)
Then, the query is generated correctly:
```sql
select "target_ats"."uid", "target_ats"."id"
from (
"public"."am_push_run_match" as "push_run_match"
join (
"public"."am_push_run" as "push_run"
join (
"public"."am_run" as "run"
join (
"public"."am" as "am"
join "public"."ats" as "target_ats"
on "am"."target_ats_uid" = "target_ats"."uid"
join "public"."ats" as "source_ats"
on "am"."source_ats_uid" = "source_ats"."uid"
)
on "run"."am_uid" = "am"."uid"
)
on "push_run"."am_run_uid" = "run"."uid"
)
on "push_run_match"."am_push_run_uid" = "push_run"."uid"
)
where (
"source_ats"."type" = 'foobar'
and "target_ats"."type" = 'foobar'
and "push_run_match"."status" = 'NEW'
)
group by "target_ats"."uid"
order by "target_ats"."uid"
fetch next 10 rows only
```
But a warning is logged, which hints at there possibly being a hidden bug somewhere, e.g. if the order of join path references is changed, etc:
```
09:24:12,929 INFO [org.jooq.impl.FieldsImpl ] - Ambiguous match found for uid. Both "am"."uid" and "source_ats"."uid" match.
java.sql.SQLWarning: null
at org.jooq.impl.FieldsImpl.field0(FieldsImpl.java:274) [jooq-3.15.3.jar:?]
at org.jooq.impl.FieldsImpl.field(FieldsImpl.java:213) [jooq-3.15.3.jar:?]
at org.jooq.impl.AbstractRow.field(AbstractRow.java:238) [jooq-3.15.3.jar:?]
at org.jooq.impl.FieldsTrait.field(FieldsTrait.java:67) [jooq-3.15.3.jar:?]
at org.jooq.impl.JoinTable.onKey(JoinTable.java:748) [jooq-3.15.3.jar:?]
at org.jooq.impl.JoinTable.onKey(JoinTable.java:732) [jooq-3.15.3.jar:?]
at org.jooq.impl.JoinTable.onKey(JoinTable.java:148) [jooq-3.15.3.jar:?]
at org.jooq.impl.AbstractContext$JoinNode.joinTree(AbstractContext.java:1049) [jooq-3.15.3.jar:?]
at org.jooq.impl.AbstractContext$JoinNode.joinTree(AbstractContext.java:1049) [jooq-3.15.3.jar:?]
at org.jooq.impl.AbstractContext$JoinNode.joinTree(AbstractContext.java:1049) [jooq-3.15.3.jar:?]
at org.jooq.impl.DefaultRenderContext.scopeEnd0(DefaultRenderContext.java:306) [jooq-3.15.3.jar:?]
at org.jooq.impl.AbstractContext.scopeEnd(AbstractContext.java:739) [jooq-3.15.3.jar:?]
at org.jooq.impl.SelectQueryImpl.accept0(SelectQueryImpl.java:1850) [jooq-3.15.3.jar:?]
at org.jooq.impl.SelectQueryImpl.accept(SelectQueryImpl.java:1435) [jooq-3.15.3.jar:?]
at org.jooq.impl.DefaultRenderContext.visit0(DefaultRenderContext.java:720) [jooq-3.15.3.jar:?]
at org.jooq.impl.AbstractContext.visit(AbstractContext.java:296) [jooq-3.15.3.jar:?]
at org.jooq.impl.AbstractQuery.getSQL0(AbstractQuery.java:469) [jooq-3.15.3.jar:?]
at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:287) [jooq-3.15.3.jar:?]
at org.jooq.impl.AbstractResultQuery.fetchLazy(AbstractResultQuery.java:295) [jooq-3.15.3.jar:?]
at org.jooq.impl.AbstractResultQuery.fetchLazyNonAutoClosing(AbstractResultQuery.java:316) [jooq-3.15.3.jar:?]
at org.jooq.impl.SelectImpl.fetchLazyNonAutoClosing(SelectImpl.java:2866) [jooq-3.15.3.jar:?]
at org.jooq.impl.ResultQueryTrait.collect(ResultQueryTrait.java:357) [jooq-3.15.3.jar:?]
at org.jooq.impl.ResultQueryTrait.fetch(ResultQueryTrait.java:1454) [jooq-3.15.3.jar:?]
at org.jooq.example.test.containers.TestContainersTest.testMultisetMappingIntoJavaRecords(TestContainersTest.java:75) [test-classes/:?]
```
There doesn't seem to be anything wrong with the query usage, just yet another issue where aliased implicit join paths may cause some trouble.
The workaround is to remove all the aliases
|
defect
|
ambiguous match found when using aliases with implicit join and joining the same table twice using the reproducer from a previous issue then this query here java test public void testmultisetmappingintojavarecords string atstype foobar ampushrunmatch push run match tables am push run match as push run match ampushrun push run push run match ampushrun as push run amrun run push run amrun as run am am run am as am ats sourceats am amsourceatsuidfkey as source ats ats targetats am amtargetatsuidfkey as target ats println ctx select targetats uid targetats id from push run match where sourceats type eq atstype targetats type eq atstype push run match status eq ampushrunmatchstatus new groupby targetats uid orderby targetats uid limit fetch record new atsref atstype record get targetats uid record get targetats id which uses aliased implicit joins see joins the same table twice using distinct aliases ats as source ats and target ats then the query is generated correctly sql select target ats uid target ats id from public am push run match as push run match join public am push run as push run join public am run as run join public am as am join public ats as target ats on am target ats uid target ats uid join public ats as source ats on am source ats uid source ats uid on run am uid am uid on push run am run uid run uid on push run match am push run uid push run uid where source ats type foobar and target ats type foobar and push run match status new group by target ats uid order by target ats uid fetch next rows only but a warning is logged which hints at there possibly being a hidden bug somewhere e g if the order of join path references is changed etc info ambiguous match found for uid both am uid and source ats uid match java sql sqlwarning null at org jooq impl fieldsimpl fieldsimpl java at org jooq impl fieldsimpl field fieldsimpl java at org jooq impl abstractrow field abstractrow java at org jooq impl fieldstrait field fieldstrait java at org jooq impl jointable onkey jointable java at org jooq impl jointable onkey jointable java at org jooq impl jointable onkey jointable java at org jooq impl abstractcontext joinnode jointree abstractcontext java at org jooq impl abstractcontext joinnode jointree abstractcontext java at org jooq impl abstractcontext joinnode jointree abstractcontext java at org jooq impl defaultrendercontext defaultrendercontext java at org jooq impl abstractcontext scopeend abstractcontext java at org jooq impl selectqueryimpl selectqueryimpl java at org jooq impl selectqueryimpl accept selectqueryimpl java at org jooq impl defaultrendercontext defaultrendercontext java at org jooq impl abstractcontext visit abstractcontext java at org jooq impl abstractquery abstractquery java at org jooq impl abstractquery execute abstractquery java at org jooq impl abstractresultquery fetchlazy abstractresultquery java at org jooq impl abstractresultquery fetchlazynonautoclosing abstractresultquery java at org jooq impl selectimpl fetchlazynonautoclosing selectimpl java at org jooq impl resultquerytrait collect resultquerytrait java at org jooq impl resultquerytrait fetch resultquerytrait java at org jooq example test containers testcontainerstest testmultisetmappingintojavarecords testcontainerstest java there doesn t seem to be anything wrong with the query usage just yet another issue where aliased implicit join paths may cause some trouble the workaround is to remove all the aliases
| 1
|
80,212
| 7,742,288,801
|
IssuesEvent
|
2018-05-29 09:04:38
|
italia/spid
|
https://api.github.com/repos/italia/spid
|
closed
|
Controllo metadati - Piattaforma Dimostrativa Datanet srl
|
Ambiente di Test
|
Salve,
vorremmo eseguire dei test di login attraverso spid su una piattaforma dimostrativa alle pubbliche amministrazioni.
Abbiamo effettuato l'installazione su piattaforma Wordpress utilizzando l'apposito plugin rilasciato da Marco Milesi.
Di seguito si trasmette link dei metadati ottenuti per la piattaforma dimostrativa (https://agid.datanetsrl.eu):
https://agid.datanetsrl.eu/wp-content/plugins/wp-spid-italia/lib/www/module.php/saml/sp/metadata.php/default-sp
In attesa di riscontro, porgo distinti saluti
Dimitri De Porzio
|
1.0
|
Controllo metadati - Piattaforma Dimostrativa Datanet srl - Salve,
vorremmo eseguire dei test di login attraverso spid su una piattaforma dimostrativa alle pubbliche amministrazioni.
Abbiamo effettuato l'installazione su piattaforma Wordpress utilizzando l'apposito plugin rilasciato da Marco Milesi.
Di seguito si trasmette link dei metadati ottenuti per la piattaforma dimostrativa (https://agid.datanetsrl.eu):
https://agid.datanetsrl.eu/wp-content/plugins/wp-spid-italia/lib/www/module.php/saml/sp/metadata.php/default-sp
In attesa di riscontro, porgo distinti saluti
Dimitri De Porzio
|
non_defect
|
controllo metadati piattaforma dimostrativa datanet srl salve vorremmo eseguire dei test di login attraverso spid su una piattaforma dimostrativa alle pubbliche amministrazioni abbiamo effettuato l installazione su piattaforma wordpress utilizzando l apposito plugin rilasciato da marco milesi di seguito si trasmette link dei metadati ottenuti per la piattaforma dimostrativa in attesa di riscontro porgo distinti saluti dimitri de porzio
| 0
|
255,936
| 27,529,639,215
|
IssuesEvent
|
2023-03-06 20:55:00
|
Enterprise-CMCS/eAPD
|
https://api.github.com/repos/Enterprise-CMCS/eAPD
|
closed
|
[Maintenance] Upgrade Knex
|
Development large security
|
### Description and related issues
There is a vulnerability with the version of knex that we are using. Synk suggests upgrading knex to version 2.4.0 or higher.
This is a major upgrade, so confirm that all of the database functionality is still working.
### This task is done when…
- [ ] knex is upgraded to a version 2.4.0 or higher
|
True
|
[Maintenance] Upgrade Knex - ### Description and related issues
There is a vulnerability with the version of knex that we are using. Synk suggests upgrading knex to version 2.4.0 or higher.
This is a major upgrade, so confirm that all of the database functionality is still working.
### This task is done when…
- [ ] knex is upgraded to a version 2.4.0 or higher
|
non_defect
|
upgrade knex description and related issues there is a vulnerability with the version of knex that we are using synk suggests upgrading knex to version or higher this is a major upgrade so confirm that all of the database functionality is still working this task is done when… knex is upgraded to a version or higher
| 0
|
28,108
| 13,532,761,582
|
IssuesEvent
|
2020-09-16 01:04:43
|
doyougnu/VSmt
|
https://api.github.com/repos/doyougnu/VSmt
|
opened
|
Use Continuation Passing Style
|
enhancement memory performance
|
After tinkering around with variational arithmetic I've concluded I need a zipper to handle deeply nested choices. Consider this formula:
```
deepChoicesLHS :: IO Result
deepChoicesLHS = flip sat Nothing $
(1 - 2 - (3 - c)) .== 23
where c = iChc "AA" (iRef ("Aleft" :: Text)) (iRef "Aright") +
iChc "BB" (iRef "Bleft") (iRef "BRight")
```
The trick for the booleans was to rotate the AST hence allows plain values to be accumulated/evaluated thereby lifting choices, but for arithmetic this does not work because `-` is neither commutative or associative, and thus the only way to get to the choice here is to crawl the tree, capturing the context until we find the choice. In Vsat I used a zipper for this before I understood the tree rotations. For VSMT CPS will likely be faster than a zipper and more memory efficient, thus it is desirable. Furthermore, I conjectured and am pretty confident that continuations are the essence of variation and so the fact that CPS sticks out to me is likely an indication of this conjecture.
|
True
|
Use Continuation Passing Style - After tinkering around with variational arithmetic I've concluded I need a zipper to handle deeply nested choices. Consider this formula:
```
deepChoicesLHS :: IO Result
deepChoicesLHS = flip sat Nothing $
(1 - 2 - (3 - c)) .== 23
where c = iChc "AA" (iRef ("Aleft" :: Text)) (iRef "Aright") +
iChc "BB" (iRef "Bleft") (iRef "BRight")
```
The trick for the booleans was to rotate the AST hence allows plain values to be accumulated/evaluated thereby lifting choices, but for arithmetic this does not work because `-` is neither commutative or associative, and thus the only way to get to the choice here is to crawl the tree, capturing the context until we find the choice. In Vsat I used a zipper for this before I understood the tree rotations. For VSMT CPS will likely be faster than a zipper and more memory efficient, thus it is desirable. Furthermore, I conjectured and am pretty confident that continuations are the essence of variation and so the fact that CPS sticks out to me is likely an indication of this conjecture.
|
non_defect
|
use continuation passing style after tinkering around with variational arithmetic i ve concluded i need a zipper to handle deeply nested choices consider this formula deepchoiceslhs io result deepchoiceslhs flip sat nothing c where c ichc aa iref aleft text iref aright ichc bb iref bleft iref bright the trick for the booleans was to rotate the ast hence allows plain values to be accumulated evaluated thereby lifting choices but for arithmetic this does not work because is neither commutative or associative and thus the only way to get to the choice here is to crawl the tree capturing the context until we find the choice in vsat i used a zipper for this before i understood the tree rotations for vsmt cps will likely be faster than a zipper and more memory efficient thus it is desirable furthermore i conjectured and am pretty confident that continuations are the essence of variation and so the fact that cps sticks out to me is likely an indication of this conjecture
| 0
|
11,518
| 2,653,053,394
|
IssuesEvent
|
2015-03-16 20:51:16
|
portah/biowardrobe
|
https://api.github.com/repos/portah/biowardrobe
|
closed
|
For some genes RPKMs don't correlate with log change
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. DEseq
2. Filter on significant genes with either removing or not non coding
What is the expected output? What do you see instead?
NM_001198868,NM_001198869,NM_005186,NR_040008 CAPN1 chr11 64949304 64979477 + 16
8.8881362 168.8581147 -0.585938255 0.004820842 0.026719021
What version of the product are you using? On what operating system?
PC
Please provide any additional information below.
note that RPKMs are essentially the same (168.8) but log chnage is -0.5 with
high significance
```
Original issue reported on code.google.com by `mroch...@gmail.com` on 22 Jan 2014 at 7:47
Attachments:
* [RNA-seq biopsies Joe my DEseq analysis no non coding.xlsx](https://storage.googleapis.com/google-code-attachments/genome-tools/issue-7/comment-0/RNA-seq biopsies Joe my DEseq analysis no non coding.xlsx)
|
1.0
|
For some genes RPKMs don't correlate with log change - ```
What steps will reproduce the problem?
1. DEseq
2. Filter on significant genes with either removing or not non coding
What is the expected output? What do you see instead?
NM_001198868,NM_001198869,NM_005186,NR_040008 CAPN1 chr11 64949304 64979477 + 16
8.8881362 168.8581147 -0.585938255 0.004820842 0.026719021
What version of the product are you using? On what operating system?
PC
Please provide any additional information below.
note that RPKMs are essentially the same (168.8) but log chnage is -0.5 with
high significance
```
Original issue reported on code.google.com by `mroch...@gmail.com` on 22 Jan 2014 at 7:47
Attachments:
* [RNA-seq biopsies Joe my DEseq analysis no non coding.xlsx](https://storage.googleapis.com/google-code-attachments/genome-tools/issue-7/comment-0/RNA-seq biopsies Joe my DEseq analysis no non coding.xlsx)
|
defect
|
for some genes rpkms don t correlate with log change what steps will reproduce the problem deseq filter on significant genes with either removing or not non coding what is the expected output what do you see instead nm nm nm nr what version of the product are you using on what operating system pc please provide any additional information below note that rpkms are essentially the same but log chnage is with high significance original issue reported on code google com by mroch gmail com on jan at attachments biopsies joe my deseq analysis no non coding xlsx
| 1
|
135,248
| 10,967,244,590
|
IssuesEvent
|
2019-11-28 09:09:38
|
BEXIS2/Core
|
https://api.github.com/repos/BEXIS2/Core
|
closed
|
during excel upload and import not all values are displayed during preview
|
Priority: Medium Status: Testing Required Type: Bug
|
during excel upload and import not all values are displayed during preview - because of:
in general, excel cells are given numberformat and styles.
but there are cases where these values are not set in an excel file.
In these cases the values are not displayed.
|
1.0
|
during excel upload and import not all values are displayed during preview - during excel upload and import not all values are displayed during preview - because of:
in general, excel cells are given numberformat and styles.
but there are cases where these values are not set in an excel file.
In these cases the values are not displayed.
|
non_defect
|
during excel upload and import not all values are displayed during preview during excel upload and import not all values are displayed during preview because of in general excel cells are given numberformat and styles but there are cases where these values are not set in an excel file in these cases the values are not displayed
| 0
|
679,852
| 23,247,587,032
|
IssuesEvent
|
2022-08-03 22:01:50
|
brave/brave-browser
|
https://api.github.com/repos/brave/brave-browser
|
closed
|
[Desktop] window size affects error message display in Rewards
|
bug feature/rewards priority/P3 QA/Yes OS/Desktop
|
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
Certain window sizes cause a problem displaying the "error" notification when trying to use Uphold rewards in a region without support (ex. NY). From what I've observed, in this case the error flashes on screen but doesn't persist. This then puts the user into a bad state because they haven't been given the error, so they have no information. If they click on "Verify Wallet" nothing happens on screen - basically it appears as though the integration is broken.
Note - in the terminal logs you do see this when clicking on `Verify Wallet`
```
[ REQUEST ]
> URL: https://api-sandbox.uphold.com/v0/me
> Method: UrlMethod::GET
[15114:775:0929/074839.490957:INFO:uphold_wallet.cc(89)] BAT not allowed
```
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Clean profile, staging, enable Rewards. Note your window size.
2. Claim UGP grant.
3. Click on Verify Wallet.
4. Login with credentials for an NY account.
5. You will be redirected back to brave://rewards page
## Actual result:
when window size is 1013px tall and 1200px wide (my default window size) the error message flashes and goes away quickly. Note how clicking on `Verify Wallet` doesn't do anything from brave://rewards and clicking on this from the panel opens a blank new tab. User is in a bad state at this point. (gif picks up at step 4 from above)

it's tough to see in the above gif, but here's a slightly different window size where you can see the flash of the error a bit better:

## Expected result:
If I size the window down (to 699px tall by 928 px wide), no issues with error message displaying:

## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
easily depending on window size
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.14.84 Chromium: 85.0.4183.121 (Official Build) (64-bit)
-- | --
Revision | a81aa729a8e1fd413943a339393c82e7b8055ddc-refs/branch-heads/4183@{#1864}
OS | macOS Version 10.14.6 (Build 18G3020)
Brave | 1.15.65 Chromium: 86.0.4240.55 (Official Build) dev (x86_64)
-- | --
Revision | a6d625ef6f7fe8ea0675f1cf759155a05ee1be40-refs/branch-heads/4240@{#953}
OS | macOS Version 10.14.6 (Build 18G3020)
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? yes, 1.14.x
- Can you reproduce this issue with the beta channel? yes, 1.15.x
- Can you reproduce this issue with the nightly channel? yes, 1.16.x
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields?
- Does the issue resolve itself when disabling Brave Rewards?
- Is the issue reproducible on the latest version of Chrome?
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
cc @NejcZdovc @brave/legacy_qa @rebron
|
1.0
|
[Desktop] window size affects error message display in Rewards - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
Certain window sizes cause a problem displaying the "error" notification when trying to use Uphold rewards in a region without support (ex. NY). From what I've observed, in this case the error flashes on screen but doesn't persist. This then puts the user into a bad state because they haven't been given the error, so they have no information. If they click on "Verify Wallet" nothing happens on screen - basically it appears as though the integration is broken.
Note - in the terminal logs you do see this when clicking on `Verify Wallet`
```
[ REQUEST ]
> URL: https://api-sandbox.uphold.com/v0/me
> Method: UrlMethod::GET
[15114:775:0929/074839.490957:INFO:uphold_wallet.cc(89)] BAT not allowed
```
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Clean profile, staging, enable Rewards. Note your window size.
2. Claim UGP grant.
3. Click on Verify Wallet.
4. Login with credentials for an NY account.
5. You will be redirected back to brave://rewards page
## Actual result:
when window size is 1013px tall and 1200px wide (my default window size) the error message flashes and goes away quickly. Note how clicking on `Verify Wallet` doesn't do anything from brave://rewards and clicking on this from the panel opens a blank new tab. User is in a bad state at this point. (gif picks up at step 4 from above)

it's tough to see in the above gif, but here's a slightly different window size where you can see the flash of the error a bit better:

## Expected result:
If I size the window down (to 699px tall by 928 px wide), no issues with error message displaying:

## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
easily depending on window size
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.14.84 Chromium: 85.0.4183.121 (Official Build) (64-bit)
-- | --
Revision | a81aa729a8e1fd413943a339393c82e7b8055ddc-refs/branch-heads/4183@{#1864}
OS | macOS Version 10.14.6 (Build 18G3020)
Brave | 1.15.65 Chromium: 86.0.4240.55 (Official Build) dev (x86_64)
-- | --
Revision | a6d625ef6f7fe8ea0675f1cf759155a05ee1be40-refs/branch-heads/4240@{#953}
OS | macOS Version 10.14.6 (Build 18G3020)
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? yes, 1.14.x
- Can you reproduce this issue with the beta channel? yes, 1.15.x
- Can you reproduce this issue with the nightly channel? yes, 1.16.x
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields?
- Does the issue resolve itself when disabling Brave Rewards?
- Is the issue reproducible on the latest version of Chrome?
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
cc @NejcZdovc @brave/legacy_qa @rebron
|
non_defect
|
window size affects error message display in rewards have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description certain window sizes cause a problem displaying the error notification when trying to use uphold rewards in a region without support ex ny from what i ve observed in this case the error flashes on screen but doesn t persist this then puts the user into a bad state because they haven t been given the error so they have no information if they click on verify wallet nothing happens on screen basically it appears as though the integration is broken note in the terminal logs you do see this when clicking on verify wallet url method urlmethod get bat not allowed steps to reproduce clean profile staging enable rewards note your window size claim ugp grant click on verify wallet login with credentials for an ny account you will be redirected back to brave rewards page actual result when window size is tall and wide my default window size the error message flashes and goes away quickly note how clicking on verify wallet doesn t do anything from brave rewards and clicking on this from the panel opens a blank new tab user is in a bad state at this point gif picks up at step from above it s tough to see in the above gif but here s a slightly different window size where you can see the flash of the error a bit better expected result if i size the window down to tall by px wide no issues with error message displaying reproduces how often easily depending on window size brave version brave version info brave chromium official build bit revision refs branch heads os macos version build brave chromium official build dev revision refs branch heads os macos version build version channel information can you reproduce this issue with the current release yes x can you reproduce this issue with the beta channel yes x can you reproduce this issue with the nightly channel yes x other additional information does the issue resolve itself when disabling brave shields does the issue resolve itself when disabling brave rewards is the issue reproducible on the latest version of chrome miscellaneous information cc nejczdovc brave legacy qa rebron
| 0
|
8,160
| 2,611,462,387
|
IssuesEvent
|
2015-02-27 05:08:18
|
chrsmith/reaver-wps
|
https://api.github.com/repos/chrsmith/reaver-wps
|
opened
|
The Code needs modifying
|
auto-migrated Priority-Low Type-Defect
|
```
In the makefile the rule for wpscrack is incorrect as it hardcodes the
dependencies but has *.o in the compilation command line.
The practical implication is that it doesn't hardcode pins.o as a dependency so
even if it's getting modified, the problem is that when pins.c is modified it
should lead to rebuilding pins.o and reaver but that's not happening.
wpscrack is a target in the makefile that should be called reaver to build the
reaver executable.
What version of the product are you using? On what operating system? r35
.
```
Original issue reported on code.google.com by `g...@watchingyou.info` on 2 Jan 2012 at 10:33
|
1.0
|
The Code needs modifying - ```
In the makefile the rule for wpscrack is incorrect as it hardcodes the
dependencies but has *.o in the compilation command line.
The practical implication is that it doesn't hardcode pins.o as a dependency so
even if it's getting modified, the problem is that when pins.c is modified it
should lead to rebuilding pins.o and reaver but that's not happening.
wpscrack is a target in the makefile that should be called reaver to build the
reaver executable.
What version of the product are you using? On what operating system? r35
.
```
Original issue reported on code.google.com by `g...@watchingyou.info` on 2 Jan 2012 at 10:33
|
defect
|
the code needs modifying in the makefile the rule for wpscrack is incorrect as it hardcodes the dependencies but has o in the compilation command line the practical implication is that it doesn t hardcode pins o as a dependency so even if it s getting modified the problem is that when pins c is modified it should lead to rebuilding pins o and reaver but that s not happening wpscrack is a target in the makefile that should be called reaver to build the reaver executable what version of the product are you using on what operating system original issue reported on code google com by g watchingyou info on jan at
| 1
|
127,510
| 10,473,132,667
|
IssuesEvent
|
2019-09-23 11:55:58
|
7mind/izumi
|
https://api.github.com/repos/7mind/izumi
|
closed
|
Support ZIO Env-based tests in distage-testkit
|
api distage (di) distage-testkit good first issue scala-reflect zio
|
Explored in this `distage-sample` PR: https://github.com/7mind/distage-sample/pull/1
Allow constructing fixture dynamically from usages of ZIO Env forwarders.
Classic `distage-testkit` test:
```scala
"my test" in dio {
(service1: Service1[IO], service2: Service2[IO]) =>
for {
_ <- service1.call1
_ <- service2.call2
} yield ()
}
```
Env based test:
```scala
"my test" in dio {
for {
_ <- service1.call1
_ <- service2.call2
} yield () // ZIO[HasService1 with HasService2, Throwable, Unit]
}
```
where:
```scala
trait HasService1 { def service1: Service1[IO] }
trait HasService2 { def service2: Service2[IO] }
val service1: Service1[ZIO[HasService1, +?, +?]]
val service2: Service2[ZIO[HasService2, +?, +?]]
TraitConstructor[HasService1 with HasService2] = macro ProviderMagnet {
(_service1: Service1[IO], _service2: Service2[IO]) =>
new HasService1 with HasService2 {
val service1 = _ service1
val service2 = _service2
}
}
```
Requires `TraitConstructor` to be faster - https://github.com/7mind/izumi/issues/585
|
1.0
|
Support ZIO Env-based tests in distage-testkit - Explored in this `distage-sample` PR: https://github.com/7mind/distage-sample/pull/1
Allow constructing fixture dynamically from usages of ZIO Env forwarders.
Classic `distage-testkit` test:
```scala
"my test" in dio {
(service1: Service1[IO], service2: Service2[IO]) =>
for {
_ <- service1.call1
_ <- service2.call2
} yield ()
}
```
Env based test:
```scala
"my test" in dio {
for {
_ <- service1.call1
_ <- service2.call2
} yield () // ZIO[HasService1 with HasService2, Throwable, Unit]
}
```
where:
```scala
trait HasService1 { def service1: Service1[IO] }
trait HasService2 { def service2: Service2[IO] }
val service1: Service1[ZIO[HasService1, +?, +?]]
val service2: Service2[ZIO[HasService2, +?, +?]]
TraitConstructor[HasService1 with HasService2] = macro ProviderMagnet {
(_service1: Service1[IO], _service2: Service2[IO]) =>
new HasService1 with HasService2 {
val service1 = _ service1
val service2 = _service2
}
}
```
Requires `TraitConstructor` to be faster - https://github.com/7mind/izumi/issues/585
|
non_defect
|
support zio env based tests in distage testkit explored in this distage sample pr allow constructing fixture dynamically from usages of zio env forwarders classic distage testkit test scala my test in dio for yield env based test scala my test in dio for yield zio where scala trait def trait def val val traitconstructor macro providermagnet new with val val requires traitconstructor to be faster
| 0
|
45,692
| 13,021,341,022
|
IssuesEvent
|
2020-07-27 06:04:36
|
hazelcast/hazelcast
|
https://api.github.com/repos/hazelcast/hazelcast
|
opened
|
HASH index on BigDecimal cannot return entries when compareTo=0, but equals=false
|
Module: Query Module: SQL Source: Internal Team: Core Type: Defect
|
Consider the following query:
```
SELECT * FROM table WHERE big_decimal_column = CAST(2.0 as REAL)
```
If there is a `HASH` index on the column `big_decimal_column`, and there is an entry with the index key "new BigDecimal(2)", the corresponding entry will not be returned. But if a `SORTED` index is used, then the entry will be returned.
Thus, the engine will produce different results depending on the index type, which is a bug. The old query engine is affected as well.
The root cause is BigDecimal semantics: if two entries have `compareTo=0`, it doesn't necessarily mean that their `equals=true`. For example, `BigDecimal(2) != BigDecimal(2.0)`.
|
1.0
|
HASH index on BigDecimal cannot return entries when compareTo=0, but equals=false - Consider the following query:
```
SELECT * FROM table WHERE big_decimal_column = CAST(2.0 as REAL)
```
If there is a `HASH` index on the column `big_decimal_column`, and there is an entry with the index key "new BigDecimal(2)", the corresponding entry will not be returned. But if a `SORTED` index is used, then the entry will be returned.
Thus, the engine will produce different results depending on the index type, which is a bug. The old query engine is affected as well.
The root cause is BigDecimal semantics: if two entries have `compareTo=0`, it doesn't necessarily mean that their `equals=true`. For example, `BigDecimal(2) != BigDecimal(2.0)`.
|
defect
|
hash index on bigdecimal cannot return entries when compareto but equals false consider the following query select from table where big decimal column cast as real if there is a hash index on the column big decimal column and there is an entry with the index key new bigdecimal the corresponding entry will not be returned but if a sorted index is used then the entry will be returned thus the engine will produce different results depending on the index type which is a bug the old query engine is affected as well the root cause is bigdecimal semantics if two entries have compareto it doesn t necessarily mean that their equals true for example bigdecimal bigdecimal
| 1
|
691,221
| 23,688,481,945
|
IssuesEvent
|
2022-08-29 08:39:31
|
datatlas-erasme/front
|
https://api.github.com/repos/datatlas-erasme/front
|
closed
|
Merge code based filter color mapped on kepler poi color
|
styling front priority high template system
|
**Describe the solution you'd like**
merge the "industrie" color mapping feature to the dev branch
|
1.0
|
Merge code based filter color mapped on kepler poi color - **Describe the solution you'd like**
merge the "industrie" color mapping feature to the dev branch
|
non_defect
|
merge code based filter color mapped on kepler poi color describe the solution you d like merge the industrie color mapping feature to the dev branch
| 0
|
130,885
| 10,675,326,238
|
IssuesEvent
|
2019-10-21 11:25:04
|
club-soda/club-soda-guide
|
https://api.github.com/repos/club-soda/club-soda-guide
|
closed
|
Update production website?
|
please-test
|
The last time the production website has been updated was the 12th of July.
Since then, some PRs have been merge and deployed into staging:

@jessyclubsoda would you like the production website to be updated now or do you prefer to wait the end of the week once a few new PRs have been merged and tested on staging?
|
1.0
|
Update production website? - The last time the production website has been updated was the 12th of July.
Since then, some PRs have been merge and deployed into staging:

@jessyclubsoda would you like the production website to be updated now or do you prefer to wait the end of the week once a few new PRs have been merged and tested on staging?
|
non_defect
|
update production website the last time the production website has been updated was the of july since then some prs have been merge and deployed into staging jessyclubsoda would you like the production website to be updated now or do you prefer to wait the end of the week once a few new prs have been merged and tested on staging
| 0
|
70,306
| 23,111,831,703
|
IssuesEvent
|
2022-07-27 13:38:32
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
closed
|
BUG: initial approximations to eigenvectors in svds should not be with all positive components
|
defect scipy.sparse.linalg
|
### Describe your issue.
https://github.com/scipy/scipy/blob/master/scipy/sparse/linalg/_eigen/_svds.py at least in one place sets initial approximations to eigenvectors with all components random uniform on [0, 1]. This is a poor choice as common sense suggests that vectors with only positive components cover the positive quadrant only, not the whole vector space, are generally not so good to approximate eigenvectors. E.g., all example in the original MATLAB/Octave code, written for the paper, https://github.com/lobpcg/blopex/blob/master/blopex_tools/matlab/lobpcg/lobpcg.m use randn, e.g., line 97.
Poor initial approximations may make the solver running longer or even fail.
The simplest solution is replacing all calls of uniform random with normal random.
### Reproducing Code Example
```python
N/A
```
### Error message
```shell
N/A
```
### SciPy/NumPy/Python version information
N/A
|
1.0
|
BUG: initial approximations to eigenvectors in svds should not be with all positive components - ### Describe your issue.
https://github.com/scipy/scipy/blob/master/scipy/sparse/linalg/_eigen/_svds.py at least in one place sets initial approximations to eigenvectors with all components random uniform on [0, 1]. This is a poor choice as common sense suggests that vectors with only positive components cover the positive quadrant only, not the whole vector space, are generally not so good to approximate eigenvectors. E.g., all example in the original MATLAB/Octave code, written for the paper, https://github.com/lobpcg/blopex/blob/master/blopex_tools/matlab/lobpcg/lobpcg.m use randn, e.g., line 97.
Poor initial approximations may make the solver running longer or even fail.
The simplest solution is replacing all calls of uniform random with normal random.
### Reproducing Code Example
```python
N/A
```
### Error message
```shell
N/A
```
### SciPy/NumPy/Python version information
N/A
|
defect
|
bug initial approximations to eigenvectors in svds should not be with all positive components describe your issue at least in one place sets initial approximations to eigenvectors with all components random uniform on this is a poor choice as common sense suggests that vectors with only positive components cover the positive quadrant only not the whole vector space are generally not so good to approximate eigenvectors e g all example in the original matlab octave code written for the paper use randn e g line poor initial approximations may make the solver running longer or even fail the simplest solution is replacing all calls of uniform random with normal random reproducing code example python n a error message shell n a scipy numpy python version information n a
| 1
|
37,055
| 5,099,476,414
|
IssuesEvent
|
2017-01-04 08:24:52
|
pandas-dev/pandas
|
https://api.github.com/repos/pandas-dev/pandas
|
closed
|
read_csv: Infers different column types in different runs
|
Bug Dtypes IO CSV Testing
|
``` python
#!/usr/bin/env python3
from io import StringIO
import pandas as pd
test_timeseries = """\
2008-02-07 09:40,1032.43
2008-02-07 09:50,1042.54
2008-02-07 10:00,1051.65
"""
df = pd.read_csv(StringIO(test_timeseries), parse_dates=[0],
usecols=['date', 'value'], index_col=0, header=None,
names=('date', 'value'))
print (df.value.dtype)
```
I run this program 10 times and the result is sometimes `float64` and sometimes `object`.
This happens with pandas 0.18.1 on Debian Jessie amd64 with Python 3.4.2 and numpy 1.11.1. I don't see it happening with Debian's packaged pandas 0.14.1.
I can work around this by specifying the `dtype` argument; but shouldn't pandas behave deterministically when it's omitted?
|
1.0
|
read_csv: Infers different column types in different runs - ``` python
#!/usr/bin/env python3
from io import StringIO
import pandas as pd
test_timeseries = """\
2008-02-07 09:40,1032.43
2008-02-07 09:50,1042.54
2008-02-07 10:00,1051.65
"""
df = pd.read_csv(StringIO(test_timeseries), parse_dates=[0],
usecols=['date', 'value'], index_col=0, header=None,
names=('date', 'value'))
print (df.value.dtype)
```
I run this program 10 times and the result is sometimes `float64` and sometimes `object`.
This happens with pandas 0.18.1 on Debian Jessie amd64 with Python 3.4.2 and numpy 1.11.1. I don't see it happening with Debian's packaged pandas 0.14.1.
I can work around this by specifying the `dtype` argument; but shouldn't pandas behave deterministically when it's omitted?
|
non_defect
|
read csv infers different column types in different runs python usr bin env from io import stringio import pandas as pd test timeseries df pd read csv stringio test timeseries parse dates usecols index col header none names date value print df value dtype i run this program times and the result is sometimes and sometimes object this happens with pandas on debian jessie with python and numpy i don t see it happening with debian s packaged pandas i can work around this by specifying the dtype argument but shouldn t pandas behave deterministically when it s omitted
| 0
|
262,146
| 19,762,132,964
|
IssuesEvent
|
2022-01-16 15:33:58
|
The-4th-Hokage/yondaime-hokage
|
https://api.github.com/repos/The-4th-Hokage/yondaime-hokage
|
closed
|
A clear description of what the Bot does is needed
|
documentation hacktoberfest-accepted
|
**Is your feature request related to a problem? Please describe.**
I saw you put forward this project asking for contributors during Hacktoberfest. Unfortunately, I cannot find any information that tells me what this Bot does.
**Describe the solution you'd like**
A better description is needed in your readme to tell potential contributors and possibly users what the Bot does.
|
1.0
|
A clear description of what the Bot does is needed - **Is your feature request related to a problem? Please describe.**
I saw you put forward this project asking for contributors during Hacktoberfest. Unfortunately, I cannot find any information that tells me what this Bot does.
**Describe the solution you'd like**
A better description is needed in your readme to tell potential contributors and possibly users what the Bot does.
|
non_defect
|
a clear description of what the bot does is needed is your feature request related to a problem please describe i saw you put forward this project asking for contributors during hacktoberfest unfortunately i cannot find any information that tells me what this bot does describe the solution you d like a better description is needed in your readme to tell potential contributors and possibly users what the bot does
| 0
|
11,721
| 2,664,570,164
|
IssuesEvent
|
2015-03-20 15:18:51
|
holahmeds/remotedroid
|
https://api.github.com/repos/holahmeds/remotedroid
|
closed
|
Current download is one year old
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. Download 'current' release
2. Run into bugs
3. Check forum and find all your issues are fixed... but not downloadable.
What is the expected output? What do you see instead?
Expected RemoteDroidServer1.4-r44.jar. Found one year old jar release full of
bugs.
What version of the product are you using? On what operating system?
Using most current version. Running Ubuntu 10.10.
Please provide any additional information below.
Currently having the same bug as described in another issue involving multiple
monitors and the mouse bugging out. Can't figure out how to compile a java
application from svn code.
```
Original issue reported on code.google.com by `Travis...@gmail.com` on 17 Feb 2011 at 5:32
|
1.0
|
Current download is one year old - ```
What steps will reproduce the problem?
1. Download 'current' release
2. Run into bugs
3. Check forum and find all your issues are fixed... but not downloadable.
What is the expected output? What do you see instead?
Expected RemoteDroidServer1.4-r44.jar. Found one year old jar release full of
bugs.
What version of the product are you using? On what operating system?
Using most current version. Running Ubuntu 10.10.
Please provide any additional information below.
Currently having the same bug as described in another issue involving multiple
monitors and the mouse bugging out. Can't figure out how to compile a java
application from svn code.
```
Original issue reported on code.google.com by `Travis...@gmail.com` on 17 Feb 2011 at 5:32
|
defect
|
current download is one year old what steps will reproduce the problem download current release run into bugs check forum and find all your issues are fixed but not downloadable what is the expected output what do you see instead expected jar found one year old jar release full of bugs what version of the product are you using on what operating system using most current version running ubuntu please provide any additional information below currently having the same bug as described in another issue involving multiple monitors and the mouse bugging out can t figure out how to compile a java application from svn code original issue reported on code google com by travis gmail com on feb at
| 1
|
84,150
| 24,241,510,495
|
IssuesEvent
|
2022-09-27 07:08:12
|
tensorflow/tensorflow
|
https://api.github.com/repos/tensorflow/tensorflow
|
closed
|
Module not Found
|
stat:awaiting response type:build/install stalled TF 2.10
|
<details><summary>Click to expand!</summary>
### Issue Type
Build/Install
### Source
source
### Tensorflow Version
tf 2.10
### Custom Code
Yes
### OS Platform and Distribution
_No response_
### Mobile device
_No response_
### Python version
_No response_
### Bazel version
_No response_
### GCC/Compiler version
_No response_
### CUDA/cuDNN version
_No response_
### GPU model and memory
_No response_
### Current Behaviour?
```shell
After installing TF 2.10, my python can't find the module. even reverting back to tf 2.9 is not helping. I'm using anaconda and I've tried installing on different environment and still not helping.
```
### Standalone code to reproduce the issue
```shell
-
```
### Relevant log output
_No response_</details>
|
1.0
|
Module not Found - <details><summary>Click to expand!</summary>
### Issue Type
Build/Install
### Source
source
### Tensorflow Version
tf 2.10
### Custom Code
Yes
### OS Platform and Distribution
_No response_
### Mobile device
_No response_
### Python version
_No response_
### Bazel version
_No response_
### GCC/Compiler version
_No response_
### CUDA/cuDNN version
_No response_
### GPU model and memory
_No response_
### Current Behaviour?
```shell
After installing TF 2.10, my python can't find the module. even reverting back to tf 2.9 is not helping. I'm using anaconda and I've tried installing on different environment and still not helping.
```
### Standalone code to reproduce the issue
```shell
-
```
### Relevant log output
_No response_</details>
|
non_defect
|
module not found click to expand issue type build install source source tensorflow version tf custom code yes os platform and distribution no response mobile device no response python version no response bazel version no response gcc compiler version no response cuda cudnn version no response gpu model and memory no response current behaviour shell after installing tf my python can t find the module even reverting back to tf is not helping i m using anaconda and i ve tried installing on different environment and still not helping standalone code to reproduce the issue shell relevant log output no response
| 0
|
283
| 2,533,975,755
|
IssuesEvent
|
2015-01-24 12:56:35
|
chocolatey/choco
|
https://api.github.com/repos/chocolatey/choco
|
closed
|
Set up Travis CI yaml
|
1 - Ready BuildAutomation Community
|
Official docs http://docs.travis-ci.com/user/languages/csharp/
Some references that are likely outdated (but might prove helpful):
* http://stackoverflow.com/a/16751773/18475
* http://pseudomuto.com/development/walkthroughs/2013/08/13/continuous-integration-for-net-with-travis-ci-and-xunit/
* https://danlimerick.wordpress.com/2013/02/03/build-your-open-source-net-project-on-travis-ci/
|
1.0
|
Set up Travis CI yaml - Official docs http://docs.travis-ci.com/user/languages/csharp/
Some references that are likely outdated (but might prove helpful):
* http://stackoverflow.com/a/16751773/18475
* http://pseudomuto.com/development/walkthroughs/2013/08/13/continuous-integration-for-net-with-travis-ci-and-xunit/
* https://danlimerick.wordpress.com/2013/02/03/build-your-open-source-net-project-on-travis-ci/
|
non_defect
|
set up travis ci yaml official docs some references that are likely outdated but might prove helpful
| 0
|
10,616
| 2,622,176,933
|
IssuesEvent
|
2015-03-04 00:17:04
|
byzhang/leveldb
|
https://api.github.com/repos/byzhang/leveldb
|
closed
|
leveldb 打不开. MANIFEST文件丢失
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
我在zeroc-ice的两个接口函数里,
频繁的打开/读写/关闭leveldb数据库.
现在leveldb不能打开了, MANIFEST文件丢失了。
有办法恢复吗?
What is the expected output? What do you see instead?
What version of the product are you using? On what operating system?
leveldb 1.1 on centos6.2
Please provide any additional information below.
```
Original issue reported on code.google.com by `liuhengl...@gmail.com` on 7 Mar 2012 at 5:49
|
1.0
|
leveldb 打不开. MANIFEST文件丢失 - ```
What steps will reproduce the problem?
我在zeroc-ice的两个接口函数里,
频繁的打开/读写/关闭leveldb数据库.
现在leveldb不能打开了, MANIFEST文件丢失了。
有办法恢复吗?
What is the expected output? What do you see instead?
What version of the product are you using? On what operating system?
leveldb 1.1 on centos6.2
Please provide any additional information below.
```
Original issue reported on code.google.com by `liuhengl...@gmail.com` on 7 Mar 2012 at 5:49
|
defect
|
leveldb 打不开 manifest文件丢失 what steps will reproduce the problem 我在zeroc ice的两个接口函数里, 频繁的打开 读写 关闭leveldb数据库 现在leveldb不能打开了 manifest文件丢失了。 有办法恢复吗? what is the expected output what do you see instead what version of the product are you using on what operating system leveldb on please provide any additional information below original issue reported on code google com by liuhengl gmail com on mar at
| 1
|
34,601
| 9,419,129,290
|
IssuesEvent
|
2019-04-10 21:02:52
|
alter-rebbe/site
|
https://api.github.com/repos/alter-rebbe/site
|
opened
|
Facsimile thumbnails
|
build
|
- [ ] make a JavaScript switch to disappear the thumbnails (see TEI-BP);
- [ ] store the state in a window or - better - in the browser;
- [ ] add a link to facsimile wrapper to the navigation bar for documents with facsimiles;
- [ ] float thumbnails out of the table in 026;
- [ ] add local CSS to disappear the borders of the table in 026;
- [ ] use JavaScript to name the initial tab `collectionViewer`;
|
1.0
|
Facsimile thumbnails - - [ ] make a JavaScript switch to disappear the thumbnails (see TEI-BP);
- [ ] store the state in a window or - better - in the browser;
- [ ] add a link to facsimile wrapper to the navigation bar for documents with facsimiles;
- [ ] float thumbnails out of the table in 026;
- [ ] add local CSS to disappear the borders of the table in 026;
- [ ] use JavaScript to name the initial tab `collectionViewer`;
|
non_defect
|
facsimile thumbnails make a javascript switch to disappear the thumbnails see tei bp store the state in a window or better in the browser add a link to facsimile wrapper to the navigation bar for documents with facsimiles float thumbnails out of the table in add local css to disappear the borders of the table in use javascript to name the initial tab collectionviewer
| 0
|
81,905
| 31,807,053,526
|
IssuesEvent
|
2023-09-13 14:31:21
|
Gogo1951/GDKPd-Classic
|
https://api.github.com/repos/Gogo1951/GDKPd-Classic
|
closed
|
Export isn't working on latest on CurseForge.
|
Priority - 3 Average Sticky - BLOCKED Type - Defect
|
Hey @xpwn3rx,
Reports that the export feature is no longer working.
https://discord.com/channels/990821777875800094/1021972047951302668/1072656696687743046
|
1.0
|
Export isn't working on latest on CurseForge. - Hey @xpwn3rx,
Reports that the export feature is no longer working.
https://discord.com/channels/990821777875800094/1021972047951302668/1072656696687743046
|
defect
|
export isn t working on latest on curseforge hey reports that the export feature is no longer working
| 1
|
28,756
| 5,348,389,304
|
IssuesEvent
|
2017-02-18 04:23:27
|
amitdholiya/vqmod
|
https://api.github.com/repos/amitdholiya/vqmod
|
reopened
|
Fatal Error
|
auto-migrated Priority-Medium Type-Defect
|
```
NOTE THAT THIS IS FOR VQMOD ENGINE ERRORS ONLY. FOR GENERAL ERRORS FROM
MODIFICATIONS CONTACT YOUR DEVELOPER
What steps will reproduce the problem?
1. When I Click on a product
2.
3.
What is the expected output? What do you see instead?
Fatal error: Cannot redeclare ControllerProductProduct::add() in
/home/arsismed/public_html/vqmod/vqcache/vq2-catalog_controller_product_product.
php on line 891
vQmod Version: 1.5.6
Server Operating System:
Please provide any additional information below.
```
Original issue reported on code.google.com by `arsisme...@gmail.com` on 1 Nov 2014 at 4:57
|
1.0
|
Fatal Error - ```
NOTE THAT THIS IS FOR VQMOD ENGINE ERRORS ONLY. FOR GENERAL ERRORS FROM
MODIFICATIONS CONTACT YOUR DEVELOPER
What steps will reproduce the problem?
1. When I Click on a product
2.
3.
What is the expected output? What do you see instead?
Fatal error: Cannot redeclare ControllerProductProduct::add() in
/home/arsismed/public_html/vqmod/vqcache/vq2-catalog_controller_product_product.
php on line 891
vQmod Version: 1.5.6
Server Operating System:
Please provide any additional information below.
```
Original issue reported on code.google.com by `arsisme...@gmail.com` on 1 Nov 2014 at 4:57
|
defect
|
fatal error note that this is for vqmod engine errors only for general errors from modifications contact your developer what steps will reproduce the problem when i click on a product what is the expected output what do you see instead fatal error cannot redeclare controllerproductproduct add in home arsismed public html vqmod vqcache catalog controller product product php on line vqmod version server operating system please provide any additional information below original issue reported on code google com by arsisme gmail com on nov at
| 1
|
59,940
| 17,023,295,276
|
IssuesEvent
|
2021-07-03 01:17:02
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
Cyclemap layer doesn't work anymore
|
Component: potlatch (flash editor) Priority: trivial Resolution: fixed Type: defect
|
**[Submitted to the original trac issue database at 11.29pm, Tuesday, 16th September 2008]**
I suppose it is because of the move to the cloudmade servers.
|
1.0
|
Cyclemap layer doesn't work anymore - **[Submitted to the original trac issue database at 11.29pm, Tuesday, 16th September 2008]**
I suppose it is because of the move to the cloudmade servers.
|
defect
|
cyclemap layer doesn t work anymore i suppose it is because of the move to the cloudmade servers
| 1
|
62,021
| 17,023,833,588
|
IssuesEvent
|
2021-07-03 04:05:11
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
Supermarket only shown at zoom level 16
|
Component: mapnik Priority: minor Resolution: invalid Type: defect
|
**[Submitted to the original trac issue database at 5.24pm, Thursday, 18th October 2012]**
Looking at a map of Hanford, Stoke-on-Trent, England using http://www.openstreetmap.org/?lat=52.9763&lon=-2.19852&zoom=16 I can see a Co-Op supermarket near the junction of Mayne Street and Bankhouse Road at zoom level 16. If I zoom in further, other nearby shops, a post box and post office are shown, but the supermarket disappears. I would expect the supermarket to remain visible as I zoom in further.
I am using Internet Explorer 9.
|
1.0
|
Supermarket only shown at zoom level 16 - **[Submitted to the original trac issue database at 5.24pm, Thursday, 18th October 2012]**
Looking at a map of Hanford, Stoke-on-Trent, England using http://www.openstreetmap.org/?lat=52.9763&lon=-2.19852&zoom=16 I can see a Co-Op supermarket near the junction of Mayne Street and Bankhouse Road at zoom level 16. If I zoom in further, other nearby shops, a post box and post office are shown, but the supermarket disappears. I would expect the supermarket to remain visible as I zoom in further.
I am using Internet Explorer 9.
|
defect
|
supermarket only shown at zoom level looking at a map of hanford stoke on trent england using i can see a co op supermarket near the junction of mayne street and bankhouse road at zoom level if i zoom in further other nearby shops a post box and post office are shown but the supermarket disappears i would expect the supermarket to remain visible as i zoom in further i am using internet explorer
| 1
|
52,829
| 13,225,109,766
|
IssuesEvent
|
2020-08-17 20:30:36
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
closed
|
clsim doesn't jive with OpenCL Framework (Trac #416)
|
Migrated from Trac combo simulation defect
|
http://builds.icecube.wisc.edu/builders/quick_simulation_osx_server_10.6.7/builds/506/steps/compile/logs/stdio
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/416">https://code.icecube.wisc.edu/projects/icecube/ticket/416</a>, reported by negaand owned by claudio.kopper</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-10-31T18:00:13",
"_ts": "1351706413000000",
"description": "http://builds.icecube.wisc.edu/builders/quick_simulation_osx_server_10.6.7/builds/506/steps/compile/logs/stdio\n\n",
"reporter": "nega",
"cc": "jvs",
"resolution": "fixed",
"time": "2012-06-12T22:08:40",
"component": "combo simulation",
"summary": "clsim doesn't jive with OpenCL Framework",
"priority": "normal",
"keywords": "clsim",
"milestone": "",
"owner": "claudio.kopper",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
clsim doesn't jive with OpenCL Framework (Trac #416) - http://builds.icecube.wisc.edu/builders/quick_simulation_osx_server_10.6.7/builds/506/steps/compile/logs/stdio
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/416">https://code.icecube.wisc.edu/projects/icecube/ticket/416</a>, reported by negaand owned by claudio.kopper</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-10-31T18:00:13",
"_ts": "1351706413000000",
"description": "http://builds.icecube.wisc.edu/builders/quick_simulation_osx_server_10.6.7/builds/506/steps/compile/logs/stdio\n\n",
"reporter": "nega",
"cc": "jvs",
"resolution": "fixed",
"time": "2012-06-12T22:08:40",
"component": "combo simulation",
"summary": "clsim doesn't jive with OpenCL Framework",
"priority": "normal",
"keywords": "clsim",
"milestone": "",
"owner": "claudio.kopper",
"type": "defect"
}
```
</p>
</details>
|
defect
|
clsim doesn t jive with opencl framework trac migrated from json status closed changetime ts description reporter nega cc jvs resolution fixed time component combo simulation summary clsim doesn t jive with opencl framework priority normal keywords clsim milestone owner claudio kopper type defect
| 1
|
28,131
| 6,953,470,217
|
IssuesEvent
|
2017-12-06 21:09:10
|
codeforboston/cliff-effects
|
https://api.github.com/repos/codeforboston/cliff-effects
|
closed
|
Turn existing suggestions on PR #280 into issues.
|
beginner friendly non-code
|
(Do this even if/when the PR has been closed) Turn these each into an issue for the 'About' page. I think they all may be beginner friendly. I'm going to fix the other problem with the page and then merge and then these improvements can be made. Some of these may be group-able.
- [ ] https://github.com/codeforboston/cliff-effects/pull/280#discussion_r155122402
- [ ] https://github.com/codeforboston/cliff-effects/pull/280#discussion_r155123692
- [ ] https://github.com/codeforboston/cliff-effects/pull/280#discussion_r155131091
- [ ] https://github.com/codeforboston/cliff-effects/pull/280#discussion_r155131324
- [ ] These three together (consult with @anhase for names for links): https://github.com/codeforboston/cliff-effects/pull/280#discussion_r155131423, https://github.com/codeforboston/cliff-effects/pull/280#discussion_r155131449, https://github.com/codeforboston/cliff-effects/pull/280#discussion_r155131469
- [ ] https://github.com/codeforboston/cliff-effects/pull/280#discussion_r155131656
|
1.0
|
Turn existing suggestions on PR #280 into issues. - (Do this even if/when the PR has been closed) Turn these each into an issue for the 'About' page. I think they all may be beginner friendly. I'm going to fix the other problem with the page and then merge and then these improvements can be made. Some of these may be group-able.
- [ ] https://github.com/codeforboston/cliff-effects/pull/280#discussion_r155122402
- [ ] https://github.com/codeforboston/cliff-effects/pull/280#discussion_r155123692
- [ ] https://github.com/codeforboston/cliff-effects/pull/280#discussion_r155131091
- [ ] https://github.com/codeforboston/cliff-effects/pull/280#discussion_r155131324
- [ ] These three together (consult with @anhase for names for links): https://github.com/codeforboston/cliff-effects/pull/280#discussion_r155131423, https://github.com/codeforboston/cliff-effects/pull/280#discussion_r155131449, https://github.com/codeforboston/cliff-effects/pull/280#discussion_r155131469
- [ ] https://github.com/codeforboston/cliff-effects/pull/280#discussion_r155131656
|
non_defect
|
turn existing suggestions on pr into issues do this even if when the pr has been closed turn these each into an issue for the about page i think they all may be beginner friendly i m going to fix the other problem with the page and then merge and then these improvements can be made some of these may be group able these three together consult with anhase for names for links
| 0
|
56,002
| 14,892,374,830
|
IssuesEvent
|
2021-01-21 02:42:07
|
SAP/fundamental-ngx
|
https://api.github.com/repos/SAP/fundamental-ngx
|
opened
|
Tabs overflow issues
|
Defect Hunting bug core
|
#### Is this a bug, enhancement, or feature request?
bug
#### Briefly describe your proposal.
tabs with overflow has some issues
#### Which versions of Angular and Fundamental Library for Angular are affected? (If this is a feature request, use current version.)
latest `main` branch
#### If this is a bug, please provide steps for reproducing it.
Icon only tabs don't work properly with overflow - maybe we don't have to support that?
<img width="282" alt="Screen Shot 2021-01-20 at 9 38 23 PM" src="https://user-images.githubusercontent.com/4380815/105272919-4c507600-5b68-11eb-8a05-f51d462fa402.png">
and when you resize the screen the overflow menu doesn't open

#### Please provide relevant source code if applicable.
#### Is there anything else we should know?
|
1.0
|
Tabs overflow issues - #### Is this a bug, enhancement, or feature request?
bug
#### Briefly describe your proposal.
tabs with overflow has some issues
#### Which versions of Angular and Fundamental Library for Angular are affected? (If this is a feature request, use current version.)
latest `main` branch
#### If this is a bug, please provide steps for reproducing it.
Icon only tabs don't work properly with overflow - maybe we don't have to support that?
<img width="282" alt="Screen Shot 2021-01-20 at 9 38 23 PM" src="https://user-images.githubusercontent.com/4380815/105272919-4c507600-5b68-11eb-8a05-f51d462fa402.png">
and when you resize the screen the overflow menu doesn't open

#### Please provide relevant source code if applicable.
#### Is there anything else we should know?
|
defect
|
tabs overflow issues is this a bug enhancement or feature request bug briefly describe your proposal tabs with overflow has some issues which versions of angular and fundamental library for angular are affected if this is a feature request use current version latest main branch if this is a bug please provide steps for reproducing it icon only tabs don t work properly with overflow maybe we don t have to support that img width alt screen shot at pm src and when you resize the screen the overflow menu doesn t open please provide relevant source code if applicable is there anything else we should know
| 1
|
93,914
| 11,835,091,648
|
IssuesEvent
|
2020-03-23 10:03:34
|
MozillaReality/FirefoxReality
|
https://api.github.com/repos/MozillaReality/FirefoxReality
|
closed
|
Settings for WebXR
|
Draft Design UX WebVR/XR
|
We'll need some settings to support WebXR.
I'm using Firefox desktop art here to get the point across, but @thenadj will deliver actually FxR specs for these. Should use the same paradigm as [popups blocking](https://github.com/MozillaReality/FirefoxReality/issues/593)
- [ ] new privacy permission for XR, defaulted to ON.
with link out to an informational page based on https://bugzilla.mozilla.org/show_bug.cgi?id=1579267

- [ ] exceptions list. users can add a sight here to prevent them from using XR

This is related to #1966 - when we teach users how to get out of immersive mode, we should also include some messaging that tells them they can turn off immersive mode from Settings.
|
1.0
|
Settings for WebXR - We'll need some settings to support WebXR.
I'm using Firefox desktop art here to get the point across, but @thenadj will deliver actually FxR specs for these. Should use the same paradigm as [popups blocking](https://github.com/MozillaReality/FirefoxReality/issues/593)
- [ ] new privacy permission for XR, defaulted to ON.
with link out to an informational page based on https://bugzilla.mozilla.org/show_bug.cgi?id=1579267

- [ ] exceptions list. users can add a sight here to prevent them from using XR

This is related to #1966 - when we teach users how to get out of immersive mode, we should also include some messaging that tells them they can turn off immersive mode from Settings.
|
non_defect
|
settings for webxr we ll need some settings to support webxr i m using firefox desktop art here to get the point across but thenadj will deliver actually fxr specs for these should use the same paradigm as new privacy permission for xr defaulted to on with link out to an informational page based on exceptions list users can add a sight here to prevent them from using xr this is related to when we teach users how to get out of immersive mode we should also include some messaging that tells them they can turn off immersive mode from settings
| 0
|
158,090
| 12,402,000,060
|
IssuesEvent
|
2020-05-21 11:01:54
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachtest: decommission/nodes=4/duration=1h0m0s failed
|
C-test-failure O-roachtest O-robot branch-master release-blocker
|
[(roachtest).decommission/nodes=4/duration=1h0m0s failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1947362&tab=buildLog) on [master@5156843cc23adecb6a70fabf19f51f46de1241ec](https://github.com/cockroachdb/cockroach/commits/5156843cc23adecb6a70fabf19f51f46de1241ec):
```
The test failed on branch=master, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/decommission/nodes=4/duration=1h0m0s/run_1
cluster.go:2138,cluster.go:2222,decommission.go:218,errgroup.go:57: pgurl for nodes [] empty: :4 from
stdout:
stderr:
```
<details><summary>More</summary><p>
Artifacts: [/decommission/nodes=4/duration=1h0m0s](https://teamcity.cockroachdb.com/viewLog.html?buildId=1947362&tab=artifacts#/decommission/nodes=4/duration=1h0m0s)
Related:
- #47738 roachtest: decommission/nodes=4/duration=1h0m0s failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Adecommission%2Fnodes%3D4%2Fduration%3D1h0m0s.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
2.0
|
roachtest: decommission/nodes=4/duration=1h0m0s failed - [(roachtest).decommission/nodes=4/duration=1h0m0s failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1947362&tab=buildLog) on [master@5156843cc23adecb6a70fabf19f51f46de1241ec](https://github.com/cockroachdb/cockroach/commits/5156843cc23adecb6a70fabf19f51f46de1241ec):
```
The test failed on branch=master, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/decommission/nodes=4/duration=1h0m0s/run_1
cluster.go:2138,cluster.go:2222,decommission.go:218,errgroup.go:57: pgurl for nodes [] empty: :4 from
stdout:
stderr:
```
<details><summary>More</summary><p>
Artifacts: [/decommission/nodes=4/duration=1h0m0s](https://teamcity.cockroachdb.com/viewLog.html?buildId=1947362&tab=artifacts#/decommission/nodes=4/duration=1h0m0s)
Related:
- #47738 roachtest: decommission/nodes=4/duration=1h0m0s failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Adecommission%2Fnodes%3D4%2Fduration%3D1h0m0s.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
non_defect
|
roachtest decommission nodes duration failed on the test failed on branch master cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts decommission nodes duration run cluster go cluster go decommission go errgroup go pgurl for nodes empty from stdout stderr more artifacts related roachtest decommission nodes duration failed powered by
| 0
|
54,144
| 13,438,779,506
|
IssuesEvent
|
2020-09-07 19:03:36
|
tschaban/AFE-Firmware
|
https://api.github.com/repos/tschaban/AFE-Firmware
|
opened
|
LED configuration form needs to be rebuilt
|
Change Defect
|
move from one form configuration to - separate ones
|
1.0
|
LED configuration form needs to be rebuilt - move from one form configuration to - separate ones
|
defect
|
led configuration form needs to be rebuilt move from one form configuration to separate ones
| 1
|
290
| 2,523,264,874
|
IssuesEvent
|
2015-01-20 08:59:23
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
closed
|
linprog() failure on 0.15.1
|
defect Documentation scipy.optimize
|
I thought I'd take the new (and exciting) `linprog()` solver for a spin. On my first attempt trying to run the example problem verbatim, I received a significantly different result. For reference, the documentation says:
```python
>>> c = [-1, 4]
>>> A = [[-3, 1], [1, 2]]
>>> b = [6, 4]
>>> x0_bounds = (None, None)
>>> x1_bounds = (-3, None)
>>> res = linprog(c, A_ub=A, b_ub=b, bounds=(x0_bounds, x1_bounds),
... options={"disp": True})
>>> print(res)
Optimization terminated successfully.
Current function value: -11.428571
Iterations: 2
status: 0
success: True
fun: -11.428571428571429
x: array([-1.14285714, 2.57142857])
message: 'Optimization terminated successfully.'
nit: 2
```
However, when I run this from miniconda python 3, I see:
```python
import scipy
print(scipy.__version__)
from scipy.optimize import linprog
c = [-1, 4]
A = [[-3, 1], [1, 2]]
b = [6, 4]
x0_bounds = (None, None)
x1_bounds = (-3, None)
res = linprog(c, A_ub=A, b_ub=b, bounds=(x0_bounds, x1_bounds),
options={"disp": True})
print(res)
```
```bash
0.15.1
Optimization terminated successfully.
Current function value: -22.000000
Iterations: 1
status: 0
x: array([ 10., -3.])
slack: array([ 39., 0.])
nit: 1
message: 'Optimization terminated successfully.'
fun: -22.0
success: True
```
It seems like m version is failing to perform an iteration. Any thoughts? Is anyone else seeing this?
|
1.0
|
linprog() failure on 0.15.1 - I thought I'd take the new (and exciting) `linprog()` solver for a spin. On my first attempt trying to run the example problem verbatim, I received a significantly different result. For reference, the documentation says:
```python
>>> c = [-1, 4]
>>> A = [[-3, 1], [1, 2]]
>>> b = [6, 4]
>>> x0_bounds = (None, None)
>>> x1_bounds = (-3, None)
>>> res = linprog(c, A_ub=A, b_ub=b, bounds=(x0_bounds, x1_bounds),
... options={"disp": True})
>>> print(res)
Optimization terminated successfully.
Current function value: -11.428571
Iterations: 2
status: 0
success: True
fun: -11.428571428571429
x: array([-1.14285714, 2.57142857])
message: 'Optimization terminated successfully.'
nit: 2
```
However, when I run this from miniconda python 3, I see:
```python
import scipy
print(scipy.__version__)
from scipy.optimize import linprog
c = [-1, 4]
A = [[-3, 1], [1, 2]]
b = [6, 4]
x0_bounds = (None, None)
x1_bounds = (-3, None)
res = linprog(c, A_ub=A, b_ub=b, bounds=(x0_bounds, x1_bounds),
options={"disp": True})
print(res)
```
```bash
0.15.1
Optimization terminated successfully.
Current function value: -22.000000
Iterations: 1
status: 0
x: array([ 10., -3.])
slack: array([ 39., 0.])
nit: 1
message: 'Optimization terminated successfully.'
fun: -22.0
success: True
```
It seems like m version is failing to perform an iteration. Any thoughts? Is anyone else seeing this?
|
defect
|
linprog failure on i thought i d take the new and exciting linprog solver for a spin on my first attempt trying to run the example problem verbatim i received a significantly different result for reference the documentation says python c a b bounds none none bounds none res linprog c a ub a b ub b bounds bounds bounds options disp true print res optimization terminated successfully current function value iterations status success true fun x array message optimization terminated successfully nit however when i run this from miniconda python i see python import scipy print scipy version from scipy optimize import linprog c a b bounds none none bounds none res linprog c a ub a b ub b bounds bounds bounds options disp true print res bash optimization terminated successfully current function value iterations status x array slack array nit message optimization terminated successfully fun success true it seems like m version is failing to perform an iteration any thoughts is anyone else seeing this
| 1
|
82,013
| 31,852,471,852
|
IssuesEvent
|
2023-09-15 03:41:13
|
idaholab/moose
|
https://api.github.com/repos/idaholab/moose
|
opened
|
Variable syntax gets duplicated into Marker, Indicator, ReactionNetwork, Variables etc
|
P: minor T: defect C: Documentation
|
## Bug Description
In modules pages, the syntax for new variables is added in very many categories, some of which are irrelevant to the module at hand
For example in NS, the NS variables show up in
- adaptivty markers and indicators
- variables and auxvariables
- [ReactionNetwork/AqueousEquilibriumReactions](https://mooseframework.inl.gov/syntax/ReactionNetwork/AqueousEquilibriumReactions/index.html) and another one
## Steps to Reproduce
look at the docs for NS module
## Impact
Untidy
Longer scrolls
|
1.0
|
Variable syntax gets duplicated into Marker, Indicator, ReactionNetwork, Variables etc - ## Bug Description
In modules pages, the syntax for new variables is added in very many categories, some of which are irrelevant to the module at hand
For example in NS, the NS variables show up in
- adaptivty markers and indicators
- variables and auxvariables
- [ReactionNetwork/AqueousEquilibriumReactions](https://mooseframework.inl.gov/syntax/ReactionNetwork/AqueousEquilibriumReactions/index.html) and another one
## Steps to Reproduce
look at the docs for NS module
## Impact
Untidy
Longer scrolls
|
defect
|
variable syntax gets duplicated into marker indicator reactionnetwork variables etc bug description in modules pages the syntax for new variables is added in very many categories some of which are irrelevant to the module at hand for example in ns the ns variables show up in adaptivty markers and indicators variables and auxvariables and another one steps to reproduce look at the docs for ns module impact untidy longer scrolls
| 1
|
97,975
| 8,673,895,306
|
IssuesEvent
|
2018-11-30 04:53:19
|
humera987/FXLabs-Test-Automation
|
https://api.github.com/repos/humera987/FXLabs-Test-Automation
|
reopened
|
FXLabs Testing : ApiV1IssuesProjectIdIdGetQueryParamPageInvalidDatatype
|
FXLabs Testing
|
Project : FXLabs Testing
Job : UAT
Env : UAT
Region : US_WEST
Result : fail
Status Code : 404
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=MmUxYTkyZmQtMjVlZS00ZmQ0LWE3NjEtNmY4YWYwNzlhYjAw; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Fri, 30 Nov 2018 04:27:38 GMT]}
Endpoint : http://13.56.210.25/api/v1/api/v1/issues/project-id/qOxKhrDk?page=n9IquP&status=qOxKhrDk
Request :
Response :
{
"timestamp" : "2018-11-30T04:27:38.546+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/issues/project-id/qOxKhrDk"
}
Logs :
Assertion [@StatusCode != 401] resolved-to [404 != 401] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed]
--- FX Bot ---
|
1.0
|
FXLabs Testing : ApiV1IssuesProjectIdIdGetQueryParamPageInvalidDatatype - Project : FXLabs Testing
Job : UAT
Env : UAT
Region : US_WEST
Result : fail
Status Code : 404
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=MmUxYTkyZmQtMjVlZS00ZmQ0LWE3NjEtNmY4YWYwNzlhYjAw; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Fri, 30 Nov 2018 04:27:38 GMT]}
Endpoint : http://13.56.210.25/api/v1/api/v1/issues/project-id/qOxKhrDk?page=n9IquP&status=qOxKhrDk
Request :
Response :
{
"timestamp" : "2018-11-30T04:27:38.546+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/issues/project-id/qOxKhrDk"
}
Logs :
Assertion [@StatusCode != 401] resolved-to [404 != 401] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed]
--- FX Bot ---
|
non_defect
|
fxlabs testing project fxlabs testing job uat env uat region us west result fail status code headers x content type options x xss protection cache control pragma expires x frame options set cookie content type transfer encoding date endpoint request response timestamp status error not found message no message available path api api issues project id qoxkhrdk logs assertion resolved to result assertion resolved to result fx bot
| 0
|
19,756
| 3,253,434,462
|
IssuesEvent
|
2015-10-19 19:04:38
|
dart-lang/sdk
|
https://api.github.com/repos/dart-lang/sdk
|
closed
|
Analysis Server runs 'pub list-package-dirs' too many times (malformed pubspecs)
|
Analyzer-Server Area-Analyzer Performance Priority-High Type-Defect
|
(SDK 1.13-dev.7.0)
- Checkout project https://github.com/MikeMitterer/dart-material-design-lite. It contains 62 pubspec.yaml files of which 54 are valid and 8 are invalid (they are templates with placeholders).
- Open this project in the IDE. Analysis Server reports that it runs 'pub list-package-dirs' 62 times which is fine. Server warm up takes just 47 seconds (I guess no 3rd party packages analysis because no pub get done).
- Now run 'pub get' for all valid pubspecs (recommended to do it from the IDE)
- Close and reopen IDE
At this moment server says that 'pub list-package-dirs' was already started **more than 2000 times** and it already took 15 minutes. I'm killing the process as I can't wait more.
Note that if I delete malformed pubspecs from the project and restart the IDE it helps a bit. Analysis Server runs pub list 150 times in this case (still more than 54 - amount of pubspecs left) and server warms up in 4 minutes.
|
1.0
|
Analysis Server runs 'pub list-package-dirs' too many times (malformed pubspecs) - (SDK 1.13-dev.7.0)
- Checkout project https://github.com/MikeMitterer/dart-material-design-lite. It contains 62 pubspec.yaml files of which 54 are valid and 8 are invalid (they are templates with placeholders).
- Open this project in the IDE. Analysis Server reports that it runs 'pub list-package-dirs' 62 times which is fine. Server warm up takes just 47 seconds (I guess no 3rd party packages analysis because no pub get done).
- Now run 'pub get' for all valid pubspecs (recommended to do it from the IDE)
- Close and reopen IDE
At this moment server says that 'pub list-package-dirs' was already started **more than 2000 times** and it already took 15 minutes. I'm killing the process as I can't wait more.
Note that if I delete malformed pubspecs from the project and restart the IDE it helps a bit. Analysis Server runs pub list 150 times in this case (still more than 54 - amount of pubspecs left) and server warms up in 4 minutes.
|
defect
|
analysis server runs pub list package dirs too many times malformed pubspecs sdk dev checkout project it contains pubspec yaml files of which are valid and are invalid they are templates with placeholders open this project in the ide analysis server reports that it runs pub list package dirs times which is fine server warm up takes just seconds i guess no party packages analysis because no pub get done now run pub get for all valid pubspecs recommended to do it from the ide close and reopen ide at this moment server says that pub list package dirs was already started more than times and it already took minutes i m killing the process as i can t wait more note that if i delete malformed pubspecs from the project and restart the ide it helps a bit analysis server runs pub list times in this case still more than amount of pubspecs left and server warms up in minutes
| 1
|
161,962
| 13,880,497,345
|
IssuesEvent
|
2020-10-17 19:03:08
|
strong-roots-capital/od
|
https://api.github.com/repos/strong-roots-capital/od
|
opened
|
Be explicit in how fractional units are added
|
documentation
|
(not just added, any other manipulation too)
e.g. what happens when adding 1.5 months? 1.5 days?
|
1.0
|
Be explicit in how fractional units are added - (not just added, any other manipulation too)
e.g. what happens when adding 1.5 months? 1.5 days?
|
non_defect
|
be explicit in how fractional units are added not just added any other manipulation too e g what happens when adding months days
| 0
|
211,772
| 16,457,673,663
|
IssuesEvent
|
2021-05-21 14:35:44
|
DataDog/helm-charts
|
https://api.github.com/repos/DataDog/helm-charts
|
closed
|
Documentation should be explicit about `site` value
|
chart-datadog documentation
|
We did an install of the chart as described, but all agents were failing:
`UTC | CLUSTER | ERROR | (pkg/forwarder/transaction.go:179 in Process) | API Key invalid, dropping transaction for https://1-3-11-app.agent.datadoghq.com/api/v1/check_run`
It took a long time for us to discover we were missing this value, as we have an EU installation.
```
datadog:
site: 'datadoghq.eu'
```
The documentation should be clearer about this setting.
|
1.0
|
Documentation should be explicit about `site` value - We did an install of the chart as described, but all agents were failing:
`UTC | CLUSTER | ERROR | (pkg/forwarder/transaction.go:179 in Process) | API Key invalid, dropping transaction for https://1-3-11-app.agent.datadoghq.com/api/v1/check_run`
It took a long time for us to discover we were missing this value, as we have an EU installation.
```
datadog:
site: 'datadoghq.eu'
```
The documentation should be clearer about this setting.
|
non_defect
|
documentation should be explicit about site value we did an install of the chart as described but all agents were failing utc cluster error pkg forwarder transaction go in process api key invalid dropping transaction for it took a long time for us to discover we were missing this value as we have an eu installation datadog site datadoghq eu the documentation should be clearer about this setting
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.