Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 757 | labels stringlengths 4 664 | body stringlengths 3 261k | index stringclasses 10 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 232k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
239,408 | 19,863,793,782 | IssuesEvent | 2022-01-22 07:30:48 | Cookie-AutoDelete/Cookie-AutoDelete | https://api.github.com/repos/Cookie-AutoDelete/Cookie-AutoDelete | opened | [Bug] Can't Delete Cookies - Help | untested bug/issue | ### Acknowledgements
- [X] I acknowledge that I have read the above items
### Describe the bug
I have firefox, latest version, 64 bit on win 10. I thought I deleted cookies for Spectrum TV through firefox and CAD. But when I restart firefox and go to Spectrum TV, I'm signed in automatically. On the CAD icon popup there are no cookies stored. Same thing in firefox's settings. Yet I find myself signed in automatically every time I boot up firefox. How do I fix this??
### To Reproduce
I just visit spectrum tv
### Expected Behavior
I expect not being able to sing in automatically into spectrum tv on firefox
### Screenshots
_No response_
### System Info - Operating System (OS)
win 10
### System Info - Browser Info
firefox 96.0.2
### System Info - CookieAutoDelete Version
3.6.0
### Additional Context
_No response_ | 1.0 | [Bug] Can't Delete Cookies - Help - ### Acknowledgements
- [X] I acknowledge that I have read the above items
### Describe the bug
I have firefox, latest version, 64 bit on win 10. I thought I deleted cookies for Spectrum TV through firefox and CAD. But when I restart firefox and go to Spectrum TV, I'm signed in automatically. On the CAD icon popup there are no cookies stored. Same thing in firefox's settings. Yet I find myself signed in automatically every time I boot up firefox. How do I fix this??
### To Reproduce
I just visit spectrum tv
### Expected Behavior
I expect not being able to sing in automatically into spectrum tv on firefox
### Screenshots
_No response_
### System Info - Operating System (OS)
win 10
### System Info - Browser Info
firefox 96.0.2
### System Info - CookieAutoDelete Version
3.6.0
### Additional Context
_No response_ | non_defect | can t delete cookies help acknowledgements i acknowledge that i have read the above items describe the bug i have firefox latest version bit on win i thought i deleted cookies for spectrum tv through firefox and cad but when i restart firefox and go to spectrum tv i m signed in automatically on the cad icon popup there are no cookies stored same thing in firefox s settings yet i find myself signed in automatically every time i boot up firefox how do i fix this to reproduce i just visit spectrum tv expected behavior i expect not being able to sing in automatically into spectrum tv on firefox screenshots no response system info operating system os win system info browser info firefox system info cookieautodelete version additional context no response | 0 |
29,468 | 5,694,740,759 | IssuesEvent | 2017-04-15 16:05:19 | PiLoT-/pcgamer-minecraft-server-usa | https://api.github.com/repos/PiLoT-/pcgamer-minecraft-server-usa | closed | Disconnect during crafting loses items. | auto-migrated Performance Priority-Medium Type-Defect | ```
Forced disconnect(network errors) while crafting appears to erase newly crafted
blocks and their source materials from inventory.
```
Original issue reported on code.google.com by `kgg...@gmail.com` on 23 Dec 2014 at 11:32
| 1.0 | Disconnect during crafting loses items. - ```
Forced disconnect(network errors) while crafting appears to erase newly crafted
blocks and their source materials from inventory.
```
Original issue reported on code.google.com by `kgg...@gmail.com` on 23 Dec 2014 at 11:32
| defect | disconnect during crafting loses items forced disconnect network errors while crafting appears to erase newly crafted blocks and their source materials from inventory original issue reported on code google com by kgg gmail com on dec at | 1 |
63,479 | 12,333,950,260 | IssuesEvent | 2020-05-14 09:23:08 | HGustavs/LenaSYS | https://api.github.com/repos/HGustavs/LenaSYS | closed | Codeviewer: iFrames when watching previews have a weird size | CodeViewer Group-1-2020 | Refering to issue #7782, when previewing a document the iFrame is sized so that it doesn't completely fill the examplebox. See:

| 1.0 | Codeviewer: iFrames when watching previews have a weird size - Refering to issue #7782, when previewing a document the iFrame is sized so that it doesn't completely fill the examplebox. See:

| non_defect | codeviewer iframes when watching previews have a weird size refering to issue when previewing a document the iframe is sized so that it doesn t completely fill the examplebox see | 0 |
11,480 | 2,652,258,650 | IssuesEvent | 2015-03-16 16:23:48 | JoseExposito/touchegg | https://api.github.com/repos/JoseExposito/touchegg | closed | Stop detecting trackpad | auto-migrated Type-Defect | ```
Yesterday I installed Touchegg on my Asus laptop and everything worked fine, I
just changed some commands to test it. Today I tried to fully configure all
commands, but now touchegg does not recognize my trackpad, when you run it from
terminal you get this:
[+] Avaliable gesture:
Name -> Flick
[+] Avaliable gesture:
Name -> Drag
[+] Avaliable gesture:
Name -> Pinch
[+] Avaliable gesture:
Name -> Rotate
[+] Avaliable gesture:
Name -> Tap
[+] Avaliable gesture:
Name -> Touch
Yesterday when I did the gesture the information appeared, right now i get
nothing, I can't get it to work again
```
Original issue reported on code.google.com by `angel.al...@gmail.com` on 26 May 2014 at 10:12 | 1.0 | Stop detecting trackpad - ```
Yesterday I installed Touchegg on my Asus laptop and everything worked fine, I
just changed some commands to test it. Today I tried to fully configure all
commands, but now touchegg does not recognize my trackpad, when you run it from
terminal you get this:
[+] Avaliable gesture:
Name -> Flick
[+] Avaliable gesture:
Name -> Drag
[+] Avaliable gesture:
Name -> Pinch
[+] Avaliable gesture:
Name -> Rotate
[+] Avaliable gesture:
Name -> Tap
[+] Avaliable gesture:
Name -> Touch
Yesterday when I did the gesture the information appeared, right now i get
nothing, I can't get it to work again
```
Original issue reported on code.google.com by `angel.al...@gmail.com` on 26 May 2014 at 10:12 | defect | stop detecting trackpad yesterday i installed touchegg on my asus laptop and everything worked fine i just changed some commands to test it today i tried to fully configure all commands but now touchegg does not recognize my trackpad when you run it from terminal you get this avaliable gesture name flick avaliable gesture name drag avaliable gesture name pinch avaliable gesture name rotate avaliable gesture name tap avaliable gesture name touch yesterday when i did the gesture the information appeared right now i get nothing i can t get it to work again original issue reported on code google com by angel al gmail com on may at | 1 |
32,882 | 7,612,629,223 | IssuesEvent | 2018-05-01 18:14:57 | joomla/joomla-cms | https://api.github.com/repos/joomla/joomla-cms | closed | Hard coupling between com_users and plg_user_joomla | No Code Attached Yet | Within [`UsersModelUser::getForm()`](https://github.com/joomla/joomla-cms/blob/49e681fc66ecdd76388fa9e9cb3a92986c075e69/administrator/components/com_users/models/user.php#L131-L138) there is logic for the send mail to users plugin capability that is hardcoded to only support that plugin. Either this capability should be fully moved to the plugin (meaning the form alterations should happen via plugin event) or be fully moved to the component (meaning the mail sending code exists in the component). | 1.0 | Hard coupling between com_users and plg_user_joomla - Within [`UsersModelUser::getForm()`](https://github.com/joomla/joomla-cms/blob/49e681fc66ecdd76388fa9e9cb3a92986c075e69/administrator/components/com_users/models/user.php#L131-L138) there is logic for the send mail to users plugin capability that is hardcoded to only support that plugin. Either this capability should be fully moved to the plugin (meaning the form alterations should happen via plugin event) or be fully moved to the component (meaning the mail sending code exists in the component). | non_defect | hard coupling between com users and plg user joomla within there is logic for the send mail to users plugin capability that is hardcoded to only support that plugin either this capability should be fully moved to the plugin meaning the form alterations should happen via plugin event or be fully moved to the component meaning the mail sending code exists in the component | 0 |
46,129 | 13,055,856,760 | IssuesEvent | 2020-07-30 02:56:28 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | can't find python libs > python 2.6.x (Trac #627) | Incomplete Migration Migrated from Trac cmake defect | Migrated from https://code.icecube.wisc.edu/ticket/627
```json
{
"status": "closed",
"changetime": "2011-05-11T19:50:31",
"description": "cmake 2.6.x only has support for python < 2.6.x in its PythonFindX.cmake modules. (hard coded values). this seems to have been fixed in cmake 2.8.x, but is currently untested by !IceCube. (i personally use cmake 2.8.x but not python 2.7.x)\n\ncurrently i solve this on the build bots by hard coding python paths in a I3_CITE_CMAKE_DIR file on akuma (rhel4 + external python 2.7) and beastie (freebsd 9RC + freebsd port install of python 2.7.1 in /usr/local).\n\nthis '''can not''' be solved by including the cmake 2.8 FindPythonX modules, as we'd get into dependency hell w/ the cmake modules.\n\nprobably be able to solve this w/ i3-tools-v4 (#278)",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"_ts": "1305143431000000",
"component": "cmake",
"summary": "can't find python libs > python 2.6.x",
"priority": "major",
"keywords": "cmake python freebsd rhel",
"time": "2011-04-30T16:22:01",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
| 1.0 | can't find python libs > python 2.6.x (Trac #627) - Migrated from https://code.icecube.wisc.edu/ticket/627
```json
{
"status": "closed",
"changetime": "2011-05-11T19:50:31",
"description": "cmake 2.6.x only has support for python < 2.6.x in its PythonFindX.cmake modules. (hard coded values). this seems to have been fixed in cmake 2.8.x, but is currently untested by !IceCube. (i personally use cmake 2.8.x but not python 2.7.x)\n\ncurrently i solve this on the build bots by hard coding python paths in a I3_CITE_CMAKE_DIR file on akuma (rhel4 + external python 2.7) and beastie (freebsd 9RC + freebsd port install of python 2.7.1 in /usr/local).\n\nthis '''can not''' be solved by including the cmake 2.8 FindPythonX modules, as we'd get into dependency hell w/ the cmake modules.\n\nprobably be able to solve this w/ i3-tools-v4 (#278)",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"_ts": "1305143431000000",
"component": "cmake",
"summary": "can't find python libs > python 2.6.x",
"priority": "major",
"keywords": "cmake python freebsd rhel",
"time": "2011-04-30T16:22:01",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
| defect | can t find python libs python x trac migrated from json status closed changetime description cmake x only has support for python x in its pythonfindx cmake modules hard coded values this seems to have been fixed in cmake x but is currently untested by icecube i personally use cmake x but not python x n ncurrently i solve this on the build bots by hard coding python paths in a cite cmake dir file on akuma external python and beastie freebsd freebsd port install of python in usr local n nthis can not be solved by including the cmake findpythonx modules as we d get into dependency hell w the cmake modules n nprobably be able to solve this w tools reporter nega cc resolution fixed ts component cmake summary can t find python libs python x priority major keywords cmake python freebsd rhel time milestone owner nega type defect | 1 |
75,874 | 14,532,157,101 | IssuesEvent | 2020-12-14 21:57:32 | NightscoutFoundation/xDrip | https://api.github.com/repos/NightscoutFoundation/xDrip | closed | Quick Question About Building xDrip | code device-watch question | I'm having a little difficulty getting my changes (improvements!!!) to the watch face to show up.
Should I be building APK's, or building Bundles? (The reason I'm not sure is I never installed the watch face APK but the main xDrip app on the phone still was able to activate and add the xDrip watch face.)
Can somebody give me a hint how I get the new, updated code to the watch? It seems I can't even uninstall applications on the watch.
Any suggestions will be much appreciated!!!
| 1.0 | Quick Question About Building xDrip - I'm having a little difficulty getting my changes (improvements!!!) to the watch face to show up.
Should I be building APK's, or building Bundles? (The reason I'm not sure is I never installed the watch face APK but the main xDrip app on the phone still was able to activate and add the xDrip watch face.)
Can somebody give me a hint how I get the new, updated code to the watch? It seems I can't even uninstall applications on the watch.
Any suggestions will be much appreciated!!!
| non_defect | quick question about building xdrip i m having a little difficulty getting my changes improvements to the watch face to show up should i be building apk s or building bundles the reason i m not sure is i never installed the watch face apk but the main xdrip app on the phone still was able to activate and add the xdrip watch face can somebody give me a hint how i get the new updated code to the watch it seems i can t even uninstall applications on the watch any suggestions will be much appreciated | 0 |
16,845 | 2,948,327,080 | IssuesEvent | 2015-07-06 01:35:28 | Winetricks/winetricks | https://api.github.com/repos/Winetricks/winetricks | closed | as easy as winetricks setup.exe ?? | auto-migrated Priority-Medium Type-Defect | ```
i'm no programmer and i banging my head try adding new verbs:
http://code.google.com/p/winetricks/wiki/AddingNewVerbs
maybe give an easier way like this:
./winetricks --isolate mfc42 setup.exe
then it setup our apps into a new prefix.
(maybe ask us for what prefix we want?)
thank you!
```
Original issue reported on code.google.com by `mail2dny@gmail.com` on 15 Apr 2011 at 1:54 | 1.0 | as easy as winetricks setup.exe ?? - ```
i'm no programmer and i banging my head try adding new verbs:
http://code.google.com/p/winetricks/wiki/AddingNewVerbs
maybe give an easier way like this:
./winetricks --isolate mfc42 setup.exe
then it setup our apps into a new prefix.
(maybe ask us for what prefix we want?)
thank you!
```
Original issue reported on code.google.com by `mail2dny@gmail.com` on 15 Apr 2011 at 1:54 | defect | as easy as winetricks setup exe i m no programmer and i banging my head try adding new verbs maybe give an easier way like this winetricks isolate setup exe then it setup our apps into a new prefix maybe ask us for what prefix we want thank you original issue reported on code google com by gmail com on apr at | 1 |
65,947 | 8,862,430,836 | IssuesEvent | 2019-01-10 05:56:52 | edgi-govdata-archiving/web-monitoring | https://api.github.com/repos/edgi-govdata-archiving/web-monitoring | closed | Create Ingest System Diagram | documentation | Create visualization of how the various modules for ingesting archived web pages work together. They will be published in the github project documentation.
Drafts: https://scene.zeplin.io/project/5ad6aa451f6e994b2acd83da
(For access to zeplin to allow for annotation contact @weatherpattern )
Ingest from IA:
Step 1. Kick off the Internet Archive Ingest
Step 2. Importer sends requests to API
Step 3. API sends job to Redis queue
Step 4. Redis sends job off queue to Import Worker
Step 5. Import Worker fetches content from Internet Archive API
Step 6. Import Worker stores metadase to Postgres database and content to S3
Ingest from Versionista:
Step 1. Kick off the Versionista Ingest
Step 2. Scraper sends requests to API
Step 3. API sends job to Redis queue
Step 4. Redis sends job off queue to Import Worker
Step 5. Import Worker fetches content from Versionista
Step 6. Import Worker stores metadase to Postgres database and content to S3
| 1.0 | Create Ingest System Diagram - Create visualization of how the various modules for ingesting archived web pages work together. They will be published in the github project documentation.
Drafts: https://scene.zeplin.io/project/5ad6aa451f6e994b2acd83da
(For access to zeplin to allow for annotation contact @weatherpattern )
Ingest from IA:
Step 1. Kick off the Internet Archive Ingest
Step 2. Importer sends requests to API
Step 3. API sends job to Redis queue
Step 4. Redis sends job off queue to Import Worker
Step 5. Import Worker fetches content from Internet Archive API
Step 6. Import Worker stores metadase to Postgres database and content to S3
Ingest from Versionista:
Step 1. Kick off the Versionista Ingest
Step 2. Scraper sends requests to API
Step 3. API sends job to Redis queue
Step 4. Redis sends job off queue to Import Worker
Step 5. Import Worker fetches content from Versionista
Step 6. Import Worker stores metadase to Postgres database and content to S3
| non_defect | create ingest system diagram create visualization of how the various modules for ingesting archived web pages work together they will be published in the github project documentation drafts for access to zeplin to allow for annotation contact weatherpattern ingest from ia step kick off the internet archive ingest step importer sends requests to api step api sends job to redis queue step redis sends job off queue to import worker step import worker fetches content from internet archive api step import worker stores metadase to postgres database and content to ingest from versionista step kick off the versionista ingest step scraper sends requests to api step api sends job to redis queue step redis sends job off queue to import worker step import worker fetches content from versionista step import worker stores metadase to postgres database and content to | 0 |
64,184 | 18,270,641,233 | IssuesEvent | 2021-10-04 13:30:13 | vector-im/element-ios | https://api.github.com/repos/vector-im/element-ios | closed | As read receipts move, they don't always update the cell data's height. | T-Defect A-Timeline | This leaves blank spaces as a conversation evolves. Easiest to reproduce by opening two accounts and having a quick back and forth 1 message at a time. Confirmed on 1.5.4 and 1.6.1. Related to #4564.

| 1.0 | As read receipts move, they don't always update the cell data's height. - This leaves blank spaces as a conversation evolves. Easiest to reproduce by opening two accounts and having a quick back and forth 1 message at a time. Confirmed on 1.5.4 and 1.6.1. Related to #4564.

| defect | as read receipts move they don t always update the cell data s height this leaves blank spaces as a conversation evolves easiest to reproduce by opening two accounts and having a quick back and forth message at a time confirmed on and related to | 1 |
44,102 | 17,796,015,763 | IssuesEvent | 2021-08-31 22:21:37 | localstack/localstack | https://api.github.com/repos/localstack/localstack | closed | S3 and DynamoDB calls sometimes return "use of closed network connection" | service:s3 | Sometimes, randomly I observe the following error when using S3 or DynamoDB when working with `localstack` locally (started in docker container) or on CI (also in docker container).
```
Error:
Received unexpected error:
RequestError: send request failed
caused by: Put http://localhost:4572/a4efd66c-f646-11e9-af26-88e9fe610200/24.0: readfrom tcp [::1]:53042->[::1]:4572: write tcp [::1]:53042->[::1]:4572: use of closed network connection
```
or
```
Error:
Received unexpected error:
RequestError: send request failed
caused by: Post http://localhost:4569/: readfrom tcp 127.0.0.1:53606->127.0.0.1:4569: write tcp 127.0.0.1:53606->127.0.0.1:4569: write: broken pipe
```
I'm using [`aws-sdk-go`](https://github.com/aws/aws-sdk-go/) version `v1.16.30` (I've also tried to newest `v1.25.18` and I get the same error).
To get you an idea of the frequency with which this happens: I spin up tests (with `go test`) with `-count 50` and I can get around 1-3 failures because of that.
An exemplar test pushes data to S3 and then deletes it and the bucket.
┆Issue is synchronized with this [Jira Bug](https://localstack.atlassian.net/browse/LOC-132) by [Unito](https://www.unito.io/learn-more)
| 1.0 | S3 and DynamoDB calls sometimes return "use of closed network connection" - Sometimes, randomly I observe the following error when using S3 or DynamoDB when working with `localstack` locally (started in docker container) or on CI (also in docker container).
```
Error:
Received unexpected error:
RequestError: send request failed
caused by: Put http://localhost:4572/a4efd66c-f646-11e9-af26-88e9fe610200/24.0: readfrom tcp [::1]:53042->[::1]:4572: write tcp [::1]:53042->[::1]:4572: use of closed network connection
```
or
```
Error:
Received unexpected error:
RequestError: send request failed
caused by: Post http://localhost:4569/: readfrom tcp 127.0.0.1:53606->127.0.0.1:4569: write tcp 127.0.0.1:53606->127.0.0.1:4569: write: broken pipe
```
I'm using [`aws-sdk-go`](https://github.com/aws/aws-sdk-go/) version `v1.16.30` (I've also tried to newest `v1.25.18` and I get the same error).
To get you an idea of the frequency with which this happens: I spin up tests (with `go test`) with `-count 50` and I can get around 1-3 failures because of that.
An exemplar test pushes data to S3 and then deletes it and the bucket.
┆Issue is synchronized with this [Jira Bug](https://localstack.atlassian.net/browse/LOC-132) by [Unito](https://www.unito.io/learn-more)
| non_defect | and dynamodb calls sometimes return use of closed network connection sometimes randomly i observe the following error when using or dynamodb when working with localstack locally started in docker container or on ci also in docker container error received unexpected error requesterror send request failed caused by put readfrom tcp write tcp use of closed network connection or error received unexpected error requesterror send request failed caused by post readfrom tcp write tcp write broken pipe i m using version i ve also tried to newest and i get the same error to get you an idea of the frequency with which this happens i spin up tests with go test with count and i can get around failures because of that an exemplar test pushes data to and then deletes it and the bucket ┆issue is synchronized with this by | 0 |
9,184 | 2,615,137,636 | IssuesEvent | 2015-03-01 06:11:16 | chrsmith/reaver-wps | https://api.github.com/repos/chrsmith/reaver-wps | closed | Reaver segmentation fault | auto-migrated Priority-Medium Type-Defect | ```
Moved from issue #2:
Comment 20 by gorilla.maguila, Today (43 minutes ago)
This is what I get with latest subversion:
[+] Waiting for beacon from C0:3F:0E:C1:DB:A7
[+] Switching mon0 to channel 2
[+] Switching mon0 to channel 4
[!] WARNING: Failed to associate with C0:3F:0E:C1:DB:A7
[!] WARNING: Failed to associate with C0:3F:0E:C1:DB:A7
[+] Associated with C0:3F:0E:C1:DB:A7
[+] Trying pin 90553301
[!] WARNING: Failed to associate with C0:3F:0E:C1:DB:A7
[+] Switching mon0 to channel 3
[!] WARNING: Receive timeout occurred
[!] WARNING: Receive timeout occurred
[!] WARNING: Receive timeout occurred
[!] WARNING: Receive timeout occurred
Segmentation fault
I'm under kernel 3.1.5 with iwlagn driver
Comment 21 by project member cheffner@tacnetsol.com, Today (33 minutes ago)
maguila,
I have not tested the iwlagn drivers, but since you were able to associate I'd
suspect that injection is working properly. The failed associations and receive
timeouts are usually an indication of poor signal strength or a lot of wireless
interference.
The segfault is troubling though. Can you give more info on your OS ?
Comment 22 by gorilla.maguila, Today (12 minutes ago)
I'm using Archlinux x64. We use almost the latest packages on everything as it
is a rolling release distro.
I have tried to run under gdb but I don't know why I don't get the segmentation
fault:
$gdb ./reaver
GNU gdb (GDB) 7.3.1
Copyright (C) 2011 Free Software Foundation, Inc.
License GPLv3+: GNU GPL version 3 or later <http://gnu.org/licenses/gpl.html>
This is free software: you are free to change and redistribute it.
There is NO WARRANTY, to the extent permitted by law. Type "show copying"
and "show warranty" for details.
This GDB was configured as "x86_64-unknown-linux-gnu".
For bug reporting instructions, please see:
<http://www.gnu.org/software/gdb/bugs/>...
Reading symbols from /home/user/reaver-wps-read-only/src/reaver...done.
(gdb) run -i mon0 -b C0:3F:0E:C1:DB:A7 -vv
Starting program: /home/user/reaver-wps-read-only/src/reaver -i mon0 -b
C0:3F:0E:C1:DB:A7 -vv
Reaver v1.0 WiFi Protected Setup Attack Tool
Copyright (c) 2011, Tactical Network Solutions, Craig Heffner
<cheffner@tacnetsol.com>
[+] Waiting for beacon from C0:3F:0E:C1:DB:A7
[+] Switching mon0 to channel 4
[!] WARNING: Failed to associate with C0:3F:0E:C1:DB:A7 (ESSID: ONODBA7)
[!] WARNING: Failed to associate with C0:3F:0E:C1:DB:A7 (ESSID: ONODBA7)
[!] WARNING: Failed to associate with C0:3F:0E:C1:DB:A7 (ESSID: ONODBA7)
[!] WARNING: Failed to associate with C0:3F:0E:C1:DB:A7 (ESSID: ONODBA7)
[!] WARNING: Failed to associate with C0:3F:0E:C1:DB:A7 (ESSID: ONODBA7)
[+] Associated with C0:3F:0E:C1:DB:A7 (ESSID: ONODBA7)
[+] Trying pin 26141367
[!] WARNING: Failed to associate with C0:3F:0E:C1:DB:A7 (ESSID: ONODBA7)
[+] Switching mon0 to channel 2
[!] WARNING: Receive timeout occurred
[!] WARNING: Receive timeout occurred
[!] WARNING: Receive timeout occurred
[!] WARNING: Receive timeout occurred
[!] WARNING: Receive timeout occurred
[!] WARNING: Receive timeout occurred
...etc
But again I get the segmentation fault without gdb.
```
Original issue reported on code.google.com by `cheff...@tacnetsol.com` on 29 Dec 2011 at 3:36 | 1.0 | Reaver segmentation fault - ```
Moved from issue #2:
Comment 20 by gorilla.maguila, Today (43 minutes ago)
This is what I get with latest subversion:
[+] Waiting for beacon from C0:3F:0E:C1:DB:A7
[+] Switching mon0 to channel 2
[+] Switching mon0 to channel 4
[!] WARNING: Failed to associate with C0:3F:0E:C1:DB:A7
[!] WARNING: Failed to associate with C0:3F:0E:C1:DB:A7
[+] Associated with C0:3F:0E:C1:DB:A7
[+] Trying pin 90553301
[!] WARNING: Failed to associate with C0:3F:0E:C1:DB:A7
[+] Switching mon0 to channel 3
[!] WARNING: Receive timeout occurred
[!] WARNING: Receive timeout occurred
[!] WARNING: Receive timeout occurred
[!] WARNING: Receive timeout occurred
Segmentation fault
I'm under kernel 3.1.5 with iwlagn driver
Comment 21 by project member cheffner@tacnetsol.com, Today (33 minutes ago)
maguila,
I have not tested the iwlagn drivers, but since you were able to associate I'd
suspect that injection is working properly. The failed associations and receive
timeouts are usually an indication of poor signal strength or a lot of wireless
interference.
The segfault is troubling though. Can you give more info on your OS ?
Comment 22 by gorilla.maguila, Today (12 minutes ago)
I'm using Archlinux x64. We use almost the latest packages on everything as it
is a rolling release distro.
I have tried to run under gdb but I don't know why I don't get the segmentation
fault:
$gdb ./reaver
GNU gdb (GDB) 7.3.1
Copyright (C) 2011 Free Software Foundation, Inc.
License GPLv3+: GNU GPL version 3 or later <http://gnu.org/licenses/gpl.html>
This is free software: you are free to change and redistribute it.
There is NO WARRANTY, to the extent permitted by law. Type "show copying"
and "show warranty" for details.
This GDB was configured as "x86_64-unknown-linux-gnu".
For bug reporting instructions, please see:
<http://www.gnu.org/software/gdb/bugs/>...
Reading symbols from /home/user/reaver-wps-read-only/src/reaver...done.
(gdb) run -i mon0 -b C0:3F:0E:C1:DB:A7 -vv
Starting program: /home/user/reaver-wps-read-only/src/reaver -i mon0 -b
C0:3F:0E:C1:DB:A7 -vv
Reaver v1.0 WiFi Protected Setup Attack Tool
Copyright (c) 2011, Tactical Network Solutions, Craig Heffner
<cheffner@tacnetsol.com>
[+] Waiting for beacon from C0:3F:0E:C1:DB:A7
[+] Switching mon0 to channel 4
[!] WARNING: Failed to associate with C0:3F:0E:C1:DB:A7 (ESSID: ONODBA7)
[!] WARNING: Failed to associate with C0:3F:0E:C1:DB:A7 (ESSID: ONODBA7)
[!] WARNING: Failed to associate with C0:3F:0E:C1:DB:A7 (ESSID: ONODBA7)
[!] WARNING: Failed to associate with C0:3F:0E:C1:DB:A7 (ESSID: ONODBA7)
[!] WARNING: Failed to associate with C0:3F:0E:C1:DB:A7 (ESSID: ONODBA7)
[+] Associated with C0:3F:0E:C1:DB:A7 (ESSID: ONODBA7)
[+] Trying pin 26141367
[!] WARNING: Failed to associate with C0:3F:0E:C1:DB:A7 (ESSID: ONODBA7)
[+] Switching mon0 to channel 2
[!] WARNING: Receive timeout occurred
[!] WARNING: Receive timeout occurred
[!] WARNING: Receive timeout occurred
[!] WARNING: Receive timeout occurred
[!] WARNING: Receive timeout occurred
[!] WARNING: Receive timeout occurred
...etc
But again I get the segmentation fault without gdb.
```
Original issue reported on code.google.com by `cheff...@tacnetsol.com` on 29 Dec 2011 at 3:36 | defect | reaver segmentation fault moved from issue comment by gorilla maguila today minutes ago this is what i get with latest subversion waiting for beacon from db switching to channel switching to channel warning failed to associate with db warning failed to associate with db associated with db trying pin warning failed to associate with db switching to channel warning receive timeout occurred warning receive timeout occurred warning receive timeout occurred warning receive timeout occurred segmentation fault i m under kernel with iwlagn driver comment by project member cheffner tacnetsol com today minutes ago maguila i have not tested the iwlagn drivers but since you were able to associate i d suspect that injection is working properly the failed associations and receive timeouts are usually an indication of poor signal strength or a lot of wireless interference the segfault is troubling though can you give more info on your os comment by gorilla maguila today minutes ago i m using archlinux we use almost the latest packages on everything as it is a rolling release distro i have tried to run under gdb but i don t know why i don t get the segmentation fault gdb reaver gnu gdb gdb copyright c free software foundation inc license gnu gpl version or later this is free software you are free to change and redistribute it there is no warranty to the extent permitted by law type show copying and show warranty for details this gdb was configured as unknown linux gnu for bug reporting instructions please see reading symbols from home user reaver wps read only src reaver done gdb run i b db vv starting program home user reaver wps read only src reaver i b db vv reaver wifi protected setup attack tool copyright c tactical network solutions craig heffner waiting for beacon from db switching to channel warning failed to associate with db essid warning failed to associate with db essid warning failed to associate with db essid warning failed to associate with db essid warning failed to associate with db essid associated with db essid trying pin warning failed to associate with db essid switching to channel warning receive timeout occurred warning receive timeout occurred warning receive timeout occurred warning receive timeout occurred warning receive timeout occurred warning receive timeout occurred etc but again i get the segmentation fault without gdb original issue reported on code google com by cheff tacnetsol com on dec at | 1 |
69,724 | 22,630,305,281 | IssuesEvent | 2022-06-30 14:09:36 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Default filename for 'save image as' is always image.png | T-Defect S-Minor A-Electron A-Media O-Occasional | ### Steps to reproduce
1. post a link to a gif such as https://media3.giphy.com/media/3o85xIO33l7RlmLR4I/giphy.gif?cid=790b7611b5238203d4dfc621261917cf4acab0e609612f97&rid=giphy.gif&ct=g
2. wait for image preview
3. click on image to view gif in lightbox
4. right click and save
### Outcome
#### What did you expect?
defaults to `something.gif`
#### What happened instead?
defaults to `image.png`
### Operating system
arch
### Application version
Element Nightly version: 2022020101 Olm version: 3.2.8
### How did you install the app?
aur/element-desktop-nightly-bin
### Homeserver
synapse 1.51
### Will you send logs?
No | 1.0 | Default filename for 'save image as' is always image.png - ### Steps to reproduce
1. post a link to a gif such as https://media3.giphy.com/media/3o85xIO33l7RlmLR4I/giphy.gif?cid=790b7611b5238203d4dfc621261917cf4acab0e609612f97&rid=giphy.gif&ct=g
2. wait for image preview
3. click on image to view gif in lightbox
4. right click and save
### Outcome
#### What did you expect?
defaults to `something.gif`
#### What happened instead?
defaults to `image.png`
### Operating system
arch
### Application version
Element Nightly version: 2022020101 Olm version: 3.2.8
### How did you install the app?
aur/element-desktop-nightly-bin
### Homeserver
synapse 1.51
### Will you send logs?
No | defect | default filename for save image as is always image png steps to reproduce post a link to a gif such as wait for image preview click on image to view gif in lightbox right click and save outcome what did you expect defaults to something gif what happened instead defaults to image png operating system arch application version element nightly version olm version how did you install the app aur element desktop nightly bin homeserver synapse will you send logs no | 1 |
219,199 | 17,070,448,596 | IssuesEvent | 2021-07-07 12:44:23 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | tests: samples: watchdog: sample.subsys.task_wdt fails on nrf platforms | area: Samples area: Tests area: Watchdog bug priority: low | **Describe the bug**
The sample from `zephyr/samples/subsys/task_wdt` fails on nrf9160 and nrf52840 (passes on nrf5340)
**To Reproduce**
1. have nrf52840 connected
2. run `./scripts/twister -T samples/subsys/task_wdt/ -p nrf52840dk_nrf52840 --device-testing --device-serial /dev/ttyACM0 -v -v`
3. See error
**Expected behavior**
The test passes
**Impact**
Not clear
**Logs and console output**
The device is printing the below msg in a loop and terminates with a timeout
```
DEBUG - DEVICE: *** Booting Zephyr OS build zephyr-v2.6.0-368-g3e492665c012 ***
DEBUG - DEVICE: Task watchdog sample application.
DEBUG - DEVICE: Main thread still alive...
DEBUG - DEVICE: Control thread started.
```
**Environment (please complete the following information):**
- OS: Ubuntu 18.04
- Toolchain zephyr-sdk 0.12.2
- Commit SHA or Version used: zephyr-v2.6.0-368-g3e492665c012
| 1.0 | tests: samples: watchdog: sample.subsys.task_wdt fails on nrf platforms - **Describe the bug**
The sample from `zephyr/samples/subsys/task_wdt` fails on nrf9160 and nrf52840 (passes on nrf5340)
**To Reproduce**
1. have nrf52840 connected
2. run `./scripts/twister -T samples/subsys/task_wdt/ -p nrf52840dk_nrf52840 --device-testing --device-serial /dev/ttyACM0 -v -v`
3. See error
**Expected behavior**
The test passes
**Impact**
Not clear
**Logs and console output**
The device is printing the below msg in a loop and terminates with a timeout
```
DEBUG - DEVICE: *** Booting Zephyr OS build zephyr-v2.6.0-368-g3e492665c012 ***
DEBUG - DEVICE: Task watchdog sample application.
DEBUG - DEVICE: Main thread still alive...
DEBUG - DEVICE: Control thread started.
```
**Environment (please complete the following information):**
- OS: Ubuntu 18.04
- Toolchain zephyr-sdk 0.12.2
- Commit SHA or Version used: zephyr-v2.6.0-368-g3e492665c012
| non_defect | tests samples watchdog sample subsys task wdt fails on nrf platforms describe the bug the sample from zephyr samples subsys task wdt fails on and passes on to reproduce have connected run scripts twister t samples subsys task wdt p device testing device serial dev v v see error expected behavior the test passes impact not clear logs and console output the device is printing the below msg in a loop and terminates with a timeout debug device booting zephyr os build zephyr debug device task watchdog sample application debug device main thread still alive debug device control thread started environment please complete the following information os ubuntu toolchain zephyr sdk commit sha or version used zephyr | 0 |
26,276 | 4,650,183,790 | IssuesEvent | 2016-10-03 01:57:06 | mercury-hpc/mercury | https://api.github.com/repos/mercury-hpc/mercury | closed | CLOCK_MONOTONIC_RAW not defined at nersc edison | defect hg minor | Looks like nersc edison's linux doesn't have CLOCK_MONOTONIC_RAW defined so mercury cannot be compiled on it.
Is it possible to use a different but close enough clock type to replace CLOCK_MONOTONIC_RAW when CLOCK_MONOTONIC_RAW is absent ? :-)
zhengq@edison10:~/deltafs-umbrella/build> gcc --version
gcc (GCC) 6.1.0 20160427 (Cray Inc.)
Copyright (C) 2016 Free Software Foundation, Inc.
This is free software; see the source for copying conditions. There is NO
warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
zhengq@edison10:~/deltafs-umbrella/build> uname -a
Linux edison10 3.0.101-0.46-default #1 SMP Wed Dec 17 11:04:10 UTC 2014 (8356111) x86_64 x86_64 x86_64 GNU/Linux
/global/homes/z/zhengq/deltafs-umbrella/build/mercury-prefix/src/mercury/src/util/mercury_time.c: In function 'hg_time_get_current':
/global/homes/z/zhengq/deltafs-umbrella/build/mercury-prefix/src/mercury/src/util/mercury_time.c:112:23: error: 'CLOCK_MONOTONIC_RAW' undeclared (first use in this function)
if (clock_gettime(CLOCK_MONOTONIC_RAW, &tp)) {
^~~~~~~~~~~~~~~~~~~
/global/homes/z/zhengq/deltafs-umbrella/build/mercury-prefix/src/mercury/src/util/mercury_time.c:112:23: note: each undeclared identifier is reported only once for each function it appears in
make[6]: *** [src/util/CMakeFiles/mercury_util.dir/mercury_time.c.o] Error 1
make[5]: *** [src/util/CMakeFiles/mercury_util.dir/all] Error 2
make[4]: *** [all] Error 2
make[3]: *** [mercury-prefix/src/mercury-stamp/mercury-build] Error 2
make[2]: *** [CMakeFiles/mercury.dir/all] Error 2
make[1]: *** [CMakeFiles/deltafs.dir/rule] Error 2
make: *** [deltafs] Error 2
| 1.0 | CLOCK_MONOTONIC_RAW not defined at nersc edison - Looks like nersc edison's linux doesn't have CLOCK_MONOTONIC_RAW defined so mercury cannot be compiled on it.
Is it possible to use a different but close enough clock type to replace CLOCK_MONOTONIC_RAW when CLOCK_MONOTONIC_RAW is absent ? :-)
zhengq@edison10:~/deltafs-umbrella/build> gcc --version
gcc (GCC) 6.1.0 20160427 (Cray Inc.)
Copyright (C) 2016 Free Software Foundation, Inc.
This is free software; see the source for copying conditions. There is NO
warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
zhengq@edison10:~/deltafs-umbrella/build> uname -a
Linux edison10 3.0.101-0.46-default #1 SMP Wed Dec 17 11:04:10 UTC 2014 (8356111) x86_64 x86_64 x86_64 GNU/Linux
/global/homes/z/zhengq/deltafs-umbrella/build/mercury-prefix/src/mercury/src/util/mercury_time.c: In function 'hg_time_get_current':
/global/homes/z/zhengq/deltafs-umbrella/build/mercury-prefix/src/mercury/src/util/mercury_time.c:112:23: error: 'CLOCK_MONOTONIC_RAW' undeclared (first use in this function)
if (clock_gettime(CLOCK_MONOTONIC_RAW, &tp)) {
^~~~~~~~~~~~~~~~~~~
/global/homes/z/zhengq/deltafs-umbrella/build/mercury-prefix/src/mercury/src/util/mercury_time.c:112:23: note: each undeclared identifier is reported only once for each function it appears in
make[6]: *** [src/util/CMakeFiles/mercury_util.dir/mercury_time.c.o] Error 1
make[5]: *** [src/util/CMakeFiles/mercury_util.dir/all] Error 2
make[4]: *** [all] Error 2
make[3]: *** [mercury-prefix/src/mercury-stamp/mercury-build] Error 2
make[2]: *** [CMakeFiles/mercury.dir/all] Error 2
make[1]: *** [CMakeFiles/deltafs.dir/rule] Error 2
make: *** [deltafs] Error 2
| defect | clock monotonic raw not defined at nersc edison looks like nersc edison s linux doesn t have clock monotonic raw defined so mercury cannot be compiled on it is it possible to use a different but close enough clock type to replace clock monotonic raw when clock monotonic raw is absent zhengq deltafs umbrella build gcc version gcc gcc cray inc copyright c free software foundation inc this is free software see the source for copying conditions there is no warranty not even for merchantability or fitness for a particular purpose zhengq deltafs umbrella build uname a linux default smp wed dec utc gnu linux global homes z zhengq deltafs umbrella build mercury prefix src mercury src util mercury time c in function hg time get current global homes z zhengq deltafs umbrella build mercury prefix src mercury src util mercury time c error clock monotonic raw undeclared first use in this function if clock gettime clock monotonic raw tp global homes z zhengq deltafs umbrella build mercury prefix src mercury src util mercury time c note each undeclared identifier is reported only once for each function it appears in make error make error make error make error make error make error make error | 1 |
194,538 | 6,895,856,045 | IssuesEvent | 2017-11-23 15:22:12 | ubuntudesign/vanilla-design | https://api.github.com/repos/ubuntudesign/vanilla-design | closed | Spec slider pattern | Priority: Medium Type: Enhancement | [Notes from Vanilla brochure theme workshop](https://github.com/ubuntudesign/meeting-notes/issues/155):
- Agreed name: **slider** (`p-slider`)
- Verify if Jaas slider is the correct/final version
- Should be in Vanilla: ubuntu.com has slider (contribute), Landscape and Jaas too
@spencerbygraves could you confirm what is the latest version of the Jaas sliders? | 1.0 | Spec slider pattern - [Notes from Vanilla brochure theme workshop](https://github.com/ubuntudesign/meeting-notes/issues/155):
- Agreed name: **slider** (`p-slider`)
- Verify if Jaas slider is the correct/final version
- Should be in Vanilla: ubuntu.com has slider (contribute), Landscape and Jaas too
@spencerbygraves could you confirm what is the latest version of the Jaas sliders? | non_defect | spec slider pattern agreed name slider p slider verify if jaas slider is the correct final version should be in vanilla ubuntu com has slider contribute landscape and jaas too spencerbygraves could you confirm what is the latest version of the jaas sliders | 0 |
73,035 | 24,414,641,683 | IssuesEvent | 2022-10-05 14:56:57 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | closed | TabMenu: items not visible on small screens | :lady_beetle: defect theme | ### Describe the bug
Items are not visible on small screens. Especially if you do not use icons. With icons, only the icons are visible on small screens.
Large (don't mind the badge.. it's a PR I'm working on):
<img width="836" alt="Screenshot 2022-08-23 at 12 48 39" src="https://user-images.githubusercontent.com/7500178/186140340-caf6b7ed-bba6-47aa-b7ad-7b681d168631.png">
Small:
<img width="580" alt="Screenshot 2022-08-23 at 12 49 09" src="https://user-images.githubusercontent.com/7500178/186140368-76b4086d-a48c-4c48-a061-a4c0afba4a99.png">
### Reproducer
Showcase without icons:
```xhtml
<p:tabMenu activeIndex="#{param.i}">
<p:menuitem value="Home" outcome="/ui/menu/tabMenu">
<f:param name="i" value="0"/>
</p:menuitem>
<p:menuitem value="Calendar" outcome="/ui/menu/tabMenu">
<f:param name="i" value="1"/>
</p:menuitem>
<p:menuitem value="Edit" outcome="/ui/menu/tabMenu"
disabled="true">
<f:param name="i" value="2"/>
</p:menuitem>
<p:menuitem value="Documentation" outcome="/ui/menu/tabMenu">
<f:param name="i" value="3"/>
</p:menuitem>
<p:menuitem value="Settings" outcome="/ui/menu/tabMenu">
<f:param name="i" value="4"/>
</p:menuitem>
</p:tabMenu>
```
### Expected behavior
Items should wrap
### PrimeFaces edition
Community
### PrimeFaces version
12.0.0-SNAPSHOT
### Theme
Saga, Diamond
### JSF implementation
All
### JSF version
2.3
### Java version
17
### Browser(s)
Chrome latest macOS | 1.0 | TabMenu: items not visible on small screens - ### Describe the bug
Items are not visible on small screens. Especially if you do not use icons. With icons, only the icons are visible on small screens.
Large (don't mind the badge.. it's a PR I'm working on):
<img width="836" alt="Screenshot 2022-08-23 at 12 48 39" src="https://user-images.githubusercontent.com/7500178/186140340-caf6b7ed-bba6-47aa-b7ad-7b681d168631.png">
Small:
<img width="580" alt="Screenshot 2022-08-23 at 12 49 09" src="https://user-images.githubusercontent.com/7500178/186140368-76b4086d-a48c-4c48-a061-a4c0afba4a99.png">
### Reproducer
Showcase without icons:
```xhtml
<p:tabMenu activeIndex="#{param.i}">
<p:menuitem value="Home" outcome="/ui/menu/tabMenu">
<f:param name="i" value="0"/>
</p:menuitem>
<p:menuitem value="Calendar" outcome="/ui/menu/tabMenu">
<f:param name="i" value="1"/>
</p:menuitem>
<p:menuitem value="Edit" outcome="/ui/menu/tabMenu"
disabled="true">
<f:param name="i" value="2"/>
</p:menuitem>
<p:menuitem value="Documentation" outcome="/ui/menu/tabMenu">
<f:param name="i" value="3"/>
</p:menuitem>
<p:menuitem value="Settings" outcome="/ui/menu/tabMenu">
<f:param name="i" value="4"/>
</p:menuitem>
</p:tabMenu>
```
### Expected behavior
Items should wrap
### PrimeFaces edition
Community
### PrimeFaces version
12.0.0-SNAPSHOT
### Theme
Saga, Diamond
### JSF implementation
All
### JSF version
2.3
### Java version
17
### Browser(s)
Chrome latest macOS | defect | tabmenu items not visible on small screens describe the bug items are not visible on small screens especially if you do not use icons with icons only the icons are visible on small screens large don t mind the badge it s a pr i m working on img width alt screenshot at src small img width alt screenshot at src reproducer showcase without icons xhtml p menuitem value edit outcome ui menu tabmenu disabled true expected behavior items should wrap primefaces edition community primefaces version snapshot theme saga diamond jsf implementation all jsf version java version browser s chrome latest macos | 1 |
9,585 | 2,615,163,133 | IssuesEvent | 2015-03-01 06:42:37 | chrsmith/reaver-wps | https://api.github.com/repos/chrsmith/reaver-wps | opened | Is WPA PSK the passphrase? | auto-migrated Priority-Triage Type-Defect | ```
Reaver has calculated both the WPS pin and WPA PSK, but I do not seem to be
able to connect to the network with the WPA PSK, is the WPA PSK the passphrase?
```
Original issue reported on code.google.com by `yanglifu90` on 24 Jun 2012 at 2:18 | 1.0 | Is WPA PSK the passphrase? - ```
Reaver has calculated both the WPS pin and WPA PSK, but I do not seem to be
able to connect to the network with the WPA PSK, is the WPA PSK the passphrase?
```
Original issue reported on code.google.com by `yanglifu90` on 24 Jun 2012 at 2:18 | defect | is wpa psk the passphrase reaver has calculated both the wps pin and wpa psk but i do not seem to be able to connect to the network with the wpa psk is the wpa psk the passphrase original issue reported on code google com by on jun at | 1 |
21,280 | 3,481,076,038 | IssuesEvent | 2015-12-29 13:35:30 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | Hz 3.4 OperationTimeoutException, op=TxnLockAndGetOperation, MapTransactionContextConflictTest | SIMULATOR Team: Core Type: Defect | Tested on Hazelcast 3.4
MapTransactionContextConflictTest reliably causes an OperationTimeoutException.
3 out of 3 runs produced the same OperationTimeoutException, each failing in approx 5min's
```
com.hazelcast.core.OperationTimeoutException: No response for 120000 ms. Aborting invocation! BasicInvocationFuture{invocation=BasicInvocati
on{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=52}, partitionId=203, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Addr
ess[10.144.243.221]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false} No response has been received! backups-expected:0 backups-completed: 0
```
The Stabilizer Test
```
txnConflict@class=com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest
txnConflict@threadCount=10
txnConflict@keyCount=50
txnConflict@maxKeysPerTxn=5
txnConflict@throwCommitException=false
txnConflict@throwRollBackException=false
txnConflict@basename=txnConflict
```
The enviroment
```
3 ec2 instances
3 Member node cluster
1 ec2 instance hosting 1 member JVM
No clients used
```
The Fail
```
Failure[
message='Worked ran into an unhandled exception'
type='Worker exception'
agentAddress=10.237.211.31
time=Wed Dec 24 14:45:07 UTC 2014
workerAddress=10.237.211.31:5701
workerId=worker-10.237.211.31-1-server
test=TestCase{
id=txnConflict
, class=com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest
, basename=txnConflict
, keyCount=50
, maxKeysPerTxn=5
, threadCount=10
, throwCommitException=false
, throwRollBackException=false
}
cause=com.hazelcast.core.OperationTimeoutException: No response for 120000 ms. Aborting invocation! BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetO
peration{timeout=120000, thread=51}, partitionId=5, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.237.211.31]:5701, backupsExpected=0, backupsComple
ted=0}, response=null, done=false} No response has been received! backups-expected:0 backups-completed: 0
at com.hazelcast.spi.impl.BasicInvocationFuture.newOperationTimeoutException(BasicInvocationFuture.java:309)
at com.hazelcast.spi.impl.BasicInvocationFuture.waitForResponse(BasicInvocationFuture.java:246)
at com.hazelcast.spi.impl.BasicInvocationFuture.get(BasicInvocationFuture.java:193)
at com.hazelcast.spi.impl.BasicInvocationFuture.get(BasicInvocationFuture.java:173)
at com.hazelcast.map.impl.tx.TransactionalMapProxySupport.lockAndGet(TransactionalMapProxySupport.java:232)
at com.hazelcast.map.impl.tx.TransactionalMapProxySupport.getForUpdateInternal(TransactionalMapProxySupport.java:126)
at com.hazelcast.map.impl.tx.TransactionalMapProxy.getForUpdate(TransactionalMapProxy.java:99)
at com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest$Worker.run(MapTransactionContextConflictTest.java:94)
at java.lang.Thread.run(Thread.java:745)
at com.hazelcast.stabilizer.tests.utils.ThreadSpawner$DefaultThread.run(ThreadSpawner.java:88)
]
```
All unique Fatal msg over 3 members, form a failing run
```
1 com.hazelcast.map.impl.tx.TxnPrepareOperation: [10.144.243.221]:5701 [workers] [3.4-RC1-SNAPSHOT] Locked: [false], key: [HeapData{type=-7, hashCode=1159565325, partitionHash=1159565325, dataSize=
4, heapCost=32}]
1 com.hazelcast.map.impl.tx.TxnPrepareOperation: [10.144.243.221]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! [<not-locked>]
1 com.hazelcast.map.impl.tx.TxnPrepareOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Locked: [false], key: [HeapData{type=-7, hashCode=-1608856668, partitionHash=-1608856668, dataSize
=4, heapCost=32}]
1 com.hazelcast.map.impl.tx.TxnPrepareOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Locked: [false], key: [HeapData{type=-7, hashCode=2076340387, partitionHash=2076340387, dataSize=4
, heapCost=32}]
1 com.hazelcast.map.impl.tx.TxnPrepareOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! [<not-locked>]
1 com.hazelcast.map.impl.tx.TxnPrepareOperation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Locked: [false], key: [HeapData{type=-7, hashCode=-1921243156, partitionHash=-1921243156, dataSize
=4, heapCost=32}]
1 com.hazelcast.map.impl.tx.TxnPrepareOperation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Locked: [false], key: [HeapData{type=-7, hashCode=1932992084, partitionHash=1932992084, dataSize=4
, heapCost=32}]
1 com.hazelcast.map.impl.tx.TxnPrepareOperation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Locked: [true], key: [HeapData{type=-7, hashCode=81381366, partitionHash=81381366, dataSize=4, hea
pCost=32}]
1 com.hazelcast.map.impl.tx.TxnPrepareOperation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! [<not-locked>]
1 com.hazelcast.map.impl.tx.TxnPrepareOperation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! [Owner: 9220a8b9-f0a7-438e-836a-668e2a7ccbc6, thread-id: 52]
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.144.243.221]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 2cd3f888-f82e-44ce-8e14-8aa9d8f02e8f, thread
-id: 49
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.144.243.221]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 852e4eb9-7131-4d57-9df6-2fd4930c0085, thread
-id: 49
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.144.243.221]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 852e4eb9-7131-4d57-9df6-2fd4930c0085, thread
-id: 52
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.144.243.221]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 9220a8b9-f0a7-438e-836a-668e2a7ccbc6, thread
-id: 51
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.144.243.221]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 9220a8b9-f0a7-438e-836a-668e2a7ccbc6, thread
-id: 52
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.144.243.221]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 9220a8b9-f0a7-438e-836a-668e2a7ccbc6, thread
-id: 57
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 2cd3f888-f82e-44ce-8e14-8aa9d8f02e8f, thread-
id: 48
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 2cd3f888-f82e-44ce-8e14-8aa9d8f02e8f, thread-
id: 49
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 2cd3f888-f82e-44ce-8e14-8aa9d8f02e8f, thread-
id: 54
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 2cd3f888-f82e-44ce-8e14-8aa9d8f02e8f, thread-
id: 56
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 852e4eb9-7131-4d57-9df6-2fd4930c0085, thread-
id: 47
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 852e4eb9-7131-4d57-9df6-2fd4930c0085, thread-
id: 48
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 852e4eb9-7131-4d57-9df6-2fd4930c0085, thread-
id: 50
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 852e4eb9-7131-4d57-9df6-2fd4930c0085, thread-
id: 53
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 852e4eb9-7131-4d57-9df6-2fd4930c0085, thread-
id: 55
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 852e4eb9-7131-4d57-9df6-2fd4930c0085, thread-
id: 56
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 9220a8b9-f0a7-438e-836a-668e2a7ccbc6, thread-
id: 56
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 2cd3f888-f82e-44ce-8e14-8aa9d8f02e8f, thread-
id: 49
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 852e4eb9-7131-4d57-9df6-2fd4930c0085, thread-
id: 49
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 9220a8b9-f0a7-438e-836a-668e2a7ccbc6, thread-
id: 53
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 9220a8b9-f0a7-438e-836a-668e2a7ccbc6, thread-
id: 56
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 9220a8b9-f0a7-438e-836a-668e2a7ccbc6, thread-
id: 58
```
All unique Warn msg over 3 members, form a failing run
```
1 com.hazelcast.instance.Node: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Terminating forcefully...
1 com.hazelcast.nio.tcp.ReadHandler: [10.144.243.221]:5701 [workers] [3.4-RC1-SNAPSHOT] hz._hzInstance_1_workers.IO.thread-in-0 Closing socket to endpoint Address[10.165.75.176]:5701, Cause:java.io.EOFException: Remote socket closed!
1 com.hazelcast.nio.tcp.ReadHandler: [10.144.243.221]:5701 [workers] [3.4-RC1-SNAPSHOT] hz._hzInstance_1_workers.IO.thread-in-1 Closing socket to endpoint Address[10.165.75.176]:5701, Cause:java.io.EOFException: Remote socket closed!
1 com.hazelcast.nio.tcp.ReadHandler: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] hz._hzInstance_1_workers.IO.thread-in-0 Closing socket to endpoint Address[10.165.75.176]:5701, Cause:java.io.EOFException: Remote socket closed!
1 com.hazelcast.nio.tcp.ReadHandler: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] hz._hzInstance_1_workers.IO.thread-in-1 Closing socket to endpoint Address[10.165.75.176]:5701, Cause:java.io.EOFException: Remote socket closed!
1 com.hazelcast.spi.impl.BasicInvocation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Asking if operation execution has been started: BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=51}, partitionId=169, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.165.75.176]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
1 com.hazelcast.spi.impl.BasicInvocation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Asking if operation execution has been started: BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=54}, partitionId=169, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.165.75.176]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
1 com.hazelcast.spi.impl.BasicInvocation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] 'is-executing': false -> BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=51}, partitionId=169, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.165.75.176]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
1 com.hazelcast.spi.impl.BasicInvocation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] 'is-executing': false -> BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=54}, partitionId=169, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.165.75.176]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
1 com.hazelcast.spi.impl.BasicInvocation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] No response for 120000 ms. BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=51}, partitionId=169, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.165.75.176]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
1 com.hazelcast.spi.impl.BasicInvocation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] No response for 120001 ms. BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=54}, partitionId=169, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.165.75.176]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
1 com.hazelcast.spi.impl.BasicInvocation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Asking if operation execution has been started: BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=51}, partitionId=5, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.237.211.31]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
1 com.hazelcast.spi.impl.BasicInvocation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Asking if operation execution has been started: BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=53}, partitionId=42, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.237.211.31]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
1 com.hazelcast.spi.impl.BasicInvocation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] 'is-executing': false -> BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=51}, partitionId=5, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.237.211.31]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
1 com.hazelcast.spi.impl.BasicInvocation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] 'is-executing': false -> BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=53}, partitionId=42, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.237.211.31]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
1 com.hazelcast.spi.impl.BasicInvocation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] No response for 120000 ms. BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=51}, partitionId=5, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.237.211.31]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
1 com.hazelcast.spi.impl.BasicInvocation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] No response for 120000 ms. BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=53}, partitionId=42, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.237.211.31]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
3 com.hazelcast.stabilizer.tests.utils.ExceptionReporter: Exception #1 detected
2 com.hazelcast.stabilizer.tests.utils.ExceptionReporter: Exception #2 detected
2 com.hazelcast.stabilizer.tests.utils.ExceptionReporter: Exception #3 detected
1 com.hazelcast.transaction.impl.TransactionImpl: [10.144.243.221]:5701 [workers] [3.4-RC1-SNAPSHOT] Error during rollback!
1 com.hazelcast.transaction.impl.TransactionImpl: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Error during rollback!
1 com.hazelcast.transaction.impl.TransactionImpl: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Error during rollback!
```
Output logged by the MapTransactionContextConflictTest with id txnConflict
```
grepall worker.log "txnConflict:"
./worker-10.237.211.31-1-server/worker.log
WARN 2014-12-24 14:45:06,457 [Thread-1] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[] Transaction couldn't obtain lock for the key:39
WARN 2014-12-24 14:47:06,126 [Thread-9] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=39, inc=442}, {key=12, inc=305}, {key=7, inc=327}, {key=47, inc=209}, {key=5, inc=276}] Transaction is timed-out!
WARN 2014-12-24 14:47:06,158 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=13, inc=596}, {key=13, inc=397}, {key=12, inc=59}, {key=5, inc=534}, {key=16, inc=709}] Transaction is timed-out!
WARN 2014-12-24 14:47:06,165 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=4, inc=714}, {key=9, inc=607}, {key=38, inc=989}, {key=19, inc=253}, {key=20, inc=624}] Transaction is timed-out!
WARN 2014-12-24 14:47:06,166 [Thread-10] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=28, inc=129}, {key=7, inc=425}, {key=7, inc=63}, {key=40, inc=707}, {key=48, inc=218}] Transaction is timed-out!
WARN 2014-12-24 14:47:06,235 [Thread-3] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=2, inc=395}, {key=44, inc=712}, {key=46, inc=289}, {key=15, inc=124}, {key=12, inc=930}] Transaction is timed-out!
WARN 2014-12-24 14:47:06,248 [Thread-2] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=11, inc=958}, {key=22, inc=312}, {key=26, inc=156}, {key=48, inc=300}, {key=22, inc=751}] Transaction is timed-out!
WARN 2014-12-24 14:47:06,481 [Thread-4] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=21, inc=434}, {key=26, inc=833}] Transaction couldn't obtain lock for the key:11
WARN 2014-12-24 14:47:07,448 [Thread-1] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=37, inc=854}] Transaction couldn't obtain lock for the key:28
INFO 2014-12-24 14:49:10,810 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: TxnCounter{committed=5, rolled=35, failedRoles=0} from 26 worker threads
INFO 2014-12-24 14:49:10,824 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=7 expected 106 != actual 1056
INFO 2014-12-24 14:49:10,828 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=12 expected 0 != actual 206
INFO 2014-12-24 14:49:10,829 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=13 expected 0 != actual 1168
INFO 2014-12-24 14:49:10,835 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=21 expected 0 != actual 776
INFO 2014-12-24 14:49:10,839 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=27 expected 544 != actual 1146
INFO 2014-12-24 14:49:10,841 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=30 expected 550 != actual 999
INFO 2014-12-24 14:49:10,849 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=42 expected 0 != actual 602
INFO 2014-12-24 14:49:10,850 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=43 expected 679 != actual 1093
java.lang.AssertionError: txnConflict: 8 key=>values have been incremented unExpected expected:<0> but was:<8>
./worker-10.165.75.176-1-server/worker.log
WARN 2014-12-24 14:45:06,130 [Thread-1] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=5, inc=305}, {key=15, inc=51}, {key=11, inc=582}, {key=32, inc=580}, {key=49, inc=91}] Transaction is timed-out!
WARN 2014-12-24 14:45:06,162 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=1, inc=43}, {key=15, inc=853}, {key=44, inc=388}, {key=23, inc=205}, {key=32, inc=519}] Transaction is timed-out!
WARN 2014-12-24 14:45:06,473 [Thread-10] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=11, inc=86}, {key=2, inc=512}] Transaction couldn't obtain lock for the key:1
WARN 2014-12-24 14:45:06,473 [Thread-9] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[] Transaction couldn't obtain lock for the key:47
WARN 2014-12-24 14:45:06,473 [Thread-3] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[] Transaction couldn't obtain lock for the key:47
WARN 2014-12-24 14:47:06,138 [Thread-2] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=47, inc=865}, {key=5, inc=959}, {key=13, inc=225}, {key=35, inc=378}, {key=21, inc=279}] Transaction is timed-out!
WARN 2014-12-24 14:47:06,142 [Thread-4] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=28, inc=479}, {key=9, inc=759}, {key=15, inc=680}, {key=44, inc=957}, {key=39, inc=94}] Transaction is timed-out!
WARN 2014-12-24 14:47:06,164 [Thread-9] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=20, inc=160}, {key=30, inc=383}, {key=25, inc=695}, {key=12, inc=418}, {key=2, inc=557}] Lock is not owned by the transaction! [<not-locked>]
WARN 2014-12-24 14:47:06,491 [Thread-10] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=11, inc=691}, {key=36, inc=64}, {key=43, inc=607}, {key=31, inc=834}, {key=24, inc=696}] Transaction is timed-out!
WARN 2014-12-24 14:47:06,505 [Thread-7] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=40, inc=1}, {key=46, inc=356}, {key=36, inc=712}, {key=31, inc=567}, {key=27, inc=379}] Transaction is timed-out!
WARN 2014-12-24 14:49:06,131 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=28, inc=722}, {key=41, inc=680}, {key=8, inc=283}, {key=13, inc=37}, {key=30, inc=515}] Transaction is timed-out!
WARN 2014-12-24 14:49:06,132 [Thread-3] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=38, inc=140}, {key=21, inc=641}, {key=20, inc=623}, {key=6, inc=637}, {key=19, inc=408}] Transaction is timed-out!
WARN 2014-12-24 14:49:06,170 [Thread-1] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=44, inc=267}, {key=6, inc=868}, {key=23, inc=730}, {key=23, inc=765}, {key=41, inc=661}] Transaction is timed-out!
INFO 2014-12-24 14:49:10,677 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: TxnCounter{committed=5, rolled=35, failedRoles=0} from 26 worker threads
INFO 2014-12-24 14:49:10,691 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=7 expected 106 != actual 1056
INFO 2014-12-24 14:49:10,693 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=12 expected 0 != actual 206
INFO 2014-12-24 14:49:10,694 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=13 expected 0 != actual 1168
INFO 2014-12-24 14:49:10,700 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=21 expected 0 != actual 776
INFO 2014-12-24 14:49:10,702 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=27 expected 544 != actual 1146
INFO 2014-12-24 14:49:10,704 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=30 expected 550 != actual 999
INFO 2014-12-24 14:49:10,711 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=42 expected 0 != actual 602
INFO 2014-12-24 14:49:10,711 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=43 expected 679 != actual 1093
java.lang.AssertionError: txnConflict: 8 key=>values have been incremented unExpected expected:<0> but was:<8>
./worker-10.144.243.221-1-server/worker.log
WARN 2014-12-24 14:45:06,307 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=17, inc=748}, {key=18, inc=297}, {key=42, inc=767}, {key=4, inc=594}, {key=35, inc=335}] Lock is not owned by the transaction! [<not-locked>]
WARN 2014-12-24 14:45:06,445 [Thread-9] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[] Transaction couldn't obtain lock for the key:5
WARN 2014-12-24 14:45:06,461 [Thread-2] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=10, inc=893}, {key=44, inc=323}] Transaction couldn't obtain lock for the key:45
WARN 2014-12-24 14:45:06,461 [Thread-4] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[] Transaction couldn't obtain lock for the key:47
WARN 2014-12-24 14:47:06,225 [Thread-5] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=33, inc=83}, {key=1, inc=181}, {key=22, inc=103}, {key=5, inc=539}, {key=22, inc=364}] Transaction is timed-out!
WARN 2014-12-24 14:47:06,259 [Thread-10] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=3, inc=655}, {key=1, inc=838}, {key=30, inc=541}, {key=10, inc=591}, {key=40, inc=258}] Lock is not owned by the transaction! [<not-locked>]
WARN 2014-12-24 14:47:06,262 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=39, inc=333}, {key=25, inc=659}, {key=8, inc=92}, {key=7, inc=758}, {key=48, inc=606}] Transaction is timed-out!
WARN 2014-12-24 14:47:06,468 [Thread-1] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=7, inc=78}, {key=0, inc=81}] Transaction couldn't obtain lock for the key:46
WARN 2014-12-24 14:47:06,469 [Thread-7] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=21, inc=955}, {key=38, inc=201}, {key=29, inc=64}] Transaction couldn't obtain lock for the key:44
WARN 2014-12-24 14:47:06,489 [Thread-9] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=35, inc=32}, {key=42, inc=32}, {key=12, inc=869}, {key=11, inc=19}, {key=15, inc=401}] Transaction is timed-out!
WARN 2014-12-24 14:49:06,114 [Thread-3] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=45, inc=434}, {key=13, inc=236}, {key=19, inc=501}, {key=44, inc=501}, {key=31, inc=198}] Transaction is timed-out!
WARN 2014-12-24 14:49:06,129 [Thread-2] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=31, inc=272}, {key=5, inc=90}, {key=43, inc=608}, {key=21, inc=810}, {key=40, inc=140}] Transaction is timed-out!
WARN 2014-12-24 14:49:06,157 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=8, inc=370}, {key=38, inc=141}, {key=1, inc=887}, {key=48, inc=436}, {key=48, inc=816}] Transaction is timed-out!
INFO 2014-12-24 14:49:10,795 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: TxnCounter{committed=5, rolled=35, failedRoles=0} from 26 worker threads
INFO 2014-12-24 14:49:10,810 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=7 expected 106 != actual 1056
INFO 2014-12-24 14:49:10,815 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=12 expected 0 != actual 206
INFO 2014-12-24 14:49:10,815 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=13 expected 0 != actual 1168
INFO 2014-12-24 14:49:10,820 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=21 expected 0 != actual 776
INFO 2014-12-24 14:49:10,826 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=27 expected 544 != actual 1146
INFO 2014-12-24 14:49:10,829 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=30 expected 550 != actual 999
INFO 2014-12-24 14:49:10,834 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=42 expected 0 != actual 602
INFO 2014-12-24 14:49:10,836 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=43 expected 679 != actual 1093
java.lang.AssertionError: txnConflict: 8 key=>values have been incremented unExpected expected:<0> but was:<8>
```
MapTransactionContextConflictTest, is catching TransactionException's and it is fulley expected that some transaction will time out, the test logs these occasions, we can see that the test catch an TransactionException with the message
```
Transaction is timed-out!
```
also we can see a TransactionException which is handled by the test with the message
```
Transaction couldn't obtain lock for the key:44
```
why should an
```
com.hazelcast.core.OperationTimeoutException: No response for 120000 ms. Aborting invocation! BasicInvocationFuture{invocation=BasicInvocati
on{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=52}, partitionId=203, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Addr
ess[10.144.243.221]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false} No response has been received! backups-expected:0 backups-completed: 0
```
be thrown, when we are already catching TransactionException: Transaction is timed-out!
| 1.0 | Hz 3.4 OperationTimeoutException, op=TxnLockAndGetOperation, MapTransactionContextConflictTest - Tested on Hazelcast 3.4
MapTransactionContextConflictTest reliably causes an OperationTimeoutException.
3 out of 3 runs produced the same OperationTimeoutException, each failing in approx 5min's
```
com.hazelcast.core.OperationTimeoutException: No response for 120000 ms. Aborting invocation! BasicInvocationFuture{invocation=BasicInvocati
on{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=52}, partitionId=203, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Addr
ess[10.144.243.221]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false} No response has been received! backups-expected:0 backups-completed: 0
```
The Stabilizer Test
```
txnConflict@class=com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest
txnConflict@threadCount=10
txnConflict@keyCount=50
txnConflict@maxKeysPerTxn=5
txnConflict@throwCommitException=false
txnConflict@throwRollBackException=false
txnConflict@basename=txnConflict
```
The enviroment
```
3 ec2 instances
3 Member node cluster
1 ec2 instance hosting 1 member JVM
No clients used
```
The Fail
```
Failure[
message='Worked ran into an unhandled exception'
type='Worker exception'
agentAddress=10.237.211.31
time=Wed Dec 24 14:45:07 UTC 2014
workerAddress=10.237.211.31:5701
workerId=worker-10.237.211.31-1-server
test=TestCase{
id=txnConflict
, class=com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest
, basename=txnConflict
, keyCount=50
, maxKeysPerTxn=5
, threadCount=10
, throwCommitException=false
, throwRollBackException=false
}
cause=com.hazelcast.core.OperationTimeoutException: No response for 120000 ms. Aborting invocation! BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetO
peration{timeout=120000, thread=51}, partitionId=5, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.237.211.31]:5701, backupsExpected=0, backupsComple
ted=0}, response=null, done=false} No response has been received! backups-expected:0 backups-completed: 0
at com.hazelcast.spi.impl.BasicInvocationFuture.newOperationTimeoutException(BasicInvocationFuture.java:309)
at com.hazelcast.spi.impl.BasicInvocationFuture.waitForResponse(BasicInvocationFuture.java:246)
at com.hazelcast.spi.impl.BasicInvocationFuture.get(BasicInvocationFuture.java:193)
at com.hazelcast.spi.impl.BasicInvocationFuture.get(BasicInvocationFuture.java:173)
at com.hazelcast.map.impl.tx.TransactionalMapProxySupport.lockAndGet(TransactionalMapProxySupport.java:232)
at com.hazelcast.map.impl.tx.TransactionalMapProxySupport.getForUpdateInternal(TransactionalMapProxySupport.java:126)
at com.hazelcast.map.impl.tx.TransactionalMapProxy.getForUpdate(TransactionalMapProxy.java:99)
at com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest$Worker.run(MapTransactionContextConflictTest.java:94)
at java.lang.Thread.run(Thread.java:745)
at com.hazelcast.stabilizer.tests.utils.ThreadSpawner$DefaultThread.run(ThreadSpawner.java:88)
]
```
All unique Fatal msg over 3 members, form a failing run
```
1 com.hazelcast.map.impl.tx.TxnPrepareOperation: [10.144.243.221]:5701 [workers] [3.4-RC1-SNAPSHOT] Locked: [false], key: [HeapData{type=-7, hashCode=1159565325, partitionHash=1159565325, dataSize=
4, heapCost=32}]
1 com.hazelcast.map.impl.tx.TxnPrepareOperation: [10.144.243.221]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! [<not-locked>]
1 com.hazelcast.map.impl.tx.TxnPrepareOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Locked: [false], key: [HeapData{type=-7, hashCode=-1608856668, partitionHash=-1608856668, dataSize
=4, heapCost=32}]
1 com.hazelcast.map.impl.tx.TxnPrepareOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Locked: [false], key: [HeapData{type=-7, hashCode=2076340387, partitionHash=2076340387, dataSize=4
, heapCost=32}]
1 com.hazelcast.map.impl.tx.TxnPrepareOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! [<not-locked>]
1 com.hazelcast.map.impl.tx.TxnPrepareOperation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Locked: [false], key: [HeapData{type=-7, hashCode=-1921243156, partitionHash=-1921243156, dataSize
=4, heapCost=32}]
1 com.hazelcast.map.impl.tx.TxnPrepareOperation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Locked: [false], key: [HeapData{type=-7, hashCode=1932992084, partitionHash=1932992084, dataSize=4
, heapCost=32}]
1 com.hazelcast.map.impl.tx.TxnPrepareOperation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Locked: [true], key: [HeapData{type=-7, hashCode=81381366, partitionHash=81381366, dataSize=4, hea
pCost=32}]
1 com.hazelcast.map.impl.tx.TxnPrepareOperation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! [<not-locked>]
1 com.hazelcast.map.impl.tx.TxnPrepareOperation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! [Owner: 9220a8b9-f0a7-438e-836a-668e2a7ccbc6, thread-id: 52]
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.144.243.221]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 2cd3f888-f82e-44ce-8e14-8aa9d8f02e8f, thread
-id: 49
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.144.243.221]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 852e4eb9-7131-4d57-9df6-2fd4930c0085, thread
-id: 49
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.144.243.221]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 852e4eb9-7131-4d57-9df6-2fd4930c0085, thread
-id: 52
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.144.243.221]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 9220a8b9-f0a7-438e-836a-668e2a7ccbc6, thread
-id: 51
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.144.243.221]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 9220a8b9-f0a7-438e-836a-668e2a7ccbc6, thread
-id: 52
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.144.243.221]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 9220a8b9-f0a7-438e-836a-668e2a7ccbc6, thread
-id: 57
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 2cd3f888-f82e-44ce-8e14-8aa9d8f02e8f, thread-
id: 48
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 2cd3f888-f82e-44ce-8e14-8aa9d8f02e8f, thread-
id: 49
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 2cd3f888-f82e-44ce-8e14-8aa9d8f02e8f, thread-
id: 54
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 2cd3f888-f82e-44ce-8e14-8aa9d8f02e8f, thread-
id: 56
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 852e4eb9-7131-4d57-9df6-2fd4930c0085, thread-
id: 47
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 852e4eb9-7131-4d57-9df6-2fd4930c0085, thread-
id: 48
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 852e4eb9-7131-4d57-9df6-2fd4930c0085, thread-
id: 50
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 852e4eb9-7131-4d57-9df6-2fd4930c0085, thread-
id: 53
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 852e4eb9-7131-4d57-9df6-2fd4930c0085, thread-
id: 55
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 852e4eb9-7131-4d57-9df6-2fd4930c0085, thread-
id: 56
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 9220a8b9-f0a7-438e-836a-668e2a7ccbc6, thread-
id: 56
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 2cd3f888-f82e-44ce-8e14-8aa9d8f02e8f, thread-
id: 49
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 852e4eb9-7131-4d57-9df6-2fd4930c0085, thread-
id: 49
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 9220a8b9-f0a7-438e-836a-668e2a7ccbc6, thread-
id: 53
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 9220a8b9-f0a7-438e-836a-668e2a7ccbc6, thread-
id: 56
1 com.hazelcast.map.impl.tx.TxnRollbackOperation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Lock is not owned by the transaction! Owner: Owner: 9220a8b9-f0a7-438e-836a-668e2a7ccbc6, thread-
id: 58
```
All unique Warn msg over 3 members, form a failing run
```
1 com.hazelcast.instance.Node: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Terminating forcefully...
1 com.hazelcast.nio.tcp.ReadHandler: [10.144.243.221]:5701 [workers] [3.4-RC1-SNAPSHOT] hz._hzInstance_1_workers.IO.thread-in-0 Closing socket to endpoint Address[10.165.75.176]:5701, Cause:java.io.EOFException: Remote socket closed!
1 com.hazelcast.nio.tcp.ReadHandler: [10.144.243.221]:5701 [workers] [3.4-RC1-SNAPSHOT] hz._hzInstance_1_workers.IO.thread-in-1 Closing socket to endpoint Address[10.165.75.176]:5701, Cause:java.io.EOFException: Remote socket closed!
1 com.hazelcast.nio.tcp.ReadHandler: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] hz._hzInstance_1_workers.IO.thread-in-0 Closing socket to endpoint Address[10.165.75.176]:5701, Cause:java.io.EOFException: Remote socket closed!
1 com.hazelcast.nio.tcp.ReadHandler: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] hz._hzInstance_1_workers.IO.thread-in-1 Closing socket to endpoint Address[10.165.75.176]:5701, Cause:java.io.EOFException: Remote socket closed!
1 com.hazelcast.spi.impl.BasicInvocation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Asking if operation execution has been started: BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=51}, partitionId=169, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.165.75.176]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
1 com.hazelcast.spi.impl.BasicInvocation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Asking if operation execution has been started: BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=54}, partitionId=169, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.165.75.176]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
1 com.hazelcast.spi.impl.BasicInvocation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] 'is-executing': false -> BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=51}, partitionId=169, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.165.75.176]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
1 com.hazelcast.spi.impl.BasicInvocation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] 'is-executing': false -> BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=54}, partitionId=169, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.165.75.176]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
1 com.hazelcast.spi.impl.BasicInvocation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] No response for 120000 ms. BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=51}, partitionId=169, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.165.75.176]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
1 com.hazelcast.spi.impl.BasicInvocation: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] No response for 120001 ms. BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=54}, partitionId=169, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.165.75.176]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
1 com.hazelcast.spi.impl.BasicInvocation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Asking if operation execution has been started: BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=51}, partitionId=5, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.237.211.31]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
1 com.hazelcast.spi.impl.BasicInvocation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Asking if operation execution has been started: BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=53}, partitionId=42, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.237.211.31]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
1 com.hazelcast.spi.impl.BasicInvocation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] 'is-executing': false -> BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=51}, partitionId=5, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.237.211.31]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
1 com.hazelcast.spi.impl.BasicInvocation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] 'is-executing': false -> BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=53}, partitionId=42, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.237.211.31]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
1 com.hazelcast.spi.impl.BasicInvocation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] No response for 120000 ms. BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=51}, partitionId=5, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.237.211.31]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
1 com.hazelcast.spi.impl.BasicInvocation: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] No response for 120000 ms. BasicInvocationFuture{invocation=BasicInvocation{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=53}, partitionId=42, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Address[10.237.211.31]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false}
3 com.hazelcast.stabilizer.tests.utils.ExceptionReporter: Exception #1 detected
2 com.hazelcast.stabilizer.tests.utils.ExceptionReporter: Exception #2 detected
2 com.hazelcast.stabilizer.tests.utils.ExceptionReporter: Exception #3 detected
1 com.hazelcast.transaction.impl.TransactionImpl: [10.144.243.221]:5701 [workers] [3.4-RC1-SNAPSHOT] Error during rollback!
1 com.hazelcast.transaction.impl.TransactionImpl: [10.165.75.176]:5701 [workers] [3.4-RC1-SNAPSHOT] Error during rollback!
1 com.hazelcast.transaction.impl.TransactionImpl: [10.237.211.31]:5701 [workers] [3.4-RC1-SNAPSHOT] Error during rollback!
```
Output logged by the MapTransactionContextConflictTest with id txnConflict
```
grepall worker.log "txnConflict:"
./worker-10.237.211.31-1-server/worker.log
WARN 2014-12-24 14:45:06,457 [Thread-1] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[] Transaction couldn't obtain lock for the key:39
WARN 2014-12-24 14:47:06,126 [Thread-9] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=39, inc=442}, {key=12, inc=305}, {key=7, inc=327}, {key=47, inc=209}, {key=5, inc=276}] Transaction is timed-out!
WARN 2014-12-24 14:47:06,158 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=13, inc=596}, {key=13, inc=397}, {key=12, inc=59}, {key=5, inc=534}, {key=16, inc=709}] Transaction is timed-out!
WARN 2014-12-24 14:47:06,165 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=4, inc=714}, {key=9, inc=607}, {key=38, inc=989}, {key=19, inc=253}, {key=20, inc=624}] Transaction is timed-out!
WARN 2014-12-24 14:47:06,166 [Thread-10] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=28, inc=129}, {key=7, inc=425}, {key=7, inc=63}, {key=40, inc=707}, {key=48, inc=218}] Transaction is timed-out!
WARN 2014-12-24 14:47:06,235 [Thread-3] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=2, inc=395}, {key=44, inc=712}, {key=46, inc=289}, {key=15, inc=124}, {key=12, inc=930}] Transaction is timed-out!
WARN 2014-12-24 14:47:06,248 [Thread-2] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=11, inc=958}, {key=22, inc=312}, {key=26, inc=156}, {key=48, inc=300}, {key=22, inc=751}] Transaction is timed-out!
WARN 2014-12-24 14:47:06,481 [Thread-4] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=21, inc=434}, {key=26, inc=833}] Transaction couldn't obtain lock for the key:11
WARN 2014-12-24 14:47:07,448 [Thread-1] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=37, inc=854}] Transaction couldn't obtain lock for the key:28
INFO 2014-12-24 14:49:10,810 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: TxnCounter{committed=5, rolled=35, failedRoles=0} from 26 worker threads
INFO 2014-12-24 14:49:10,824 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=7 expected 106 != actual 1056
INFO 2014-12-24 14:49:10,828 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=12 expected 0 != actual 206
INFO 2014-12-24 14:49:10,829 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=13 expected 0 != actual 1168
INFO 2014-12-24 14:49:10,835 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=21 expected 0 != actual 776
INFO 2014-12-24 14:49:10,839 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=27 expected 544 != actual 1146
INFO 2014-12-24 14:49:10,841 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=30 expected 550 != actual 999
INFO 2014-12-24 14:49:10,849 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=42 expected 0 != actual 602
INFO 2014-12-24 14:49:10,850 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=43 expected 679 != actual 1093
java.lang.AssertionError: txnConflict: 8 key=>values have been incremented unExpected expected:<0> but was:<8>
./worker-10.165.75.176-1-server/worker.log
WARN 2014-12-24 14:45:06,130 [Thread-1] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=5, inc=305}, {key=15, inc=51}, {key=11, inc=582}, {key=32, inc=580}, {key=49, inc=91}] Transaction is timed-out!
WARN 2014-12-24 14:45:06,162 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=1, inc=43}, {key=15, inc=853}, {key=44, inc=388}, {key=23, inc=205}, {key=32, inc=519}] Transaction is timed-out!
WARN 2014-12-24 14:45:06,473 [Thread-10] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=11, inc=86}, {key=2, inc=512}] Transaction couldn't obtain lock for the key:1
WARN 2014-12-24 14:45:06,473 [Thread-9] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[] Transaction couldn't obtain lock for the key:47
WARN 2014-12-24 14:45:06,473 [Thread-3] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[] Transaction couldn't obtain lock for the key:47
WARN 2014-12-24 14:47:06,138 [Thread-2] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=47, inc=865}, {key=5, inc=959}, {key=13, inc=225}, {key=35, inc=378}, {key=21, inc=279}] Transaction is timed-out!
WARN 2014-12-24 14:47:06,142 [Thread-4] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=28, inc=479}, {key=9, inc=759}, {key=15, inc=680}, {key=44, inc=957}, {key=39, inc=94}] Transaction is timed-out!
WARN 2014-12-24 14:47:06,164 [Thread-9] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=20, inc=160}, {key=30, inc=383}, {key=25, inc=695}, {key=12, inc=418}, {key=2, inc=557}] Lock is not owned by the transaction! [<not-locked>]
WARN 2014-12-24 14:47:06,491 [Thread-10] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=11, inc=691}, {key=36, inc=64}, {key=43, inc=607}, {key=31, inc=834}, {key=24, inc=696}] Transaction is timed-out!
WARN 2014-12-24 14:47:06,505 [Thread-7] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=40, inc=1}, {key=46, inc=356}, {key=36, inc=712}, {key=31, inc=567}, {key=27, inc=379}] Transaction is timed-out!
WARN 2014-12-24 14:49:06,131 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=28, inc=722}, {key=41, inc=680}, {key=8, inc=283}, {key=13, inc=37}, {key=30, inc=515}] Transaction is timed-out!
WARN 2014-12-24 14:49:06,132 [Thread-3] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=38, inc=140}, {key=21, inc=641}, {key=20, inc=623}, {key=6, inc=637}, {key=19, inc=408}] Transaction is timed-out!
WARN 2014-12-24 14:49:06,170 [Thread-1] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=44, inc=267}, {key=6, inc=868}, {key=23, inc=730}, {key=23, inc=765}, {key=41, inc=661}] Transaction is timed-out!
INFO 2014-12-24 14:49:10,677 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: TxnCounter{committed=5, rolled=35, failedRoles=0} from 26 worker threads
INFO 2014-12-24 14:49:10,691 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=7 expected 106 != actual 1056
INFO 2014-12-24 14:49:10,693 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=12 expected 0 != actual 206
INFO 2014-12-24 14:49:10,694 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=13 expected 0 != actual 1168
INFO 2014-12-24 14:49:10,700 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=21 expected 0 != actual 776
INFO 2014-12-24 14:49:10,702 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=27 expected 544 != actual 1146
INFO 2014-12-24 14:49:10,704 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=30 expected 550 != actual 999
INFO 2014-12-24 14:49:10,711 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=42 expected 0 != actual 602
INFO 2014-12-24 14:49:10,711 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=43 expected 679 != actual 1093
java.lang.AssertionError: txnConflict: 8 key=>values have been incremented unExpected expected:<0> but was:<8>
./worker-10.144.243.221-1-server/worker.log
WARN 2014-12-24 14:45:06,307 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=17, inc=748}, {key=18, inc=297}, {key=42, inc=767}, {key=4, inc=594}, {key=35, inc=335}] Lock is not owned by the transaction! [<not-locked>]
WARN 2014-12-24 14:45:06,445 [Thread-9] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[] Transaction couldn't obtain lock for the key:5
WARN 2014-12-24 14:45:06,461 [Thread-2] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=10, inc=893}, {key=44, inc=323}] Transaction couldn't obtain lock for the key:45
WARN 2014-12-24 14:45:06,461 [Thread-4] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[] Transaction couldn't obtain lock for the key:47
WARN 2014-12-24 14:47:06,225 [Thread-5] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=33, inc=83}, {key=1, inc=181}, {key=22, inc=103}, {key=5, inc=539}, {key=22, inc=364}] Transaction is timed-out!
WARN 2014-12-24 14:47:06,259 [Thread-10] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=3, inc=655}, {key=1, inc=838}, {key=30, inc=541}, {key=10, inc=591}, {key=40, inc=258}] Lock is not owned by the transaction! [<not-locked>]
WARN 2014-12-24 14:47:06,262 [Thread-6] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=39, inc=333}, {key=25, inc=659}, {key=8, inc=92}, {key=7, inc=758}, {key=48, inc=606}] Transaction is timed-out!
WARN 2014-12-24 14:47:06,468 [Thread-1] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=7, inc=78}, {key=0, inc=81}] Transaction couldn't obtain lock for the key:46
WARN 2014-12-24 14:47:06,469 [Thread-7] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=21, inc=955}, {key=38, inc=201}, {key=29, inc=64}] Transaction couldn't obtain lock for the key:44
WARN 2014-12-24 14:47:06,489 [Thread-9] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=35, inc=32}, {key=42, inc=32}, {key=12, inc=869}, {key=11, inc=19}, {key=15, inc=401}] Transaction is timed-out!
WARN 2014-12-24 14:49:06,114 [Thread-3] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=45, inc=434}, {key=13, inc=236}, {key=19, inc=501}, {key=44, inc=501}, {key=31, inc=198}] Transaction is timed-out!
WARN 2014-12-24 14:49:06,129 [Thread-2] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=31, inc=272}, {key=5, inc=90}, {key=43, inc=608}, {key=21, inc=810}, {key=40, inc=140}] Transaction is timed-out!
WARN 2014-12-24 14:49:06,157 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: commit fail. tried key increments=[{key=8, inc=370}, {key=38, inc=141}, {key=1, inc=887}, {key=48, inc=436}, {key=48, inc=816}] Transaction is timed-out!
INFO 2014-12-24 14:49:10,795 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: TxnCounter{committed=5, rolled=35, failedRoles=0} from 26 worker threads
INFO 2014-12-24 14:49:10,810 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=7 expected 106 != actual 1056
INFO 2014-12-24 14:49:10,815 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=12 expected 0 != actual 206
INFO 2014-12-24 14:49:10,815 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=13 expected 0 != actual 1168
INFO 2014-12-24 14:49:10,820 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=21 expected 0 != actual 776
INFO 2014-12-24 14:49:10,826 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=27 expected 544 != actual 1146
INFO 2014-12-24 14:49:10,829 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=30 expected 550 != actual 999
INFO 2014-12-24 14:49:10,834 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=42 expected 0 != actual 602
INFO 2014-12-24 14:49:10,836 [Thread-8] com.hazelcast.stabilizer.tests.map.MapTransactionContextConflictTest: txnConflict: key=43 expected 679 != actual 1093
java.lang.AssertionError: txnConflict: 8 key=>values have been incremented unExpected expected:<0> but was:<8>
```
MapTransactionContextConflictTest, is catching TransactionException's and it is fulley expected that some transaction will time out, the test logs these occasions, we can see that the test catch an TransactionException with the message
```
Transaction is timed-out!
```
also we can see a TransactionException which is handled by the test with the message
```
Transaction couldn't obtain lock for the key:44
```
why should an
```
com.hazelcast.core.OperationTimeoutException: No response for 120000 ms. Aborting invocation! BasicInvocationFuture{invocation=BasicInvocati
on{ serviceName='hz:impl:mapService', op=TxnLockAndGetOperation{timeout=120000, thread=52}, partitionId=203, replicaIndex=0, tryCount=250, tryPauseMillis=500, invokeCount=1, callTimeout=60000, target=Addr
ess[10.144.243.221]:5701, backupsExpected=0, backupsCompleted=0}, response=null, done=false} No response has been received! backups-expected:0 backups-completed: 0
```
be thrown, when we are already catching TransactionException: Transaction is timed-out!
| defect | hz operationtimeoutexception op txnlockandgetoperation maptransactioncontextconflicttest tested on hazelcast maptransactioncontextconflicttest reliably causes an operationtimeoutexception out of runs produced the same operationtimeoutexception each failing in approx s com hazelcast core operationtimeoutexception no response for ms aborting invocation basicinvocationfuture invocation basicinvocati on servicename hz impl mapservice op txnlockandgetoperation timeout thread partitionid replicaindex trycount trypausemillis invokecount calltimeout target addr ess backupsexpected backupscompleted response null done false no response has been received backups expected backups completed the stabilizer test txnconflict class com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict threadcount txnconflict keycount txnconflict maxkeyspertxn txnconflict throwcommitexception false txnconflict throwrollbackexception false txnconflict basename txnconflict the enviroment instances member node cluster instance hosting member jvm no clients used the fail failure message worked ran into an unhandled exception type worker exception agentaddress time wed dec utc workeraddress workerid worker server test testcase id txnconflict class com hazelcast stabilizer tests map maptransactioncontextconflicttest basename txnconflict keycount maxkeyspertxn threadcount throwcommitexception false throwrollbackexception false cause com hazelcast core operationtimeoutexception no response for ms aborting invocation basicinvocationfuture invocation basicinvocation servicename hz impl mapservice op txnlockandgeto peration timeout thread partitionid replicaindex trycount trypausemillis invokecount calltimeout target address backupsexpected backupscomple ted response null done false no response has been received backups expected backups completed at com hazelcast spi impl basicinvocationfuture newoperationtimeoutexception basicinvocationfuture java at com hazelcast spi impl basicinvocationfuture waitforresponse basicinvocationfuture java at com hazelcast spi impl basicinvocationfuture get basicinvocationfuture java at com hazelcast spi impl basicinvocationfuture get basicinvocationfuture java at com hazelcast map impl tx transactionalmapproxysupport lockandget transactionalmapproxysupport java at com hazelcast map impl tx transactionalmapproxysupport getforupdateinternal transactionalmapproxysupport java at com hazelcast map impl tx transactionalmapproxy getforupdate transactionalmapproxy java at com hazelcast stabilizer tests map maptransactioncontextconflicttest worker run maptransactioncontextconflicttest java at java lang thread run thread java at com hazelcast stabilizer tests utils threadspawner defaultthread run threadspawner java all unique fatal msg over members form a failing run com hazelcast map impl tx txnprepareoperation locked key heapdata type hashcode partitionhash datasize heapcost com hazelcast map impl tx txnprepareoperation lock is not owned by the transaction com hazelcast map impl tx txnprepareoperation locked key heapdata type hashcode partitionhash datasize heapcost com hazelcast map impl tx txnprepareoperation locked key heapdata type hashcode partitionhash datasize heapcost com hazelcast map impl tx txnprepareoperation lock is not owned by the transaction com hazelcast map impl tx txnprepareoperation locked key heapdata type hashcode partitionhash datasize heapcost com hazelcast map impl tx txnprepareoperation locked key heapdata type hashcode partitionhash datasize heapcost com hazelcast map impl tx txnprepareoperation locked key heapdata type hashcode partitionhash datasize hea pcost com hazelcast map impl tx txnprepareoperation lock is not owned by the transaction com hazelcast map impl tx txnprepareoperation lock is not owned by the transaction com hazelcast map impl tx txnrollbackoperation lock is not owned by the transaction owner owner thread id com hazelcast map impl tx txnrollbackoperation lock is not owned by the transaction owner owner thread id com hazelcast map impl tx txnrollbackoperation lock is not owned by the transaction owner owner thread id com hazelcast map impl tx txnrollbackoperation lock is not owned by the transaction owner owner thread id com hazelcast map impl tx txnrollbackoperation lock is not owned by the transaction owner owner thread id com hazelcast map impl tx txnrollbackoperation lock is not owned by the transaction owner owner thread id com hazelcast map impl tx txnrollbackoperation lock is not owned by the transaction owner owner thread id com hazelcast map impl tx txnrollbackoperation lock is not owned by the transaction owner owner thread id com hazelcast map impl tx txnrollbackoperation lock is not owned by the transaction owner owner thread id com hazelcast map impl tx txnrollbackoperation lock is not owned by the transaction owner owner thread id com hazelcast map impl tx txnrollbackoperation lock is not owned by the transaction owner owner thread id com hazelcast map impl tx txnrollbackoperation lock is not owned by the transaction owner owner thread id com hazelcast map impl tx txnrollbackoperation lock is not owned by the transaction owner owner thread id com hazelcast map impl tx txnrollbackoperation lock is not owned by the transaction owner owner thread id com hazelcast map impl tx txnrollbackoperation lock is not owned by the transaction owner owner thread id com hazelcast map impl tx txnrollbackoperation lock is not owned by the transaction owner owner thread id com hazelcast map impl tx txnrollbackoperation lock is not owned by the transaction owner owner thread id com hazelcast map impl tx txnrollbackoperation lock is not owned by the transaction owner owner thread id com hazelcast map impl tx txnrollbackoperation lock is not owned by the transaction owner owner thread id com hazelcast map impl tx txnrollbackoperation lock is not owned by the transaction owner owner thread id com hazelcast map impl tx txnrollbackoperation lock is not owned by the transaction owner owner thread id com hazelcast map impl tx txnrollbackoperation lock is not owned by the transaction owner owner thread id all unique warn msg over members form a failing run com hazelcast instance node terminating forcefully com hazelcast nio tcp readhandler hz hzinstance workers io thread in closing socket to endpoint address cause java io eofexception remote socket closed com hazelcast nio tcp readhandler hz hzinstance workers io thread in closing socket to endpoint address cause java io eofexception remote socket closed com hazelcast nio tcp readhandler hz hzinstance workers io thread in closing socket to endpoint address cause java io eofexception remote socket closed com hazelcast nio tcp readhandler hz hzinstance workers io thread in closing socket to endpoint address cause java io eofexception remote socket closed com hazelcast spi impl basicinvocation asking if operation execution has been started basicinvocationfuture invocation basicinvocation servicename hz impl mapservice op txnlockandgetoperation timeout thread partitionid replicaindex trycount trypausemillis invokecount calltimeout target address backupsexpected backupscompleted response null done false com hazelcast spi impl basicinvocation asking if operation execution has been started basicinvocationfuture invocation basicinvocation servicename hz impl mapservice op txnlockandgetoperation timeout thread partitionid replicaindex trycount trypausemillis invokecount calltimeout target address backupsexpected backupscompleted response null done false com hazelcast spi impl basicinvocation is executing false basicinvocationfuture invocation basicinvocation servicename hz impl mapservice op txnlockandgetoperation timeout thread partitionid replicaindex trycount trypausemillis invokecount calltimeout target address backupsexpected backupscompleted response null done false com hazelcast spi impl basicinvocation is executing false basicinvocationfuture invocation basicinvocation servicename hz impl mapservice op txnlockandgetoperation timeout thread partitionid replicaindex trycount trypausemillis invokecount calltimeout target address backupsexpected backupscompleted response null done false com hazelcast spi impl basicinvocation no response for ms basicinvocationfuture invocation basicinvocation servicename hz impl mapservice op txnlockandgetoperation timeout thread partitionid replicaindex trycount trypausemillis invokecount calltimeout target address backupsexpected backupscompleted response null done false com hazelcast spi impl basicinvocation no response for ms basicinvocationfuture invocation basicinvocation servicename hz impl mapservice op txnlockandgetoperation timeout thread partitionid replicaindex trycount trypausemillis invokecount calltimeout target address backupsexpected backupscompleted response null done false com hazelcast spi impl basicinvocation asking if operation execution has been started basicinvocationfuture invocation basicinvocation servicename hz impl mapservice op txnlockandgetoperation timeout thread partitionid replicaindex trycount trypausemillis invokecount calltimeout target address backupsexpected backupscompleted response null done false com hazelcast spi impl basicinvocation asking if operation execution has been started basicinvocationfuture invocation basicinvocation servicename hz impl mapservice op txnlockandgetoperation timeout thread partitionid replicaindex trycount trypausemillis invokecount calltimeout target address backupsexpected backupscompleted response null done false com hazelcast spi impl basicinvocation is executing false basicinvocationfuture invocation basicinvocation servicename hz impl mapservice op txnlockandgetoperation timeout thread partitionid replicaindex trycount trypausemillis invokecount calltimeout target address backupsexpected backupscompleted response null done false com hazelcast spi impl basicinvocation is executing false basicinvocationfuture invocation basicinvocation servicename hz impl mapservice op txnlockandgetoperation timeout thread partitionid replicaindex trycount trypausemillis invokecount calltimeout target address backupsexpected backupscompleted response null done false com hazelcast spi impl basicinvocation no response for ms basicinvocationfuture invocation basicinvocation servicename hz impl mapservice op txnlockandgetoperation timeout thread partitionid replicaindex trycount trypausemillis invokecount calltimeout target address backupsexpected backupscompleted response null done false com hazelcast spi impl basicinvocation no response for ms basicinvocationfuture invocation basicinvocation servicename hz impl mapservice op txnlockandgetoperation timeout thread partitionid replicaindex trycount trypausemillis invokecount calltimeout target address backupsexpected backupscompleted response null done false com hazelcast stabilizer tests utils exceptionreporter exception detected com hazelcast stabilizer tests utils exceptionreporter exception detected com hazelcast stabilizer tests utils exceptionreporter exception detected com hazelcast transaction impl transactionimpl error during rollback com hazelcast transaction impl transactionimpl error during rollback com hazelcast transaction impl transactionimpl error during rollback output logged by the maptransactioncontextconflicttest with id txnconflict grepall worker log txnconflict worker server worker log warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction couldn t obtain lock for the key warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction is timed out warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction is timed out warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction is timed out warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction is timed out warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction is timed out warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction is timed out warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction couldn t obtain lock for the key warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction couldn t obtain lock for the key info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict txncounter committed rolled failedroles from worker threads info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual java lang assertionerror txnconflict key values have been incremented unexpected expected but was worker server worker log warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction is timed out warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction is timed out warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction couldn t obtain lock for the key warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction couldn t obtain lock for the key warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction couldn t obtain lock for the key warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction is timed out warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction is timed out warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments lock is not owned by the transaction warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction is timed out warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction is timed out warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction is timed out warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction is timed out warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction is timed out info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict txncounter committed rolled failedroles from worker threads info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual java lang assertionerror txnconflict key values have been incremented unexpected expected but was worker server worker log warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments lock is not owned by the transaction warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction couldn t obtain lock for the key warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction couldn t obtain lock for the key warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction couldn t obtain lock for the key warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction is timed out warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments lock is not owned by the transaction warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction is timed out warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction couldn t obtain lock for the key warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction couldn t obtain lock for the key warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction is timed out warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction is timed out warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction is timed out warn com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict commit fail tried key increments transaction is timed out info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict txncounter committed rolled failedroles from worker threads info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual info com hazelcast stabilizer tests map maptransactioncontextconflicttest txnconflict key expected actual java lang assertionerror txnconflict key values have been incremented unexpected expected but was maptransactioncontextconflicttest is catching transactionexception s and it is fulley expected that some transaction will time out the test logs these occasions we can see that the test catch an transactionexception with the message transaction is timed out also we can see a transactionexception which is handled by the test with the message transaction couldn t obtain lock for the key why should an com hazelcast core operationtimeoutexception no response for ms aborting invocation basicinvocationfuture invocation basicinvocati on servicename hz impl mapservice op txnlockandgetoperation timeout thread partitionid replicaindex trycount trypausemillis invokecount calltimeout target addr ess backupsexpected backupscompleted response null done false no response has been received backups expected backups completed be thrown when we are already catching transactionexception transaction is timed out | 1 |
31,692 | 6,586,074,720 | IssuesEvent | 2017-09-13 15:57:34 | scholarslab/Neatline | https://api.github.com/repos/scholarslab/Neatline | closed | Image-layer-only exhibits | defect priority-high | Currently, users must select a spatial layer from a dropdown menu to serve as the base layer for a new exhibit. This means that an image layer can't be the first layer to load when the exhibit is viewed. Per #428, some users want to have exhibits where the image layer is the default view for the map — this may have been unintentionally possible prior to the last release due to an issue preventing the successful creation of Google Maps layers. Users should be able to choose to have no spatial layer as the base, or an option specifically to use the exhibit's image layer as its base in place of a spatial layer.
Related forum post: https://forum.omeka.org/t/neatline-image-layer-default-view/4333 | 1.0 | Image-layer-only exhibits - Currently, users must select a spatial layer from a dropdown menu to serve as the base layer for a new exhibit. This means that an image layer can't be the first layer to load when the exhibit is viewed. Per #428, some users want to have exhibits where the image layer is the default view for the map — this may have been unintentionally possible prior to the last release due to an issue preventing the successful creation of Google Maps layers. Users should be able to choose to have no spatial layer as the base, or an option specifically to use the exhibit's image layer as its base in place of a spatial layer.
Related forum post: https://forum.omeka.org/t/neatline-image-layer-default-view/4333 | defect | image layer only exhibits currently users must select a spatial layer from a dropdown menu to serve as the base layer for a new exhibit this means that an image layer can t be the first layer to load when the exhibit is viewed per some users want to have exhibits where the image layer is the default view for the map — this may have been unintentionally possible prior to the last release due to an issue preventing the successful creation of google maps layers users should be able to choose to have no spatial layer as the base or an option specifically to use the exhibit s image layer as its base in place of a spatial layer related forum post | 1 |
100,688 | 8,752,748,771 | IssuesEvent | 2018-12-14 04:59:21 | humera987/FXLabs-Test-Automation | https://api.github.com/repos/humera987/FXLabs-Test-Automation | closed | Testing 14 : ApiV1TestSuitesProjectIdIdCoverageGetPathParamIdMysqlSqlInjectionTimebound | Testing 14 | Project : Testing 14
Job : UAT
Env : UAT
Region : US_WEST
Result : fail
Status Code : 404
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=MTdjNGVkOTctMGRlZC00NTRmLThjNTYtZDY0ODcyZTY2MDZj; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Fri, 14 Dec 2018 04:57:10 GMT]}
Endpoint : http://13.56.210.25/api/v1/api/v1/test-suites/project-id//coverage
Request :
Response :
{
"timestamp" : "2018-12-14T04:57:10.805+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/test-suites/project-id/coverage"
}
Logs :
Assertion [@ResponseTime < 7000 OR @ResponseTime > 10000] resolved-to [445 < 7000 OR 445 > 10000] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed]
--- FX Bot --- | 1.0 | Testing 14 : ApiV1TestSuitesProjectIdIdCoverageGetPathParamIdMysqlSqlInjectionTimebound - Project : Testing 14
Job : UAT
Env : UAT
Region : US_WEST
Result : fail
Status Code : 404
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=MTdjNGVkOTctMGRlZC00NTRmLThjNTYtZDY0ODcyZTY2MDZj; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Fri, 14 Dec 2018 04:57:10 GMT]}
Endpoint : http://13.56.210.25/api/v1/api/v1/test-suites/project-id//coverage
Request :
Response :
{
"timestamp" : "2018-12-14T04:57:10.805+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/test-suites/project-id/coverage"
}
Logs :
Assertion [@ResponseTime < 7000 OR @ResponseTime > 10000] resolved-to [445 < 7000 OR 445 > 10000] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed]
--- FX Bot --- | non_defect | testing project testing job uat env uat region us west result fail status code headers x content type options x xss protection cache control pragma expires x frame options set cookie content type transfer encoding date endpoint request response timestamp status error not found message no message available path api api test suites project id coverage logs assertion resolved to result assertion resolved to result fx bot | 0 |
351,015 | 10,511,907,736 | IssuesEvent | 2019-09-27 16:31:44 | freeorion/freeorion | https://api.github.com/repos/freeorion/freeorion | closed | Scylior OP, add some bad traits | category:bug component:content scripting priority:low | ### Description
Scylor have great research (200%) and good population (125%), no more traits, which makes them rather OP compared to other playable species.
Also, they start on a medium (instead of large) planet (this is undocumented, should be fixed), but that is not enough to balance them.
I suggest adding them these bad traits:
- Bad supply.
- Bad detection.
- Bad stockpiling.
None of these are crippling, and all can be circumvented using their great research, so that they are just a bit of a slowdown for expansion.
I don't think these would really make Scylior balanced, just less OP.
Discussion thread: http://www.freeorion.org/forum/viewtopic.php?f=28&t=10985 | 1.0 | Scylior OP, add some bad traits - ### Description
Scylor have great research (200%) and good population (125%), no more traits, which makes them rather OP compared to other playable species.
Also, they start on a medium (instead of large) planet (this is undocumented, should be fixed), but that is not enough to balance them.
I suggest adding them these bad traits:
- Bad supply.
- Bad detection.
- Bad stockpiling.
None of these are crippling, and all can be circumvented using their great research, so that they are just a bit of a slowdown for expansion.
I don't think these would really make Scylior balanced, just less OP.
Discussion thread: http://www.freeorion.org/forum/viewtopic.php?f=28&t=10985 | non_defect | scylior op add some bad traits description scylor have great research and good population no more traits which makes them rather op compared to other playable species also they start on a medium instead of large planet this is undocumented should be fixed but that is not enough to balance them i suggest adding them these bad traits bad supply bad detection bad stockpiling none of these are crippling and all can be circumvented using their great research so that they are just a bit of a slowdown for expansion i don t think these would really make scylior balanced just less op discussion thread | 0 |
71,534 | 23,681,205,771 | IssuesEvent | 2022-08-28 20:34:08 | BOINC/boinc | https://api.github.com/repos/BOINC/boinc | closed | Boinc client won't start with cc_config, exclude_gpu, device_num=99 | C: Client - Daemon P: Minor R: fixed T: Defect | **Describe the bug**
Boinc client fails to start when cc_config.xml have exclude_gpu, device_num=99. The only error message is from BOINC manager, "The BOINC client has exited unexpectedly 3 times within the last 3 minutes".
**Steps To Reproduce**
1. Open cc_config.xml
2. `<cc_config><exclude_gpu><url>http://milkyway.cs.rpi.edu/milkyway/</url><device_num>99</device_num><type>ATI</type></exclude_gpu></cc_config>`
**Expected behavior**
Starts up and work, maybe with a warning about invalid device_num
**System Information**
- OS: Windows 10
- BOINC Version: 7.20.2
**Additional context**
Old Boinc version 7.16.20 did start ok, but may have some random stability problems with invalid device_num.
[cc_config.txt](https://github.com/BOINC/boinc/files/9438900/cc_config.txt) | 1.0 | Boinc client won't start with cc_config, exclude_gpu, device_num=99 - **Describe the bug**
Boinc client fails to start when cc_config.xml have exclude_gpu, device_num=99. The only error message is from BOINC manager, "The BOINC client has exited unexpectedly 3 times within the last 3 minutes".
**Steps To Reproduce**
1. Open cc_config.xml
2. `<cc_config><exclude_gpu><url>http://milkyway.cs.rpi.edu/milkyway/</url><device_num>99</device_num><type>ATI</type></exclude_gpu></cc_config>`
**Expected behavior**
Starts up and work, maybe with a warning about invalid device_num
**System Information**
- OS: Windows 10
- BOINC Version: 7.20.2
**Additional context**
Old Boinc version 7.16.20 did start ok, but may have some random stability problems with invalid device_num.
[cc_config.txt](https://github.com/BOINC/boinc/files/9438900/cc_config.txt) | defect | boinc client won t start with cc config exclude gpu device num describe the bug boinc client fails to start when cc config xml have exclude gpu device num the only error message is from boinc manager the boinc client has exited unexpectedly times within the last minutes steps to reproduce open cc config xml expected behavior starts up and work maybe with a warning about invalid device num system information os windows boinc version additional context old boinc version did start ok but may have some random stability problems with invalid device num | 1 |
47,344 | 13,056,132,149 | IssuesEvent | 2020-07-30 03:45:19 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | closed | NuGen Bug (Trac #384) | Migrated from Trac combo simulation defect | http://internal.icecube.wisc.edu/simulation/job/10876569
Fix Me.
Migrated from https://code.icecube.wisc.edu/ticket/384
```json
{
"status": "closed",
"changetime": "2014-11-22T22:53:31",
"description": "http://internal.icecube.wisc.edu/simulation/job/10876569\n\nFix Me.",
"reporter": "olivas",
"cc": "",
"resolution": "worksforme",
"_ts": "1416696811687977",
"component": "combo simulation",
"summary": "NuGen Bug",
"priority": "normal",
"keywords": "",
"time": "2012-03-23T19:33:55",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
| 1.0 | NuGen Bug (Trac #384) - http://internal.icecube.wisc.edu/simulation/job/10876569
Fix Me.
Migrated from https://code.icecube.wisc.edu/ticket/384
```json
{
"status": "closed",
"changetime": "2014-11-22T22:53:31",
"description": "http://internal.icecube.wisc.edu/simulation/job/10876569\n\nFix Me.",
"reporter": "olivas",
"cc": "",
"resolution": "worksforme",
"_ts": "1416696811687977",
"component": "combo simulation",
"summary": "NuGen Bug",
"priority": "normal",
"keywords": "",
"time": "2012-03-23T19:33:55",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
| defect | nugen bug trac fix me migrated from json status closed changetime description me reporter olivas cc resolution worksforme ts component combo simulation summary nugen bug priority normal keywords time milestone owner olivas type defect | 1 |
3,141 | 2,607,985,429 | IssuesEvent | 2015-02-26 00:51:28 | chrsmithdemos/zen-coding | https://api.github.com/repos/chrsmithdemos/zen-coding | closed | How to move cursor at the next '|' postion | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
Actually, this is a question.
I put following in the snippets, with two |'s.
'fori' : 'for (int count = 0; count < |SIZE ; count++) {|}',
When expanded, the cursor positions at the first | place. My question is how to
move the cursor to the next | position? I am using Windows 7, 64bit.
What is the expected output? What do you see instead?
What version of the product are you using? On what operating system?
0.7 for NP++, Windows7, 64
Please provide any additional information below.
I like zen coding. It rocks!
```
-----
Original issue reported on code.google.com by `yds...@gmail.com` on 24 Feb 2012 at 4:12 | 1.0 | How to move cursor at the next '|' postion - ```
What steps will reproduce the problem?
Actually, this is a question.
I put following in the snippets, with two |'s.
'fori' : 'for (int count = 0; count < |SIZE ; count++) {|}',
When expanded, the cursor positions at the first | place. My question is how to
move the cursor to the next | position? I am using Windows 7, 64bit.
What is the expected output? What do you see instead?
What version of the product are you using? On what operating system?
0.7 for NP++, Windows7, 64
Please provide any additional information below.
I like zen coding. It rocks!
```
-----
Original issue reported on code.google.com by `yds...@gmail.com` on 24 Feb 2012 at 4:12 | defect | how to move cursor at the next postion what steps will reproduce the problem actually this is a question i put following in the snippets with two s fori for int count count size count when expanded the cursor positions at the first place my question is how to move the cursor to the next position i am using windows what is the expected output what do you see instead what version of the product are you using on what operating system for np please provide any additional information below i like zen coding it rocks original issue reported on code google com by yds gmail com on feb at | 1 |
348,220 | 24,908,889,910 | IssuesEvent | 2022-10-29 16:05:56 | DominicThorpe/iridium_assembler | https://api.github.com/repos/DominicThorpe/iridium_assembler | opened | Improperly Documented ADDC and SUBC | documentation | The ADDC and SUBC instructions take 2 operands, not 1.
Currently, the documentation states that the carry flag is added to and stored in the 1st operand, but this is not true. Instead, they add the carry flag to the 2nd operand and store it in the 1st. | 1.0 | Improperly Documented ADDC and SUBC - The ADDC and SUBC instructions take 2 operands, not 1.
Currently, the documentation states that the carry flag is added to and stored in the 1st operand, but this is not true. Instead, they add the carry flag to the 2nd operand and store it in the 1st. | non_defect | improperly documented addc and subc the addc and subc instructions take operands not currently the documentation states that the carry flag is added to and stored in the operand but this is not true instead they add the carry flag to the operand and store it in the | 0 |
75,204 | 25,586,624,618 | IssuesEvent | 2022-12-01 09:49:48 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Unable to demote myself in a space. | T-Defect S-Minor O-Uncommon A-Power-Levels | ### Steps to reproduce
1. Give an alt account administrator in your space
2. Try to demote alt account (from the alt account)
3. The dialogue is greyed out
### Outcome
#### What did you expect?
I expected to be able to demote myself.
#### What happened instead?
I could not demote myself.
### Operating system
Windows
### Browser information
Librewolf 107.0-1
### URL for webapp
https://app.element.io/
### Application version
Element version: 1.11.15 Olm version: 3.2.12
### Homeserver
the-apothecary.club
### Will you send logs?
No
My alt account I'm trying to demote is the @averythefurry:matrix.org one.

| 1.0 | Unable to demote myself in a space. - ### Steps to reproduce
1. Give an alt account administrator in your space
2. Try to demote alt account (from the alt account)
3. The dialogue is greyed out
### Outcome
#### What did you expect?
I expected to be able to demote myself.
#### What happened instead?
I could not demote myself.
### Operating system
Windows
### Browser information
Librewolf 107.0-1
### URL for webapp
https://app.element.io/
### Application version
Element version: 1.11.15 Olm version: 3.2.12
### Homeserver
the-apothecary.club
### Will you send logs?
No
My alt account I'm trying to demote is the @averythefurry:matrix.org one.

| defect | unable to demote myself in a space steps to reproduce give an alt account administrator in your space try to demote alt account from the alt account the dialogue is greyed out outcome what did you expect i expected to be able to demote myself what happened instead i could not demote myself operating system windows browser information librewolf url for webapp application version element version olm version homeserver the apothecary club will you send logs no my alt account i m trying to demote is the averythefurry matrix org one | 1 |
29,004 | 5,476,189,593 | IssuesEvent | 2017-03-11 18:23:58 | richgel999/miniz | https://api.github.com/repos/richgel999/miniz | closed | Wrong number of files returned for more that 65535 entries | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Open a zip file with more than 65535 entries
2. Use the mz_zip_reader_get_num_files function to read it and you'll see it
returns 65535 entries, instead of the correct value.
What version of the product are you using? On what operating system?
trunk
```
Original issue reported on code.google.com by `mingo...@gmail.com` on 5 Jul 2014 at 12:10
| 1.0 | Wrong number of files returned for more that 65535 entries - ```
What steps will reproduce the problem?
1. Open a zip file with more than 65535 entries
2. Use the mz_zip_reader_get_num_files function to read it and you'll see it
returns 65535 entries, instead of the correct value.
What version of the product are you using? On what operating system?
trunk
```
Original issue reported on code.google.com by `mingo...@gmail.com` on 5 Jul 2014 at 12:10
| defect | wrong number of files returned for more that entries what steps will reproduce the problem open a zip file with more than entries use the mz zip reader get num files function to read it and you ll see it returns entries instead of the correct value what version of the product are you using on what operating system trunk original issue reported on code google com by mingo gmail com on jul at | 1 |
263,966 | 19,984,516,051 | IssuesEvent | 2022-01-30 12:55:10 | FsCatamarca2021-2/grupo-4-grupo4catamarca | https://api.github.com/repos/FsCatamarca2021-2/grupo-4-grupo4catamarca | opened | US#crear Botón de Ingreso | documentation | * Yo como desarrollador
* crearé un boton
* para ingresar a la aplicación | 1.0 | US#crear Botón de Ingreso - * Yo como desarrollador
* crearé un boton
* para ingresar a la aplicación | non_defect | us crear botón de ingreso yo como desarrollador crearé un boton para ingresar a la aplicación | 0 |
3,753 | 2,690,063,286 | IssuesEvent | 2015-03-31 14:19:15 | backbee/BbCoreJs | https://api.github.com/repos/backbee/BbCoreJs | closed | clicking on users in a group | Fixed To test | everytime i click on "users" in a group is displays the admin user form
I should see the list of users of my group | 1.0 | clicking on users in a group - everytime i click on "users" in a group is displays the admin user form
I should see the list of users of my group | non_defect | clicking on users in a group everytime i click on users in a group is displays the admin user form i should see the list of users of my group | 0 |
61,395 | 17,023,683,280 | IssuesEvent | 2021-07-03 03:17:03 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | Many addresses in The Netherlands are registered by nominatim as Bottrop Germany | Component: nominatim Priority: major Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 11.30am, Wednesday, 23rd February 2011]**
There are 2 addresses in The Netherlands that are registered by nominatim as Bottrop Germany.
Adress 1:
Zutphensestraat, Apeldoorn
http://open.mapquestapi.com/nominatim/v1/details.php?place_id=53735767
Adress 2:
Vroomshoopseweg, Den Ham
http://open.mapquestapi.com/nominatim/v1/details.php?place_id=15762072
The polygon of bottrop is not on the same location as those 2 addresses. So I don't know what is going on, a bug in the software maybe? | 1.0 | Many addresses in The Netherlands are registered by nominatim as Bottrop Germany - **[Submitted to the original trac issue database at 11.30am, Wednesday, 23rd February 2011]**
There are 2 addresses in The Netherlands that are registered by nominatim as Bottrop Germany.
Adress 1:
Zutphensestraat, Apeldoorn
http://open.mapquestapi.com/nominatim/v1/details.php?place_id=53735767
Adress 2:
Vroomshoopseweg, Den Ham
http://open.mapquestapi.com/nominatim/v1/details.php?place_id=15762072
The polygon of bottrop is not on the same location as those 2 addresses. So I don't know what is going on, a bug in the software maybe? | defect | many addresses in the netherlands are registered by nominatim as bottrop germany there are addresses in the netherlands that are registered by nominatim as bottrop germany adress zutphensestraat apeldoorn adress vroomshoopseweg den ham the polygon of bottrop is not on the same location as those addresses so i don t know what is going on a bug in the software maybe | 1 |
375,814 | 26,178,862,712 | IssuesEvent | 2023-01-02 13:11:58 | k-saito-en/engineer-study-app | https://api.github.com/repos/k-saito-en/engineer-study-app | closed | [future task] ファイル構成、ソースコードのリファクタリング | documentation | # 現状
- メモアプリとしての大枠は完成したが、コメントアウトが乱立し、コードが読みづらい
# 実現したいこと
- [x] ディレクトリ構成の整理
- [x] ついでにREADME.mdにディレクトリ構成を追記
- [x] Developブランチを削除し、ブランチを整理する
- [x] 「問題」をできるだけ消去し、改善が必要な部分を洗い出す | 1.0 | [future task] ファイル構成、ソースコードのリファクタリング - # 現状
- メモアプリとしての大枠は完成したが、コメントアウトが乱立し、コードが読みづらい
# 実現したいこと
- [x] ディレクトリ構成の整理
- [x] ついでにREADME.mdにディレクトリ構成を追記
- [x] Developブランチを削除し、ブランチを整理する
- [x] 「問題」をできるだけ消去し、改善が必要な部分を洗い出す | non_defect | ファイル構成、ソースコードのリファクタリング 現状 メモアプリとしての大枠は完成したが、コメントアウトが乱立し、コードが読みづらい 実現したいこと ディレクトリ構成の整理 ついでにreadme mdにディレクトリ構成を追記 developブランチを削除し、ブランチを整理する 「問題」をできるだけ消去し、改善が必要な部分を洗い出す | 0 |
69,231 | 22,287,878,874 | IssuesEvent | 2022-06-11 23:18:04 | comtrya/comtrya | https://api.github.com/repos/comtrya/comtrya | closed | FR: Properly deal with `apt-key add` deprecation | Type::Defect OS::Linux Actions::Package Component::Actions PackageProvider::Aptitude | ## The Background
`apt-key add` is being deprecated.
Quoting the [The Debian WIKI](https://wiki.debian.org/DebianRepository/UseThirdParty):
> ... The key MUST NOT be placed in /etc/apt/trusted.gpg.d or loaded by apt-key add.
Here is a rather good article that explains how to add keys properly:
> https://www.linuxuprising.com/2021/01/apt-key-is-deprecated-how-to-add.html
It is also reflected in the more-up-to-date vendor's documentation, like google-cloud-cli ([link](https://cloud.google.com/sdk/docs/install#debianubuntu)):
> If your distribution (Debian 11+ or Ubuntu 21.10+) doesn't support apt-key, run the following command:
> ```
> curl https://packages.cloud.google.com/apt/doc/apt-key.gpg | sudo tee /usr/share/keyrings/cl
> ```
The above is just to state, that this has '_become a thing_'.
## The request:
Using `apt-key` should be replaced with the process of adding a key to `/usr/share/keyrings` and then referring to it in the '/etc/apt/sources.d/repo-name.list` file.
Maybe even adding an option to mention a GPG fingerprint (or better yet, a URL to a FINGERPRINT) to be downloaded.
## Some examples
An example of grabbing the gpg file and storing it (note: will require sudo)
```
wget -O- <https://example.com/key/repo-key.gpg> | gpg --dearmor | sudo tee /usr/share/keyrings/<myrepository>-archive-keyring.gpg
```
followed by
```
sudo apt-add-repository "deb [arch=$(dpkg --print-architecture) signed-by=/usr/share/keyrings/<myrepository>-archive-keyring.gpg] https://repository.example.com/debian/ $(lsb_release -cs) stable main "
```
| 1.0 | FR: Properly deal with `apt-key add` deprecation - ## The Background
`apt-key add` is being deprecated.
Quoting the [The Debian WIKI](https://wiki.debian.org/DebianRepository/UseThirdParty):
> ... The key MUST NOT be placed in /etc/apt/trusted.gpg.d or loaded by apt-key add.
Here is a rather good article that explains how to add keys properly:
> https://www.linuxuprising.com/2021/01/apt-key-is-deprecated-how-to-add.html
It is also reflected in the more-up-to-date vendor's documentation, like google-cloud-cli ([link](https://cloud.google.com/sdk/docs/install#debianubuntu)):
> If your distribution (Debian 11+ or Ubuntu 21.10+) doesn't support apt-key, run the following command:
> ```
> curl https://packages.cloud.google.com/apt/doc/apt-key.gpg | sudo tee /usr/share/keyrings/cl
> ```
The above is just to state, that this has '_become a thing_'.
## The request:
Using `apt-key` should be replaced with the process of adding a key to `/usr/share/keyrings` and then referring to it in the '/etc/apt/sources.d/repo-name.list` file.
Maybe even adding an option to mention a GPG fingerprint (or better yet, a URL to a FINGERPRINT) to be downloaded.
## Some examples
An example of grabbing the gpg file and storing it (note: will require sudo)
```
wget -O- <https://example.com/key/repo-key.gpg> | gpg --dearmor | sudo tee /usr/share/keyrings/<myrepository>-archive-keyring.gpg
```
followed by
```
sudo apt-add-repository "deb [arch=$(dpkg --print-architecture) signed-by=/usr/share/keyrings/<myrepository>-archive-keyring.gpg] https://repository.example.com/debian/ $(lsb_release -cs) stable main "
```
| defect | fr properly deal with apt key add deprecation the background apt key add is being deprecated quoting the the key must not be placed in etc apt trusted gpg d or loaded by apt key add here is a rather good article that explains how to add keys properly it is also reflected in the more up to date vendor s documentation like google cloud cli if your distribution debian or ubuntu doesn t support apt key run the following command curl sudo tee usr share keyrings cl the above is just to state that this has become a thing the request using apt key should be replaced with the process of adding a key to usr share keyrings and then referring to it in the etc apt sources d repo name list file maybe even adding an option to mention a gpg fingerprint or better yet a url to a fingerprint to be downloaded some examples an example of grabbing the gpg file and storing it note will require sudo wget o archive keyring gpg followed by sudo apt add repository deb lsb release cs stable main | 1 |
686,687 | 23,501,168,395 | IssuesEvent | 2022-08-18 08:33:31 | kubesphere/console | https://api.github.com/repos/kubesphere/console | closed | An error is reported when a user with cluster viewing permission views the storage class | kind/bug priority/low | **Describe the bug**
1、There is a user cc, who has the permission to view the platform cluster
2、Use this user to log in, enter the cluster management->storage->storage class, click the existing storage class, enter the details page, the page reports an error



**Versions used(KubeSphere/Kubernetes)**
KubeSphere: v3.3.0
/kind bug
/priority low
/assign @weili520 | 1.0 | An error is reported when a user with cluster viewing permission views the storage class - **Describe the bug**
1、There is a user cc, who has the permission to view the platform cluster
2、Use this user to log in, enter the cluster management->storage->storage class, click the existing storage class, enter the details page, the page reports an error



**Versions used(KubeSphere/Kubernetes)**
KubeSphere: v3.3.0
/kind bug
/priority low
/assign @weili520 | non_defect | an error is reported when a user with cluster viewing permission views the storage class describe the bug 、there is a user cc who has the permission to view the platform cluster 、use this user to log in enter the cluster management storage storage class click the existing storage class enter the details page the page reports an error versions used kubesphere kubernetes kubesphere kind bug priority low assign | 0 |
50,757 | 13,187,720,095 | IssuesEvent | 2020-08-13 04:21:09 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | closed | paraboloid - coverage of private interfaces is weak (Trac #1305) | Migrated from Trac combo reconstruction defect | a serialization and a "known failure" test will quickly bump coverage
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1305">https://code.icecube.wisc.edu/ticket/1305</a>, reported by nega and owned by kjmeagher</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-02-10T10:59:05",
"description": "a serialization and a \"known failure\" test will quickly bump coverage",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"_ts": "1455101945934200",
"component": "combo reconstruction",
"summary": "paraboloid - coverage of private interfaces is weak",
"priority": "normal",
"keywords": "",
"time": "2015-08-28T23:07:43",
"milestone": "",
"owner": "kjmeagher",
"type": "defect"
}
```
</p>
</details>
| 1.0 | paraboloid - coverage of private interfaces is weak (Trac #1305) - a serialization and a "known failure" test will quickly bump coverage
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1305">https://code.icecube.wisc.edu/ticket/1305</a>, reported by nega and owned by kjmeagher</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-02-10T10:59:05",
"description": "a serialization and a \"known failure\" test will quickly bump coverage",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"_ts": "1455101945934200",
"component": "combo reconstruction",
"summary": "paraboloid - coverage of private interfaces is weak",
"priority": "normal",
"keywords": "",
"time": "2015-08-28T23:07:43",
"milestone": "",
"owner": "kjmeagher",
"type": "defect"
}
```
</p>
</details>
| defect | paraboloid coverage of private interfaces is weak trac a serialization and a known failure test will quickly bump coverage migrated from json status closed changetime description a serialization and a known failure test will quickly bump coverage reporter nega cc resolution fixed ts component combo reconstruction summary paraboloid coverage of private interfaces is weak priority normal keywords time milestone owner kjmeagher type defect | 1 |
33,168 | 7,045,680,301 | IssuesEvent | 2018-01-01 23:12:18 | STEllAR-GROUP/hpx | https://api.github.com/repos/STEllAR-GROUP/hpx | closed | Assertion failure when using bulk hpx::new_ in asynchronous mode | category: components type: defect | I stumbled upon a weird assertion failure in `hpx/runtime/components/default_distribution_policy.hpp`.
You can reproduce it by adding the following lines to `tests/regressions/components/new_2848.cpp`
```c++
{
auto locs = hpx::find_all_localities();
std::vector<hpx::id_type> ids =
hpx::new_<test_server[]>(hpx::default_layout(locs), 10, 42).get();
hpx::future<std::vector<hpx::id_type>> ids_f =
hpx::new_<test_server[]>(hpx::default_layout(locs), 10, 42);
}
```
and running `new_2848_test` on more than one locality.
```
new_2848_test: /home/igor/hpx/hpx.git/hpx/runtime/components/default_distribution_policy.hpp:163: hpx::components::default_distribution_policy::bulk_create(std::size_t, Ts&& ...) const::<lambda(std::vector<hpx::lcos::future<std::vector<hpx::naming::id_type> > >&&)> mutable [with Component = test_server; Ts = {int}]: Assertion `localities_.size() == v.size()' failed.
```
The weird part here is that `ids` are returned normally, and only the second call to `hpx::new_` crashes. | 1.0 | Assertion failure when using bulk hpx::new_ in asynchronous mode - I stumbled upon a weird assertion failure in `hpx/runtime/components/default_distribution_policy.hpp`.
You can reproduce it by adding the following lines to `tests/regressions/components/new_2848.cpp`
```c++
{
auto locs = hpx::find_all_localities();
std::vector<hpx::id_type> ids =
hpx::new_<test_server[]>(hpx::default_layout(locs), 10, 42).get();
hpx::future<std::vector<hpx::id_type>> ids_f =
hpx::new_<test_server[]>(hpx::default_layout(locs), 10, 42);
}
```
and running `new_2848_test` on more than one locality.
```
new_2848_test: /home/igor/hpx/hpx.git/hpx/runtime/components/default_distribution_policy.hpp:163: hpx::components::default_distribution_policy::bulk_create(std::size_t, Ts&& ...) const::<lambda(std::vector<hpx::lcos::future<std::vector<hpx::naming::id_type> > >&&)> mutable [with Component = test_server; Ts = {int}]: Assertion `localities_.size() == v.size()' failed.
```
The weird part here is that `ids` are returned normally, and only the second call to `hpx::new_` crashes. | defect | assertion failure when using bulk hpx new in asynchronous mode i stumbled upon a weird assertion failure in hpx runtime components default distribution policy hpp you can reproduce it by adding the following lines to tests regressions components new cpp c auto locs hpx find all localities std vector ids hpx new hpx default layout locs get hpx future ids f hpx new hpx default layout locs and running new test on more than one locality new test home igor hpx hpx git hpx runtime components default distribution policy hpp hpx components default distribution policy bulk create std size t ts const mutable assertion localities size v size failed the weird part here is that ids are returned normally and only the second call to hpx new crashes | 1 |
126,690 | 26,897,367,357 | IssuesEvent | 2023-02-06 13:25:13 | foundry-rs/foundry | https://api.github.com/repos/foundry-rs/foundry | closed | JSON: unable to decode arrays | T-bug C-forge A-cheatcodes | ### Component
Forge
### Have you ensured that all of these are up to date?
- [X] Foundry
- [X] Foundryup
### What version of Foundry are you on?
forge 0.2.0 (cd7850b 2023-02-01T00:04:58.420758Z)
### What command(s) is the bug in?
forge script
### Operating System
macOS (Apple Silicon)
### Describe the bug
`jsonParse` seems to be returning improperly encoded data.
```solidity
# this would fail
uint256 ids = jsonRoles.readUintArray(".roles[*].id");
# now this is what happens
bytes memory idBytes = jsonRoles.parseRaw(".roles[*].id");
console2.log("ids bytes:", vm.toString(idBytes));
```
The json looks more or less like this:
```json
{ "roles": [ { "id": 1 }, { "id": 2 } ] }
```
So we want to get an array of `[1, 2]`.
Encoded bytes we got look like this:
```
0x00000000000000000000000000000000000000000000000000000000000000010000000000000000000000000000000000000000000000000000000000000002
```
This is obviously not the right ABI-encoded version of that array.
I have the same issues with using `readStringArray` and `reaedAddressArray`. | 1.0 | JSON: unable to decode arrays - ### Component
Forge
### Have you ensured that all of these are up to date?
- [X] Foundry
- [X] Foundryup
### What version of Foundry are you on?
forge 0.2.0 (cd7850b 2023-02-01T00:04:58.420758Z)
### What command(s) is the bug in?
forge script
### Operating System
macOS (Apple Silicon)
### Describe the bug
`jsonParse` seems to be returning improperly encoded data.
```solidity
# this would fail
uint256 ids = jsonRoles.readUintArray(".roles[*].id");
# now this is what happens
bytes memory idBytes = jsonRoles.parseRaw(".roles[*].id");
console2.log("ids bytes:", vm.toString(idBytes));
```
The json looks more or less like this:
```json
{ "roles": [ { "id": 1 }, { "id": 2 } ] }
```
So we want to get an array of `[1, 2]`.
Encoded bytes we got look like this:
```
0x00000000000000000000000000000000000000000000000000000000000000010000000000000000000000000000000000000000000000000000000000000002
```
This is obviously not the right ABI-encoded version of that array.
I have the same issues with using `readStringArray` and `reaedAddressArray`. | non_defect | json unable to decode arrays component forge have you ensured that all of these are up to date foundry foundryup what version of foundry are you on forge what command s is the bug in forge script operating system macos apple silicon describe the bug jsonparse seems to be returning improperly encoded data solidity this would fail ids jsonroles readuintarray roles id now this is what happens bytes memory idbytes jsonroles parseraw roles id log ids bytes vm tostring idbytes the json looks more or less like this json roles so we want to get an array of encoded bytes we got look like this this is obviously not the right abi encoded version of that array i have the same issues with using readstringarray and reaedaddressarray | 0 |
125,109 | 16,725,027,288 | IssuesEvent | 2021-06-10 12:00:36 | elastic/observability-design | https://api.github.com/repos/elastic/observability-design | closed | [APM] Mobile APM entry points to Exploratory view | Team:apm design v7.14.0 | ## Summary of the problem
As we're adding support for Mobile APM with the new iOS agent in 7.14, we want to also add the ability to utilize the Exploratory view for analyzing user demographic data such as device types, locations, and more. This means we need to add entry points to the Exploratory view from APM on the service charts that will exist for iOS.
**Telemetry**
We want to measure what charts users tend to use for analyzing their metrics further, so we can gauge which types of charts are used the most.
## User stories
Take each of the user stories and describe this specific use case in more detail and the benefit that this feature will bring to the user.
### Example user story 1
As a user, I want to analyze my iOS data by breaking down performance data by specific dimensions available such as device type, location, and more.
----
**This section should be filled in collaboration and agreement with a designer.**
## Deliverables
Example screens that define the entry points into Exploratory view from the APM service views.
### Low-fi wireframes
<details>
<summary>View screens</summary>


</details>
| 1.0 | [APM] Mobile APM entry points to Exploratory view - ## Summary of the problem
As we're adding support for Mobile APM with the new iOS agent in 7.14, we want to also add the ability to utilize the Exploratory view for analyzing user demographic data such as device types, locations, and more. This means we need to add entry points to the Exploratory view from APM on the service charts that will exist for iOS.
**Telemetry**
We want to measure what charts users tend to use for analyzing their metrics further, so we can gauge which types of charts are used the most.
## User stories
Take each of the user stories and describe this specific use case in more detail and the benefit that this feature will bring to the user.
### Example user story 1
As a user, I want to analyze my iOS data by breaking down performance data by specific dimensions available such as device type, location, and more.
----
**This section should be filled in collaboration and agreement with a designer.**
## Deliverables
Example screens that define the entry points into Exploratory view from the APM service views.
### Low-fi wireframes
<details>
<summary>View screens</summary>


</details>
| non_defect | mobile apm entry points to exploratory view summary of the problem as we re adding support for mobile apm with the new ios agent in we want to also add the ability to utilize the exploratory view for analyzing user demographic data such as device types locations and more this means we need to add entry points to the exploratory view from apm on the service charts that will exist for ios telemetry we want to measure what charts users tend to use for analyzing their metrics further so we can gauge which types of charts are used the most user stories take each of the user stories and describe this specific use case in more detail and the benefit that this feature will bring to the user example user story as a user i want to analyze my ios data by breaking down performance data by specific dimensions available such as device type location and more this section should be filled in collaboration and agreement with a designer deliverables example screens that define the entry points into exploratory view from the apm service views low fi wireframes view screens | 0 |
150,030 | 13,308,007,338 | IssuesEvent | 2020-08-25 23:40:45 | great-expectations/great_expectations | https://api.github.com/repos/great-expectations/great_expectations | closed | Broken links/images throughout "Getting Started" docs | bug documentation | **Describe the bug**
There are missing images throughout the getting started guide. Seems like there might be a bad reference somewhere that's effecting several docs pages
## Quickstart guide
The [quickstart-guide](https://docs.greatexpectations.io/en/latest/guides/tutorials/quick_start.html) in the docs is missing the following items:
- The link to the expecations notebook
- the gif

## Connect to Data
Also found missing images in the [Connect to Data docs page](https://docs.greatexpectations.io/en/latest/guides/tutorials/getting_started/connect_to_data.html)

## Set up Data Docs
[Set up Data docs](https://docs.greatexpectations.io/en/latest/guides/tutorials/getting_started/set_up_data_docs.html) is missing a few images as well.

| 1.0 | Broken links/images throughout "Getting Started" docs - **Describe the bug**
There are missing images throughout the getting started guide. Seems like there might be a bad reference somewhere that's effecting several docs pages
## Quickstart guide
The [quickstart-guide](https://docs.greatexpectations.io/en/latest/guides/tutorials/quick_start.html) in the docs is missing the following items:
- The link to the expecations notebook
- the gif

## Connect to Data
Also found missing images in the [Connect to Data docs page](https://docs.greatexpectations.io/en/latest/guides/tutorials/getting_started/connect_to_data.html)

## Set up Data Docs
[Set up Data docs](https://docs.greatexpectations.io/en/latest/guides/tutorials/getting_started/set_up_data_docs.html) is missing a few images as well.

| non_defect | broken links images throughout getting started docs describe the bug there are missing images throughout the getting started guide seems like there might be a bad reference somewhere that s effecting several docs pages quickstart guide the in the docs is missing the following items the link to the expecations notebook the gif connect to data also found missing images in the set up data docs is missing a few images as well | 0 |
384,605 | 11,395,232,746 | IssuesEvent | 2020-01-30 10:58:22 | unitystation/unitystation | https://api.github.com/repos/unitystation/unitystation | closed | Make shuttle to not use the same power-grid as the station B: 30 | Bounty Bug Electrical High Priority | Actually, if the station lacks power, the shuttles lights are off.
1. Lose station power (wait for plasma to be burnt or turn off power)
2. Go inside a shuttle
3. No lights!
That also happen if you're flying a shuttle and power stops in the station at a distance.
| 1.0 | Make shuttle to not use the same power-grid as the station B: 30 - Actually, if the station lacks power, the shuttles lights are off.
1. Lose station power (wait for plasma to be burnt or turn off power)
2. Go inside a shuttle
3. No lights!
That also happen if you're flying a shuttle and power stops in the station at a distance.
| non_defect | make shuttle to not use the same power grid as the station b actually if the station lacks power the shuttles lights are off lose station power wait for plasma to be burnt or turn off power go inside a shuttle no lights that also happen if you re flying a shuttle and power stops in the station at a distance | 0 |
225,911 | 17,292,272,779 | IssuesEvent | 2021-07-25 02:06:39 | BulletTech/BulletTech | https://api.github.com/repos/BulletTech/BulletTech | closed | [Main Course] - Job crawler of PayPal | documentation medium priority | # Outline
- 分析网页结构
- 找出定位元素
- 爬取数据
- 整理数据
# Milestone
- 7/19 - Outline complete
- 7/23 - Draft complete
- 7/23 - Review complete
- 7/26 - Release
| 1.0 | [Main Course] - Job crawler of PayPal - # Outline
- 分析网页结构
- 找出定位元素
- 爬取数据
- 整理数据
# Milestone
- 7/19 - Outline complete
- 7/23 - Draft complete
- 7/23 - Review complete
- 7/26 - Release
| non_defect | job crawler of paypal outline 分析网页结构 找出定位元素 爬取数据 整理数据 milestone outline complete draft complete review complete release | 0 |
67,331 | 20,961,606,059 | IssuesEvent | 2022-03-27 21:48:15 | abedmaatalla/sipdroid | https://api.github.com/repos/abedmaatalla/sipdroid | closed | Scrambled void when calling, but voice ok when being called. | Priority-Medium Type-Defect auto-migrated | ```
Before you begin please read the FAQ!
Attention!
Don't post doubles!
Why is it important to find doubles?
Sipdroid is a community project. This is no dropbox for your personal
issue. You need to find others having the same problem to get it fixed. If
the problem has already been reported you add more details and ideas to the
existing discussion. If it has not been reported yet you include as many
background info as possible to give the community a chance to follow up on
it. No one will just sit there and try to replicate your setup. The root
cause of the problem will need to be identified to get it fixed.
How can I find doubles?
When clicking on issues tab there is a search box. Enter keywords for your
issue to find previous reports matching yours. Choose find all to match
closed tickets as well.
How can I investigate an incompatiblity?
Just mentioning a SIP server or device that does not interoperate with
Sipdroid won't help much because the developers might not have that
particular setup on hands. So you will try to add a SIP or logcat trace
from your network/phone to make a fix at all possible.
Did you carefully read above?
If yes, please fill out the following form and provide all necessary
details.
If not, you may continue in the Forum where you can discuss your problem to
clear things up before filing an issue report.
What steps will reproduce the problem?
1. Call a mobile phone with SipDroid 3.4 beta
2. Connection established. Voice totally scrambled. Can't understand a word.
3. Do the call other way round (from mobile phone to SipDroid 3.4) voice is ok.
What is the expected output? What do you see instead?
Don't see any error. Terrible voice when SipDroid calls "out". When receiving
calls, voice is ok.
What version of the product are you using? On what device/operating system?
Using SipDroid 3.4 beta on a Google Nexus 10 with 4.4 (KitKat).
Which SIP server are you using? What happens with PBXes?
VOIP is local FritzBox Cable 6360.
Not yet tested with PBXes
Which type of network are you using?
Using my local WLAN for testing. SipDroid conneted directly via WLAN to VOIP
server
Please provide any additional information below.
Device is a present for my girlfriend. I have to do secret testing until next
week. Will try different VOIP client to check, if there is the same problem.
Kind regards,
Robert
```
Original issue reported on code.google.com by `pasch...@googlemail.com` on 24 Nov 2013 at 4:23
| 1.0 | Scrambled void when calling, but voice ok when being called. - ```
Before you begin please read the FAQ!
Attention!
Don't post doubles!
Why is it important to find doubles?
Sipdroid is a community project. This is no dropbox for your personal
issue. You need to find others having the same problem to get it fixed. If
the problem has already been reported you add more details and ideas to the
existing discussion. If it has not been reported yet you include as many
background info as possible to give the community a chance to follow up on
it. No one will just sit there and try to replicate your setup. The root
cause of the problem will need to be identified to get it fixed.
How can I find doubles?
When clicking on issues tab there is a search box. Enter keywords for your
issue to find previous reports matching yours. Choose find all to match
closed tickets as well.
How can I investigate an incompatiblity?
Just mentioning a SIP server or device that does not interoperate with
Sipdroid won't help much because the developers might not have that
particular setup on hands. So you will try to add a SIP or logcat trace
from your network/phone to make a fix at all possible.
Did you carefully read above?
If yes, please fill out the following form and provide all necessary
details.
If not, you may continue in the Forum where you can discuss your problem to
clear things up before filing an issue report.
What steps will reproduce the problem?
1. Call a mobile phone with SipDroid 3.4 beta
2. Connection established. Voice totally scrambled. Can't understand a word.
3. Do the call other way round (from mobile phone to SipDroid 3.4) voice is ok.
What is the expected output? What do you see instead?
Don't see any error. Terrible voice when SipDroid calls "out". When receiving
calls, voice is ok.
What version of the product are you using? On what device/operating system?
Using SipDroid 3.4 beta on a Google Nexus 10 with 4.4 (KitKat).
Which SIP server are you using? What happens with PBXes?
VOIP is local FritzBox Cable 6360.
Not yet tested with PBXes
Which type of network are you using?
Using my local WLAN for testing. SipDroid conneted directly via WLAN to VOIP
server
Please provide any additional information below.
Device is a present for my girlfriend. I have to do secret testing until next
week. Will try different VOIP client to check, if there is the same problem.
Kind regards,
Robert
```
Original issue reported on code.google.com by `pasch...@googlemail.com` on 24 Nov 2013 at 4:23
| defect | scrambled void when calling but voice ok when being called before you begin please read the faq attention don t post doubles why is it important to find doubles sipdroid is a community project this is no dropbox for your personal issue you need to find others having the same problem to get it fixed if the problem has already been reported you add more details and ideas to the existing discussion if it has not been reported yet you include as many background info as possible to give the community a chance to follow up on it no one will just sit there and try to replicate your setup the root cause of the problem will need to be identified to get it fixed how can i find doubles when clicking on issues tab there is a search box enter keywords for your issue to find previous reports matching yours choose find all to match closed tickets as well how can i investigate an incompatiblity just mentioning a sip server or device that does not interoperate with sipdroid won t help much because the developers might not have that particular setup on hands so you will try to add a sip or logcat trace from your network phone to make a fix at all possible did you carefully read above if yes please fill out the following form and provide all necessary details if not you may continue in the forum where you can discuss your problem to clear things up before filing an issue report what steps will reproduce the problem call a mobile phone with sipdroid beta connection established voice totally scrambled can t understand a word do the call other way round from mobile phone to sipdroid voice is ok what is the expected output what do you see instead don t see any error terrible voice when sipdroid calls out when receiving calls voice is ok what version of the product are you using on what device operating system using sipdroid beta on a google nexus with kitkat which sip server are you using what happens with pbxes voip is local fritzbox cable not yet tested with pbxes which type of network are you using using my local wlan for testing sipdroid conneted directly via wlan to voip server please provide any additional information below device is a present for my girlfriend i have to do secret testing until next week will try different voip client to check if there is the same problem kind regards robert original issue reported on code google com by pasch googlemail com on nov at | 1 |
145,879 | 19,359,009,484 | IssuesEvent | 2021-12-16 01:21:32 | tctc008/WebGoat-develop | https://api.github.com/repos/tctc008/WebGoat-develop | opened | WS-2021-0491 (Medium) detected in logback-classic-1.2.5.jar | security vulnerability | ## WS-2021-0491 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>logback-classic-1.2.5.jar</b></p></summary>
<p>logback-classic module</p>
<p>Library home page: <a href="http://logback.qos.ch">http://logback.qos.ch</a></p>
<p>Path to dependency file: WebGoat-develop/webgoat-server/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/ch/qos/logback/logback-classic/1.2.5/logback-classic-1.2.5.jar,/home/wss-scanner/.m2/repository/ch/qos/logback/logback-classic/1.2.5/logback-classic-1.2.5.jar,/home/wss-scanner/.m2/repository/ch/qos/logback/logback-classic/1.2.5/logback-classic-1.2.5.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-validation-2.5.4.jar (Root Library)
- spring-boot-starter-2.5.4.jar
- spring-boot-starter-logging-2.5.4.jar
- :x: **logback-classic-1.2.5.jar** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
LOGBack before 1.2.8 is vulnerable to Remote-Code-Execution (RCE) when the write access to 'logback.xml' and JNDI lookup are enabled.
<p>Publish Date: 2021-12-13
<p>URL: <a href=https://github.com/qos-ch/logback/commit/21d772f2bc2ed780b01b4fe108df7e29707763f1>WS-2021-0491</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://logback.qos.ch/news.html">http://logback.qos.ch/news.html</a></p>
<p>Release Date: 2021-12-13</p>
<p>Fix Resolution: ch.qos.logback:logback-classic:1.2.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2021-0491 (Medium) detected in logback-classic-1.2.5.jar - ## WS-2021-0491 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>logback-classic-1.2.5.jar</b></p></summary>
<p>logback-classic module</p>
<p>Library home page: <a href="http://logback.qos.ch">http://logback.qos.ch</a></p>
<p>Path to dependency file: WebGoat-develop/webgoat-server/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/ch/qos/logback/logback-classic/1.2.5/logback-classic-1.2.5.jar,/home/wss-scanner/.m2/repository/ch/qos/logback/logback-classic/1.2.5/logback-classic-1.2.5.jar,/home/wss-scanner/.m2/repository/ch/qos/logback/logback-classic/1.2.5/logback-classic-1.2.5.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-validation-2.5.4.jar (Root Library)
- spring-boot-starter-2.5.4.jar
- spring-boot-starter-logging-2.5.4.jar
- :x: **logback-classic-1.2.5.jar** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
LOGBack before 1.2.8 is vulnerable to Remote-Code-Execution (RCE) when the write access to 'logback.xml' and JNDI lookup are enabled.
<p>Publish Date: 2021-12-13
<p>URL: <a href=https://github.com/qos-ch/logback/commit/21d772f2bc2ed780b01b4fe108df7e29707763f1>WS-2021-0491</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://logback.qos.ch/news.html">http://logback.qos.ch/news.html</a></p>
<p>Release Date: 2021-12-13</p>
<p>Fix Resolution: ch.qos.logback:logback-classic:1.2.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | ws medium detected in logback classic jar ws medium severity vulnerability vulnerable library logback classic jar logback classic module library home page a href path to dependency file webgoat develop webgoat server pom xml path to vulnerable library home wss scanner repository ch qos logback logback classic logback classic jar home wss scanner repository ch qos logback logback classic logback classic jar home wss scanner repository ch qos logback logback classic logback classic jar dependency hierarchy spring boot starter validation jar root library spring boot starter jar spring boot starter logging jar x logback classic jar vulnerable library found in base branch main vulnerability details logback before is vulnerable to remote code execution rce when the write access to logback xml and jndi lookup are enabled publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ch qos logback logback classic step up your open source security game with whitesource | 0 |
72,714 | 9,605,573,602 | IssuesEvent | 2019-05-11 01:20:22 | madou/yubaba | https://api.github.com/repos/madou/yubaba | closed | Fix UI/UX problems in docs | tag: documentation | - Johhny pic should act more like a modal
- email threads zindex on banner overlaps with navigation
- capitals search back button isn't visible enough
- conceal/reveal back buttons isn't visible enough
- maybe animations should trigger automatically 🤔 | 1.0 | Fix UI/UX problems in docs - - Johhny pic should act more like a modal
- email threads zindex on banner overlaps with navigation
- capitals search back button isn't visible enough
- conceal/reveal back buttons isn't visible enough
- maybe animations should trigger automatically 🤔 | non_defect | fix ui ux problems in docs johhny pic should act more like a modal email threads zindex on banner overlaps with navigation capitals search back button isn t visible enough conceal reveal back buttons isn t visible enough maybe animations should trigger automatically 🤔 | 0 |
292,952 | 22,040,147,157 | IssuesEvent | 2022-05-29 08:10:47 | NotSujal/Neptune.js | https://api.github.com/repos/NotSujal/Neptune.js | closed | Try it Yourself | documentation enhancement | Add a try it yourself editor in the documentation.
Which can quickly run the code on the output panel/frame. | 1.0 | Try it Yourself - Add a try it yourself editor in the documentation.
Which can quickly run the code on the output panel/frame. | non_defect | try it yourself add a try it yourself editor in the documentation which can quickly run the code on the output panel frame | 0 |
179,694 | 14,709,104,291 | IssuesEvent | 2021-01-05 01:28:57 | microsoft/Project-Santa-Cruz-Preview | https://api.github.com/repos/microsoft/Project-Santa-Cruz-Preview | closed | cable not supplied | documentation | In https://github.com/microsoft/Project-Santa-Cruz-Preview/blob/main/user-guides/updating/usb_updating.md
in Method 2 Step 8 says that the usb-c to usb-A cable is supplied, that cable doesn't come with the dev kit (at least not with mine), also a picture showing the connection would be very useful.
| 1.0 | cable not supplied - In https://github.com/microsoft/Project-Santa-Cruz-Preview/blob/main/user-guides/updating/usb_updating.md
in Method 2 Step 8 says that the usb-c to usb-A cable is supplied, that cable doesn't come with the dev kit (at least not with mine), also a picture showing the connection would be very useful.
| non_defect | cable not supplied in in method step says that the usb c to usb a cable is supplied that cable doesn t come with the dev kit at least not with mine also a picture showing the connection would be very useful | 0 |
80,509 | 30,311,011,525 | IssuesEvent | 2023-07-10 12:49:48 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Possible race condition in leave room slash command | T-Defect P2 A-Slash-Commands X-Needs-Investigation |
```
2020-11-09T18:28:56.259Z E Command failure: %s Cannot read property 'message' of null
TypeError: Cannot read property 'message' of null
at leaveRoomBehaviour (webpack-internal:///544:144:35)
at async SendMessageComposer._runSlashCommand (webpack-internal:///1655:336:9)
``` | 1.0 | Possible race condition in leave room slash command -
```
2020-11-09T18:28:56.259Z E Command failure: %s Cannot read property 'message' of null
TypeError: Cannot read property 'message' of null
at leaveRoomBehaviour (webpack-internal:///544:144:35)
at async SendMessageComposer._runSlashCommand (webpack-internal:///1655:336:9)
``` | defect | possible race condition in leave room slash command e command failure s cannot read property message of null typeerror cannot read property message of null at leaveroombehaviour webpack internal at async sendmessagecomposer runslashcommand webpack internal | 1 |
9,185 | 8,554,065,337 | IssuesEvent | 2018-11-08 04:05:08 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | Number in strictFilter Value results in "no good match" | cognitive-services/svc cxp product-question triaged | we sync the qna maker with an internal system. so we added a metadata syncitemid with a number as value, e.g. "name":"syncitemid", "value":"123"
after we added this metadata to all our questions in the qnamaker, we were getting always "no good match" result from the qna REST service - with and without using a strictFilter in query. Changing the value to "_123" fixed that and we where getting the question/answer with and without using a strictFilter. Changing it back to "123" and we are getting "no good match".
Does the qna maker REST service check if the value is a number or try to parse it?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: cb1d6a8c-954c-9ce8-6101-c19b20724c91
* Version Independent ID: 5c219f6c-260b-c17a-7f41-1bf4095aa521
* Content: [Metadata with GenerateAnswer API - QnA Maker - Azure Cognitive Services](https://docs.microsoft.com/en-us/azure//cognitive-services/qnamaker/how-to/metadata-generateanswer-usage)
* Content Source: [articles/cognitive-services/QnAMaker/How-To/metadata-generateanswer-usage.md](https://github.com/Microsoft/azure-docs/blob/master/articles/cognitive-services/QnAMaker/How-To/metadata-generateanswer-usage.md)
* Service: **cognitive-services**
* GitHub Login: @tulasim88
* Microsoft Alias: **tulasim88** | 1.0 | Number in strictFilter Value results in "no good match" - we sync the qna maker with an internal system. so we added a metadata syncitemid with a number as value, e.g. "name":"syncitemid", "value":"123"
after we added this metadata to all our questions in the qnamaker, we were getting always "no good match" result from the qna REST service - with and without using a strictFilter in query. Changing the value to "_123" fixed that and we where getting the question/answer with and without using a strictFilter. Changing it back to "123" and we are getting "no good match".
Does the qna maker REST service check if the value is a number or try to parse it?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: cb1d6a8c-954c-9ce8-6101-c19b20724c91
* Version Independent ID: 5c219f6c-260b-c17a-7f41-1bf4095aa521
* Content: [Metadata with GenerateAnswer API - QnA Maker - Azure Cognitive Services](https://docs.microsoft.com/en-us/azure//cognitive-services/qnamaker/how-to/metadata-generateanswer-usage)
* Content Source: [articles/cognitive-services/QnAMaker/How-To/metadata-generateanswer-usage.md](https://github.com/Microsoft/azure-docs/blob/master/articles/cognitive-services/QnAMaker/How-To/metadata-generateanswer-usage.md)
* Service: **cognitive-services**
* GitHub Login: @tulasim88
* Microsoft Alias: **tulasim88** | non_defect | number in strictfilter value results in no good match we sync the qna maker with an internal system so we added a metadata syncitemid with a number as value e g name syncitemid value after we added this metadata to all our questions in the qnamaker we were getting always no good match result from the qna rest service with and without using a strictfilter in query changing the value to fixed that and we where getting the question answer with and without using a strictfilter changing it back to and we are getting no good match does the qna maker rest service check if the value is a number or try to parse it document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service cognitive services github login microsoft alias | 0 |
477,430 | 13,762,182,257 | IssuesEvent | 2020-10-07 08:47:40 | Assetory/landingpage | https://api.github.com/repos/Assetory/landingpage | opened | Find header slogans | Priority: High | The landing page needs a slogan. It should meet these criteria:
* Reflect the identity of the brand
* Reflect the functionality of the product
* Fit the header image
* Be concise
* Be memorable
Example might be: We store worlds. | 1.0 | Find header slogans - The landing page needs a slogan. It should meet these criteria:
* Reflect the identity of the brand
* Reflect the functionality of the product
* Fit the header image
* Be concise
* Be memorable
Example might be: We store worlds. | non_defect | find header slogans the landing page needs a slogan it should meet these criteria reflect the identity of the brand reflect the functionality of the product fit the header image be concise be memorable example might be we store worlds | 0 |
65,233 | 19,278,480,144 | IssuesEvent | 2021-12-10 14:36:43 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Desktop app crashes on logout | T-Defect S-Major Z-Rageshake A-Widgets Z-Soft-Crash A-Logout O-Uncommon | ### Steps to reproduce
1. Where are you starting? What can you see?
2. What do you click?
3. More steps…
### Outcome
What happened?
I was logging out when a issue screen appeared.
### Operating system
Windows
### Application version
Element version 1.9.6
### How did you install the app?
From the official page of Element
### Homeserver
matrix.org
### Will you send logs?
Yes | 1.0 | Desktop app crashes on logout - ### Steps to reproduce
1. Where are you starting? What can you see?
2. What do you click?
3. More steps…
### Outcome
What happened?
I was logging out when a issue screen appeared.
### Operating system
Windows
### Application version
Element version 1.9.6
### How did you install the app?
From the official page of Element
### Homeserver
matrix.org
### Will you send logs?
Yes | defect | desktop app crashes on logout steps to reproduce where are you starting what can you see what do you click more steps… outcome what happened i was logging out when a issue screen appeared operating system windows application version element version how did you install the app from the official page of element homeserver matrix org will you send logs yes | 1 |
786,665 | 27,661,626,030 | IssuesEvent | 2023-03-12 15:38:04 | marcus-grant/markipedia | https://api.github.com/repos/marcus-grant/markipedia | closed | Add TailwindCSS typography plugin and apply to prose like notes | enhancement priority | There's a [prose plugin for TailwindCSS](https://tailwindcss.com/docs/typography-plugin) that can be used to get jump-started with some pretty fantastic typographic defaults | 1.0 | Add TailwindCSS typography plugin and apply to prose like notes - There's a [prose plugin for TailwindCSS](https://tailwindcss.com/docs/typography-plugin) that can be used to get jump-started with some pretty fantastic typographic defaults | non_defect | add tailwindcss typography plugin and apply to prose like notes there s a that can be used to get jump started with some pretty fantastic typographic defaults | 0 |
28,891 | 5,416,430,789 | IssuesEvent | 2017-03-02 00:28:27 | jccastillo0007/eFacturaT | https://api.github.com/repos/jccastillo0007/eFacturaT | opened | Error al timbrar nómina por conector para asimilados - detenida la nómina del cliente | bug defect | Tipo de comprobante es NOMINA, pero no se cuenta con la información necesaria para generar el complemento
ES QUE SI MANDAS ASIMILADOS A SALARIOS EN EL ARCHIVO DE TEXTO, NO DEBE INCLUIR EL REGISTRO PATRONAL, Y AHÍ ES DONDE ESTÁ TRONANDO…
SI LO INCLUYES MARCA ERROR PARA ASMILADOS, YA QUE NO LO DEBE LLEVAR (SEGÚN YO DEL SAT)….
SI NO LO INCLUYES MANDA EL ERROR DE ARRIBA AL LOG… (SEGÚN YO MENSAJE DE ERROR DE FACTURAT)
YO CREO QUE TENDRÍAMOS QUE SABER SI ES ASMILADOS, CUÁLES DATOS TOMAR Y CUÁLES IGNORAR, QUE NO???
| 1.0 | Error al timbrar nómina por conector para asimilados - detenida la nómina del cliente - Tipo de comprobante es NOMINA, pero no se cuenta con la información necesaria para generar el complemento
ES QUE SI MANDAS ASIMILADOS A SALARIOS EN EL ARCHIVO DE TEXTO, NO DEBE INCLUIR EL REGISTRO PATRONAL, Y AHÍ ES DONDE ESTÁ TRONANDO…
SI LO INCLUYES MARCA ERROR PARA ASMILADOS, YA QUE NO LO DEBE LLEVAR (SEGÚN YO DEL SAT)….
SI NO LO INCLUYES MANDA EL ERROR DE ARRIBA AL LOG… (SEGÚN YO MENSAJE DE ERROR DE FACTURAT)
YO CREO QUE TENDRÍAMOS QUE SABER SI ES ASMILADOS, CUÁLES DATOS TOMAR Y CUÁLES IGNORAR, QUE NO???
| defect | error al timbrar nómina por conector para asimilados detenida la nómina del cliente tipo de comprobante es nomina pero no se cuenta con la información necesaria para generar el complemento es que si mandas asimilados a salarios en el archivo de texto no debe incluir el registro patronal y ahí es donde está tronando… si lo incluyes marca error para asmilados ya que no lo debe llevar según yo del sat … si no lo incluyes manda el error de arriba al log… según yo mensaje de error de facturat yo creo que tendríamos que saber si es asmilados cuáles datos tomar y cuáles ignorar que no | 1 |
25,758 | 4,440,470,452 | IssuesEvent | 2016-08-19 03:47:54 | zaproxy/zaproxy | https://api.github.com/repos/zaproxy/zaproxy | closed | Check for Updates not able to download the new version from Help | Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1.Install older version;say 2.2.2
2.When i click or go to help and check for updates,it shows new version updates 2.3.1
3.A dialog box appears to launch the file and close zap
What is the expected output? What do you see instead?
Zap New version should be updates,but not able to download the new version
What version of the product are you using? On what operating system?
ZAP 2.2.2,win 7
Please provide any additional information below.
attached screenshot
```
Original issue reported on code.google.com by `nawaz.khan@mpowerglobal.com` on 2014-06-01 05:33:32
<hr>
* *Attachment: [zap_updates.jpg](https://storage.googleapis.com/google-code-attachments/zaproxy/issue-1228/comment-0/zap_updates.jpg)* | 1.0 | Check for Updates not able to download the new version from Help - ```
What steps will reproduce the problem?
1.Install older version;say 2.2.2
2.When i click or go to help and check for updates,it shows new version updates 2.3.1
3.A dialog box appears to launch the file and close zap
What is the expected output? What do you see instead?
Zap New version should be updates,but not able to download the new version
What version of the product are you using? On what operating system?
ZAP 2.2.2,win 7
Please provide any additional information below.
attached screenshot
```
Original issue reported on code.google.com by `nawaz.khan@mpowerglobal.com` on 2014-06-01 05:33:32
<hr>
* *Attachment: [zap_updates.jpg](https://storage.googleapis.com/google-code-attachments/zaproxy/issue-1228/comment-0/zap_updates.jpg)* | defect | check for updates not able to download the new version from help what steps will reproduce the problem install older version say when i click or go to help and check for updates it shows new version updates a dialog box appears to launch the file and close zap what is the expected output what do you see instead zap new version should be updates but not able to download the new version what version of the product are you using on what operating system zap win please provide any additional information below attached screenshot original issue reported on code google com by nawaz khan mpowerglobal com on attachment | 1 |
14,212 | 2,793,819,826 | IssuesEvent | 2015-05-11 13:37:16 | elecoest/allevents-3-2 | https://api.github.com/repos/elecoest/allevents-3-2 | closed | Blocage sur script jquery | auto-migrated Priority-Low Type-Defect | ```
AE version 3.2.13
1. Ouverture de la fiche d'un événement à partir d'un calendrier
2. Fermeture de la fiche d'un événement (avec ou sans modification)
3. Clic sur un événement dans la liste et blocage d'un script (voir image
jointe)
```
Original issue reported on code.google.com by `motos.A...@gmail.com` on 9 Mar 2015 at 8:41
Attachments:
* [Blocage script jquery.PNG](https://storage.googleapis.com/google-code-attachments/allevents-3-2/issue-473/comment-0/Blocage script jquery.PNG)
| 1.0 | Blocage sur script jquery - ```
AE version 3.2.13
1. Ouverture de la fiche d'un événement à partir d'un calendrier
2. Fermeture de la fiche d'un événement (avec ou sans modification)
3. Clic sur un événement dans la liste et blocage d'un script (voir image
jointe)
```
Original issue reported on code.google.com by `motos.A...@gmail.com` on 9 Mar 2015 at 8:41
Attachments:
* [Blocage script jquery.PNG](https://storage.googleapis.com/google-code-attachments/allevents-3-2/issue-473/comment-0/Blocage script jquery.PNG)
| defect | blocage sur script jquery ae version ouverture de la fiche d un événement à partir d un calendrier fermeture de la fiche d un événement avec ou sans modification clic sur un événement dans la liste et blocage d un script voir image jointe original issue reported on code google com by motos a gmail com on mar at attachments script jquery png | 1 |
39,439 | 9,459,749,618 | IssuesEvent | 2019-04-17 09:16:00 | primefaces/primeng | https://api.github.com/repos/primefaces/primeng | closed | Selecting row doesn't work with inplace | defect | **I'm submitting a ...** (check one with "x")
```
[ x] bug report => Search github for a similar issue or PR before submitting
[ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap
[ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35
```
**Plunkr Case (Bug Reports)**
https://stackblitz.com/edit/github-tnjgvt
**Current behavior**
1. Add inplace edit to table with selectable rows.
2. Click edit.
3. An error pops up in the console:
> ERROR TypeError: Cannot read property 'nodeName' of null
> at Table.handleRowClick (table.ts:761)
> at SelectableRow.onClick (table.ts:2608)
> at Object.eval [as handleEvent] (AppComponent.html:10)
> at handleEvent (view.ts:138)
> at callWithDebugContext (services.ts:629)
> at Object.debugHandleEvent [as handleEvent] (services.ts:376)
> at dispatchEvent (util.ts:134)
> at eval (element.ts:200)
> at HTMLTableRowElement.eval (dom_renderer.ts:52)
> at ZoneDelegate.invokeTask (zone.js:423)
**Expected behavior**
no error
**Minimal reproduction of the problem with instructions**
as above
* **Angular version:** 7.2.6
<!-- Check whether this is still an issue in the most recent Angular version -->
* **PrimeNG version:** 7.0.5
<!-- Check whether this is still an issue in the most recent Angular version -->
* **Browser:** all
<!-- All browsers where this could be reproduced -->
* **Language:** TS
| 1.0 | Selecting row doesn't work with inplace - **I'm submitting a ...** (check one with "x")
```
[ x] bug report => Search github for a similar issue or PR before submitting
[ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap
[ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35
```
**Plunkr Case (Bug Reports)**
https://stackblitz.com/edit/github-tnjgvt
**Current behavior**
1. Add inplace edit to table with selectable rows.
2. Click edit.
3. An error pops up in the console:
> ERROR TypeError: Cannot read property 'nodeName' of null
> at Table.handleRowClick (table.ts:761)
> at SelectableRow.onClick (table.ts:2608)
> at Object.eval [as handleEvent] (AppComponent.html:10)
> at handleEvent (view.ts:138)
> at callWithDebugContext (services.ts:629)
> at Object.debugHandleEvent [as handleEvent] (services.ts:376)
> at dispatchEvent (util.ts:134)
> at eval (element.ts:200)
> at HTMLTableRowElement.eval (dom_renderer.ts:52)
> at ZoneDelegate.invokeTask (zone.js:423)
**Expected behavior**
no error
**Minimal reproduction of the problem with instructions**
as above
* **Angular version:** 7.2.6
<!-- Check whether this is still an issue in the most recent Angular version -->
* **PrimeNG version:** 7.0.5
<!-- Check whether this is still an issue in the most recent Angular version -->
* **Browser:** all
<!-- All browsers where this could be reproduced -->
* **Language:** TS
| defect | selecting row doesn t work with inplace i m submitting a check one with x bug report search github for a similar issue or pr before submitting feature request please check if request is not on the roadmap already support request please do not submit support request here instead see plunkr case bug reports current behavior add inplace edit to table with selectable rows click edit an error pops up in the console error typeerror cannot read property nodename of null at table handlerowclick table ts at selectablerow onclick table ts at object eval appcomponent html at handleevent view ts at callwithdebugcontext services ts at object debughandleevent services ts at dispatchevent util ts at eval element ts at htmltablerowelement eval dom renderer ts at zonedelegate invoketask zone js expected behavior no error minimal reproduction of the problem with instructions as above angular version primeng version browser all language ts | 1 |
299,509 | 22,610,572,109 | IssuesEvent | 2022-06-29 16:47:32 | spring-gradle-plugins/dependency-management-plugin | https://api.github.com/repos/spring-gradle-plugins/dependency-management-plugin | closed | Modernise the look and feel of the reference documentation | type: documentation | We should switch to the Spring Asciidoctor backend. | 1.0 | Modernise the look and feel of the reference documentation - We should switch to the Spring Asciidoctor backend. | non_defect | modernise the look and feel of the reference documentation we should switch to the spring asciidoctor backend | 0 |
107,579 | 11,565,248,512 | IssuesEvent | 2020-02-20 10:09:46 | balajirama/Text-Parser | https://api.github.com/repos/balajirama/Text-Parser | closed | Make Text::Parser::RuleSpec module | documentation enhancement test | - [x] Make class attributes here. `use MooseX::ClassAttribute`
- [x] Make an exported function named `applies_rule` in `Text::Parser::RuleSpec`
- [x] Rules can be "inherited" as if they are methods.
- [x] When you inherit a class from another class, all the rules of the base class should be cloned. | 1.0 | Make Text::Parser::RuleSpec module - - [x] Make class attributes here. `use MooseX::ClassAttribute`
- [x] Make an exported function named `applies_rule` in `Text::Parser::RuleSpec`
- [x] Rules can be "inherited" as if they are methods.
- [x] When you inherit a class from another class, all the rules of the base class should be cloned. | non_defect | make text parser rulespec module make class attributes here use moosex classattribute make an exported function named applies rule in text parser rulespec rules can be inherited as if they are methods when you inherit a class from another class all the rules of the base class should be cloned | 0 |
72,403 | 24,106,948,552 | IssuesEvent | 2022-09-20 08:16:14 | vector-im/element-ios | https://api.github.com/repos/vector-im/element-ios | closed | Bug: Logout and login as different user shows the previous users all chats. | T-Defect | ### Steps to reproduce
1. Login as user 1
2. You can see All chat list
3. Logout user1
4. login as user2
5. you can see the user1 All chat list
6. App crashes.
### Outcome
#### What did you expect?
User 2 All chats should be displayed
#### What happened instead?
User1 chats is listed.
### Your phone model
iPhone 11
### Operating system version
16 , 15.3.1
### Application version
New UI
### Homeserver
_No response_
### Will you send logs?
No | 1.0 | Bug: Logout and login as different user shows the previous users all chats. - ### Steps to reproduce
1. Login as user 1
2. You can see All chat list
3. Logout user1
4. login as user2
5. you can see the user1 All chat list
6. App crashes.
### Outcome
#### What did you expect?
User 2 All chats should be displayed
#### What happened instead?
User1 chats is listed.
### Your phone model
iPhone 11
### Operating system version
16 , 15.3.1
### Application version
New UI
### Homeserver
_No response_
### Will you send logs?
No | defect | bug logout and login as different user shows the previous users all chats steps to reproduce login as user you can see all chat list logout login as you can see the all chat list app crashes outcome what did you expect user all chats should be displayed what happened instead chats is listed your phone model iphone operating system version application version new ui homeserver no response will you send logs no | 1 |
17,833 | 3,013,296,464 | IssuesEvent | 2015-07-29 08:00:02 | playframework/playframework | https://api.github.com/repos/playframework/playframework | closed | Play should support HikariCP connectionInitSql setting | defect has-pr | Hi, this is an issue [from the mailing list](https://groups.google.com/forum/#!topic/play-framework/9o5HqozBDQg):
##### Support for the `connectionInitSql` setting
HikariCP does support a [`connectionInitSql` setting](https://github.com/brettwooldridge/HikariCP#infrequently-used),
which allows to let a command run before each connection is given to the application. This allows e.g. to switch the default schema or something else.
Play does support this as well with the `db.default.initSQL` [configuration key](https://www.playframework.com/documentation/2.4.2/SettingsJDBC), e.g. with BoneCP.
It looks like it doesn't support it currently with HikariCP, because there is no mapping for it in the [Play Configuration to HikariCP configuration mapper](https://github.com/playframework/playframework/blob/master/framework/src/play-jdbc/src/main/scala/play/api/db/HikariCPModule.scala#L128).
###### Sidenode: Split the ConnectionPool from the Configuration mapper?
It is more difficult to exchange this mapper than it should be, imho. What about letting the `HikariCPConnectionPool` depend on a `Provider<HikariConfig>` instead and bind it to the `HikariCPConfig` class (with minimal changes to fit the `Provider` interface).
This would allow to exchange the mapper just exchanging `HikariCPModule` to bind it differently without the need to reimplement the whole `ConnectionPool` class. | 1.0 | Play should support HikariCP connectionInitSql setting - Hi, this is an issue [from the mailing list](https://groups.google.com/forum/#!topic/play-framework/9o5HqozBDQg):
##### Support for the `connectionInitSql` setting
HikariCP does support a [`connectionInitSql` setting](https://github.com/brettwooldridge/HikariCP#infrequently-used),
which allows to let a command run before each connection is given to the application. This allows e.g. to switch the default schema or something else.
Play does support this as well with the `db.default.initSQL` [configuration key](https://www.playframework.com/documentation/2.4.2/SettingsJDBC), e.g. with BoneCP.
It looks like it doesn't support it currently with HikariCP, because there is no mapping for it in the [Play Configuration to HikariCP configuration mapper](https://github.com/playframework/playframework/blob/master/framework/src/play-jdbc/src/main/scala/play/api/db/HikariCPModule.scala#L128).
###### Sidenode: Split the ConnectionPool from the Configuration mapper?
It is more difficult to exchange this mapper than it should be, imho. What about letting the `HikariCPConnectionPool` depend on a `Provider<HikariConfig>` instead and bind it to the `HikariCPConfig` class (with minimal changes to fit the `Provider` interface).
This would allow to exchange the mapper just exchanging `HikariCPModule` to bind it differently without the need to reimplement the whole `ConnectionPool` class. | defect | play should support hikaricp connectioninitsql setting hi this is an issue support for the connectioninitsql setting hikaricp does support a which allows to let a command run before each connection is given to the application this allows e g to switch the default schema or something else play does support this as well with the db default initsql e g with bonecp it looks like it doesn t support it currently with hikaricp because there is no mapping for it in the sidenode split the connectionpool from the configuration mapper it is more difficult to exchange this mapper than it should be imho what about letting the hikaricpconnectionpool depend on a provider instead and bind it to the hikaricpconfig class with minimal changes to fit the provider interface this would allow to exchange the mapper just exchanging hikaricpmodule to bind it differently without the need to reimplement the whole connectionpool class | 1 |
59,287 | 17,018,082,825 | IssuesEvent | 2021-07-02 14:42:31 | department-of-veterans-affairs/va.gov-team | https://api.github.com/repos/department-of-veterans-affairs/va.gov-team | opened | [Link and button labels] Button and link language isn't descriptive and/or their purpose isn't clear. | 508-defect-2 508/Accessibility Staging collab-cycle-feedback vsa-decision-tools | ### General Information
#### VFS team name
Decision Tools
#### VFS product name
Find a VA Form, Add Relevance, Popularity, Related to
#### Point of Contact/Reviewers
Angela Fowler
---
### Platform Issue
Button and link language isn't descriptive and/or their purpose isn't clear.
### Issue Details
On the form detail page About VA Form 10-10EZ, the text of the link to access the online form simply says "Go to the online tool." Since there are other forms listed on the page, this may cause confusion as to which online tool the user is going to.
### Link, screenshot or steps to recreate
### Supporting Guidance
Category 08, Issue 03
### Other References
WCAG SC 2.4.4_A
---
### Platform Recommendation
Ensure the test of the "Go to online tool" link on form detail pages include the form title.
### VFS Team Tasks to Complete
- [ ] Comment on the ticket if there are questions or concerns
- [ ] VFS team closes the ticket when the issue has been resolved | 1.0 | [Link and button labels] Button and link language isn't descriptive and/or their purpose isn't clear. - ### General Information
#### VFS team name
Decision Tools
#### VFS product name
Find a VA Form, Add Relevance, Popularity, Related to
#### Point of Contact/Reviewers
Angela Fowler
---
### Platform Issue
Button and link language isn't descriptive and/or their purpose isn't clear.
### Issue Details
On the form detail page About VA Form 10-10EZ, the text of the link to access the online form simply says "Go to the online tool." Since there are other forms listed on the page, this may cause confusion as to which online tool the user is going to.
### Link, screenshot or steps to recreate
### Supporting Guidance
Category 08, Issue 03
### Other References
WCAG SC 2.4.4_A
---
### Platform Recommendation
Ensure the test of the "Go to online tool" link on form detail pages include the form title.
### VFS Team Tasks to Complete
- [ ] Comment on the ticket if there are questions or concerns
- [ ] VFS team closes the ticket when the issue has been resolved | defect | button and link language isn t descriptive and or their purpose isn t clear general information vfs team name decision tools vfs product name find a va form add relevance popularity related to point of contact reviewers angela fowler platform issue button and link language isn t descriptive and or their purpose isn t clear issue details on the form detail page about va form the text of the link to access the online form simply says go to the online tool since there are other forms listed on the page this may cause confusion as to which online tool the user is going to link screenshot or steps to recreate supporting guidance category issue other references wcag sc a platform recommendation ensure the test of the go to online tool link on form detail pages include the form title vfs team tasks to complete comment on the ticket if there are questions or concerns vfs team closes the ticket when the issue has been resolved | 1 |
62,808 | 17,200,370,367 | IssuesEvent | 2021-07-17 05:02:48 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | opened | Element defaults to white theme on startup | T-Defect | <!-- A picture's worth a thousand words: PLEASE INCLUDE A SCREENSHOT :P -->

#### Description
Locale change happening on startup every time, causing the theme to not stay set, always resetting to white when restarting element.
#### Steps to reproduce
I don't know what exactly i've done, because I can't fix it either, but, it appears that it's trying to set the language to en-us when it should be en_US
Describe how what happens differs from what you expected.
`No update_base_url is defined: auto update is disabled
Fetching translation json for locale: en_EN
Changing application language to en-us
Fetching translation json for locale: en-us
Could not fetch translation json for locale: 'en-us' Error: Cannot find module './i18n/strings/en-us.json'
Require stack:
- /usr/lib/element/app.asar/src/language-helper.js
- /usr/lib/element/app.asar/src/tray.js
- /usr/lib/element/app.asar/src/electron-main.js
- /usr/lib/electron12/resources/default_app.asar/main.js
-
at Module._resolveFilename (internal/modules/cjs/loader.js:887:15)
at Function.n._resolveFilename (electron/js2c/browser_init.js:261:1128)
at Module._load (internal/modules/cjs/loader.js:732:27)
at Function.f._load (electron/js2c/asar_bundle.js:5:12913)
at Module.require (internal/modules/cjs/loader.js:959:19)
at require (internal/modules/cjs/helpers.js:88:18)
at AppLocalization.fetchTranslationJson (/usr/lib/element/app.asar/src/language-helper.js:83:20)
at /usr/lib/element/app.asar/src/language-helper.js:102:39
at Array.forEach (<anonymous>)
at AppLocalization.setAppLocale (/usr/lib/element/app.asar/src/language-helper.js:101:17) {
code: 'MODULE_NOT_FOUND',
requireStack: [
'/usr/lib/element/app.asar/src/language-helper.js',
'/usr/lib/element/app.asar/src/tray.js',
'/usr/lib/element/app.asar/src/electron-main.js',
'/usr/lib/electron12/resources/default_app.asar/main.js',
undefined
]
}
Resetting the UI components after locale change
Resetting the UI components after locale change
Changing application language to en-us
Fetching translation json for locale: en-us
Could not fetch translation json for locale: 'en-us' Error: Cannot find module './i18n/strings/en-us.json'
Require stack:
- /usr/lib/element/app.asar/src/language-helper.js
- /usr/lib/element/app.asar/src/tray.js
- /usr/lib/element/app.asar/src/electron-main.js
- /usr/lib/electron12/resources/default_app.asar/main.js
-
at Module._resolveFilename (internal/modules/cjs/loader.js:887:15)
at Function.n._resolveFilename (electron/js2c/browser_init.js:261:1128)
at Module._load (internal/modules/cjs/loader.js:732:27)
at Function.f._load (electron/js2c/asar_bundle.js:5:12913)
at Module.require (internal/modules/cjs/loader.js:959:19)
at require (internal/modules/cjs/helpers.js:88:18)
at AppLocalization.fetchTranslationJson (/usr/lib/element/app.asar/src/language-helper.js:83:20)
at /usr/lib/element/app.asar/src/language-helper.js:102:39
at Array.forEach (<anonymous>)
at AppLocalization.setAppLocale (/usr/lib/element/app.asar/src/language-helper.js:101:17) {
code: 'MODULE_NOT_FOUND',
requireStack: [
'/usr/lib/element/app.asar/src/language-helper.js',
'/usr/lib/element/app.asar/src/tray.js',
'/usr/lib/element/app.asar/src/electron-main.js',
'/usr/lib/electron12/resources/default_app.asar/main.js',
undefined
]
}
Resetting the UI components after locale change
Error getting the event index passphrase out of the secret store [Error: The name is not activatable]
`
<!-- Include screenshots if possible: you can drag and drop images below. -->
#### Version information
i've tried latest arch linux package, element on the web, and it seems to be account-wide.
<!-- IMPORTANT: please answer the following questions, to help us narrow down the problem -->
- **Platform**: linux
- web (in-browser) or desktop? both the desktop app and firefox 90 were tested
For the web app:
- **OS**: Arch
- **URL**: develop.element.io / app.element.io / somewhere else? If a private server, what version of Element Web?
For the desktop app:
- **OS**: arch
| 1.0 | Element defaults to white theme on startup - <!-- A picture's worth a thousand words: PLEASE INCLUDE A SCREENSHOT :P -->

#### Description
Locale change happening on startup every time, causing the theme to not stay set, always resetting to white when restarting element.
#### Steps to reproduce
I don't know what exactly i've done, because I can't fix it either, but, it appears that it's trying to set the language to en-us when it should be en_US
Describe how what happens differs from what you expected.
`No update_base_url is defined: auto update is disabled
Fetching translation json for locale: en_EN
Changing application language to en-us
Fetching translation json for locale: en-us
Could not fetch translation json for locale: 'en-us' Error: Cannot find module './i18n/strings/en-us.json'
Require stack:
- /usr/lib/element/app.asar/src/language-helper.js
- /usr/lib/element/app.asar/src/tray.js
- /usr/lib/element/app.asar/src/electron-main.js
- /usr/lib/electron12/resources/default_app.asar/main.js
-
at Module._resolveFilename (internal/modules/cjs/loader.js:887:15)
at Function.n._resolveFilename (electron/js2c/browser_init.js:261:1128)
at Module._load (internal/modules/cjs/loader.js:732:27)
at Function.f._load (electron/js2c/asar_bundle.js:5:12913)
at Module.require (internal/modules/cjs/loader.js:959:19)
at require (internal/modules/cjs/helpers.js:88:18)
at AppLocalization.fetchTranslationJson (/usr/lib/element/app.asar/src/language-helper.js:83:20)
at /usr/lib/element/app.asar/src/language-helper.js:102:39
at Array.forEach (<anonymous>)
at AppLocalization.setAppLocale (/usr/lib/element/app.asar/src/language-helper.js:101:17) {
code: 'MODULE_NOT_FOUND',
requireStack: [
'/usr/lib/element/app.asar/src/language-helper.js',
'/usr/lib/element/app.asar/src/tray.js',
'/usr/lib/element/app.asar/src/electron-main.js',
'/usr/lib/electron12/resources/default_app.asar/main.js',
undefined
]
}
Resetting the UI components after locale change
Resetting the UI components after locale change
Changing application language to en-us
Fetching translation json for locale: en-us
Could not fetch translation json for locale: 'en-us' Error: Cannot find module './i18n/strings/en-us.json'
Require stack:
- /usr/lib/element/app.asar/src/language-helper.js
- /usr/lib/element/app.asar/src/tray.js
- /usr/lib/element/app.asar/src/electron-main.js
- /usr/lib/electron12/resources/default_app.asar/main.js
-
at Module._resolveFilename (internal/modules/cjs/loader.js:887:15)
at Function.n._resolveFilename (electron/js2c/browser_init.js:261:1128)
at Module._load (internal/modules/cjs/loader.js:732:27)
at Function.f._load (electron/js2c/asar_bundle.js:5:12913)
at Module.require (internal/modules/cjs/loader.js:959:19)
at require (internal/modules/cjs/helpers.js:88:18)
at AppLocalization.fetchTranslationJson (/usr/lib/element/app.asar/src/language-helper.js:83:20)
at /usr/lib/element/app.asar/src/language-helper.js:102:39
at Array.forEach (<anonymous>)
at AppLocalization.setAppLocale (/usr/lib/element/app.asar/src/language-helper.js:101:17) {
code: 'MODULE_NOT_FOUND',
requireStack: [
'/usr/lib/element/app.asar/src/language-helper.js',
'/usr/lib/element/app.asar/src/tray.js',
'/usr/lib/element/app.asar/src/electron-main.js',
'/usr/lib/electron12/resources/default_app.asar/main.js',
undefined
]
}
Resetting the UI components after locale change
Error getting the event index passphrase out of the secret store [Error: The name is not activatable]
`
<!-- Include screenshots if possible: you can drag and drop images below. -->
#### Version information
i've tried latest arch linux package, element on the web, and it seems to be account-wide.
<!-- IMPORTANT: please answer the following questions, to help us narrow down the problem -->
- **Platform**: linux
- web (in-browser) or desktop? both the desktop app and firefox 90 were tested
For the web app:
- **OS**: Arch
- **URL**: develop.element.io / app.element.io / somewhere else? If a private server, what version of Element Web?
For the desktop app:
- **OS**: arch
| defect | element defaults to white theme on startup description locale change happening on startup every time causing the theme to not stay set always resetting to white when restarting element steps to reproduce i don t know what exactly i ve done because i can t fix it either but it appears that it s trying to set the language to en us when it should be en us describe how what happens differs from what you expected no update base url is defined auto update is disabled fetching translation json for locale en en changing application language to en us fetching translation json for locale en us could not fetch translation json for locale en us error cannot find module strings en us json require stack usr lib element app asar src language helper js usr lib element app asar src tray js usr lib element app asar src electron main js usr lib resources default app asar main js at module resolvefilename internal modules cjs loader js at function n resolvefilename electron browser init js at module load internal modules cjs loader js at function f load electron asar bundle js at module require internal modules cjs loader js at require internal modules cjs helpers js at applocalization fetchtranslationjson usr lib element app asar src language helper js at usr lib element app asar src language helper js at array foreach at applocalization setapplocale usr lib element app asar src language helper js code module not found requirestack usr lib element app asar src language helper js usr lib element app asar src tray js usr lib element app asar src electron main js usr lib resources default app asar main js undefined resetting the ui components after locale change resetting the ui components after locale change changing application language to en us fetching translation json for locale en us could not fetch translation json for locale en us error cannot find module strings en us json require stack usr lib element app asar src language helper js usr lib element app asar src tray js usr lib element app asar src electron main js usr lib resources default app asar main js at module resolvefilename internal modules cjs loader js at function n resolvefilename electron browser init js at module load internal modules cjs loader js at function f load electron asar bundle js at module require internal modules cjs loader js at require internal modules cjs helpers js at applocalization fetchtranslationjson usr lib element app asar src language helper js at usr lib element app asar src language helper js at array foreach at applocalization setapplocale usr lib element app asar src language helper js code module not found requirestack usr lib element app asar src language helper js usr lib element app asar src tray js usr lib element app asar src electron main js usr lib resources default app asar main js undefined resetting the ui components after locale change error getting the event index passphrase out of the secret store version information i ve tried latest arch linux package element on the web and it seems to be account wide platform linux web in browser or desktop both the desktop app and firefox were tested for the web app os arch url develop element io app element io somewhere else if a private server what version of element web for the desktop app os arch | 1 |
40,513 | 10,026,546,514 | IssuesEvent | 2019-07-17 06:58:53 | primefaces/primeng | https://api.github.com/repos/primefaces/primeng | reopened | Calendar needs double click when minDate/ MaxDate property is set. | defect pending-review | ### There is no guarantee in receiving a response in GitHub Issue Tracker, If you'd like to secure our response, you may consider *PrimeNG PRO Support* where support is provided within 4 business hours
**I'm submitting a ...** (check one with "x")
```
[x] bug report => Search github for a similar issue or PR before submitting
[ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap
[ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35
```
**Plunkr Case (Bug Reports)**
Please fork the plunkr below and create a case demonstrating your bug report. Issues without a plunkr have much less possibility to be reviewed.
http://plnkr.co/edit/ZXkmtf1OFKmQokyEdmwa?p=preview
**Current behavior**
<!-- Describe how the bug manifests. -->
When minDate or maxDate property is set, calendar needs two clicks to select date.
**Expected behavior**
<!-- Describe what the behavior would be without the bug. -->
It should select date in single click.
**Minimal reproduction of the problem with instructions**
<!--
If the current behavior is a bug or you can illustrate your feature request better with an example,
please provide the *STEPS TO REPRODUCE* and if possible a *MINIMAL DEMO* of the problem via
https://plnkr.co or similar (you can use this template as a starting point: http://plnkr.co/edit/tpl:AvJOMERrnz94ekVua0u5).
-->
**What is the motivation / use case for changing the behavior?**
<!-- Describe the motivation or the concrete use case -->
**Please tell us about your environment:**
<!-- Operating system, IDE, package manager, HTTP server, ... -->
* **Angular version:** 4.2.4
<!-- Check whether this is still an issue in the most recent Angular version -->
* **PrimeNG version:** 4.3.0
<!-- Check whether this is still an issue in the most recent Angular version -->
* **Browser:** [ Chrome 62.0.3202.94]
<!-- All browsers where this could be reproduced -->
* **Language:** [all]
* **Node (for AoT issues):** `node --version` =
| 1.0 | Calendar needs double click when minDate/ MaxDate property is set. - ### There is no guarantee in receiving a response in GitHub Issue Tracker, If you'd like to secure our response, you may consider *PrimeNG PRO Support* where support is provided within 4 business hours
**I'm submitting a ...** (check one with "x")
```
[x] bug report => Search github for a similar issue or PR before submitting
[ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap
[ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35
```
**Plunkr Case (Bug Reports)**
Please fork the plunkr below and create a case demonstrating your bug report. Issues without a plunkr have much less possibility to be reviewed.
http://plnkr.co/edit/ZXkmtf1OFKmQokyEdmwa?p=preview
**Current behavior**
<!-- Describe how the bug manifests. -->
When minDate or maxDate property is set, calendar needs two clicks to select date.
**Expected behavior**
<!-- Describe what the behavior would be without the bug. -->
It should select date in single click.
**Minimal reproduction of the problem with instructions**
<!--
If the current behavior is a bug or you can illustrate your feature request better with an example,
please provide the *STEPS TO REPRODUCE* and if possible a *MINIMAL DEMO* of the problem via
https://plnkr.co or similar (you can use this template as a starting point: http://plnkr.co/edit/tpl:AvJOMERrnz94ekVua0u5).
-->
**What is the motivation / use case for changing the behavior?**
<!-- Describe the motivation or the concrete use case -->
**Please tell us about your environment:**
<!-- Operating system, IDE, package manager, HTTP server, ... -->
* **Angular version:** 4.2.4
<!-- Check whether this is still an issue in the most recent Angular version -->
* **PrimeNG version:** 4.3.0
<!-- Check whether this is still an issue in the most recent Angular version -->
* **Browser:** [ Chrome 62.0.3202.94]
<!-- All browsers where this could be reproduced -->
* **Language:** [all]
* **Node (for AoT issues):** `node --version` =
| defect | calendar needs double click when mindate maxdate property is set there is no guarantee in receiving a response in github issue tracker if you d like to secure our response you may consider primeng pro support where support is provided within business hours i m submitting a check one with x bug report search github for a similar issue or pr before submitting feature request please check if request is not on the roadmap already support request please do not submit support request here instead see plunkr case bug reports please fork the plunkr below and create a case demonstrating your bug report issues without a plunkr have much less possibility to be reviewed current behavior when mindate or maxdate property is set calendar needs two clicks to select date expected behavior it should select date in single click minimal reproduction of the problem with instructions if the current behavior is a bug or you can illustrate your feature request better with an example please provide the steps to reproduce and if possible a minimal demo of the problem via or similar you can use this template as a starting point what is the motivation use case for changing the behavior please tell us about your environment angular version primeng version browser language node for aot issues node version | 1 |
313,512 | 26,934,330,761 | IssuesEvent | 2023-02-07 19:16:56 | MPMG-DCC-UFMG/F01 | https://api.github.com/repos/MPMG-DCC-UFMG/F01 | closed | Teste de generalizacao para a tag Obras Públicas - Dados para Acompanhamento - Careaçu | generalization test development template - PT (45) tag - Obras Públicas subtag - Dados para acompanhamento | DoD: Realizar o teste de Generalização do validador da tag Obras Públicas - Dados para Acompanhamento para o Município de Careaçu. | 1.0 | Teste de generalizacao para a tag Obras Públicas - Dados para Acompanhamento - Careaçu - DoD: Realizar o teste de Generalização do validador da tag Obras Públicas - Dados para Acompanhamento para o Município de Careaçu. | non_defect | teste de generalizacao para a tag obras públicas dados para acompanhamento careaçu dod realizar o teste de generalização do validador da tag obras públicas dados para acompanhamento para o município de careaçu | 0 |
343,020 | 10,324,455,370 | IssuesEvent | 2019-09-01 09:18:14 | wso2/product-is | https://api.github.com/repos/wso2/product-is | opened | Audience and recipient validation failed when not using the port (443) | Affected/5.9.0-Alpha Priority/Normal Severity/Minor | Audience and recipient validation failed when not using the port (443) https://is590.qa.com:443/oauth2/token
Steps to reproduce:
- Configure saml bearer grant using the using [1]
- Identity Server is started with 443 port instead of 9443.
- Ideally Audience and Recipient should be configured as follows.
Audience : https://localhost/oauth2/token
Recipient : https://localhost/oauth2/token
But it won't as above. It is mandatory to provide port although it is the default HTTPS port.
[1] https://is.docs.wso2.com/en/5.9.0/using-wso2-identity-server/setting-up-a-saml2-bearer-assertion-profile-for-oauth-2.0/
| 1.0 | Audience and recipient validation failed when not using the port (443) - Audience and recipient validation failed when not using the port (443) https://is590.qa.com:443/oauth2/token
Steps to reproduce:
- Configure saml bearer grant using the using [1]
- Identity Server is started with 443 port instead of 9443.
- Ideally Audience and Recipient should be configured as follows.
Audience : https://localhost/oauth2/token
Recipient : https://localhost/oauth2/token
But it won't as above. It is mandatory to provide port although it is the default HTTPS port.
[1] https://is.docs.wso2.com/en/5.9.0/using-wso2-identity-server/setting-up-a-saml2-bearer-assertion-profile-for-oauth-2.0/
| non_defect | audience and recipient validation failed when not using the port audience and recipient validation failed when not using the port steps to reproduce configure saml bearer grant using the using identity server is started with port instead of ideally audience and recipient should be configured as follows audience recipient but it won t as above it is mandatory to provide port although it is the default https port | 0 |
37,167 | 8,273,011,434 | IssuesEvent | 2018-09-17 02:27:02 | cakephp/cakephp | https://api.github.com/repos/cakephp/cakephp | closed | Unexpected Cake\Filesystem\File::copy behavior | Defect | This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.5.17
* Platform and Target: Ubuntu 18.04 LTS, Apache/2.4.33, PHP 7.0.31-1+ubuntu18.04.1+deb.sury.org+1
### What you did
I had expected a file copy to create a copy in the the files current directory, instead the copy was created at the webroot. While I have no problem specifying a full path, I find this to be odd behavior. Not sure if a change is worth it or even possible due to BC.
```php
$objFile = new File('/var/some/path/to/image.jpg');
$objFile->copy('image_copy.jpg');
```
### What happened
Copy was created at `webroot/`
### What you expected to happen
Copy to be created at same working directory as the original.
I don't actually expect this to be addressed because of BC, but at a minimum the documentation should specify File::copy() expects an absolute path as its argument. | 1.0 | Unexpected Cake\Filesystem\File::copy behavior - This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.5.17
* Platform and Target: Ubuntu 18.04 LTS, Apache/2.4.33, PHP 7.0.31-1+ubuntu18.04.1+deb.sury.org+1
### What you did
I had expected a file copy to create a copy in the the files current directory, instead the copy was created at the webroot. While I have no problem specifying a full path, I find this to be odd behavior. Not sure if a change is worth it or even possible due to BC.
```php
$objFile = new File('/var/some/path/to/image.jpg');
$objFile->copy('image_copy.jpg');
```
### What happened
Copy was created at `webroot/`
### What you expected to happen
Copy to be created at same working directory as the original.
I don't actually expect this to be addressed because of BC, but at a minimum the documentation should specify File::copy() expects an absolute path as its argument. | defect | unexpected cake filesystem file copy behavior this is a multiple allowed bug enhancement feature discussion rfc cakephp version platform and target ubuntu lts apache php deb sury org what you did i had expected a file copy to create a copy in the the files current directory instead the copy was created at the webroot while i have no problem specifying a full path i find this to be odd behavior not sure if a change is worth it or even possible due to bc php objfile new file var some path to image jpg objfile copy image copy jpg what happened copy was created at webroot what you expected to happen copy to be created at same working directory as the original i don t actually expect this to be addressed because of bc but at a minimum the documentation should specify file copy expects an absolute path as its argument | 1 |
75,606 | 25,942,960,946 | IssuesEvent | 2022-12-16 20:29:16 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | opened | [heads-up | FYI] zfs-tests-functional + zloop / tests (push) - within the last few days all seem to fail | Type: Defect | ### System information
Distribution Name | Ubuntu 20.04 , Ubuntu 22.04
Distribution Version |
Kernel Version |
Architecture | amd64
OpenZFS Version | upstream
Since I had to update my system, I went with the most recent appearing "stable" commit, namely "Bypass metaslab throttle for removal allocations", 7bf4c97a3696663959d1891c5890b2667761dd58
Commit: https://github.com/openzfs/zfs/commit/242a5b748cbf02001a1324f3e195df56f8f28f5a "Fix dereference after null check in enqueue_range" from December 8th - is the latest commit which doesn't show any errors - at all
CC: @behlendorf @ryao | 1.0 | [heads-up | FYI] zfs-tests-functional + zloop / tests (push) - within the last few days all seem to fail - ### System information
Distribution Name | Ubuntu 20.04 , Ubuntu 22.04
Distribution Version |
Kernel Version |
Architecture | amd64
OpenZFS Version | upstream
Since I had to update my system, I went with the most recent appearing "stable" commit, namely "Bypass metaslab throttle for removal allocations", 7bf4c97a3696663959d1891c5890b2667761dd58
Commit: https://github.com/openzfs/zfs/commit/242a5b748cbf02001a1324f3e195df56f8f28f5a "Fix dereference after null check in enqueue_range" from December 8th - is the latest commit which doesn't show any errors - at all
CC: @behlendorf @ryao | defect | zfs tests functional zloop tests push within the last few days all seem to fail system information distribution name ubuntu ubuntu distribution version kernel version architecture openzfs version upstream since i had to update my system i went with the most recent appearing stable commit namely bypass metaslab throttle for removal allocations commit fix dereference after null check in enqueue range from december is the latest commit which doesn t show any errors at all cc behlendorf ryao | 1 |
15,311 | 2,611,454,184 | IssuesEvent | 2015-02-27 05:01:18 | chrsmith/hedgewars | https://api.github.com/repos/chrsmith/hedgewars | closed | healthbar to indicate individual hog health | auto-migrated Component-UI Priority-Low Type-Enhancement | ```
there should be some small, non intrusive way to determine the health of each
member of the team.
See attachment for an example rendering.
```
Original issue reported on code.google.com by `vittorio...@gmail.com` on 26 Dec 2010 at 12:41
Attachments:
* [like.png](https://storage.googleapis.com/google-code-attachments/hedgewars/issue-133/comment-0/like.png)
| 1.0 | healthbar to indicate individual hog health - ```
there should be some small, non intrusive way to determine the health of each
member of the team.
See attachment for an example rendering.
```
Original issue reported on code.google.com by `vittorio...@gmail.com` on 26 Dec 2010 at 12:41
Attachments:
* [like.png](https://storage.googleapis.com/google-code-attachments/hedgewars/issue-133/comment-0/like.png)
| non_defect | healthbar to indicate individual hog health there should be some small non intrusive way to determine the health of each member of the team see attachment for an example rendering original issue reported on code google com by vittorio gmail com on dec at attachments | 0 |
7,261 | 2,610,360,783 | IssuesEvent | 2015-02-26 19:56:43 | chrsmith/scribefire-chrome | https://api.github.com/repos/chrsmith/scribefire-chrome | closed | word press login problem | auto-migrated Priority-Medium Type-Defect | ```
What's the problem?
my wordpress blog cant login in scribe fire next in my computer. after i
authorize the login its coming bad login password
What browser are you using?
fire fox 10
What version of ScribeFire are you running?
scribe fire next 4.0
```
-----
Original issue reported on code.google.com by `dmadanku...@gmail.com` on 10 Mar 2012 at 8:44 | 1.0 | word press login problem - ```
What's the problem?
my wordpress blog cant login in scribe fire next in my computer. after i
authorize the login its coming bad login password
What browser are you using?
fire fox 10
What version of ScribeFire are you running?
scribe fire next 4.0
```
-----
Original issue reported on code.google.com by `dmadanku...@gmail.com` on 10 Mar 2012 at 8:44 | defect | word press login problem what s the problem my wordpress blog cant login in scribe fire next in my computer after i authorize the login its coming bad login password what browser are you using fire fox what version of scribefire are you running scribe fire next original issue reported on code google com by dmadanku gmail com on mar at | 1 |
50,975 | 13,188,009,952 | IssuesEvent | 2020-08-13 05:17:44 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | closed | Software documentation links almost all lead to 403 - Forbidden (Trac #1758) | Migrated from Trac defect other | http://software.icecube.wisc.edu/documentation
Almost all links I've tried from "Project Details" heading and below go to
403 - Forbidden
clsim and children work, though, as well as links above "Project Details." I suspect this is because I visited these pages and my web browser cached them; trying from a browser from a server, all documentation links (including the documentation link itself) return 403 errors.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1758">https://code.icecube.wisc.edu/ticket/1758</a>, reported by jlanfranchi and owned by </em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-06-23T07:37:44",
"description": "http://software.icecube.wisc.edu/documentation\n\nAlmost all links I've tried from \"Project Details\" heading and below go to\n403 - Forbidden\nclsim and children work, though, as well as links above \"Project Details.\" I suspect this is because I visited these pages and my web browser cached them; trying from a browser from a server, all documentation links (including the documentation link itself) return 403 errors.",
"reporter": "jlanfranchi",
"cc": "",
"resolution": "invalid",
"_ts": "1466667464194375",
"component": "other",
"summary": "Software documentation links almost all lead to 403 - Forbidden",
"priority": "normal",
"keywords": "",
"time": "2016-06-23T04:42:28",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
| 1.0 | Software documentation links almost all lead to 403 - Forbidden (Trac #1758) - http://software.icecube.wisc.edu/documentation
Almost all links I've tried from "Project Details" heading and below go to
403 - Forbidden
clsim and children work, though, as well as links above "Project Details." I suspect this is because I visited these pages and my web browser cached them; trying from a browser from a server, all documentation links (including the documentation link itself) return 403 errors.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1758">https://code.icecube.wisc.edu/ticket/1758</a>, reported by jlanfranchi and owned by </em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-06-23T07:37:44",
"description": "http://software.icecube.wisc.edu/documentation\n\nAlmost all links I've tried from \"Project Details\" heading and below go to\n403 - Forbidden\nclsim and children work, though, as well as links above \"Project Details.\" I suspect this is because I visited these pages and my web browser cached them; trying from a browser from a server, all documentation links (including the documentation link itself) return 403 errors.",
"reporter": "jlanfranchi",
"cc": "",
"resolution": "invalid",
"_ts": "1466667464194375",
"component": "other",
"summary": "Software documentation links almost all lead to 403 - Forbidden",
"priority": "normal",
"keywords": "",
"time": "2016-06-23T04:42:28",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
| defect | software documentation links almost all lead to forbidden trac almost all links i ve tried from project details heading and below go to forbidden clsim and children work though as well as links above project details i suspect this is because i visited these pages and my web browser cached them trying from a browser from a server all documentation links including the documentation link itself return errors migrated from json status closed changetime description all links i ve tried from project details heading and below go to forbidden nclsim and children work though as well as links above project details i suspect this is because i visited these pages and my web browser cached them trying from a browser from a server all documentation links including the documentation link itself return errors reporter jlanfranchi cc resolution invalid ts component other summary software documentation links almost all lead to forbidden priority normal keywords time milestone owner type defect | 1 |
59,346 | 8,358,838,925 | IssuesEvent | 2018-10-03 05:33:16 | fga-eps-mds/2018.2-NaturalSearch | https://api.github.com/repos/fga-eps-mds/2018.2-NaturalSearch | closed | Definir Roadmap do arquiteto | Architect documentation eps | Definição do Roadmap relacionado ao papel de arquiteto do produto. | 1.0 | Definir Roadmap do arquiteto - Definição do Roadmap relacionado ao papel de arquiteto do produto. | non_defect | definir roadmap do arquiteto definição do roadmap relacionado ao papel de arquiteto do produto | 0 |
35,462 | 7,748,433,446 | IssuesEvent | 2018-05-30 08:19:19 | cakephp/cakephp | https://api.github.com/repos/cakephp/cakephp | opened | 3.6: BC Break: Shell in command runner gone | Defect | This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.6 latest
### What you did
Switched out the cake.php etc, and use the new command runner.
I also have an existing shell for some stuff to run on build:
```
bin/cake app run
```
### What happened
The shell is not found and therefore cannot be executed
### What you expected to happen
```
class AppShell extends Shell
```
should work.
Renaming it to e.g. MyAppShell works.
| 1.0 | 3.6: BC Break: Shell in command runner gone - This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.6 latest
### What you did
Switched out the cake.php etc, and use the new command runner.
I also have an existing shell for some stuff to run on build:
```
bin/cake app run
```
### What happened
The shell is not found and therefore cannot be executed
### What you expected to happen
```
class AppShell extends Shell
```
should work.
Renaming it to e.g. MyAppShell works.
| defect | bc break shell in command runner gone this is a multiple allowed bug enhancement feature discussion rfc cakephp version latest what you did switched out the cake php etc and use the new command runner i also have an existing shell for some stuff to run on build bin cake app run what happened the shell is not found and therefore cannot be executed what you expected to happen class appshell extends shell should work renaming it to e g myappshell works | 1 |
39,762 | 9,651,734,718 | IssuesEvent | 2019-05-18 10:44:47 | boxbackup/boxbackup | https://api.github.com/repos/boxbackup/boxbackup | closed | Updated bbackupd.conf file for Windows (Trac #23) | Migrated from Trac bbackupd ben defect | The bbackupd.conf file included in Chris' Windows clients (for example, boxbackup-chris_general_1569-backup-client-mingw32.zip) has some issues.
I attach an updated one for your consideration.
Migrated from https://www.boxbackup.org/ticket/23
```json
{
"status": "closed",
"changetime": "2007-11-05T00:35:01",
"description": "The bbackupd.conf file included in Chris' Windows clients (for example, boxbackup-chris_general_1569-backup-client-mingw32.zip) has some issues. \n\nI attach an updated one for your consideration. \n\n",
"reporter": "petej",
"cc": "",
"resolution": "fixed",
"_ts": "1194222901000000",
"component": "bbackupd",
"summary": "Updated bbackupd.conf file for Windows",
"priority": "trivial",
"keywords": "win32 bbackupd configuration sample notifyscript",
"version": "0.10",
"time": "2007-05-04T22:19:57",
"milestone": "0.11",
"owner": "ben",
"type": "defect"
}
```
| 1.0 | Updated bbackupd.conf file for Windows (Trac #23) - The bbackupd.conf file included in Chris' Windows clients (for example, boxbackup-chris_general_1569-backup-client-mingw32.zip) has some issues.
I attach an updated one for your consideration.
Migrated from https://www.boxbackup.org/ticket/23
```json
{
"status": "closed",
"changetime": "2007-11-05T00:35:01",
"description": "The bbackupd.conf file included in Chris' Windows clients (for example, boxbackup-chris_general_1569-backup-client-mingw32.zip) has some issues. \n\nI attach an updated one for your consideration. \n\n",
"reporter": "petej",
"cc": "",
"resolution": "fixed",
"_ts": "1194222901000000",
"component": "bbackupd",
"summary": "Updated bbackupd.conf file for Windows",
"priority": "trivial",
"keywords": "win32 bbackupd configuration sample notifyscript",
"version": "0.10",
"time": "2007-05-04T22:19:57",
"milestone": "0.11",
"owner": "ben",
"type": "defect"
}
```
| defect | updated bbackupd conf file for windows trac the bbackupd conf file included in chris windows clients for example boxbackup chris general backup client zip has some issues i attach an updated one for your consideration migrated from json status closed changetime description the bbackupd conf file included in chris windows clients for example boxbackup chris general backup client zip has some issues n ni attach an updated one for your consideration n n reporter petej cc resolution fixed ts component bbackupd summary updated bbackupd conf file for windows priority trivial keywords bbackupd configuration sample notifyscript version time milestone owner ben type defect | 1 |
61,043 | 17,023,586,620 | IssuesEvent | 2021-07-03 02:47:37 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | make clean is unable to clean all generated files if sudo make install was used | Component: tirex Priority: minor Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 3.06pm, Saturday, 8th May 2010]**
`sudo make install` generates files owned by root in the man-generated -directory, so when running `make clean` as normal user, the files cannot be removed. | 1.0 | make clean is unable to clean all generated files if sudo make install was used - **[Submitted to the original trac issue database at 3.06pm, Saturday, 8th May 2010]**
`sudo make install` generates files owned by root in the man-generated -directory, so when running `make clean` as normal user, the files cannot be removed. | defect | make clean is unable to clean all generated files if sudo make install was used sudo make install generates files owned by root in the man generated directory so when running make clean as normal user the files cannot be removed | 1 |
521,861 | 15,144,055,189 | IssuesEvent | 2021-02-11 00:21:44 | aims-group/metagrid | https://api.github.com/repos/aims-group/metagrid | closed | Facet option dropdown forms stays fixed while scrolling | Misc: Feedback Platform: React Priority: Low Type: Bug | **Describe the bug**
A clear and concise description of what the bug is.
The facet option dropdown forms stays fixed while scrolling.
**Desktop (please complete the following information):**
- OS: MacOS
- Browser: chrome
- Version: 86
**To Reproduce**
Steps to reproduce the behavior:
1. Select a project on the sidebar
2. Open a facet collapsable
3. Select the dropdown form
4. Scroll up and down (responsive to the size of your browser)
5. See form stays in fixed position and covers other elements
**Expected behavior**
A clear and concise description of what you expected to happen.
The select drop down should automatically minimize when it is out of the view.
Fix: https://github.com/ant-design/ant-design/issues/8268
**Screenshots**
If applicable, add screenshots to help explain your problem.

**Additional context**
Add any other context about the problem here.
| 1.0 | Facet option dropdown forms stays fixed while scrolling - **Describe the bug**
A clear and concise description of what the bug is.
The facet option dropdown forms stays fixed while scrolling.
**Desktop (please complete the following information):**
- OS: MacOS
- Browser: chrome
- Version: 86
**To Reproduce**
Steps to reproduce the behavior:
1. Select a project on the sidebar
2. Open a facet collapsable
3. Select the dropdown form
4. Scroll up and down (responsive to the size of your browser)
5. See form stays in fixed position and covers other elements
**Expected behavior**
A clear and concise description of what you expected to happen.
The select drop down should automatically minimize when it is out of the view.
Fix: https://github.com/ant-design/ant-design/issues/8268
**Screenshots**
If applicable, add screenshots to help explain your problem.

**Additional context**
Add any other context about the problem here.
| non_defect | facet option dropdown forms stays fixed while scrolling describe the bug a clear and concise description of what the bug is the facet option dropdown forms stays fixed while scrolling desktop please complete the following information os macos browser chrome version to reproduce steps to reproduce the behavior select a project on the sidebar open a facet collapsable select the dropdown form scroll up and down responsive to the size of your browser see form stays in fixed position and covers other elements expected behavior a clear and concise description of what you expected to happen the select drop down should automatically minimize when it is out of the view fix screenshots if applicable add screenshots to help explain your problem additional context add any other context about the problem here | 0 |
9,488 | 7,733,822,983 | IssuesEvent | 2018-05-26 16:26:38 | cmaurer/relative.time.parser | https://api.github.com/repos/cmaurer/relative.time.parser | closed | Security vulnerability through npm-check-updates | security | Hi,
thanks for this great plugin. I'm using it for a HomeKit/homebridge plugin (my homebridge-calendar) and one of its users has reported a security vulnerability for it: It turns out that the source is a broken version of the debug module, which is imported through npm-check-updates. Please upgrade the dependencies of relative-time-parser and publish a new version.
See https://nodesecurity.io/advisories/534 for more details.
Thanks!
grover | True | Security vulnerability through npm-check-updates - Hi,
thanks for this great plugin. I'm using it for a HomeKit/homebridge plugin (my homebridge-calendar) and one of its users has reported a security vulnerability for it: It turns out that the source is a broken version of the debug module, which is imported through npm-check-updates. Please upgrade the dependencies of relative-time-parser and publish a new version.
See https://nodesecurity.io/advisories/534 for more details.
Thanks!
grover | non_defect | security vulnerability through npm check updates hi thanks for this great plugin i m using it for a homekit homebridge plugin my homebridge calendar and one of its users has reported a security vulnerability for it it turns out that the source is a broken version of the debug module which is imported through npm check updates please upgrade the dependencies of relative time parser and publish a new version see for more details thanks grover | 0 |
69,543 | 22,470,606,783 | IssuesEvent | 2022-06-22 07:47:28 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | closed | The <expression/> element in <forcedType/> is deprecated | T: Defect | Noting that this may be a bug in the gradle plugin, not the jOOQ generator, sorry if this is the wrong place.
### Expected behavior
Warning, and hard error
### Actual behavior
Deprecation warning
### Steps to reproduce the problem
Use something like:
forcedTypes {
forcedType {
expression = ".*_HZ"
}
in gradle script
This now appears to be ignored.
Which should result in a hard error, because it breaks our build, because the forcedType is not applied.
Deprecation normally means : "this is currently supported, but will be going away" :-)
### Versions
- jOOQ: 3.16.5
- Java: 11
- Database (include vendor): H2 1.4+
- OS: All
- JDBC Driver (include name if inofficial driver): H2
| 1.0 | The <expression/> element in <forcedType/> is deprecated - Noting that this may be a bug in the gradle plugin, not the jOOQ generator, sorry if this is the wrong place.
### Expected behavior
Warning, and hard error
### Actual behavior
Deprecation warning
### Steps to reproduce the problem
Use something like:
forcedTypes {
forcedType {
expression = ".*_HZ"
}
in gradle script
This now appears to be ignored.
Which should result in a hard error, because it breaks our build, because the forcedType is not applied.
Deprecation normally means : "this is currently supported, but will be going away" :-)
### Versions
- jOOQ: 3.16.5
- Java: 11
- Database (include vendor): H2 1.4+
- OS: All
- JDBC Driver (include name if inofficial driver): H2
| defect | the element in is deprecated noting that this may be a bug in the gradle plugin not the jooq generator sorry if this is the wrong place expected behavior warning and hard error actual behavior deprecation warning steps to reproduce the problem use something like forcedtypes forcedtype expression hz in gradle script this now appears to be ignored which should result in a hard error because it breaks our build because the forcedtype is not applied deprecation normally means this is currently supported but will be going away versions jooq java database include vendor os all jdbc driver include name if inofficial driver | 1 |
215,359 | 16,667,831,745 | IssuesEvent | 2021-06-07 07:15:04 | linrunner/TLP | https://api.github.com/repos/linrunner/TLP | closed | Platform profile support | feature request needs testing | To be able to set the preferred [platform_profile](https://www.kernel.org/doc/html/v5.13-rc3/userspace-api/sysfs-platform_profile.html) depending on power state (charging and ac).
The profiles available on my Thinkpad X1 Carbon Gen 9 are `low-power balanced performance`.
Currently I have to run `echo performance | sudo tee /sys/firmware/acpi/platform_profile` manually to set the profile. Would be nice if TLP would do that automatically.
To get all of the profiles, I run `cat /sys/firmware/acpi/platform_profile_choices`.
At the moment some Lenovo laptops support this feature in 5.12 and some HP laptops in 5.13
https://www.phoronix.com/scan.php?page=news_item&px=Lenovo-Platform-Profile-For-512
https://www.phoronix.com/scan.php?page=news_item&px=Linux-5.13-HP-Platform-Profile | 1.0 | Platform profile support - To be able to set the preferred [platform_profile](https://www.kernel.org/doc/html/v5.13-rc3/userspace-api/sysfs-platform_profile.html) depending on power state (charging and ac).
The profiles available on my Thinkpad X1 Carbon Gen 9 are `low-power balanced performance`.
Currently I have to run `echo performance | sudo tee /sys/firmware/acpi/platform_profile` manually to set the profile. Would be nice if TLP would do that automatically.
To get all of the profiles, I run `cat /sys/firmware/acpi/platform_profile_choices`.
At the moment some Lenovo laptops support this feature in 5.12 and some HP laptops in 5.13
https://www.phoronix.com/scan.php?page=news_item&px=Lenovo-Platform-Profile-For-512
https://www.phoronix.com/scan.php?page=news_item&px=Linux-5.13-HP-Platform-Profile | non_defect | platform profile support to be able to set the preferred depending on power state charging and ac the profiles available on my thinkpad carbon gen are low power balanced performance currently i have to run echo performance sudo tee sys firmware acpi platform profile manually to set the profile would be nice if tlp would do that automatically to get all of the profiles i run cat sys firmware acpi platform profile choices at the moment some lenovo laptops support this feature in and some hp laptops in | 0 |
34,546 | 7,453,587,946 | IssuesEvent | 2018-03-29 12:31:58 | kerdokullamae/test_koik_issued | https://api.github.com/repos/kerdokullamae/test_koik_issued | closed | perioodi facetid ei teki | P: highest R: fixed T: defect | **Reported by katrin vesterblom on 23 Aug 2013 07:27 UTC**
Kui teha päring, nt kirjeldustasand=Arhiiv, siis peaks otsingutulemuse piiramiseks tulema teise hulgas (lisaks isik ja märksõna/valdkonna) facetitele ka perioodide facetid. Neid aga ei tule üldse.
samas, perioodid päringukriteeriumina töötavad.
Tegelikult peaks perioodi-facetid päringutulemuse piirajana tulema igasuguse päringu järel, mille tulemuseks tulnud valimi elementidest kasvõi ühel on piirdaatumid sisestatud. | 1.0 | perioodi facetid ei teki - **Reported by katrin vesterblom on 23 Aug 2013 07:27 UTC**
Kui teha päring, nt kirjeldustasand=Arhiiv, siis peaks otsingutulemuse piiramiseks tulema teise hulgas (lisaks isik ja märksõna/valdkonna) facetitele ka perioodide facetid. Neid aga ei tule üldse.
samas, perioodid päringukriteeriumina töötavad.
Tegelikult peaks perioodi-facetid päringutulemuse piirajana tulema igasuguse päringu järel, mille tulemuseks tulnud valimi elementidest kasvõi ühel on piirdaatumid sisestatud. | defect | perioodi facetid ei teki reported by katrin vesterblom on aug utc kui teha päring nt kirjeldustasand arhiiv siis peaks otsingutulemuse piiramiseks tulema teise hulgas lisaks isik ja märksõna valdkonna facetitele ka perioodide facetid neid aga ei tule üldse samas perioodid päringukriteeriumina töötavad tegelikult peaks perioodi facetid päringutulemuse piirajana tulema igasuguse päringu järel mille tulemuseks tulnud valimi elementidest kasvõi ühel on piirdaatumid sisestatud | 1 |
58,399 | 16,525,239,914 | IssuesEvent | 2021-05-26 19:11:00 | idaholab/moose | https://api.github.com/repos/idaholab/moose | closed | Failed to detect PETSc | P: normal T: defect | ## Bug Description
<!--A clear and concise description of the problem (Note: A missing feature is not a bug).-->When I try to make the makefile in Moduls, It always show that PETSc can not be detected. However, I have successfully installed PETSc. Is there something I should menstion? Thanks.
## Steps to Reproduce
<!--Steps to reproduce the behavior (input file, or modifications to an existing input file, etc.)-->
## Impact
<!--Does this prevent you from getting your work done, or is it more of an annoyance?-->
| 1.0 | Failed to detect PETSc - ## Bug Description
<!--A clear and concise description of the problem (Note: A missing feature is not a bug).-->When I try to make the makefile in Moduls, It always show that PETSc can not be detected. However, I have successfully installed PETSc. Is there something I should menstion? Thanks.
## Steps to Reproduce
<!--Steps to reproduce the behavior (input file, or modifications to an existing input file, etc.)-->
## Impact
<!--Does this prevent you from getting your work done, or is it more of an annoyance?-->
| defect | failed to detect petsc bug description when i try to make the makefile in moduls it always show that petsc can not be detected however i have successfully installed petsc is there something i should menstion thanks steps to reproduce impact | 1 |
38,580 | 8,920,440,424 | IssuesEvent | 2019-01-21 06:46:07 | netty/netty | https://api.github.com/repos/netty/netty | closed | Allow IP_FREEBIND option for UDP epoll | defect | ### Expected behavior
while adding IP_FREEBIND option to boostrap it should work with EpollDatagramChannel
### Actual behavior
notification is logged that IP_FREEBIND is unknown option for this type of channel
### Steps to reproduce
set EpollChannelOption.IP_FREEBIND to true on ConnectionlessBootstrap
### Minimal yet complete reproducer code (or URL to code)
group = new EpollEventLoopGroup(poolSize);
connectionlessBootstrap=new Bootstrap();
connectionlessBootstrap.option(EpollChannelOption.SO_REUSEPORT, true);
connectionlessBootstrap.option(EpollChannelOption.IP_RECVORIGDSTADDR, true);
connectionlessBootstrap.option(EpollChannelOption.IP_FREEBIND, true);
connectionlessBootstrap.channel(EpollDatagramChannel.class);
connectionlessBootstrap.group(group);
connectionlessBootstrap.bind(new InetSocketAddress("0.0.0.0", port));
### Netty version
4.1.25
### JVM version (e.g. `java -version`)
1.8.0
### OS version (e.g. `uname -a`)
ubuntu | 1.0 | Allow IP_FREEBIND option for UDP epoll - ### Expected behavior
while adding IP_FREEBIND option to boostrap it should work with EpollDatagramChannel
### Actual behavior
notification is logged that IP_FREEBIND is unknown option for this type of channel
### Steps to reproduce
set EpollChannelOption.IP_FREEBIND to true on ConnectionlessBootstrap
### Minimal yet complete reproducer code (or URL to code)
group = new EpollEventLoopGroup(poolSize);
connectionlessBootstrap=new Bootstrap();
connectionlessBootstrap.option(EpollChannelOption.SO_REUSEPORT, true);
connectionlessBootstrap.option(EpollChannelOption.IP_RECVORIGDSTADDR, true);
connectionlessBootstrap.option(EpollChannelOption.IP_FREEBIND, true);
connectionlessBootstrap.channel(EpollDatagramChannel.class);
connectionlessBootstrap.group(group);
connectionlessBootstrap.bind(new InetSocketAddress("0.0.0.0", port));
### Netty version
4.1.25
### JVM version (e.g. `java -version`)
1.8.0
### OS version (e.g. `uname -a`)
ubuntu | defect | allow ip freebind option for udp epoll expected behavior while adding ip freebind option to boostrap it should work with epolldatagramchannel actual behavior notification is logged that ip freebind is unknown option for this type of channel steps to reproduce set epollchanneloption ip freebind to true on connectionlessbootstrap minimal yet complete reproducer code or url to code group new epolleventloopgroup poolsize connectionlessbootstrap new bootstrap connectionlessbootstrap option epollchanneloption so reuseport true connectionlessbootstrap option epollchanneloption ip recvorigdstaddr true connectionlessbootstrap option epollchanneloption ip freebind true connectionlessbootstrap channel epolldatagramchannel class connectionlessbootstrap group group connectionlessbootstrap bind new inetsocketaddress port netty version jvm version e g java version os version e g uname a ubuntu | 1 |
127,420 | 18,010,459,471 | IssuesEvent | 2021-09-16 08:00:13 | maddyCode23/linux-4.1.15 | https://api.github.com/repos/maddyCode23/linux-4.1.15 | opened | CVE-2015-8785 (Medium) detected in linux-stable-rtv4.1.33 | security vulnerability | ## CVE-2015-8785 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The fuse_fill_write_pages function in fs/fuse/file.c in the Linux kernel before 4.4 allows local users to cause a denial of service (infinite loop) via a writev system call that triggers a zero length for the first segment of an iov.
<p>Publish Date: 2016-02-08
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-8785>CVE-2015-8785</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-8785">https://nvd.nist.gov/vuln/detail/CVE-2015-8785</a></p>
<p>Release Date: 2016-02-08</p>
<p>Fix Resolution: 4.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2015-8785 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2015-8785 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The fuse_fill_write_pages function in fs/fuse/file.c in the Linux kernel before 4.4 allows local users to cause a denial of service (infinite loop) via a writev system call that triggers a zero length for the first segment of an iov.
<p>Publish Date: 2016-02-08
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-8785>CVE-2015-8785</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-8785">https://nvd.nist.gov/vuln/detail/CVE-2015-8785</a></p>
<p>Release Date: 2016-02-08</p>
<p>Fix Resolution: 4.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href vulnerable source files vulnerability details the fuse fill write pages function in fs fuse file c in the linux kernel before allows local users to cause a denial of service infinite loop via a writev system call that triggers a zero length for the first segment of an iov publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
68,661 | 21,779,462,516 | IssuesEvent | 2022-05-13 17:06:15 | scipy/scipy | https://api.github.com/repos/scipy/scipy | closed | signal.convolve2d performs unexpectedly with boundary='wrap' | defect scipy.signal | The signal.convolve2d(sig_1, sig_2, boundary='wrap') function behaves unexpectedly when the input signals match a certain criteria. Let's call the shape of sig_1 AxB and the shape of sig_2 CxD. If C > A+1 or D > B+1, then the output of signal.convolve2d tends to take on absurd values at certain points.
Notes: After a bit of experimentation, these absurd values are still linearly related to the input. See the reproducing code example below:
### Reproducing code example:
```
>>> sig_1 = np.array([[1, 2], [-2, 0]])
>>> sig_2 = np.array([[1, 2, 3, 1], [0, 0, 1, 2]])
>>> convolve2d(sig_1, sig_2, boundary='wrap')
array([[-1878980598, 7, 8, 7, 5],
[ 537006075, -2, -3, -2, 48],
[-1878980598, 7, 8, 7, 5]])
>>> sig_2[0,3] -= 1
>>> convolve2d(sig_1, sig_2, boundary='wrap')
array([[ 10, 6, 6, 6, 3],
[537006073, 0, -3, 0, 48],
[ 10, 6, 6, 6, 3]])
>>> # The absurd values in the first and third row increased by 1878980608
>>> # The absurd value in the second row decreased by 2
>>> sig_2[0,3] -= 1
>>> convolve2d(sig_1, sig_2, boundary='wrap')
array([[1878980618, 5, 4, 5, 1],
[ 537006071, 2, -3, 2, 48],
[1878980618, 5, 4, 5, 1]])
>>> # The absurd values in the first and third row increased by 1878980608
>>> # The absurd value in the second row decreased by 2
>>> # There is linear consistency in changing sig_2[0,3]
```
My presumption is that, internally, the algorithm is only tiling sig_1 twice and then performing convolution. I believe that this insufficient tiling is leading to a buffer over-read, or something of the sort. For example, based on the above, I expect the internal algorithm is really only calculating the non-wrapped convolution of the following:
sig_1 (tiled):
```
1878980608 | 1 2 | 1 2
? |-2 0 |-2 0
--------------------------
1878980608 | 1 2 | 1 2
? |-2 0 |-2 0
```
sig_2:
```
1 2 3 1
0 0 1 2
```
Or something to this effect. When experimented with further, I'm not sure this is truly the case.
### Scipy/Numpy/Python version information:
```
>>> import sys, scipy, numpy; print(scipy.__version__, numpy.__version__, sys.version_info)
1.0.0 1.14.0 sys.version_info(major=3, minor=6, micro=4, releaselevel='final', serial=0)
```
I've attached some code that I used to gain insight on the issue.
```
import numpy as np
from scipy.signal import convolve2d
# How I believe the algorithm should behave
def my_conv2d_wrap(sig_1, sig_2):
A, B = sig_1.shape
C, D = sig_2.shape
sig_out = np.zeros((A + C - 1, B + D - 1))
# For the sig_out domain
for t_m in range(A + C - 1):
for t_n in range(B + D - 1):
conv_sum = 0
# For the convolution summation domain
for tau_m in range(C):
for tau_n in range(D):
# Using the flipping orientation that scipy uses
conv_sum += sig_2[tau_m, tau_n] * \
sig_1[(t_m - tau_m) % A, (t_n - tau_n) % B]
sig_out[t_m, t_n] = conv_sum
return sig_out
for i in range(10000):
if 0 == i % 100:
print('Done with:', i, 'runs')
# Make random signals up to 5x5 large
A = 1 + np.random.randint(5)
B = 1 + np.random.randint(5)
C = 1 + np.random.randint(5)
D = 1 + np.random.randint(5)
sig_1 = np.random.randint(10, size=(A, B)) - 5
sig_2 = np.random.randint(10, size=(C, D)) - 5
# Generate the outputs
conv_scipy = convolve2d(sig_1, sig_2, boundary='wrap')
conv_mine = my_conv2d_wrap(sig_1, sig_2)
# Ensure no sizing errors ever occur
assert conv_mine.shape[0] == conv_scipy.shape[0] and conv_mine.shape[1] \
== conv_scipy.shape[1]
if np.any(conv_mine != conv_scipy):
# Assume either C > A+1 or D > B+1 for inconsistencies to occur
if not ((C > A+1) or (D > B+1)):
print("The C > A+1 or D > B+1 assumption is wrong!")
```
| 1.0 | signal.convolve2d performs unexpectedly with boundary='wrap' - The signal.convolve2d(sig_1, sig_2, boundary='wrap') function behaves unexpectedly when the input signals match a certain criteria. Let's call the shape of sig_1 AxB and the shape of sig_2 CxD. If C > A+1 or D > B+1, then the output of signal.convolve2d tends to take on absurd values at certain points.
Notes: After a bit of experimentation, these absurd values are still linearly related to the input. See the reproducing code example below:
### Reproducing code example:
```
>>> sig_1 = np.array([[1, 2], [-2, 0]])
>>> sig_2 = np.array([[1, 2, 3, 1], [0, 0, 1, 2]])
>>> convolve2d(sig_1, sig_2, boundary='wrap')
array([[-1878980598, 7, 8, 7, 5],
[ 537006075, -2, -3, -2, 48],
[-1878980598, 7, 8, 7, 5]])
>>> sig_2[0,3] -= 1
>>> convolve2d(sig_1, sig_2, boundary='wrap')
array([[ 10, 6, 6, 6, 3],
[537006073, 0, -3, 0, 48],
[ 10, 6, 6, 6, 3]])
>>> # The absurd values in the first and third row increased by 1878980608
>>> # The absurd value in the second row decreased by 2
>>> sig_2[0,3] -= 1
>>> convolve2d(sig_1, sig_2, boundary='wrap')
array([[1878980618, 5, 4, 5, 1],
[ 537006071, 2, -3, 2, 48],
[1878980618, 5, 4, 5, 1]])
>>> # The absurd values in the first and third row increased by 1878980608
>>> # The absurd value in the second row decreased by 2
>>> # There is linear consistency in changing sig_2[0,3]
```
My presumption is that, internally, the algorithm is only tiling sig_1 twice and then performing convolution. I believe that this insufficient tiling is leading to a buffer over-read, or something of the sort. For example, based on the above, I expect the internal algorithm is really only calculating the non-wrapped convolution of the following:
sig_1 (tiled):
```
1878980608 | 1 2 | 1 2
? |-2 0 |-2 0
--------------------------
1878980608 | 1 2 | 1 2
? |-2 0 |-2 0
```
sig_2:
```
1 2 3 1
0 0 1 2
```
Or something to this effect. When experimented with further, I'm not sure this is truly the case.
### Scipy/Numpy/Python version information:
```
>>> import sys, scipy, numpy; print(scipy.__version__, numpy.__version__, sys.version_info)
1.0.0 1.14.0 sys.version_info(major=3, minor=6, micro=4, releaselevel='final', serial=0)
```
I've attached some code that I used to gain insight on the issue.
```
import numpy as np
from scipy.signal import convolve2d
# How I believe the algorithm should behave
def my_conv2d_wrap(sig_1, sig_2):
A, B = sig_1.shape
C, D = sig_2.shape
sig_out = np.zeros((A + C - 1, B + D - 1))
# For the sig_out domain
for t_m in range(A + C - 1):
for t_n in range(B + D - 1):
conv_sum = 0
# For the convolution summation domain
for tau_m in range(C):
for tau_n in range(D):
# Using the flipping orientation that scipy uses
conv_sum += sig_2[tau_m, tau_n] * \
sig_1[(t_m - tau_m) % A, (t_n - tau_n) % B]
sig_out[t_m, t_n] = conv_sum
return sig_out
for i in range(10000):
if 0 == i % 100:
print('Done with:', i, 'runs')
# Make random signals up to 5x5 large
A = 1 + np.random.randint(5)
B = 1 + np.random.randint(5)
C = 1 + np.random.randint(5)
D = 1 + np.random.randint(5)
sig_1 = np.random.randint(10, size=(A, B)) - 5
sig_2 = np.random.randint(10, size=(C, D)) - 5
# Generate the outputs
conv_scipy = convolve2d(sig_1, sig_2, boundary='wrap')
conv_mine = my_conv2d_wrap(sig_1, sig_2)
# Ensure no sizing errors ever occur
assert conv_mine.shape[0] == conv_scipy.shape[0] and conv_mine.shape[1] \
== conv_scipy.shape[1]
if np.any(conv_mine != conv_scipy):
# Assume either C > A+1 or D > B+1 for inconsistencies to occur
if not ((C > A+1) or (D > B+1)):
print("The C > A+1 or D > B+1 assumption is wrong!")
```
| defect | signal performs unexpectedly with boundary wrap the signal sig sig boundary wrap function behaves unexpectedly when the input signals match a certain criteria let s call the shape of sig axb and the shape of sig cxd if c a or d b then the output of signal tends to take on absurd values at certain points notes after a bit of experimentation these absurd values are still linearly related to the input see the reproducing code example below reproducing code example sig np array sig np array sig sig boundary wrap array sig sig sig boundary wrap array the absurd values in the first and third row increased by the absurd value in the second row decreased by sig sig sig boundary wrap array the absurd values in the first and third row increased by the absurd value in the second row decreased by there is linear consistency in changing sig my presumption is that internally the algorithm is only tiling sig twice and then performing convolution i believe that this insufficient tiling is leading to a buffer over read or something of the sort for example based on the above i expect the internal algorithm is really only calculating the non wrapped convolution of the following sig tiled sig or something to this effect when experimented with further i m not sure this is truly the case scipy numpy python version information import sys scipy numpy print scipy version numpy version sys version info sys version info major minor micro releaselevel final serial i ve attached some code that i used to gain insight on the issue import numpy as np from scipy signal import how i believe the algorithm should behave def my wrap sig sig a b sig shape c d sig shape sig out np zeros a c b d for the sig out domain for t m in range a c for t n in range b d conv sum for the convolution summation domain for tau m in range c for tau n in range d using the flipping orientation that scipy uses conv sum sig sig sig out conv sum return sig out for i in range if i print done with i runs make random signals up to large a np random randint b np random randint c np random randint d np random randint sig np random randint size a b sig np random randint size c d generate the outputs conv scipy sig sig boundary wrap conv mine my wrap sig sig ensure no sizing errors ever occur assert conv mine shape conv scipy shape and conv mine shape conv scipy shape if np any conv mine conv scipy assume either c a or d b for inconsistencies to occur if not c a or d b print the c a or d b assumption is wrong | 1 |
208,385 | 16,112,550,277 | IssuesEvent | 2021-04-28 00:15:23 | cms-opendata-workshop/2021-07-19-cms-open-data-workshop | https://api.github.com/repos/cms-opendata-workshop/2021-07-19-cms-open-data-workshop | closed | Simplify README and add template style stuff | documentation | I think the README can be very much simplified. I will get rid of most of the repetitive stuff that is already in carpentry and go directly to the key instructions.
Also, add the option of picking the workshop style repo as template, so any changes in style can be easily propagated to other repos (made out of this one) with
`git fetch -all`
`git merge template/gh-pages`
But be aware that whatever is in the boilerplate does not get properly propagated. | 1.0 | Simplify README and add template style stuff - I think the README can be very much simplified. I will get rid of most of the repetitive stuff that is already in carpentry and go directly to the key instructions.
Also, add the option of picking the workshop style repo as template, so any changes in style can be easily propagated to other repos (made out of this one) with
`git fetch -all`
`git merge template/gh-pages`
But be aware that whatever is in the boilerplate does not get properly propagated. | non_defect | simplify readme and add template style stuff i think the readme can be very much simplified i will get rid of most of the repetitive stuff that is already in carpentry and go directly to the key instructions also add the option of picking the workshop style repo as template so any changes in style can be easily propagated to other repos made out of this one with git fetch all git merge template gh pages but be aware that whatever is in the boilerplate does not get properly propagated | 0 |
60,576 | 17,023,461,947 | IssuesEvent | 2021-07-03 02:09:15 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | RTL text word-wraps in reverse order | Component: mapnik Priority: minor Resolution: duplicate Type: defect | **[Submitted to the original trac issue database at 4.07pm, Saturday, 15th August 2009]**
When a hebrew string is broken into several lines, the lines appear in reverse.
For example, the text (in English for demonstration purposes) "FED CBA" wraps as:
```
FED
CBA
```
When it should wrap as:
```
CBA
FED
```
And a real-world example:
http://www.openstreetmap.org/?lat=32.1724&lon=34.8106&zoom=14&layers=B000FTF
This is " "
It is shown as
```
```
instead of
```
``` | 1.0 | RTL text word-wraps in reverse order - **[Submitted to the original trac issue database at 4.07pm, Saturday, 15th August 2009]**
When a hebrew string is broken into several lines, the lines appear in reverse.
For example, the text (in English for demonstration purposes) "FED CBA" wraps as:
```
FED
CBA
```
When it should wrap as:
```
CBA
FED
```
And a real-world example:
http://www.openstreetmap.org/?lat=32.1724&lon=34.8106&zoom=14&layers=B000FTF
This is " "
It is shown as
```
```
instead of
```
``` | defect | rtl text word wraps in reverse order when a hebrew string is broken into several lines the lines appear in reverse for example the text in english for demonstration purposes fed cba wraps as fed cba when it should wrap as cba fed and a real world example this is it is shown as instead of | 1 |
649,686 | 21,317,562,725 | IssuesEvent | 2022-04-16 14:53:25 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [YSQL] Update with expression pushdown may cause data inconsistency | kind/bug area/ysql priority/high status/awaiting-triage | ### Description
The following test fails on latest master
```
TEST_F(PgMiniTest, YB_DISABLE_TEST_IN_TSAN(ConcurrentSingleRowUpdate)) {
auto conn = ASSERT_RESULT(Connect());
ASSERT_OK(conn.Execute("CREATE TABLE t(k INT PRIMARY KEY, counter INT)"));
ASSERT_OK(conn.Execute("INSERT INTO t VALUES(1, 0)"));
const size_t thread_count = 10;
const size_t increment_per_thread = 5;
{
CountDownLatch latch(thread_count);
TestThreadHolder thread_holder;
for (size_t i = 0; i < thread_count; ++i) {
thread_holder.AddThreadFunctor([this, &stop = thread_holder.stop_flag(), &latch] {
auto thread_conn = ASSERT_RESULT(Connect());
ASSERT_OK(thread_conn.Execute("SET yb_enable_expression_pushdown TO true"));
latch.CountDown();
latch.Wait();
for (size_t j = 0; j < increment_per_thread; ++j) {
ASSERT_OK(thread_conn.Execute("UPDATE t SET counter = counter + 1 WHERE k = 1"));
}
});
}
}
auto res = ASSERT_RESULT(conn.Fetch("SELECT counter FROM t WHERE k = 1"));
ASSERT_EQ(1, PQnfields(res.get()));
ASSERT_EQ(1, PQntuples(res.get()));
auto counter = ASSERT_RESULT(GetInt32(res.get(), 0, 0));
ASSERT_EQ(thread_count * increment_per_thread, counter);
}
```
Reason: Non transactional write operation uses read-time as a result this write operation may see non latest state of the DocDB. | 1.0 | [YSQL] Update with expression pushdown may cause data inconsistency - ### Description
The following test fails on latest master
```
TEST_F(PgMiniTest, YB_DISABLE_TEST_IN_TSAN(ConcurrentSingleRowUpdate)) {
auto conn = ASSERT_RESULT(Connect());
ASSERT_OK(conn.Execute("CREATE TABLE t(k INT PRIMARY KEY, counter INT)"));
ASSERT_OK(conn.Execute("INSERT INTO t VALUES(1, 0)"));
const size_t thread_count = 10;
const size_t increment_per_thread = 5;
{
CountDownLatch latch(thread_count);
TestThreadHolder thread_holder;
for (size_t i = 0; i < thread_count; ++i) {
thread_holder.AddThreadFunctor([this, &stop = thread_holder.stop_flag(), &latch] {
auto thread_conn = ASSERT_RESULT(Connect());
ASSERT_OK(thread_conn.Execute("SET yb_enable_expression_pushdown TO true"));
latch.CountDown();
latch.Wait();
for (size_t j = 0; j < increment_per_thread; ++j) {
ASSERT_OK(thread_conn.Execute("UPDATE t SET counter = counter + 1 WHERE k = 1"));
}
});
}
}
auto res = ASSERT_RESULT(conn.Fetch("SELECT counter FROM t WHERE k = 1"));
ASSERT_EQ(1, PQnfields(res.get()));
ASSERT_EQ(1, PQntuples(res.get()));
auto counter = ASSERT_RESULT(GetInt32(res.get(), 0, 0));
ASSERT_EQ(thread_count * increment_per_thread, counter);
}
```
Reason: Non transactional write operation uses read-time as a result this write operation may see non latest state of the DocDB. | non_defect | update with expression pushdown may cause data inconsistency description the following test fails on latest master test f pgminitest yb disable test in tsan concurrentsinglerowupdate auto conn assert result connect assert ok conn execute create table t k int primary key counter int assert ok conn execute insert into t values const size t thread count const size t increment per thread countdownlatch latch thread count testthreadholder thread holder for size t i i thread count i thread holder addthreadfunctor auto thread conn assert result connect assert ok thread conn execute set yb enable expression pushdown to true latch countdown latch wait for size t j j increment per thread j assert ok thread conn execute update t set counter counter where k auto res assert result conn fetch select counter from t where k assert eq pqnfields res get assert eq pqntuples res get auto counter assert result res get assert eq thread count increment per thread counter reason non transactional write operation uses read time as a result this write operation may see non latest state of the docdb | 0 |
38,899 | 9,011,431,462 | IssuesEvent | 2019-02-05 14:41:10 | Automattic/wp-calypso | https://api.github.com/repos/Automattic/wp-calypso | closed | `FeatureExample` component has blue fade that doesn't match with page background | Activity Log Color Schemes Components [Pri] Low [Type] Defect | On the _Activity log_ page, when viewing the log with simple site that doesn't have any recent events, I can see slight off-colour fade in the feature preview component:
<img width="1125" alt="screenshot 2019-01-23 at 12 14 03" src="https://user-images.githubusercontent.com/87168/51599590-a7fbb080-1f08-11e9-8a11-c0b8f4d7b7db.png">
Note how it has too blue hue compared to background and you can see slight border between the component and background:
<img width="110" alt="image" src="https://user-images.githubusercontent.com/87168/51599618-c06bcb00-1f08-11e9-8d1c-b5e22306e32b.png">
Component:
https://github.com/Automattic/wp-calypso/tree/cdcc997b9118bb998071fa095e2806993f5a2273/client/components/feature-example
Component is used here:
https://github.com/Automattic/wp-calypso/blob/cdcc997b9118bb998071fa095e2806993f5a2273/client/my-sites/activity/activity-log-example/index.jsx#L67-L78
FYI @Automattic/team-calypso
| 1.0 | `FeatureExample` component has blue fade that doesn't match with page background - On the _Activity log_ page, when viewing the log with simple site that doesn't have any recent events, I can see slight off-colour fade in the feature preview component:
<img width="1125" alt="screenshot 2019-01-23 at 12 14 03" src="https://user-images.githubusercontent.com/87168/51599590-a7fbb080-1f08-11e9-8a11-c0b8f4d7b7db.png">
Note how it has too blue hue compared to background and you can see slight border between the component and background:
<img width="110" alt="image" src="https://user-images.githubusercontent.com/87168/51599618-c06bcb00-1f08-11e9-8d1c-b5e22306e32b.png">
Component:
https://github.com/Automattic/wp-calypso/tree/cdcc997b9118bb998071fa095e2806993f5a2273/client/components/feature-example
Component is used here:
https://github.com/Automattic/wp-calypso/blob/cdcc997b9118bb998071fa095e2806993f5a2273/client/my-sites/activity/activity-log-example/index.jsx#L67-L78
FYI @Automattic/team-calypso
| defect | featureexample component has blue fade that doesn t match with page background on the activity log page when viewing the log with simple site that doesn t have any recent events i can see slight off colour fade in the feature preview component img width alt screenshot at src note how it has too blue hue compared to background and you can see slight border between the component and background img width alt image src component component is used here fyi automattic team calypso | 1 |
141,920 | 21,640,431,633 | IssuesEvent | 2022-05-05 18:11:28 | bcgov/cas-public-reporting | https://api.github.com/repos/bcgov/cas-public-reporting | closed | Get overview of CleanBC & GoElectric Project | Service Design Task | Acceptance Criteria:
- [x] Schedule session with Jenna from CleanBC/GoElectric Project
- [x] Hold session
- [x] Record takeaways in comments below | 1.0 | Get overview of CleanBC & GoElectric Project - Acceptance Criteria:
- [x] Schedule session with Jenna from CleanBC/GoElectric Project
- [x] Hold session
- [x] Record takeaways in comments below | non_defect | get overview of cleanbc goelectric project acceptance criteria schedule session with jenna from cleanbc goelectric project hold session record takeaways in comments below | 0 |
77,407 | 26,975,771,252 | IssuesEvent | 2023-02-09 09:26:57 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | opened | WHERE clause issue with streaming | Type: Defect | The WHERE clause seems to be misbehaving in a streaming query.
If I create this streaming view:
```
CREATE OR REPLACE VIEW trades AS
SELECT __key,
CASE WHEN tickRand BETWEEN 0 AND 0.1 THEN 'APPL'
WHEN tickRand BETWEEN 0.2 AND 0.3 THEN 'GOOG'
WHEN tickRand BETWEEN 0.3 AND 0.4 THEN 'META'
WHEN tickRand BETWEEN 0.4 AND 0.5 THEN 'NFLX'
WHEN tickRand BETWEEN 0.5 AND 0.6 THEN 'AMZ'
ELSE 'VOO'
END as ticker,
price,
tradetime,
amount
FROM
(SELECT v as __key,
RAND() as tickRand,
RAND()*v as price,
TO_TIMESTAMP_TZ(v*1000 + 1645484400000) as tradetime,
ROUND(RAND()*100, 0) as amount
FROM TABLE(generate_stream(10)));
```
and then if I want to filter on the ticker:
```
SELECT ticker, price, amount AS shares_sold
FROM trades
WHERE ticker = 'GOOG';
```
It IGNORES the filter:
<img width="554" alt="image" src="https://user-images.githubusercontent.com/113477375/217771547-e81c9e16-5fcf-459b-9af2-318f06b407d7.png">
I would not have expected any tickers in the result to have values other than GOOG
Here is the EXPLAIN of the query:
<img width="1511" alt="image" src="https://user-images.githubusercontent.com/113477375/217771742-ad7628b9-e4cd-4676-a59c-fbcd3793bb4b.png">
| 1.0 | WHERE clause issue with streaming - The WHERE clause seems to be misbehaving in a streaming query.
If I create this streaming view:
```
CREATE OR REPLACE VIEW trades AS
SELECT __key,
CASE WHEN tickRand BETWEEN 0 AND 0.1 THEN 'APPL'
WHEN tickRand BETWEEN 0.2 AND 0.3 THEN 'GOOG'
WHEN tickRand BETWEEN 0.3 AND 0.4 THEN 'META'
WHEN tickRand BETWEEN 0.4 AND 0.5 THEN 'NFLX'
WHEN tickRand BETWEEN 0.5 AND 0.6 THEN 'AMZ'
ELSE 'VOO'
END as ticker,
price,
tradetime,
amount
FROM
(SELECT v as __key,
RAND() as tickRand,
RAND()*v as price,
TO_TIMESTAMP_TZ(v*1000 + 1645484400000) as tradetime,
ROUND(RAND()*100, 0) as amount
FROM TABLE(generate_stream(10)));
```
and then if I want to filter on the ticker:
```
SELECT ticker, price, amount AS shares_sold
FROM trades
WHERE ticker = 'GOOG';
```
It IGNORES the filter:
<img width="554" alt="image" src="https://user-images.githubusercontent.com/113477375/217771547-e81c9e16-5fcf-459b-9af2-318f06b407d7.png">
I would not have expected any tickers in the result to have values other than GOOG
Here is the EXPLAIN of the query:
<img width="1511" alt="image" src="https://user-images.githubusercontent.com/113477375/217771742-ad7628b9-e4cd-4676-a59c-fbcd3793bb4b.png">
| defect | where clause issue with streaming the where clause seems to be misbehaving in a streaming query if i create this streaming view create or replace view trades as select key case when tickrand between and then appl when tickrand between and then goog when tickrand between and then meta when tickrand between and then nflx when tickrand between and then amz else voo end as ticker price tradetime amount from select v as key rand as tickrand rand v as price to timestamp tz v as tradetime round rand as amount from table generate stream and then if i want to filter on the ticker select ticker price amount as shares sold from trades where ticker goog it ignores the filter img width alt image src i would not have expected any tickers in the result to have values other than goog here is the explain of the query img width alt image src | 1 |
56,637 | 14,078,466,996 | IssuesEvent | 2020-11-04 13:36:49 | themagicalmammal/android_kernel_samsung_a5xelte | https://api.github.com/repos/themagicalmammal/android_kernel_samsung_a5xelte | opened | CVE-2018-16658 (Medium) detected in linuxv3.11 | security vulnerability | ## CVE-2018-16658 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.11</b></p></summary>
<p>
<p>Dave Jiang fork of linux.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/djiang/linux.git>https://git.kernel.org/pub/scm/linux/kernel/git/djiang/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/themagicalmammal/android_kernel_samsung_a5xelte/commit/738375813823cb33918102af385bdd5d82225e17">738375813823cb33918102af385bdd5d82225e17</a></p>
<p>Found in base branch: <b>cosmic-1.6-experimental</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>android_kernel_samsung_a5xelte/drivers/cdrom/cdrom.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>android_kernel_samsung_a5xelte/drivers/cdrom/cdrom.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel before 4.18.6. An information leak in cdrom_ioctl_drive_status in drivers/cdrom/cdrom.c could be used by local attackers to read kernel memory because a cast from unsigned long to int interferes with bounds checking. This is similar to CVE-2018-10940.
<p>Publish Date: 2018-09-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16658>CVE-2018-16658</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-16658">https://nvd.nist.gov/vuln/detail/CVE-2018-16658</a></p>
<p>Release Date: 2018-09-07</p>
<p>Fix Resolution: 4.18.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-16658 (Medium) detected in linuxv3.11 - ## CVE-2018-16658 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.11</b></p></summary>
<p>
<p>Dave Jiang fork of linux.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/djiang/linux.git>https://git.kernel.org/pub/scm/linux/kernel/git/djiang/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/themagicalmammal/android_kernel_samsung_a5xelte/commit/738375813823cb33918102af385bdd5d82225e17">738375813823cb33918102af385bdd5d82225e17</a></p>
<p>Found in base branch: <b>cosmic-1.6-experimental</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>android_kernel_samsung_a5xelte/drivers/cdrom/cdrom.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>android_kernel_samsung_a5xelte/drivers/cdrom/cdrom.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel before 4.18.6. An information leak in cdrom_ioctl_drive_status in drivers/cdrom/cdrom.c could be used by local attackers to read kernel memory because a cast from unsigned long to int interferes with bounds checking. This is similar to CVE-2018-10940.
<p>Publish Date: 2018-09-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16658>CVE-2018-16658</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-16658">https://nvd.nist.gov/vuln/detail/CVE-2018-16658</a></p>
<p>Release Date: 2018-09-07</p>
<p>Fix Resolution: 4.18.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve medium detected in cve medium severity vulnerability vulnerable library dave jiang fork of linux git library home page a href found in head commit a href found in base branch cosmic experimental vulnerable source files android kernel samsung drivers cdrom cdrom c android kernel samsung drivers cdrom cdrom c vulnerability details an issue was discovered in the linux kernel before an information leak in cdrom ioctl drive status in drivers cdrom cdrom c could be used by local attackers to read kernel memory because a cast from unsigned long to int interferes with bounds checking this is similar to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
13,588 | 2,772,065,359 | IssuesEvent | 2015-05-02 09:05:58 | cultibox/cultibox | https://api.github.com/repos/cultibox/cultibox | opened | [cultipi] information du bouton :"afficher les differences" | Component-cultipi Priority-Low Type-Defect | v2.0.48-r1505011427 GPL-V3
j'ai beau faire ma synchro, des que je change d'onglet, le logiciel m'indique que je ne suis plus synchronisé.
En écart il me donne :
Only in /etc/cultipi/01_defaultConf_RPi/serverCultibox: conf.xml
Only in /etc/cultipi/01_defaultConf_RPi: serverIrrigation
Je dois faire une copie à la main à la fin?
Sur winscp:
* je vois bien que j'ai le dossier "serverIrrigation" dans ma conf appliqué (01_defaultConf_RPi) mais le dossier n'existe pas dans ma conf temporaire (conf_tmp)
*idem pour le fichier "conf.xml" dans le dossier "serverCultibox"
*j'ai un nouveau dossier dans /etc/cultipi/ qui s'apelle 01_defaultConf_RPi.new
| 1.0 | [cultipi] information du bouton :"afficher les differences" - v2.0.48-r1505011427 GPL-V3
j'ai beau faire ma synchro, des que je change d'onglet, le logiciel m'indique que je ne suis plus synchronisé.
En écart il me donne :
Only in /etc/cultipi/01_defaultConf_RPi/serverCultibox: conf.xml
Only in /etc/cultipi/01_defaultConf_RPi: serverIrrigation
Je dois faire une copie à la main à la fin?
Sur winscp:
* je vois bien que j'ai le dossier "serverIrrigation" dans ma conf appliqué (01_defaultConf_RPi) mais le dossier n'existe pas dans ma conf temporaire (conf_tmp)
*idem pour le fichier "conf.xml" dans le dossier "serverCultibox"
*j'ai un nouveau dossier dans /etc/cultipi/ qui s'apelle 01_defaultConf_RPi.new
| defect | information du bouton afficher les differences gpl j ai beau faire ma synchro des que je change d onglet le logiciel m indique que je ne suis plus synchronisé en écart il me donne only in etc cultipi defaultconf rpi servercultibox conf xml only in etc cultipi defaultconf rpi serverirrigation je dois faire une copie à la main à la fin sur winscp je vois bien que j ai le dossier serverirrigation dans ma conf appliqué defaultconf rpi mais le dossier n existe pas dans ma conf temporaire conf tmp idem pour le fichier conf xml dans le dossier servercultibox j ai un nouveau dossier dans etc cultipi qui s apelle defaultconf rpi new | 1 |
231,603 | 7,641,187,760 | IssuesEvent | 2018-05-08 03:14:43 | the-blue-alliance/the-blue-alliance-ios | https://api.github.com/repos/the-blue-alliance/the-blue-alliance-ios | opened | Implement myTBA Preferences model/endpoint | enhancement high priority | This should include implementing convenience methods on MyTBAFavorite/MyTBASubscription | 1.0 | Implement myTBA Preferences model/endpoint - This should include implementing convenience methods on MyTBAFavorite/MyTBASubscription | non_defect | implement mytba preferences model endpoint this should include implementing convenience methods on mytbafavorite mytbasubscription | 0 |
325,838 | 9,936,599,119 | IssuesEvent | 2019-07-02 19:33:55 | ChainSafe/lodestar | https://api.github.com/repos/ChainSafe/lodestar | closed | Change db getter return types to Promise<T | null> | priority: P3 important type: enhancement type: good first issue | Currently, for database getters, eg: `await db.getBlock(root)`, we either return the block, or we `throw` an error if the block is not found.
Instead of `throw`ing an error if the value is not found, return `null`.
There are many places where we currently rely on try/catch behavior that will need to be updated as well. | 1.0 | Change db getter return types to Promise<T | null> - Currently, for database getters, eg: `await db.getBlock(root)`, we either return the block, or we `throw` an error if the block is not found.
Instead of `throw`ing an error if the value is not found, return `null`.
There are many places where we currently rely on try/catch behavior that will need to be updated as well. | non_defect | change db getter return types to promise currently for database getters eg await db getblock root we either return the block or we throw an error if the block is not found instead of throw ing an error if the value is not found return null there are many places where we currently rely on try catch behavior that will need to be updated as well | 0 |
112,289 | 9,559,630,212 | IssuesEvent | 2019-05-03 17:15:36 | knative/serving | https://api.github.com/repos/knative/serving | closed | Label: "serving.knative.dev/release" should have release name | area/test-and-release kind/bug | ## In what area(s)?
/area test-and-release
## What version of Knative?
0.4.x
HEAD
## Expected Behavior
label `serving.knative.dev/release: devel` should have the release name/number instead of `devel`
## Actual Behavior
it's set to `devel`
## Steps to Reproduce the Problem
Install v0.4.0
/cc @evankanderson | 1.0 | Label: "serving.knative.dev/release" should have release name - ## In what area(s)?
/area test-and-release
## What version of Knative?
0.4.x
HEAD
## Expected Behavior
label `serving.knative.dev/release: devel` should have the release name/number instead of `devel`
## Actual Behavior
it's set to `devel`
## Steps to Reproduce the Problem
Install v0.4.0
/cc @evankanderson | non_defect | label serving knative dev release should have release name in what area s area test and release what version of knative x head expected behavior label serving knative dev release devel should have the release name number instead of devel actual behavior it s set to devel steps to reproduce the problem install cc evankanderson | 0 |
18,684 | 3,080,680,547 | IssuesEvent | 2015-08-22 00:32:16 | jameslh/pagedown | https://api.github.com/repos/jameslh/pagedown | closed | Relative links don't display correctly in-line unless in a list | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Create a link of the form [link](/relative/path)
What is the expected output? What do you see instead?
In page the preview will display [link](/relative/path) however the same link
in a list will display as a normal hyperlink.
What version of the product are you using? On what operating system?
Current version running on LAMP server.
Please provide any additional information below.
Appears to be a regex problem
```
Original issue reported on code.google.com by `tgoza-...@imagehawk.com` on 9 Oct 2013 at 8:42 | 1.0 | Relative links don't display correctly in-line unless in a list - ```
What steps will reproduce the problem?
1. Create a link of the form [link](/relative/path)
What is the expected output? What do you see instead?
In page the preview will display [link](/relative/path) however the same link
in a list will display as a normal hyperlink.
What version of the product are you using? On what operating system?
Current version running on LAMP server.
Please provide any additional information below.
Appears to be a regex problem
```
Original issue reported on code.google.com by `tgoza-...@imagehawk.com` on 9 Oct 2013 at 8:42 | defect | relative links don t display correctly in line unless in a list what steps will reproduce the problem create a link of the form relative path what is the expected output what do you see instead in page the preview will display relative path however the same link in a list will display as a normal hyperlink what version of the product are you using on what operating system current version running on lamp server please provide any additional information below appears to be a regex problem original issue reported on code google com by tgoza imagehawk com on oct at | 1 |
33,777 | 16,109,104,216 | IssuesEvent | 2021-04-27 18:35:28 | microsoft/msquic | https://api.github.com/repos/microsoft/msquic | opened | Generating retry token uses too much CPU | Area: Core Area: Performance | Retry token generation uses a single instance of the hash, and a lock. There is a lot of contention on that lock.
The solution is to have a per-CPU hash to remove the lock, and then update all the instances at the same time. | True | Generating retry token uses too much CPU - Retry token generation uses a single instance of the hash, and a lock. There is a lot of contention on that lock.
The solution is to have a per-CPU hash to remove the lock, and then update all the instances at the same time. | non_defect | generating retry token uses too much cpu retry token generation uses a single instance of the hash and a lock there is a lot of contention on that lock the solution is to have a per cpu hash to remove the lock and then update all the instances at the same time | 0 |
77,216 | 26,854,229,259 | IssuesEvent | 2023-02-03 13:27:33 | matrix-org/synapse | https://api.github.com/repos/matrix-org/synapse | closed | Logging in Rust code not working (in unit tests?) | S-Minor T-Defect O-Occasional | We use [pyo3-log](https://docs.rs/pyo3-log/latest/pyo3_log/) to get Rust logs into the Python logging framework; this seems to not work properly. If you put e.g. a `warn!(...)` then it doesn't print out anywhere.
We initialize the logging module in `lib.rs`:
https://github.com/matrix-org/synapse/blob/42d261c32f13e2de7494a0ade77c1f7b646af1fe/rust/src/lib.rs#L23
And we import the synapse rust code *very early* in the Synapse process:
https://github.com/matrix-org/synapse/blob/dc37b68a25754240243cdca6f521919abfe71db0/synapse/__init__.py#L83
----
I think we're importing the rust code before the loggers are configured, which then causes `pyo3-log` to initialize and [cache the Python logging configuration](https://docs.rs/pyo3-log/latest/pyo3_log/#performance-filtering-and-caching), then by the time we actually call Rust code we've reconfigured the python loggers, but that doesn't propagate into Rust.
I ran into this in unittests, but I think it would happen during normal usage as well. | 1.0 | Logging in Rust code not working (in unit tests?) - We use [pyo3-log](https://docs.rs/pyo3-log/latest/pyo3_log/) to get Rust logs into the Python logging framework; this seems to not work properly. If you put e.g. a `warn!(...)` then it doesn't print out anywhere.
We initialize the logging module in `lib.rs`:
https://github.com/matrix-org/synapse/blob/42d261c32f13e2de7494a0ade77c1f7b646af1fe/rust/src/lib.rs#L23
And we import the synapse rust code *very early* in the Synapse process:
https://github.com/matrix-org/synapse/blob/dc37b68a25754240243cdca6f521919abfe71db0/synapse/__init__.py#L83
----
I think we're importing the rust code before the loggers are configured, which then causes `pyo3-log` to initialize and [cache the Python logging configuration](https://docs.rs/pyo3-log/latest/pyo3_log/#performance-filtering-and-caching), then by the time we actually call Rust code we've reconfigured the python loggers, but that doesn't propagate into Rust.
I ran into this in unittests, but I think it would happen during normal usage as well. | defect | logging in rust code not working in unit tests we use to get rust logs into the python logging framework this seems to not work properly if you put e g a warn then it doesn t print out anywhere we initialize the logging module in lib rs and we import the synapse rust code very early in the synapse process i think we re importing the rust code before the loggers are configured which then causes log to initialize and then by the time we actually call rust code we ve reconfigured the python loggers but that doesn t propagate into rust i ran into this in unittests but i think it would happen during normal usage as well | 1 |
34,468 | 7,451,795,891 | IssuesEvent | 2018-03-29 05:21:52 | kerdokullamae/test_koik_issued | https://api.github.com/repos/kerdokullamae/test_koik_issued | closed | nimistu liigutamine teise arhiivi alla on ebaselge | P: high R: fixed T: defect | **Reported by katrin vesterblom on 22 Apr 2013 12:01 UTC**
rahvusarhiiv.tietotest.ee , kasutaja katrinv
Sisestasin arhiivid tla.230 ja tla.9999 .Mõlema alla tegin Nimistu=1.
tla.230 sisaldab all-arhiivi, see omakorda sarja, milles omakorda on 2 säilikut (mõlemad nimistust 1).
Muutsin arhiivi tla.230 nimistu 1 ülemat, ehk eet liigutasin ta arhiivi TLA.9999 alla. Nüüd on tulemuseks, et arhiivi tla.9999 kuulub kaks nimistut, mõlemal tähis=1. Ja ühes nendest on säilikud, mis kuuluvad arhiivi TLA.230 struktuuri (säiliku andmetes olev asukoht kirjeldushierarhias on tla.230 oma, ülemad kü-d viivad tla.230 juurde jne.).
Minu meelest peaks nimistu teise arhiivi alla tõstmise tulemusena vähemalt muutuma sellesse nimistusse kuuluvate säilikute leidandmed (vastavaks sellele arhiivile, mille alla nad tõsteti). Võimalik, et säilikute jaoks peaks tekkima ka sarnane struktuurihierarhia, nagu algses arhiivis oli. Ka peaks tulema vähemalt hoiatus, kui minu poolt liigutatava nimistu tähis langeb kokku selles arhiivis juba olemas oleva nimistu tähisega, kuhu nimistut liigutada tahan. | 1.0 | nimistu liigutamine teise arhiivi alla on ebaselge - **Reported by katrin vesterblom on 22 Apr 2013 12:01 UTC**
rahvusarhiiv.tietotest.ee , kasutaja katrinv
Sisestasin arhiivid tla.230 ja tla.9999 .Mõlema alla tegin Nimistu=1.
tla.230 sisaldab all-arhiivi, see omakorda sarja, milles omakorda on 2 säilikut (mõlemad nimistust 1).
Muutsin arhiivi tla.230 nimistu 1 ülemat, ehk eet liigutasin ta arhiivi TLA.9999 alla. Nüüd on tulemuseks, et arhiivi tla.9999 kuulub kaks nimistut, mõlemal tähis=1. Ja ühes nendest on säilikud, mis kuuluvad arhiivi TLA.230 struktuuri (säiliku andmetes olev asukoht kirjeldushierarhias on tla.230 oma, ülemad kü-d viivad tla.230 juurde jne.).
Minu meelest peaks nimistu teise arhiivi alla tõstmise tulemusena vähemalt muutuma sellesse nimistusse kuuluvate säilikute leidandmed (vastavaks sellele arhiivile, mille alla nad tõsteti). Võimalik, et säilikute jaoks peaks tekkima ka sarnane struktuurihierarhia, nagu algses arhiivis oli. Ka peaks tulema vähemalt hoiatus, kui minu poolt liigutatava nimistu tähis langeb kokku selles arhiivis juba olemas oleva nimistu tähisega, kuhu nimistut liigutada tahan. | defect | nimistu liigutamine teise arhiivi alla on ebaselge reported by katrin vesterblom on apr utc rahvusarhiiv tietotest ee kasutaja katrinv sisestasin arhiivid tla ja tla mõlema alla tegin nimistu tla sisaldab all arhiivi see omakorda sarja milles omakorda on säilikut mõlemad nimistust muutsin arhiivi tla nimistu ülemat ehk eet liigutasin ta arhiivi tla alla nüüd on tulemuseks et arhiivi tla kuulub kaks nimistut mõlemal tähis ja ühes nendest on säilikud mis kuuluvad arhiivi tla struktuuri säiliku andmetes olev asukoht kirjeldushierarhias on tla oma ülemad kü d viivad tla juurde jne minu meelest peaks nimistu teise arhiivi alla tõstmise tulemusena vähemalt muutuma sellesse nimistusse kuuluvate säilikute leidandmed vastavaks sellele arhiivile mille alla nad tõsteti võimalik et säilikute jaoks peaks tekkima ka sarnane struktuurihierarhia nagu algses arhiivis oli ka peaks tulema vähemalt hoiatus kui minu poolt liigutatava nimistu tähis langeb kokku selles arhiivis juba olemas oleva nimistu tähisega kuhu nimistut liigutada tahan | 1 |
59,996 | 17,023,307,250 | IssuesEvent | 2021-07-03 01:20:43 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | color settings in site.css make site invisible with non-standard color scheme | Component: website Priority: minor Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 2.36pm, Wednesday, 8th October 2008]**
This bug is rally nasty but should be very easy to fix, i wonder why nobody has noticed it yet.
The site.css file for your entire website (including the mainpage www.openstreetmap.org) includes the following paragraph:
body {
font-family: Arial,sans-serif;
color: Black;
margin: 0;
padding: 0;
}
as you can see the textcolor for the whole page is set to black while there is no background color defined. This is critical for users having their desktop style configured to show a black background with, say white or red text on it.
While the users foreground color (which would be viewable on the users background color) is overridden the users background color is still displayed by the browser. Well, this makes the site appear with black text on black background, which is naturally unreadable.
The fault lies neither in the users settings as they would look nice if not overridden by the css nor does it lie in the webbrowsers rendering engine as it behaves exactly the way one would expect. The webdesigner simply assumed that every visitors would use the default (even worse, the MS Windows default) color scheme to display the Website.
Well, shouldn't be that hard to insert the line
background: White;
into the body-paragraph of your site.css ;-)
Even better style with hex colors:
body {
font-family: Arial,sans-serif;
color: #000000;
background: #FFFFFF;
margin: 0;
padding: 0;
}
Critical thing about this bug is: it's so awkward for the guys at openstreetmap.
Attached is a screenshot, in case you dont believe me ;-) | 1.0 | color settings in site.css make site invisible with non-standard color scheme - **[Submitted to the original trac issue database at 2.36pm, Wednesday, 8th October 2008]**
This bug is rally nasty but should be very easy to fix, i wonder why nobody has noticed it yet.
The site.css file for your entire website (including the mainpage www.openstreetmap.org) includes the following paragraph:
body {
font-family: Arial,sans-serif;
color: Black;
margin: 0;
padding: 0;
}
as you can see the textcolor for the whole page is set to black while there is no background color defined. This is critical for users having their desktop style configured to show a black background with, say white or red text on it.
While the users foreground color (which would be viewable on the users background color) is overridden the users background color is still displayed by the browser. Well, this makes the site appear with black text on black background, which is naturally unreadable.
The fault lies neither in the users settings as they would look nice if not overridden by the css nor does it lie in the webbrowsers rendering engine as it behaves exactly the way one would expect. The webdesigner simply assumed that every visitors would use the default (even worse, the MS Windows default) color scheme to display the Website.
Well, shouldn't be that hard to insert the line
background: White;
into the body-paragraph of your site.css ;-)
Even better style with hex colors:
body {
font-family: Arial,sans-serif;
color: #000000;
background: #FFFFFF;
margin: 0;
padding: 0;
}
Critical thing about this bug is: it's so awkward for the guys at openstreetmap.
Attached is a screenshot, in case you dont believe me ;-) | defect | color settings in site css make site invisible with non standard color scheme this bug is rally nasty but should be very easy to fix i wonder why nobody has noticed it yet the site css file for your entire website including the mainpage includes the following paragraph body font family arial sans serif color black margin padding as you can see the textcolor for the whole page is set to black while there is no background color defined this is critical for users having their desktop style configured to show a black background with say white or red text on it while the users foreground color which would be viewable on the users background color is overridden the users background color is still displayed by the browser well this makes the site appear with black text on black background which is naturally unreadable the fault lies neither in the users settings as they would look nice if not overridden by the css nor does it lie in the webbrowsers rendering engine as it behaves exactly the way one would expect the webdesigner simply assumed that every visitors would use the default even worse the ms windows default color scheme to display the website well shouldn t be that hard to insert the line background white into the body paragraph of your site css even better style with hex colors body font family arial sans serif color background ffffff margin padding critical thing about this bug is it s so awkward for the guys at openstreetmap attached is a screenshot in case you dont believe me | 1 |
17,528 | 3,011,227,761 | IssuesEvent | 2015-07-28 16:49:39 | dart-lang/sdk | https://api.github.com/repos/dart-lang/sdk | closed | Confusing error message for "C?.x", where C is a class name | Area-Analyzer Priority-Medium Triaged Type-Defect | Analyzer currently reports an error for this situation, but the error message is confusing. It should be cleaned up so that it is clear to the user that to fix it, they simply need to replace "?." with ".". | 1.0 | Confusing error message for "C?.x", where C is a class name - Analyzer currently reports an error for this situation, but the error message is confusing. It should be cleaned up so that it is clear to the user that to fix it, they simply need to replace "?." with ".". | defect | confusing error message for c x where c is a class name analyzer currently reports an error for this situation but the error message is confusing it should be cleaned up so that it is clear to the user that to fix it they simply need to replace quot quot with quot quot | 1 |
27,440 | 5,018,864,354 | IssuesEvent | 2016-12-14 09:50:47 | ndasjowo/gCode | https://api.github.com/repos/ndasjowo/gCode | closed | tproxy single interface/ethernet | auto-migrated Priority-Medium Type-Defect | ```
http://www.forummikrotik.com/general-networking/21391-tproxy-single-interface-et
hernet.html#post245156
```
Original issue reported on code.google.com by `hedy.joe@gmail.com` on 2 Apr 2012 at 4:56
| 1.0 | tproxy single interface/ethernet - ```
http://www.forummikrotik.com/general-networking/21391-tproxy-single-interface-et
hernet.html#post245156
```
Original issue reported on code.google.com by `hedy.joe@gmail.com` on 2 Apr 2012 at 4:56
| defect | tproxy single interface ethernet hernet html original issue reported on code google com by hedy joe gmail com on apr at | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.