Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
123,816
10,291,647,780
IssuesEvent
2019-08-27 12:57:19
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
teamcity: failed test: _size_out_of_range_direct=false
C-test-failure O-robot
The following tests appear to have failed on master (testrace): _size_out_of_range_direct=false You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+_size_out_of_range_direct=false). [#1451701](https://teamcity.cockroachdb.com/viewLog.html?buildId=1451701): ``` _size_out_of_range_direct=false --- FAIL: testrace/TestImportData/PGDUMP:_size_out_of_range_direct=false (0.000s) Test ended in panic. ------- Stdout: ------- I190823 18:30:37.305015 17834 storage/replica_command.go:598 [n1,merge,s1,r63/1:/Table/9{0-2}] initiating a merge of r65:/Table/9{2-4} [(n1,s1):1, next=2, gen=26] into this range (lhs+rhs has (size=0 B+0 B qps=0.00+0.89 --> 0.89qps) below threshold (size=0 B, qps=0.89)) I190823 18:30:37.388071 769 sql/event_log.go:130 [n1,client=127.0.0.1:49446,user=root] Event: "create_database", target: 120, info: {DatabaseName:d34 Statement:CREATE DATABASE d34 User:root} I190823 18:30:37.414493 105 storage/store.go:2593 [n1,s1,r63/1:/Table/9{0-2}] removing replica r65/1 I190823 18:30:37.588024 769 sql/event_log.go:130 [n1,client=127.0.0.1:49446,user=root] Event: "drop_database", target: 120, info: {DatabaseName:d34 Statement:DROP DATABASE d34 User:root DroppedSchemaObjects:[]} _size_out_of_range_direct=false --- FAIL: testrace/TestImportData/PGCOPY:_size_out_of_range_direct=false (0.000s) Test ended in panic. ------- Stdout: ------- I190823 18:30:31.158536 769 sql/event_log.go:130 [n1,client=127.0.0.1:49446,user=root] Event: "create_database", target: 107, info: {DatabaseName:d27 Statement:CREATE DATABASE d27 User:root} I190823 18:30:31.308078 769 sql/event_log.go:130 [n1,client=127.0.0.1:49446,user=root] Event: "drop_database", target: 107, info: {DatabaseName:d27 Statement:DROP DATABASE d27 User:root DroppedSchemaObjects:[]} ``` Please assign, take a look and update the issue accordingly.
1.0
teamcity: failed test: _size_out_of_range_direct=false - The following tests appear to have failed on master (testrace): _size_out_of_range_direct=false You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+_size_out_of_range_direct=false). [#1451701](https://teamcity.cockroachdb.com/viewLog.html?buildId=1451701): ``` _size_out_of_range_direct=false --- FAIL: testrace/TestImportData/PGDUMP:_size_out_of_range_direct=false (0.000s) Test ended in panic. ------- Stdout: ------- I190823 18:30:37.305015 17834 storage/replica_command.go:598 [n1,merge,s1,r63/1:/Table/9{0-2}] initiating a merge of r65:/Table/9{2-4} [(n1,s1):1, next=2, gen=26] into this range (lhs+rhs has (size=0 B+0 B qps=0.00+0.89 --> 0.89qps) below threshold (size=0 B, qps=0.89)) I190823 18:30:37.388071 769 sql/event_log.go:130 [n1,client=127.0.0.1:49446,user=root] Event: "create_database", target: 120, info: {DatabaseName:d34 Statement:CREATE DATABASE d34 User:root} I190823 18:30:37.414493 105 storage/store.go:2593 [n1,s1,r63/1:/Table/9{0-2}] removing replica r65/1 I190823 18:30:37.588024 769 sql/event_log.go:130 [n1,client=127.0.0.1:49446,user=root] Event: "drop_database", target: 120, info: {DatabaseName:d34 Statement:DROP DATABASE d34 User:root DroppedSchemaObjects:[]} _size_out_of_range_direct=false --- FAIL: testrace/TestImportData/PGCOPY:_size_out_of_range_direct=false (0.000s) Test ended in panic. ------- Stdout: ------- I190823 18:30:31.158536 769 sql/event_log.go:130 [n1,client=127.0.0.1:49446,user=root] Event: "create_database", target: 107, info: {DatabaseName:d27 Statement:CREATE DATABASE d27 User:root} I190823 18:30:31.308078 769 sql/event_log.go:130 [n1,client=127.0.0.1:49446,user=root] Event: "drop_database", target: 107, info: {DatabaseName:d27 Statement:DROP DATABASE d27 User:root DroppedSchemaObjects:[]} ``` Please assign, take a look and update the issue accordingly.
non_defect
teamcity failed test size out of range direct false the following tests appear to have failed on master testrace size out of range direct false you may want to check size out of range direct false fail testrace testimportdata pgdump size out of range direct false test ended in panic stdout storage replica command go initiating a merge of table into this range lhs rhs has size b b qps below threshold size b qps sql event log go event create database target info databasename statement create database user root storage store go removing replica sql event log go event drop database target info databasename statement drop database user root droppedschemaobjects size out of range direct false fail testrace testimportdata pgcopy size out of range direct false test ended in panic stdout sql event log go event create database target info databasename statement create database user root sql event log go event drop database target info databasename statement drop database user root droppedschemaobjects please assign take a look and update the issue accordingly
0
48,816
13,184,749,271
IssuesEvent
2020-08-12 20:01:26
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
Gain-dependent saturation in PMT simulator not on? (Trac #263)
Incomplete Migration Migrated from Trac combo simulation defect
<details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/263 , reported by icecube and owned by olivas_</summary> <p> ```json { "status": "closed", "changetime": "2011-05-19T05:16:10", "description": "IceTop LG DOM saturation doesn't match data in latest IceSim RC (V02-05-04-RC). Simulation seems to create 7 Volt waveforms while IceTop IT73 data doesn't. IceTop HG DOMs operate at 1e6 gain, while IceTop LG DOMs operate at 5e5 gain, so maybe this is related to the implementation of the gain-dependent saturation? Or to recent changes in DOMsimulator?", "reporter": "icecube", "cc": "", "resolution": "fixed", "_ts": "1305782170000000", "component": "combo simulation", "summary": "Gain-dependent saturation in PMT simulator not on?", "priority": "major", "keywords": "PMT saturation", "time": "2011-05-17T15:53:51", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
1.0
Gain-dependent saturation in PMT simulator not on? (Trac #263) - <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/263 , reported by icecube and owned by olivas_</summary> <p> ```json { "status": "closed", "changetime": "2011-05-19T05:16:10", "description": "IceTop LG DOM saturation doesn't match data in latest IceSim RC (V02-05-04-RC). Simulation seems to create 7 Volt waveforms while IceTop IT73 data doesn't. IceTop HG DOMs operate at 1e6 gain, while IceTop LG DOMs operate at 5e5 gain, so maybe this is related to the implementation of the gain-dependent saturation? Or to recent changes in DOMsimulator?", "reporter": "icecube", "cc": "", "resolution": "fixed", "_ts": "1305782170000000", "component": "combo simulation", "summary": "Gain-dependent saturation in PMT simulator not on?", "priority": "major", "keywords": "PMT saturation", "time": "2011-05-17T15:53:51", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
defect
gain dependent saturation in pmt simulator not on trac migrated from reported by icecube and owned by olivas json status closed changetime description icetop lg dom saturation doesn t match data in latest icesim rc rc simulation seems to create volt waveforms while icetop data doesn t icetop hg doms operate at gain while icetop lg doms operate at gain so maybe this is related to the implementation of the gain dependent saturation or to recent changes in domsimulator reporter icecube cc resolution fixed ts component combo simulation summary gain dependent saturation in pmt simulator not on priority major keywords pmt saturation time milestone owner olivas type defect
1
828,753
31,841,388,469
IssuesEvent
2023-09-14 16:33:59
Souchy/Celebi
https://api.github.com/repos/Souchy/Celebi
opened
Task: Jolteon improve load speed
task api/ui priority: high
1. Use aggregations 2. Don't call API for things that are not in the current page (creature tab loads only creatures) 3. Don't load things twice 4. Cache
1.0
Task: Jolteon improve load speed - 1. Use aggregations 2. Don't call API for things that are not in the current page (creature tab loads only creatures) 3. Don't load things twice 4. Cache
non_defect
task jolteon improve load speed use aggregations don t call api for things that are not in the current page creature tab loads only creatures don t load things twice cache
0
261,872
19,749,388,825
IssuesEvent
2022-01-15 00:02:58
schmouk/ObjectGL
https://api.github.com/repos/schmouk/ObjectGL
opened
OpenGL Programming Guide 9th- chapt.2 Shader Fundamentals
documentation
The reading of this chapter will lead to a first rough specification of the `shaders`part of the software achitecture.
1.0
OpenGL Programming Guide 9th- chapt.2 Shader Fundamentals - The reading of this chapter will lead to a first rough specification of the `shaders`part of the software achitecture.
non_defect
opengl programming guide chapt shader fundamentals the reading of this chapter will lead to a first rough specification of the shaders part of the software achitecture
0
9,616
2,615,163,725
IssuesEvent
2015-03-01 06:43:28
chrsmith/reaver-wps
https://api.github.com/repos/chrsmith/reaver-wps
opened
Is it password?
auto-migrated Priority-Triage Type-Defect
``` A few things to consider before submitting an issue: 0. We write documentation for a reason, if you have not read it and are having problems with Reaver these pages are required reading before submitting an issue: http://code.google.com/p/reaver-wps/wiki/HintsAndTips http://code.google.com/p/reaver-wps/wiki/README http://code.google.com/p/reaver-wps/wiki/FAQ http://code.google.com/p/reaver-wps/wiki/SupportedWirelessDrivers 1. Reaver will only work if your card is in monitor mode. If you do not know what monitor mode is then you should learn more about 802.11 hacking in linux before using Reaver. 2. Using Reaver against access points you do not own or have permission to attack is illegal. If you cannot answer basic questions (i.e. model number, distance away, etc) about the device you are attacking then do not post your issue here. We will not help you break the law. 3. Please look through issues that have already been posted and make sure your question has not already been asked here: http://code.google.com/p /reaver-wps/issues/list 4. Often times we need packet captures of mon0 while Reaver is running to troubleshoot the issue (tcpdump -i mon0 -s0 -w broken_reaver.pcap). Issue reports with pcap files attached will receive more serious consideration. Answer the following questions for every issue submitted: 0. What version of Reaver are you using? (Only defects against the latest version will be considered.) reaver 1.4 1. What operating system are you using (Linux is the only supported OS)? Ubuntu 12.04 64bit 2. Is your wireless card in monitor mode (yes/no)? Yes 3. What is the signal strength of the Access Point you are trying to crack? -61 4. What is the manufacturer and model # of the device you are trying to crack? Alfa awus o36H rtl8187 5. What is the entire command line string you are supplying to reaver? reaver - i mon0 -b bssid -c x -vv 6. Please describe what you think the issue is. no idea 7. Paste the output from Reaver below. [+] Trying pin 99466978 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received M5 message [+] Sending M6 message [+] Received M7 message [+] Sending WSC NACK [+] Sending WSC NACK [+] Pin cracked in 2696 seconds [+] WPS PIN: '99466978' [+] WPA PSK: 'elizabeth1989' [+] AP SSID: 'Dios es amor' kasper@kasper:~$ sudo reaver -i mon0 -b F4:EC:38:FC:24:EE -c 1 -vv -p 99466978 Reaver v1.4 WiFi Protected Setup Attack Tool Copyright (c) 2011, Tactical Network Solutions, Craig Heffner <cheffner@tacnetsol.com> [+] Switching mon0 to channel 1 [+] Waiting for beacon from F4:EC:38:FC:24:EE [+] Associated with F4:EC:38:FC:24:EE (ESSID: Dios es amor) [+] Trying pin 99466978 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received M5 message [+] Sending M6 message [+] Received M7 message [+] Sending WSC NACK [+] Sending WSC NACK [+] Pin cracked in 4 seconds [+] WPS PIN: '99466978' [+] WPA PSK: 'elizabeth1989' [+] AP SSID: 'Dios es amor' CH 6 ][ Elapsed: 8 s ][ 2012-08-19 16:36 ][ fixed channel mon0: -1 BSSID PWR RXQ Beacons #Data, #/s CH MB ENC CIPHER AUTH ESSID 14:D6:4D:B1:D5:C8 -56 11 13 0 0 6 54e WPA2 CCMP PSK Luisana Evansio kasper@kasper:~$ sudo reaver -i mon0 -b 14:D6:4D:B1:D5:C8 -c 6 -vv -p 56635836 Reaver v1.4 WiFi Protected Setup Attack Tool Copyright (c) 2011, Tactical Network Solutions, Craig Heffner <cheffner@tacnetsol.com> [+] Switching mon0 to channel 6 [+] Waiting for beacon from 14:D6:4D:B1:D5:C8 [+] Associated with 14:D6:4D:B1:D5:C8 (ESSID: Luisana Evansio) [+] Trying pin 56635836 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received M5 message [+] Sending M6 message [+] Received M7 message [+] Sending WSC NACK [+] Sending WSC NACK [+] Pin cracked in 6 seconds [+] WPS PIN: '56635836' [+] WPA PSK: 'luisana15831815' [+] AP SSID: 'Luisana Evansio' BSSID PWR RXQ Beacons #Data, #/s CH MB ENC CIPHER AUTH ESSID F4:EC:38:AC:BF:3C -61 2 70 67 8 1 54e. WPA2 CCMP PSK TP-LINK_ACBF3C kasper@kasper:~$ sudo reaver -i mon0 -b F4:EC:38:AC:BF:3C -c 1 -p 20050092 -vv Reaver v1.4 WiFi Protected Setup Attack Tool Copyright (c) 2011, Tactical Network Solutions, Craig Heffner <cheffner@tacnetsol.com> [+] Switching mon0 to channel 1 [+] Waiting for beacon from F4:EC:38:AC:BF:3C [+] Associated with F4:EC:38:AC:BF:3C (ESSID: TP-LINK_ACBF3C) [+] Trying pin 20050092 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received M5 message [+] Sending M6 message [+] Received M7 message [+] Sending WSC NACK [+] Sending WSC NACK [+] Pin cracked in 4 seconds [+] WPS PIN: '20050092' [+] WPA PSK: '3BCD8597FD2E61A2827C44FD625C6848443C2BD6C3EEE5DA710F1B10C17950EC' [+] AP SSID: 'Network-f4ec38acbf3c' [+] WPS PIN: '20050092' [+] WPA PSK: '4AE1DA1D9BF0733C4A0C1B7D16D5F4E5E4F69C270211B0A3594EC5B4679AFBC7' [+] AP SSID: 'Network-f4ec38acbf3c' +] WPS PIN: '20050092' [+] WPA PSK: 'C18852B2501F7BAC93CADCFA2814E88C12485946B26313A8B844231C7E78AED9' [+] AP SSID: 'Network-f4ec38acbf3c' [+] Nothing done, nothing to save. kasper@kasper:~$ sudo reaver -i mon0 -b F4:EC:38:AC:BF:3C -c 1 -p 20050092 -vv Reaver v1.4 WiFi Protected Setup Attack Tool Copyright (c) 2011, Tactical Network Solutions, Craig Heffner <cheffner@tacnetsol.com> [+] Switching mon0 to channel 1 [+] Waiting for beacon from F4:EC:38:AC:BF:3C [+] Associated with F4:EC:38:AC:BF:3C (ESSID: TP-LINK_ACBF3C) [+] Trying pin 20050092 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received M5 message [+] Sending M6 message [+] Received M7 message [+] Sending WSC NACK [+] Sending WSC NACK [+] Pin cracked in 4 seconds [+] WPS PIN: '20050092' [+] WPA PSK: 'AE4AA94825D00D9E7CE068BF2BD409536ADAA85149AFEA212FE183D3848EBF27' [+] AP SSID: 'Network-f4ec38acbf3c' [+] Nothing done, nothing to save any idea whats problems? i know this question was several times,but no good idea. Like you see reaver properly work,why it give to me unknown psk key thks ``` Original issue reported on code.google.com by `karpe...@gmail.com` on 19 Aug 2012 at 9:31
1.0
Is it password? - ``` A few things to consider before submitting an issue: 0. We write documentation for a reason, if you have not read it and are having problems with Reaver these pages are required reading before submitting an issue: http://code.google.com/p/reaver-wps/wiki/HintsAndTips http://code.google.com/p/reaver-wps/wiki/README http://code.google.com/p/reaver-wps/wiki/FAQ http://code.google.com/p/reaver-wps/wiki/SupportedWirelessDrivers 1. Reaver will only work if your card is in monitor mode. If you do not know what monitor mode is then you should learn more about 802.11 hacking in linux before using Reaver. 2. Using Reaver against access points you do not own or have permission to attack is illegal. If you cannot answer basic questions (i.e. model number, distance away, etc) about the device you are attacking then do not post your issue here. We will not help you break the law. 3. Please look through issues that have already been posted and make sure your question has not already been asked here: http://code.google.com/p /reaver-wps/issues/list 4. Often times we need packet captures of mon0 while Reaver is running to troubleshoot the issue (tcpdump -i mon0 -s0 -w broken_reaver.pcap). Issue reports with pcap files attached will receive more serious consideration. Answer the following questions for every issue submitted: 0. What version of Reaver are you using? (Only defects against the latest version will be considered.) reaver 1.4 1. What operating system are you using (Linux is the only supported OS)? Ubuntu 12.04 64bit 2. Is your wireless card in monitor mode (yes/no)? Yes 3. What is the signal strength of the Access Point you are trying to crack? -61 4. What is the manufacturer and model # of the device you are trying to crack? Alfa awus o36H rtl8187 5. What is the entire command line string you are supplying to reaver? reaver - i mon0 -b bssid -c x -vv 6. Please describe what you think the issue is. no idea 7. Paste the output from Reaver below. [+] Trying pin 99466978 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received M5 message [+] Sending M6 message [+] Received M7 message [+] Sending WSC NACK [+] Sending WSC NACK [+] Pin cracked in 2696 seconds [+] WPS PIN: '99466978' [+] WPA PSK: 'elizabeth1989' [+] AP SSID: 'Dios es amor' kasper@kasper:~$ sudo reaver -i mon0 -b F4:EC:38:FC:24:EE -c 1 -vv -p 99466978 Reaver v1.4 WiFi Protected Setup Attack Tool Copyright (c) 2011, Tactical Network Solutions, Craig Heffner <cheffner@tacnetsol.com> [+] Switching mon0 to channel 1 [+] Waiting for beacon from F4:EC:38:FC:24:EE [+] Associated with F4:EC:38:FC:24:EE (ESSID: Dios es amor) [+] Trying pin 99466978 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received M5 message [+] Sending M6 message [+] Received M7 message [+] Sending WSC NACK [+] Sending WSC NACK [+] Pin cracked in 4 seconds [+] WPS PIN: '99466978' [+] WPA PSK: 'elizabeth1989' [+] AP SSID: 'Dios es amor' CH 6 ][ Elapsed: 8 s ][ 2012-08-19 16:36 ][ fixed channel mon0: -1 BSSID PWR RXQ Beacons #Data, #/s CH MB ENC CIPHER AUTH ESSID 14:D6:4D:B1:D5:C8 -56 11 13 0 0 6 54e WPA2 CCMP PSK Luisana Evansio kasper@kasper:~$ sudo reaver -i mon0 -b 14:D6:4D:B1:D5:C8 -c 6 -vv -p 56635836 Reaver v1.4 WiFi Protected Setup Attack Tool Copyright (c) 2011, Tactical Network Solutions, Craig Heffner <cheffner@tacnetsol.com> [+] Switching mon0 to channel 6 [+] Waiting for beacon from 14:D6:4D:B1:D5:C8 [+] Associated with 14:D6:4D:B1:D5:C8 (ESSID: Luisana Evansio) [+] Trying pin 56635836 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received M5 message [+] Sending M6 message [+] Received M7 message [+] Sending WSC NACK [+] Sending WSC NACK [+] Pin cracked in 6 seconds [+] WPS PIN: '56635836' [+] WPA PSK: 'luisana15831815' [+] AP SSID: 'Luisana Evansio' BSSID PWR RXQ Beacons #Data, #/s CH MB ENC CIPHER AUTH ESSID F4:EC:38:AC:BF:3C -61 2 70 67 8 1 54e. WPA2 CCMP PSK TP-LINK_ACBF3C kasper@kasper:~$ sudo reaver -i mon0 -b F4:EC:38:AC:BF:3C -c 1 -p 20050092 -vv Reaver v1.4 WiFi Protected Setup Attack Tool Copyright (c) 2011, Tactical Network Solutions, Craig Heffner <cheffner@tacnetsol.com> [+] Switching mon0 to channel 1 [+] Waiting for beacon from F4:EC:38:AC:BF:3C [+] Associated with F4:EC:38:AC:BF:3C (ESSID: TP-LINK_ACBF3C) [+] Trying pin 20050092 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received M5 message [+] Sending M6 message [+] Received M7 message [+] Sending WSC NACK [+] Sending WSC NACK [+] Pin cracked in 4 seconds [+] WPS PIN: '20050092' [+] WPA PSK: '3BCD8597FD2E61A2827C44FD625C6848443C2BD6C3EEE5DA710F1B10C17950EC' [+] AP SSID: 'Network-f4ec38acbf3c' [+] WPS PIN: '20050092' [+] WPA PSK: '4AE1DA1D9BF0733C4A0C1B7D16D5F4E5E4F69C270211B0A3594EC5B4679AFBC7' [+] AP SSID: 'Network-f4ec38acbf3c' +] WPS PIN: '20050092' [+] WPA PSK: 'C18852B2501F7BAC93CADCFA2814E88C12485946B26313A8B844231C7E78AED9' [+] AP SSID: 'Network-f4ec38acbf3c' [+] Nothing done, nothing to save. kasper@kasper:~$ sudo reaver -i mon0 -b F4:EC:38:AC:BF:3C -c 1 -p 20050092 -vv Reaver v1.4 WiFi Protected Setup Attack Tool Copyright (c) 2011, Tactical Network Solutions, Craig Heffner <cheffner@tacnetsol.com> [+] Switching mon0 to channel 1 [+] Waiting for beacon from F4:EC:38:AC:BF:3C [+] Associated with F4:EC:38:AC:BF:3C (ESSID: TP-LINK_ACBF3C) [+] Trying pin 20050092 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received M5 message [+] Sending M6 message [+] Received M7 message [+] Sending WSC NACK [+] Sending WSC NACK [+] Pin cracked in 4 seconds [+] WPS PIN: '20050092' [+] WPA PSK: 'AE4AA94825D00D9E7CE068BF2BD409536ADAA85149AFEA212FE183D3848EBF27' [+] AP SSID: 'Network-f4ec38acbf3c' [+] Nothing done, nothing to save any idea whats problems? i know this question was several times,but no good idea. Like you see reaver properly work,why it give to me unknown psk key thks ``` Original issue reported on code.google.com by `karpe...@gmail.com` on 19 Aug 2012 at 9:31
defect
is it password a few things to consider before submitting an issue we write documentation for a reason if you have not read it and are having problems with reaver these pages are required reading before submitting an issue reaver will only work if your card is in monitor mode if you do not know what monitor mode is then you should learn more about hacking in linux before using reaver using reaver against access points you do not own or have permission to attack is illegal if you cannot answer basic questions i e model number distance away etc about the device you are attacking then do not post your issue here we will not help you break the law please look through issues that have already been posted and make sure your question has not already been asked here reaver wps issues list often times we need packet captures of while reaver is running to troubleshoot the issue tcpdump i w broken reaver pcap issue reports with pcap files attached will receive more serious consideration answer the following questions for every issue submitted what version of reaver are you using only defects against the latest version will be considered reaver what operating system are you using linux is the only supported os ubuntu is your wireless card in monitor mode yes no yes what is the signal strength of the access point you are trying to crack what is the manufacturer and model of the device you are trying to crack alfa awus what is the entire command line string you are supplying to reaver reaver i b bssid c x vv please describe what you think the issue is no idea paste the output from reaver below trying pin sending eapol start request received identity request sending identity response received message sending message received message sending message received message sending message received message sending wsc nack sending wsc nack pin cracked in seconds wps pin wpa psk ap ssid dios es amor kasper kasper sudo reaver i b ec fc ee c vv p reaver wifi protected setup attack tool copyright c tactical network solutions craig heffner switching to channel waiting for beacon from ec fc ee associated with ec fc ee essid dios es amor trying pin sending eapol start request received identity request sending identity response received message sending message received message sending message received message sending message received message sending wsc nack sending wsc nack pin cracked in seconds wps pin wpa psk ap ssid dios es amor ch fixed channel bssid pwr rxq beacons data s ch mb enc cipher auth essid ccmp psk luisana evansio kasper kasper sudo reaver i b c vv p reaver wifi protected setup attack tool copyright c tactical network solutions craig heffner switching to channel waiting for beacon from associated with essid luisana evansio trying pin sending eapol start request received identity request sending identity response received message sending message received message sending message received message sending message received message sending wsc nack sending wsc nack pin cracked in seconds wps pin wpa psk ap ssid luisana evansio bssid pwr rxq beacons data s ch mb enc cipher auth essid ec ac bf ccmp psk tp link kasper kasper sudo reaver i b ec ac bf c p vv reaver wifi protected setup attack tool copyright c tactical network solutions craig heffner switching to channel waiting for beacon from ec ac bf associated with ec ac bf essid tp link trying pin sending eapol start request received identity request sending identity response received message sending message received message sending message received message sending message received message sending wsc nack sending wsc nack pin cracked in seconds wps pin wpa psk ap ssid network wps pin wpa psk ap ssid network wps pin wpa psk ap ssid network nothing done nothing to save kasper kasper sudo reaver i b ec ac bf c p vv reaver wifi protected setup attack tool copyright c tactical network solutions craig heffner switching to channel waiting for beacon from ec ac bf associated with ec ac bf essid tp link trying pin sending eapol start request received identity request sending identity response received message sending message received message sending message received message sending message received message sending wsc nack sending wsc nack pin cracked in seconds wps pin wpa psk ap ssid network nothing done nothing to save any idea whats problems i know this question was several times but no good idea like you see reaver properly work why it give to me unknown psk key thks original issue reported on code google com by karpe gmail com on aug at
1
73,309
24,556,996,208
IssuesEvent
2022-10-12 16:40:42
scipy/scipy
https://api.github.com/repos/scipy/scipy
opened
BUG: scipy.stats.mode: inconsistent shape with `axis=None` and `keepdims=True`
defect
### Describe your issue. Depending on whether an array is empty or full, `scipy.stats.mode` with `axis=None` and `keepdims=True` returns inconsistent results: ```python >>> from scipy.stats import mode >>> import numpy as np >>> mode(np.empty((0, 1, 2)), axis=None, keepdims=True)[0].shape (1, 1, 1) >>> mode(np.empty((1, 1, 2)), axis=None, keepdims=True)[0].shape (1,) ``` I'm not sure which is more correct here (the first is correct if `axis=None` should be thought of as a reduction over all axes, while the second is correct if `axis=None` should be thought of as equivalent to ravelling the input) but in either case, I think the outputs should match. ### Reproducing Code Example ```python N/A ``` ### Error message ```shell N/A ``` ### SciPy/NumPy/Python version information 1.9.2 1.23.2 sys.version_info(major=3, minor=8, micro=2, releaselevel='final', serial=0)
1.0
BUG: scipy.stats.mode: inconsistent shape with `axis=None` and `keepdims=True` - ### Describe your issue. Depending on whether an array is empty or full, `scipy.stats.mode` with `axis=None` and `keepdims=True` returns inconsistent results: ```python >>> from scipy.stats import mode >>> import numpy as np >>> mode(np.empty((0, 1, 2)), axis=None, keepdims=True)[0].shape (1, 1, 1) >>> mode(np.empty((1, 1, 2)), axis=None, keepdims=True)[0].shape (1,) ``` I'm not sure which is more correct here (the first is correct if `axis=None` should be thought of as a reduction over all axes, while the second is correct if `axis=None` should be thought of as equivalent to ravelling the input) but in either case, I think the outputs should match. ### Reproducing Code Example ```python N/A ``` ### Error message ```shell N/A ``` ### SciPy/NumPy/Python version information 1.9.2 1.23.2 sys.version_info(major=3, minor=8, micro=2, releaselevel='final', serial=0)
defect
bug scipy stats mode inconsistent shape with axis none and keepdims true describe your issue depending on whether an array is empty or full scipy stats mode with axis none and keepdims true returns inconsistent results python from scipy stats import mode import numpy as np mode np empty axis none keepdims true shape mode np empty axis none keepdims true shape i m not sure which is more correct here the first is correct if axis none should be thought of as a reduction over all axes while the second is correct if axis none should be thought of as equivalent to ravelling the input but in either case i think the outputs should match reproducing code example python n a error message shell n a scipy numpy python version information sys version info major minor micro releaselevel final serial
1
249,397
7,961,374,877
IssuesEvent
2018-07-13 10:31:31
qutech/qc-toolkit
https://api.github.com/repos/qutech/qc-toolkit
closed
None default pulse registry
Priority: Optional
Set the default pulse registry to be None and add a method to conveniently set it to something else.
1.0
None default pulse registry - Set the default pulse registry to be None and add a method to conveniently set it to something else.
non_defect
none default pulse registry set the default pulse registry to be none and add a method to conveniently set it to something else
0
41,639
10,550,926,763
IssuesEvent
2019-10-03 12:16:13
PowerDNS/pdns
https://api.github.com/repos/PowerDNS/pdns
reopened
Records duplication when changing them concurrently via web API
auth defect
<!-- Hi! Thanks for filing an issue. It will be read with care by human beings. Can we ask you to please fill out this template and not simply demand new features or send in complaints? Thanks! --> <!-- Also please search the existing issues (both open and closed) to see if your report might be duplicate --> <!-- Please don't file an issue when you have a support question, send support questions to the mailinglist or ask them on IRC (https://www.powerdns.com/opensource.html) --> <!-- Tell us what is issue is about --> - Program: Authoritative <!-- delete the ones that do not apply --> - Issue type: Bug report <!-- delete the one that does not apply --> ### Short description <!-- Explain in a few sentences what the issue/request is --> Concurrent web API requests `PATCH /servers/{server_id}/zones/{zone_id}`, results in records duplication. Both updated record and SOA record is duplicated. <!-- If this is a bug report, use the following part of the the template and delete the part at the bottom --> ### Environment <!-- Tell us about the environment --> - Operating system: Linux Debian Stretch - Software version: 4.0.3-1+deb9u2 - Software source: Debian repository <!-- e.g. Operating system repository, PowerDNS repository, compiled yourself --> ### Steps to reproduce <!-- Tell us step-by-step how the issue can be triggered. Please include your configuration files and any (Lua) scripts that are loaded. --> 1. pdns authoritative server with web API enabled and gpgsql module enabled, otherwise the default pdns configuration. 2. postgresql server runs on the same host machine 3. <!-- step 3 --> ``` for i in {1..5}; do curl -X PATCH "http://127.1/api/v1/servers/localhost/zones/test.com" \ -H "accept: application/json" -H "X-API-Key: secret" -H "Content-Type: application/json" \ -d "{ \"rrsets\": [ { \"name\": \"a.test.com.\", \"type\": \"CNAME\", \"ttl\": 3600, \"changetype\": \"REPLACE\", \"records\": [ { \"content\": \"test.com.\", \"disabled\": false, \"set-ptr\": false} ] } ] }" & done ``` ### Expected behaviour <!-- What would you expect to happen when the reproduction steps are run --> Expected results of postresql queries ``` pdns=# select type, content from records where domain_id = 1 and name = 'a.test.com' and type='CNAME'; type | content -------+----------- CNAME | test.com (1 rows) ``` ``` pdns=# select type, content from records where domain_id = 1 and type='SOA'; type | content -----+----------------------------------------------------------- SOA | ns.test.com admin.test.com 2018103164 10800 180 604800 60 (1 rows) ``` ### Actual behaviour <!-- What did happen? Please (if possible) provide logs, output from `dig` and/or tcpdump/wireshark data --> ``` pdns=# select type, content from records where domain_id = 1 and name = 'a.test.com' and type='CNAME'; type | content -------+----------- CNAME | test.com CNAME | test.com CNAME | test.com CNAME | test.com (4 rows) ``` ``` pdns=# select type, content from records where domain_id = 1 and type='SOA'; type | content -----+----------------------------------------------------------- SOA | ns.test.com admin.test.com 2018103164 10800 180 604800 60 SOA | ns.test.com admin.test.com 2018103164 10800 180 604800 60 SOA | ns.test.com admin.test.com 2018103164 10800 180 604800 60 SOA | ns.test.com admin.test.com 2018103164 10800 180 604800 60 (4 rows) ``` ### Other information <!-- if you already did more digging into the issue, please provide all the information you gathered --> There should be a transaction in `PacketHandler::performUpdate` ``` BEGIN; SELECT * FROM records WHERE type='CNAME' AND name='a.test.com' AND domain_id=1; DELETE FROM records WHERE type='CNAME' AND name='a.test.com' AND domain_id=1; INSERT ...; COMMIT; ``` with repeatable read isolation level, but I found no transaction in `PacketHandler::performUpdate` or `GSQLBackend::replaceRRSet`.
1.0
Records duplication when changing them concurrently via web API - <!-- Hi! Thanks for filing an issue. It will be read with care by human beings. Can we ask you to please fill out this template and not simply demand new features or send in complaints? Thanks! --> <!-- Also please search the existing issues (both open and closed) to see if your report might be duplicate --> <!-- Please don't file an issue when you have a support question, send support questions to the mailinglist or ask them on IRC (https://www.powerdns.com/opensource.html) --> <!-- Tell us what is issue is about --> - Program: Authoritative <!-- delete the ones that do not apply --> - Issue type: Bug report <!-- delete the one that does not apply --> ### Short description <!-- Explain in a few sentences what the issue/request is --> Concurrent web API requests `PATCH /servers/{server_id}/zones/{zone_id}`, results in records duplication. Both updated record and SOA record is duplicated. <!-- If this is a bug report, use the following part of the the template and delete the part at the bottom --> ### Environment <!-- Tell us about the environment --> - Operating system: Linux Debian Stretch - Software version: 4.0.3-1+deb9u2 - Software source: Debian repository <!-- e.g. Operating system repository, PowerDNS repository, compiled yourself --> ### Steps to reproduce <!-- Tell us step-by-step how the issue can be triggered. Please include your configuration files and any (Lua) scripts that are loaded. --> 1. pdns authoritative server with web API enabled and gpgsql module enabled, otherwise the default pdns configuration. 2. postgresql server runs on the same host machine 3. <!-- step 3 --> ``` for i in {1..5}; do curl -X PATCH "http://127.1/api/v1/servers/localhost/zones/test.com" \ -H "accept: application/json" -H "X-API-Key: secret" -H "Content-Type: application/json" \ -d "{ \"rrsets\": [ { \"name\": \"a.test.com.\", \"type\": \"CNAME\", \"ttl\": 3600, \"changetype\": \"REPLACE\", \"records\": [ { \"content\": \"test.com.\", \"disabled\": false, \"set-ptr\": false} ] } ] }" & done ``` ### Expected behaviour <!-- What would you expect to happen when the reproduction steps are run --> Expected results of postresql queries ``` pdns=# select type, content from records where domain_id = 1 and name = 'a.test.com' and type='CNAME'; type | content -------+----------- CNAME | test.com (1 rows) ``` ``` pdns=# select type, content from records where domain_id = 1 and type='SOA'; type | content -----+----------------------------------------------------------- SOA | ns.test.com admin.test.com 2018103164 10800 180 604800 60 (1 rows) ``` ### Actual behaviour <!-- What did happen? Please (if possible) provide logs, output from `dig` and/or tcpdump/wireshark data --> ``` pdns=# select type, content from records where domain_id = 1 and name = 'a.test.com' and type='CNAME'; type | content -------+----------- CNAME | test.com CNAME | test.com CNAME | test.com CNAME | test.com (4 rows) ``` ``` pdns=# select type, content from records where domain_id = 1 and type='SOA'; type | content -----+----------------------------------------------------------- SOA | ns.test.com admin.test.com 2018103164 10800 180 604800 60 SOA | ns.test.com admin.test.com 2018103164 10800 180 604800 60 SOA | ns.test.com admin.test.com 2018103164 10800 180 604800 60 SOA | ns.test.com admin.test.com 2018103164 10800 180 604800 60 (4 rows) ``` ### Other information <!-- if you already did more digging into the issue, please provide all the information you gathered --> There should be a transaction in `PacketHandler::performUpdate` ``` BEGIN; SELECT * FROM records WHERE type='CNAME' AND name='a.test.com' AND domain_id=1; DELETE FROM records WHERE type='CNAME' AND name='a.test.com' AND domain_id=1; INSERT ...; COMMIT; ``` with repeatable read isolation level, but I found no transaction in `PacketHandler::performUpdate` or `GSQLBackend::replaceRRSet`.
defect
records duplication when changing them concurrently via web api program authoritative issue type bug report short description concurrent web api requests patch servers server id zones zone id results in records duplication both updated record and soa record is duplicated if this is a bug report use the following part of the the template and delete the part at the bottom environment operating system linux debian stretch software version software source debian repository steps to reproduce pdns authoritative server with web api enabled and gpgsql module enabled otherwise the default pdns configuration postgresql server runs on the same host machine for i in do curl x patch h accept application json h x api key secret h content type application json d rrsets done expected behaviour expected results of postresql queries pdns select type content from records where domain id and name a test com and type cname type content cname test com rows pdns select type content from records where domain id and type soa type content soa ns test com admin test com rows actual behaviour pdns select type content from records where domain id and name a test com and type cname type content cname test com cname test com cname test com cname test com rows pdns select type content from records where domain id and type soa type content soa ns test com admin test com soa ns test com admin test com soa ns test com admin test com soa ns test com admin test com rows other information there should be a transaction in packethandler performupdate begin select from records where type cname and name a test com and domain id delete from records where type cname and name a test com and domain id insert commit with repeatable read isolation level but i found no transaction in packethandler performupdate or gsqlbackend replacerrset
1
73,255
9,654,490,102
IssuesEvent
2019-05-19 14:30:11
gii-is-psg2/PSG2-1819-G2-21
https://api.github.com/repos/gii-is-psg2/PSG2-1819-G2-21
closed
SLA & iTop
documentation
Introduction of the document and a screenshot of the SLA generated in iTop, and its association to the maintenance service as created in the iTop tool.
1.0
SLA & iTop - Introduction of the document and a screenshot of the SLA generated in iTop, and its association to the maintenance service as created in the iTop tool.
non_defect
sla itop introduction of the document and a screenshot of the sla generated in itop and its association to the maintenance service as created in the itop tool
0
63,188
17,420,378,065
IssuesEvent
2021-08-04 00:12:13
extnet/Ext.NET
https://api.github.com/repos/extnet/Ext.NET
opened
TabCloseMenu: plug in no longer works as of Ext.NET 5
5.x defect
Found: 5.3.0 Ext.NET Forums' thread: [Plugin TabCloseMenu - closing the tab does not work](https://forums.ext.net/showthread.php?63172) The plug in's functionality has stopped working, possibly between 4.x and 5.x upgrade as, per the override provided along the issue report, now the tab close method should be run over the individual tab panel's `tab` sub property. The same applies to both **Close Tab** and **Close Other Tabs** options, as well as **Close All Tabs**. The change provided in the thread works as a work around for the **Close Tab** menu option. The example for Ext.NET 5 is [Tab Panel > Plugin > TabCloseMenu](https://examples5.ext.net/#/TabPanel/Plugin/TabCloseMenu/) and no option from the plugin, to close tabs, works. In the [equivalent v4 example](https://examples4.ext.net/#/TabPanel/Plugin/TabCloseMenu/) though, the plug in closes all tabs on **Close Other Tabs** and **Close All Tabs**, and closes none on **Close Tab**. So it's behavior in v4 is incomplete yet acts differently than v5's. The support for this plugin is still not implemented for Ext.NET 7, so it only applies to Ext.NET 5. This issue was first fixed in #1648.
1.0
TabCloseMenu: plug in no longer works as of Ext.NET 5 - Found: 5.3.0 Ext.NET Forums' thread: [Plugin TabCloseMenu - closing the tab does not work](https://forums.ext.net/showthread.php?63172) The plug in's functionality has stopped working, possibly between 4.x and 5.x upgrade as, per the override provided along the issue report, now the tab close method should be run over the individual tab panel's `tab` sub property. The same applies to both **Close Tab** and **Close Other Tabs** options, as well as **Close All Tabs**. The change provided in the thread works as a work around for the **Close Tab** menu option. The example for Ext.NET 5 is [Tab Panel > Plugin > TabCloseMenu](https://examples5.ext.net/#/TabPanel/Plugin/TabCloseMenu/) and no option from the plugin, to close tabs, works. In the [equivalent v4 example](https://examples4.ext.net/#/TabPanel/Plugin/TabCloseMenu/) though, the plug in closes all tabs on **Close Other Tabs** and **Close All Tabs**, and closes none on **Close Tab**. So it's behavior in v4 is incomplete yet acts differently than v5's. The support for this plugin is still not implemented for Ext.NET 7, so it only applies to Ext.NET 5. This issue was first fixed in #1648.
defect
tabclosemenu plug in no longer works as of ext net found ext net forums thread the plug in s functionality has stopped working possibly between x and x upgrade as per the override provided along the issue report now the tab close method should be run over the individual tab panel s tab sub property the same applies to both close tab and close other tabs options as well as close all tabs the change provided in the thread works as a work around for the close tab menu option the example for ext net is and no option from the plugin to close tabs works in the though the plug in closes all tabs on close other tabs and close all tabs and closes none on close tab so it s behavior in is incomplete yet acts differently than s the support for this plugin is still not implemented for ext net so it only applies to ext net this issue was first fixed in
1
17,476
3,008,882,379
IssuesEvent
2015-07-28 00:07:07
belangeo/cecilia5
https://api.github.com/repos/belangeo/cecilia5
closed
Donations?
auto-migrated Priority-Medium Type-Defect
``` I think this piece of software is so incredible I'd like to make a donation towards the project. Is that possible? ``` Original issue reported on code.google.com by `oivind.i...@gmail.com` on 9 Apr 2013 at 12:32
1.0
Donations? - ``` I think this piece of software is so incredible I'd like to make a donation towards the project. Is that possible? ``` Original issue reported on code.google.com by `oivind.i...@gmail.com` on 9 Apr 2013 at 12:32
defect
donations i think this piece of software is so incredible i d like to make a donation towards the project is that possible original issue reported on code google com by oivind i gmail com on apr at
1
161,236
13,820,751,948
IssuesEvent
2020-10-13 00:24:50
aws/amazon-chime-sdk-js
https://api.github.com/repos/aws/amazon-chime-sdk-js
closed
"Cannot read property 'addEventListener' of undefined" error on HTTP
Bug Documentation Minor Issue Triaged
**Important:** Please attach INFO-level logs with timestamps to the report from *before* the session started up until the point at which the unexpected behavior occured. For multi-participant calls, attach a log file for each participant if possible. **Meeting and attendee IDs:** List the meeting and attendee IDs that the attached logs apply to. N/A **Timeline** Outline the main actions the user performed and when: N/A **Describe the bug** Describe the bug and the timestamp at which it happened relative to the timestamp in the log file. After deploying the browser demo to a non-HTTPS host, SDK displays this error message when attempting to join a meeting. <img src="https://user-images.githubusercontent.com/36240708/83927697-8dfbf000-a742-11ea-836d-a9cb4135babc.png" width="300"></img> The cause is that media device APIs are only available in secure contexts, e.g. HTTPS/TLS in a browser. The `navigator.mediaDevices` is `undefined` unless a customer explicitly overrides the option. https://developer.mozilla.org/en-US/docs/Web/API/MediaDevices/getUserMedia The SDK demo needs to display a better error message. **Screenshots** If applicable, add screenshots to help explain the problem. **Platform** - OS: macOS/Windows - Browser: Chrome/FF - Browser Version: Tested in Chrome 83.0.4103.61 but recent versions of all modern browsers don't allow the media API access. - Amazon Chime SDK GitHub Commit: c610e1222af3c60cdadac193d7aba3e6afdc0e5a
1.0
"Cannot read property 'addEventListener' of undefined" error on HTTP - **Important:** Please attach INFO-level logs with timestamps to the report from *before* the session started up until the point at which the unexpected behavior occured. For multi-participant calls, attach a log file for each participant if possible. **Meeting and attendee IDs:** List the meeting and attendee IDs that the attached logs apply to. N/A **Timeline** Outline the main actions the user performed and when: N/A **Describe the bug** Describe the bug and the timestamp at which it happened relative to the timestamp in the log file. After deploying the browser demo to a non-HTTPS host, SDK displays this error message when attempting to join a meeting. <img src="https://user-images.githubusercontent.com/36240708/83927697-8dfbf000-a742-11ea-836d-a9cb4135babc.png" width="300"></img> The cause is that media device APIs are only available in secure contexts, e.g. HTTPS/TLS in a browser. The `navigator.mediaDevices` is `undefined` unless a customer explicitly overrides the option. https://developer.mozilla.org/en-US/docs/Web/API/MediaDevices/getUserMedia The SDK demo needs to display a better error message. **Screenshots** If applicable, add screenshots to help explain the problem. **Platform** - OS: macOS/Windows - Browser: Chrome/FF - Browser Version: Tested in Chrome 83.0.4103.61 but recent versions of all modern browsers don't allow the media API access. - Amazon Chime SDK GitHub Commit: c610e1222af3c60cdadac193d7aba3e6afdc0e5a
non_defect
cannot read property addeventlistener of undefined error on http important please attach info level logs with timestamps to the report from before the session started up until the point at which the unexpected behavior occured for multi participant calls attach a log file for each participant if possible meeting and attendee ids list the meeting and attendee ids that the attached logs apply to n a timeline outline the main actions the user performed and when n a describe the bug describe the bug and the timestamp at which it happened relative to the timestamp in the log file after deploying the browser demo to a non https host sdk displays this error message when attempting to join a meeting the cause is that media device apis are only available in secure contexts e g https tls in a browser the navigator mediadevices is undefined unless a customer explicitly overrides the option the sdk demo needs to display a better error message screenshots if applicable add screenshots to help explain the problem platform os macos windows browser chrome ff browser version tested in chrome but recent versions of all modern browsers don t allow the media api access amazon chime sdk github commit
0
36,913
8,194,780,543
IssuesEvent
2018-08-31 01:41:19
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
fourier_uniform returning strange result
defect duplicate scipy.ndimage
When running `fourier_uniform` from SciPy's `ndimage`, seem to be getting an inaccurate result with SciPy 1.0.0+. The problem is not present in SciPy pre-1.0.0 (e.g. 0.19.1). ### Reproducing code example: From SciPy 0.19.1: ```python In [1]: import numpy as np In [2]: import scipy.ndimage as spimg In [3]: spimg.fourier_uniform(np.arange(5, dtype=float), 1) Out[3]: array([ 0. , 0.93548928, 1.51365346, 2.27048019, 3.74195714]) ``` From SciPy 1.1.0: ```python In [1]: import numpy as np In [2]: import scipy.ndimage as spimg In [3]: spimg.fourier_uniform(np.arange(5, dtype=float), 1) Out[3]: array([0.00000000e+000, 5.25819693e-315, 5.28483274e-315, 5.31059453e-315, 5.34108739e-315]) ``` ### Scipy/Numpy/Python version information: ``` 1.1.0 1.15.1 sys.version_info(major=3, minor=6, micro=6, releaselevel='final', serial=0) ```
1.0
fourier_uniform returning strange result - When running `fourier_uniform` from SciPy's `ndimage`, seem to be getting an inaccurate result with SciPy 1.0.0+. The problem is not present in SciPy pre-1.0.0 (e.g. 0.19.1). ### Reproducing code example: From SciPy 0.19.1: ```python In [1]: import numpy as np In [2]: import scipy.ndimage as spimg In [3]: spimg.fourier_uniform(np.arange(5, dtype=float), 1) Out[3]: array([ 0. , 0.93548928, 1.51365346, 2.27048019, 3.74195714]) ``` From SciPy 1.1.0: ```python In [1]: import numpy as np In [2]: import scipy.ndimage as spimg In [3]: spimg.fourier_uniform(np.arange(5, dtype=float), 1) Out[3]: array([0.00000000e+000, 5.25819693e-315, 5.28483274e-315, 5.31059453e-315, 5.34108739e-315]) ``` ### Scipy/Numpy/Python version information: ``` 1.1.0 1.15.1 sys.version_info(major=3, minor=6, micro=6, releaselevel='final', serial=0) ```
defect
fourier uniform returning strange result when running fourier uniform from scipy s ndimage seem to be getting an inaccurate result with scipy the problem is not present in scipy pre e g reproducing code example from scipy python in import numpy as np in import scipy ndimage as spimg in spimg fourier uniform np arange dtype float out array from scipy python in import numpy as np in import scipy ndimage as spimg in spimg fourier uniform np arange dtype float out array scipy numpy python version information sys version info major minor micro releaselevel final serial
1
49,835
13,187,278,733
IssuesEvent
2020-08-13 02:54:35
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
wimpsim-reader - default options are invalid (Trac #2155)
Incomplete Migration Migrated from Trac combo simulation defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2155">https://code.icecube.wisc.edu/ticket/2155</a>, reported by grenzi and owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:15:23", "description": "In [http://software.icecube.wisc.edu/documentation/inspect/wimpsim_reader.html?highlight=i3wimpsim#I3WimpSimReader I3WimpSimReader] we can read:\n\n Param EndMJD:\tDefault = nan, MJD to end simulation; if unspecified: read everything\n\nBut if I try not to set it (and take the NAN default) I receive this error\n\n\n{{{\nERROR (dataclasses): Calling with NAN not possible; will do nothing (I3Time.cxx:142 in void I3Time::SetModJulianTimeDouble(double))\n}}}\n\nThe same is for `StartMJD`.\n\n", "reporter": "grenzi", "cc": "", "resolution": "fixed", "_ts": "1550067323910946", "component": "combo simulation", "summary": "wimpsim-reader - default options are invalid", "priority": "normal", "keywords": "", "time": "2018-05-17T15:48:52", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
1.0
wimpsim-reader - default options are invalid (Trac #2155) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2155">https://code.icecube.wisc.edu/ticket/2155</a>, reported by grenzi and owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:15:23", "description": "In [http://software.icecube.wisc.edu/documentation/inspect/wimpsim_reader.html?highlight=i3wimpsim#I3WimpSimReader I3WimpSimReader] we can read:\n\n Param EndMJD:\tDefault = nan, MJD to end simulation; if unspecified: read everything\n\nBut if I try not to set it (and take the NAN default) I receive this error\n\n\n{{{\nERROR (dataclasses): Calling with NAN not possible; will do nothing (I3Time.cxx:142 in void I3Time::SetModJulianTimeDouble(double))\n}}}\n\nThe same is for `StartMJD`.\n\n", "reporter": "grenzi", "cc": "", "resolution": "fixed", "_ts": "1550067323910946", "component": "combo simulation", "summary": "wimpsim-reader - default options are invalid", "priority": "normal", "keywords": "", "time": "2018-05-17T15:48:52", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
defect
wimpsim reader default options are invalid trac migrated from json status closed changetime description in we can read n n param endmjd tdefault nan mjd to end simulation if unspecified read everything n nbut if i try not to set it and take the nan default i receive this error n n n nerror dataclasses calling with nan not possible will do nothing cxx in void setmodjuliantimedouble double n n nthe same is for startmjd n n reporter grenzi cc resolution fixed ts component combo simulation summary wimpsim reader default options are invalid priority normal keywords time milestone owner nega type defect
1
7,001
2,610,321,414
IssuesEvent
2015-02-26 19:43:40
chrsmith/republic-at-war
https://api.github.com/repos/chrsmith/republic-at-war
closed
Graphics Glitch
auto-migrated Priority-Medium Type-Defect
``` he tail-fin of Acclomators cast shadows on ships around them. I have recreated this bug with fighters, venators, and corvettes. (meaning acclomators cast shadows on them as well) ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 9 May 2011 at 12:11
1.0
Graphics Glitch - ``` he tail-fin of Acclomators cast shadows on ships around them. I have recreated this bug with fighters, venators, and corvettes. (meaning acclomators cast shadows on them as well) ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 9 May 2011 at 12:11
defect
graphics glitch he tail fin of acclomators cast shadows on ships around them i have recreated this bug with fighters venators and corvettes meaning acclomators cast shadows on them as well original issue reported on code google com by gmail com on may at
1
173,842
13,447,464,601
IssuesEvent
2020-09-08 14:16:41
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: tpccbench/nodes=3/cpu=4 failed
C-test-failure O-roachtest O-robot branch-release-19.2 release-blocker
[(roachtest).tpccbench/nodes=3/cpu=4 failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2249658&tab=buildLog) on [release-19.2@5196ff6a149e140aaae8cbbb55df2fc1ca51380a](https://github.com/cockroachdb/cockroach/commits/5196ff6a149e140aaae8cbbb55df2fc1ca51380a): ``` The test failed on branch=release-19.2, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/tpccbench/nodes=3/cpu=4/run_1 cluster.go:1651,context.go:135,cluster.go:1640,test_runner.go:823: dead node detection: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor teamcity-2249658-1599284871-04-n4cpu4 --oneshot --ignore-empty-nodes: exit status 1 4: skipped 1: 20126 2: dead 3: 19134 Error: UNCLASSIFIED_PROBLEM: 2: dead (1) UNCLASSIFIED_PROBLEM Wraps: (2) attached stack trace -- stack trace: | main.glob..func14 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1143 | main.wrap.func1 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:267 | github.com/spf13/cobra.(*Command).execute | /home/agent/work/.go/pkg/mod/github.com/spf13/cobra@v0.0.5/command.go:830 | github.com/spf13/cobra.(*Command).ExecuteC | /home/agent/work/.go/pkg/mod/github.com/spf13/cobra@v0.0.5/command.go:914 | github.com/spf13/cobra.(*Command).Execute | /home/agent/work/.go/pkg/mod/github.com/spf13/cobra@v0.0.5/command.go:864 | main.main | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1839 | runtime.main | /usr/local/go/src/runtime/proc.go:203 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1357 Wraps: (3) 2: dead Error types: (1) errors.Unclassified (2) *withstack.withStack (3) *errutil.leafError ``` <details><summary>More</summary><p> Artifacts: [/tpccbench/nodes=3/cpu=4](https://teamcity.cockroachdb.com/viewLog.html?buildId=2249658&tab=artifacts#/tpccbench/nodes=3/cpu=4) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Atpccbench%2Fnodes%3D3%2Fcpu%3D4.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
2.0
roachtest: tpccbench/nodes=3/cpu=4 failed - [(roachtest).tpccbench/nodes=3/cpu=4 failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2249658&tab=buildLog) on [release-19.2@5196ff6a149e140aaae8cbbb55df2fc1ca51380a](https://github.com/cockroachdb/cockroach/commits/5196ff6a149e140aaae8cbbb55df2fc1ca51380a): ``` The test failed on branch=release-19.2, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/tpccbench/nodes=3/cpu=4/run_1 cluster.go:1651,context.go:135,cluster.go:1640,test_runner.go:823: dead node detection: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor teamcity-2249658-1599284871-04-n4cpu4 --oneshot --ignore-empty-nodes: exit status 1 4: skipped 1: 20126 2: dead 3: 19134 Error: UNCLASSIFIED_PROBLEM: 2: dead (1) UNCLASSIFIED_PROBLEM Wraps: (2) attached stack trace -- stack trace: | main.glob..func14 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1143 | main.wrap.func1 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:267 | github.com/spf13/cobra.(*Command).execute | /home/agent/work/.go/pkg/mod/github.com/spf13/cobra@v0.0.5/command.go:830 | github.com/spf13/cobra.(*Command).ExecuteC | /home/agent/work/.go/pkg/mod/github.com/spf13/cobra@v0.0.5/command.go:914 | github.com/spf13/cobra.(*Command).Execute | /home/agent/work/.go/pkg/mod/github.com/spf13/cobra@v0.0.5/command.go:864 | main.main | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1839 | runtime.main | /usr/local/go/src/runtime/proc.go:203 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1357 Wraps: (3) 2: dead Error types: (1) errors.Unclassified (2) *withstack.withStack (3) *errutil.leafError ``` <details><summary>More</summary><p> Artifacts: [/tpccbench/nodes=3/cpu=4](https://teamcity.cockroachdb.com/viewLog.html?buildId=2249658&tab=artifacts#/tpccbench/nodes=3/cpu=4) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Atpccbench%2Fnodes%3D3%2Fcpu%3D4.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
non_defect
roachtest tpccbench nodes cpu failed on the test failed on branch release cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts tpccbench nodes cpu run cluster go context go cluster go test runner go dead node detection home agent work go src github com cockroachdb cockroach bin roachprod monitor teamcity oneshot ignore empty nodes exit status skipped dead error unclassified problem dead unclassified problem wraps attached stack trace stack trace main glob home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go main wrap home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go github com cobra command execute home agent work go pkg mod github com cobra command go github com cobra command executec home agent work go pkg mod github com cobra command go github com cobra command execute home agent work go pkg mod github com cobra command go main main home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps dead error types errors unclassified withstack withstack errutil leaferror more artifacts powered by
0
626,674
19,830,693,894
IssuesEvent
2022-01-20 11:39:28
GoldenSoftwareLtd/gedemin
https://api.github.com/repos/GoldenSoftwareLtd/gedemin
closed
В документах ККЦ и ОТ добавить в выборе подразделения ограничение по типу подразделения
Type-Enhancement Priority-Low Meat
Originally reported on Google Code with ID 2216 ``` В шапках документов ККЦ и ОТ добавить в лукапе выбора подразделения ограничение по типу подразделения - производство. ``` Reported by `stasgm` on 2010-10-29 12:15:57
1.0
В документах ККЦ и ОТ добавить в выборе подразделения ограничение по типу подразделения - Originally reported on Google Code with ID 2216 ``` В шапках документов ККЦ и ОТ добавить в лукапе выбора подразделения ограничение по типу подразделения - производство. ``` Reported by `stasgm` on 2010-10-29 12:15:57
non_defect
в документах ккц и от добавить в выборе подразделения ограничение по типу подразделения originally reported on google code with id в шапках документов ккц и от добавить в лукапе выбора подразделения ограничение по типу подразделения производство reported by stasgm on
0
9,415
3,039,849,010
IssuesEvent
2015-08-07 11:56:13
ThaDafinser/ZfcDatagrid
https://api.github.com/repos/ThaDafinser/ZfcDatagrid
closed
GridCell shows NaN when using formatter=number together with Type\Number
short term Verify/test needed
My data contains decimal values from my database that I would like to show with `jqgrid` locale formatter number for locale **de**: https://github.com/tonytomov/jqGrid/blob/master/js/i18n/grid.locale-de.js#L102 ```javascript``` number : {decimalSeparator:",", thousandsSeparator: ".", decimalPlaces: 2, defaultValue: '0,00'}, ``` This is my column config using the `rendererParameter` setting **formatter** to **number**: ```php $col = new Column\Select('workDuration_week'); $col->setLabel('Total'); $col->setRendererParameter('formatter', 'number'); $grid->addColumn($col); ``` Given from my `dataSource` is the value **11.50**. The grid shows **11,50** as expected. Of course I would like to align the number that's why I add the *ZfcDatagrid\Column\Type\Number*: ```php $col->setType(new Type\Number(NumberFormatter::DECIMAL, NumberFormatter::TYPE_DEFAULT, 'de_DE')); ``` Now the `gridcell` shows **NaN** for my **11.50** value. But it seems to work fine using data like **3.00** which correctely produces **3,00**! I then removed the **formatter** and left the *Type\Number* only. Now the **11.50** value is displayed as **11,5** with `align="right"`. How come the `gridcell` become a **NaN** value? Since it is a javascript parsing problem it looks like `ZfcDataGrid` is incorrectely parsing the value for an **xx,5x* data value.
1.0
GridCell shows NaN when using formatter=number together with Type\Number - My data contains decimal values from my database that I would like to show with `jqgrid` locale formatter number for locale **de**: https://github.com/tonytomov/jqGrid/blob/master/js/i18n/grid.locale-de.js#L102 ```javascript``` number : {decimalSeparator:",", thousandsSeparator: ".", decimalPlaces: 2, defaultValue: '0,00'}, ``` This is my column config using the `rendererParameter` setting **formatter** to **number**: ```php $col = new Column\Select('workDuration_week'); $col->setLabel('Total'); $col->setRendererParameter('formatter', 'number'); $grid->addColumn($col); ``` Given from my `dataSource` is the value **11.50**. The grid shows **11,50** as expected. Of course I would like to align the number that's why I add the *ZfcDatagrid\Column\Type\Number*: ```php $col->setType(new Type\Number(NumberFormatter::DECIMAL, NumberFormatter::TYPE_DEFAULT, 'de_DE')); ``` Now the `gridcell` shows **NaN** for my **11.50** value. But it seems to work fine using data like **3.00** which correctely produces **3,00**! I then removed the **formatter** and left the *Type\Number* only. Now the **11.50** value is displayed as **11,5** with `align="right"`. How come the `gridcell` become a **NaN** value? Since it is a javascript parsing problem it looks like `ZfcDataGrid` is incorrectely parsing the value for an **xx,5x* data value.
non_defect
gridcell shows nan when using formatter number together with type number my data contains decimal values from my database that i would like to show with jqgrid locale formatter number for locale de javascript number decimalseparator thousandsseparator decimalplaces defaultvalue this is my column config using the rendererparameter setting formatter to number php col new column select workduration week col setlabel total col setrendererparameter formatter number grid addcolumn col given from my datasource is the value the grid shows as expected of course i would like to align the number that s why i add the zfcdatagrid column type number php col settype new type number numberformatter decimal numberformatter type default de de now the gridcell shows nan for my value but it seems to work fine using data like which correctely produces i then removed the formatter and left the type number only now the value is displayed as with align right how come the gridcell become a nan value since it is a javascript parsing problem it looks like zfcdatagrid is incorrectely parsing the value for an xx data value
0
90,697
26,171,722,079
IssuesEvent
2023-01-02 01:15:09
CGNS/CGNS
https://api.github.com/repos/CGNS/CGNS
opened
[CGNS-94] configure fails to find zlib
bug Build Major To Do
> This issue has been imported from JIRA. Read the [original ticket here](https://cgnsorg.atlassian.net/browse/CGNS-94). - _**Created at:**_ Mon, 18 Apr 2016 14:40:51 -0500 <p>configure.ac looks for the zlib header in the zlib dir<br/> if test -f "$zlibdir/zlib.h"; then<br/> but it should be zlibdir/include/zlib.h<br/> for the lib, it should be zlibdir/lib/libz.*</p>
1.0
[CGNS-94] configure fails to find zlib - > This issue has been imported from JIRA. Read the [original ticket here](https://cgnsorg.atlassian.net/browse/CGNS-94). - _**Created at:**_ Mon, 18 Apr 2016 14:40:51 -0500 <p>configure.ac looks for the zlib header in the zlib dir<br/> if test -f "$zlibdir/zlib.h"; then<br/> but it should be zlibdir/include/zlib.h<br/> for the lib, it should be zlibdir/lib/libz.*</p>
non_defect
configure fails to find zlib this issue has been imported from jira read the created at mon apr configure ac looks for the zlib header in the zlib dir if test f zlibdir zlib h then but it should be zlibdir include zlib h for the lib it should be zlibdir lib libz
0
23,291
11,901,081,435
IssuesEvent
2020-03-30 11:49:42
Altinn/altinn-studio
https://api.github.com/repos/Altinn/altinn-studio
opened
Enable pod-to-pod communication in Platform
kind/user-story ops/performance solution/platform
## Description Performance testing shows that communication between components in Platform (as expected) goes through both Application Gateway and API Management. This causes problems: - Complex call stacks that is hard to reason about - High latency and bad performance because of unneccesary network calls and TLS-termination - Scaling/cost of network infrastructure (gateway and APIM) and unneccessary logging ## Acceptance criteria All calls from Platform to Platform is using linkerd service mesh (is already installed) instead of external adresses. ## Development tasks - [ ] All Platform-components should use linkerd service mesh for inter-platform communication. - [ ] Storage - [ ] Authorisation - [ ] Authentication - [ ] Profile - [ ] Register
True
Enable pod-to-pod communication in Platform - ## Description Performance testing shows that communication between components in Platform (as expected) goes through both Application Gateway and API Management. This causes problems: - Complex call stacks that is hard to reason about - High latency and bad performance because of unneccesary network calls and TLS-termination - Scaling/cost of network infrastructure (gateway and APIM) and unneccessary logging ## Acceptance criteria All calls from Platform to Platform is using linkerd service mesh (is already installed) instead of external adresses. ## Development tasks - [ ] All Platform-components should use linkerd service mesh for inter-platform communication. - [ ] Storage - [ ] Authorisation - [ ] Authentication - [ ] Profile - [ ] Register
non_defect
enable pod to pod communication in platform description performance testing shows that communication between components in platform as expected goes through both application gateway and api management this causes problems complex call stacks that is hard to reason about high latency and bad performance because of unneccesary network calls and tls termination scaling cost of network infrastructure gateway and apim and unneccessary logging acceptance criteria all calls from platform to platform is using linkerd service mesh is already installed instead of external adresses development tasks all platform components should use linkerd service mesh for inter platform communication storage authorisation authentication profile register
0
64,526
18,724,551,141
IssuesEvent
2021-11-03 15:05:29
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Riot/Chrome state can result in inability to load new messages
T-Defect P1 S-Major
### Description Riot in Chrome has gotten into a sate where only the messages that were received up to a certain point can be loaded. Trying to view a room with unread messages results in a loading spinner and eventual error: > "Connectivity to the server has been lost. > Sent messages will be stored until your connection has returned." Trying to post a message seems to work, but the message is gone upon refreshing the page. Viewing a room with new messages (so that they are marked as read and the room isn't in bold anymore) is reset when refreshing the page. It seems nothing can be saved or loaded from after a certain point in time. When logging in from an incognito window, the messages that were attempted to send but not visible after refreshing the page *did* in fact go through, but riot couldn't save them locally. Have also had strange behavior in Firefox #5173 ### Steps to reproduce - Reproducibility unknown Log: sent ### Version information **Platform**: web (in-browser) For the web app: - **Browser**: Chrome Beta v62 - **OS**: Fedora 27 - **URL**: affecting /app /experimental and /develop
1.0
Riot/Chrome state can result in inability to load new messages - ### Description Riot in Chrome has gotten into a sate where only the messages that were received up to a certain point can be loaded. Trying to view a room with unread messages results in a loading spinner and eventual error: > "Connectivity to the server has been lost. > Sent messages will be stored until your connection has returned." Trying to post a message seems to work, but the message is gone upon refreshing the page. Viewing a room with new messages (so that they are marked as read and the room isn't in bold anymore) is reset when refreshing the page. It seems nothing can be saved or loaded from after a certain point in time. When logging in from an incognito window, the messages that were attempted to send but not visible after refreshing the page *did* in fact go through, but riot couldn't save them locally. Have also had strange behavior in Firefox #5173 ### Steps to reproduce - Reproducibility unknown Log: sent ### Version information **Platform**: web (in-browser) For the web app: - **Browser**: Chrome Beta v62 - **OS**: Fedora 27 - **URL**: affecting /app /experimental and /develop
defect
riot chrome state can result in inability to load new messages description riot in chrome has gotten into a sate where only the messages that were received up to a certain point can be loaded trying to view a room with unread messages results in a loading spinner and eventual error connectivity to the server has been lost sent messages will be stored until your connection has returned trying to post a message seems to work but the message is gone upon refreshing the page viewing a room with new messages so that they are marked as read and the room isn t in bold anymore is reset when refreshing the page it seems nothing can be saved or loaded from after a certain point in time when logging in from an incognito window the messages that were attempted to send but not visible after refreshing the page did in fact go through but riot couldn t save them locally have also had strange behavior in firefox steps to reproduce reproducibility unknown log sent version information platform web in browser for the web app browser chrome beta os fedora url affecting app experimental and develop
1
255,192
21,911,615,574
IssuesEvent
2022-05-21 06:05:19
MohistMC/Mohist
https://api.github.com/repos/MohistMC/Mohist
closed
[1.16.5] The bug that has been developing for some time has not been fixed.
1.16.5 Wait Needs Testing
## Informations **The Minecraft Version** is 1.16.5 **The Mohist Version** is 990 **Attachments : **https://cdn.discordapp.com/attachments/968819857124626462/977153501153079316/latest.log, https://cdn.discordapp.com/attachments/968819857124626462/977153556211699742/latest.log **Reported by miku1881 with id 320886072416665602** ## Description ## How to reproduce Using the locate command
1.0
[1.16.5] The bug that has been developing for some time has not been fixed. - ## Informations **The Minecraft Version** is 1.16.5 **The Mohist Version** is 990 **Attachments : **https://cdn.discordapp.com/attachments/968819857124626462/977153501153079316/latest.log, https://cdn.discordapp.com/attachments/968819857124626462/977153556211699742/latest.log **Reported by miku1881 with id 320886072416665602** ## Description ## How to reproduce Using the locate command
non_defect
the bug that has been developing for some time has not been fixed informations the minecraft version is the mohist version is attachments reported by with id description how to reproduce using the locate command
0
112,271
14,234,514,838
IssuesEvent
2020-11-18 13:42:07
gnosis/conditional-tokens-explorer
https://api.github.com/repos/gnosis/conditional-tokens-explorer
opened
Position id is not pre-selected when navigate to the Split position page from Positions list/Position details page
Medium priority New design bug
Related to #583, #515 See the video: position is not pre-selected when navigate to the Split position page (like it works on the current app version) https://drive.google.com/file/d/1PPTw8Ky5f6eBSWBAM4BGf2utifJ_RozR/view **Expected Result:** the position should be preselected when navigate to the lit position page from Positions list/Position details page
1.0
Position id is not pre-selected when navigate to the Split position page from Positions list/Position details page - Related to #583, #515 See the video: position is not pre-selected when navigate to the Split position page (like it works on the current app version) https://drive.google.com/file/d/1PPTw8Ky5f6eBSWBAM4BGf2utifJ_RozR/view **Expected Result:** the position should be preselected when navigate to the lit position page from Positions list/Position details page
non_defect
position id is not pre selected when navigate to the split position page from positions list position details page related to see the video position is not pre selected when navigate to the split position page like it works on the current app version expected result the position should be preselected when navigate to the lit position page from positions list position details page
0
41,870
10,685,750,498
IssuesEvent
2019-10-22 13:16:39
mozilla-lockwise/lockwise-android
https://api.github.com/repos/mozilla-lockwise/lockwise-android
opened
Dismissing opening website from long tap in website field takes user to Entry List view
type: defect
## Steps to reproduce Open an entry detail view Long tap on the url Dismiss opening the website by tapping on any other part of the screen ### Expected behavior User is taken back to the entry detail ### Actual behavior User is taken to entry list view ### Device & build information * Device: Pixel 3 * Build version: v3.0.0 RC-2 ### Notes Attachments: See [video](https://youtu.be/ue-D5GbZebs)
1.0
Dismissing opening website from long tap in website field takes user to Entry List view - ## Steps to reproduce Open an entry detail view Long tap on the url Dismiss opening the website by tapping on any other part of the screen ### Expected behavior User is taken back to the entry detail ### Actual behavior User is taken to entry list view ### Device & build information * Device: Pixel 3 * Build version: v3.0.0 RC-2 ### Notes Attachments: See [video](https://youtu.be/ue-D5GbZebs)
defect
dismissing opening website from long tap in website field takes user to entry list view steps to reproduce open an entry detail view long tap on the url dismiss opening the website by tapping on any other part of the screen expected behavior user is taken back to the entry detail actual behavior user is taken to entry list view device build information device pixel build version rc notes attachments see
1
200,176
7,001,034,631
IssuesEvent
2017-12-18 08:39:56
opentx/opentx
https://api.github.com/repos/opentx/opentx
closed
Inconsistent behavior for model conversions
Bug Companion Priority-Low
Scenario: A model window is open. The model window has the same setting for radio type as the current profile (The normal case). If the current profile is switched to a profile for another radio the models are automatically converted to the new radio type. If the radio type in the current profiles is switched, the models are however not converted. I can not see that either alternative is more logical than the other, but they are inconsistent.
1.0
Inconsistent behavior for model conversions - Scenario: A model window is open. The model window has the same setting for radio type as the current profile (The normal case). If the current profile is switched to a profile for another radio the models are automatically converted to the new radio type. If the radio type in the current profiles is switched, the models are however not converted. I can not see that either alternative is more logical than the other, but they are inconsistent.
non_defect
inconsistent behavior for model conversions scenario a model window is open the model window has the same setting for radio type as the current profile the normal case if the current profile is switched to a profile for another radio the models are automatically converted to the new radio type if the radio type in the current profiles is switched the models are however not converted i can not see that either alternative is more logical than the other but they are inconsistent
0
744
4,153,979,996
IssuesEvent
2016-06-16 09:48:37
ElderByte-/Warden
https://api.github.com/repos/ElderByte-/Warden
closed
Don't expose H2 console in embedded mode
type: architecture
In embedded mode, there is no guarantee which datasource the spring data repositories are using as they get picked up dynamically.
1.0
Don't expose H2 console in embedded mode - In embedded mode, there is no guarantee which datasource the spring data repositories are using as they get picked up dynamically.
non_defect
don t expose console in embedded mode in embedded mode there is no guarantee which datasource the spring data repositories are using as they get picked up dynamically
0
85,662
15,755,125,467
IssuesEvent
2021-03-31 01:13:46
tamirverthim/NodeGoat
https://api.github.com/repos/tamirverthim/NodeGoat
closed
CVE-2011-4969 (Medium) detected in jquery-1.4.4.min.js - autoclosed
security vulnerability
## CVE-2011-4969 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.4.4.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.4/jquery.min.js</a></p> <p>Path to dependency file: NodeGoat/node_modules/selenium-webdriver/lib/test/data/droppableItems.html</p> <p>Path to vulnerable library: NodeGoat/node_modules/selenium-webdriver/lib/test/data/js/jquery-1.4.4.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.4.4.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/tamirverthim/NodeGoat/commit/b122231e3322e09af113633cfa74ba5f9014c375">b122231e3322e09af113633cfa74ba5f9014c375</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Cross-site scripting (XSS) vulnerability in jQuery before 1.6.3, when using location.hash to select elements, allows remote attackers to inject arbitrary web script or HTML via a crafted tag. <p>Publish Date: 2013-03-08 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2011-4969>CVE-2011-4969</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>4.3</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2011-4969">https://nvd.nist.gov/vuln/detail/CVE-2011-4969</a></p> <p>Release Date: 2013-03-08</p> <p>Fix Resolution: 1.6.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.4.4","isTransitiveDependency":false,"dependencyTree":"jquery:1.4.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.6.3"}],"vulnerabilityIdentifier":"CVE-2011-4969","vulnerabilityDetails":"Cross-site scripting (XSS) vulnerability in jQuery before 1.6.3, when using location.hash to select elements, allows remote attackers to inject arbitrary web script or HTML via a crafted tag.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2011-4969","cvss2Severity":"medium","cvss2Score":"4.3","extraData":{}}</REMEDIATE> -->
True
CVE-2011-4969 (Medium) detected in jquery-1.4.4.min.js - autoclosed - ## CVE-2011-4969 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.4.4.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.4/jquery.min.js</a></p> <p>Path to dependency file: NodeGoat/node_modules/selenium-webdriver/lib/test/data/droppableItems.html</p> <p>Path to vulnerable library: NodeGoat/node_modules/selenium-webdriver/lib/test/data/js/jquery-1.4.4.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.4.4.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/tamirverthim/NodeGoat/commit/b122231e3322e09af113633cfa74ba5f9014c375">b122231e3322e09af113633cfa74ba5f9014c375</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Cross-site scripting (XSS) vulnerability in jQuery before 1.6.3, when using location.hash to select elements, allows remote attackers to inject arbitrary web script or HTML via a crafted tag. <p>Publish Date: 2013-03-08 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2011-4969>CVE-2011-4969</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>4.3</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2011-4969">https://nvd.nist.gov/vuln/detail/CVE-2011-4969</a></p> <p>Release Date: 2013-03-08</p> <p>Fix Resolution: 1.6.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.4.4","isTransitiveDependency":false,"dependencyTree":"jquery:1.4.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.6.3"}],"vulnerabilityIdentifier":"CVE-2011-4969","vulnerabilityDetails":"Cross-site scripting (XSS) vulnerability in jQuery before 1.6.3, when using location.hash to select elements, allows remote attackers to inject arbitrary web script or HTML via a crafted tag.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2011-4969","cvss2Severity":"medium","cvss2Score":"4.3","extraData":{}}</REMEDIATE> -->
non_defect
cve medium detected in jquery min js autoclosed cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file nodegoat node modules selenium webdriver lib test data droppableitems html path to vulnerable library nodegoat node modules selenium webdriver lib test data js jquery min js dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details cross site scripting xss vulnerability in jquery before when using location hash to select elements allows remote attackers to inject arbitrary web script or html via a crafted tag publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails cross site scripting xss vulnerability in jquery before when using location hash to select elements allows remote attackers to inject arbitrary web script or html via a crafted tag vulnerabilityurl
0
369,233
10,894,230,826
IssuesEvent
2019-11-19 08:11:40
threefoldtech/jumpscaleX_core
https://api.github.com/repos/threefoldtech/jumpscaleX_core
opened
release Jumpscale 10.1
priority_critical type_feature
- [ ] make sure there are no branches left which are not on stories in home - [ ] make sure all branches for 10.1 are merged into development - [ ] run the autotests we have on development - [ ] use jsx threebotbuilder ... to see if it works (builds) - [ ] stories which cannot be finished in 10.1 should be discussed with stakeholders & moved to 10.2 - [ ] once tests done -> development -> master so we are ready for new development branch for 10.2
1.0
release Jumpscale 10.1 - - [ ] make sure there are no branches left which are not on stories in home - [ ] make sure all branches for 10.1 are merged into development - [ ] run the autotests we have on development - [ ] use jsx threebotbuilder ... to see if it works (builds) - [ ] stories which cannot be finished in 10.1 should be discussed with stakeholders & moved to 10.2 - [ ] once tests done -> development -> master so we are ready for new development branch for 10.2
non_defect
release jumpscale make sure there are no branches left which are not on stories in home make sure all branches for are merged into development run the autotests we have on development use jsx threebotbuilder to see if it works builds stories which cannot be finished in should be discussed with stakeholders moved to once tests done development master so we are ready for new development branch for
0
158,693
13,745,784,746
IssuesEvent
2020-10-06 03:53:17
msandfor/10-Easy-Steps
https://api.github.com/repos/msandfor/10-Easy-Steps
closed
[Hacktoberfest]: Add "JSNation" to the Conferences Section
beginner beginner-friendly documentation :memo: first-contribution good first issue hacktoberfest :children_crossing: help wanted :hand: up-for-grabs
🆕🐥☝ Beginners Only: This issue is reserved for people who are new to Open Source. We know that the process of creating a pull request is the biggest barrier for new contributors. This issue is for you 💝 To do: * Add [JSNation](https://jsnation.com/) to the Conferences Section * See below for link to step-by-step tutorial * Go down to Conferences section - it starts at around line 173 * Using the following template, put in the correct info for that conference (Search for it on the internet to find the correct details) underneath the last row in the table, ensuring there is still an empty row before the next table. `| [Conference Name](Website Address) | Conference Name or Organising Org | 10 | Description of the Conferences, found on their website |` * Add your details in the contributors section at the bottom of the page using the same formatting as the last one 📋 *Step by Step* [Step-By-Step Tutorial](https://github.com/msandfor/10-Easy-Steps/blob/master/docs/GitHub-browser-tutorial.md) Remember to observe the [Code of Conduct](https://github.com/msandfor/10-Easy-Steps/blob/master/code_of_conduct.md) - by which we mean, in short - Be Kind, Be Positive, Be Helpful
1.0
[Hacktoberfest]: Add "JSNation" to the Conferences Section - 🆕🐥☝ Beginners Only: This issue is reserved for people who are new to Open Source. We know that the process of creating a pull request is the biggest barrier for new contributors. This issue is for you 💝 To do: * Add [JSNation](https://jsnation.com/) to the Conferences Section * See below for link to step-by-step tutorial * Go down to Conferences section - it starts at around line 173 * Using the following template, put in the correct info for that conference (Search for it on the internet to find the correct details) underneath the last row in the table, ensuring there is still an empty row before the next table. `| [Conference Name](Website Address) | Conference Name or Organising Org | 10 | Description of the Conferences, found on their website |` * Add your details in the contributors section at the bottom of the page using the same formatting as the last one 📋 *Step by Step* [Step-By-Step Tutorial](https://github.com/msandfor/10-Easy-Steps/blob/master/docs/GitHub-browser-tutorial.md) Remember to observe the [Code of Conduct](https://github.com/msandfor/10-Easy-Steps/blob/master/code_of_conduct.md) - by which we mean, in short - Be Kind, Be Positive, Be Helpful
non_defect
add jsnation to the conferences section 🆕🐥☝ beginners only this issue is reserved for people who are new to open source we know that the process of creating a pull request is the biggest barrier for new contributors this issue is for you 💝 to do add to the conferences section see below for link to step by step tutorial go down to conferences section it starts at around line using the following template put in the correct info for that conference search for it on the internet to find the correct details underneath the last row in the table ensuring there is still an empty row before the next table website address conference name or organising org description of the conferences found on their website add your details in the contributors section at the bottom of the page using the same formatting as the last one 📋 step by step remember to observe the by which we mean in short be kind be positive be helpful
0
276,450
23,992,930,780
IssuesEvent
2022-09-14 04:03:50
Tencent/bk-ci
https://api.github.com/repos/Tencent/bk-ci
closed
支持插件通过 BK_CI_START_TYPE 获取 stream 触发的流水线的启动方式
for test grayed streams/for test streams/for gray streams/grayed streams/done area/stream/backend
- 插件通过 BK_CI_START_TYPE 获取启动方式时,无论是 蓝盾触发还是 stream 触发,都应该能获取到启动方式 - 构建启动方式,MANUAL/TIME_TRIGGER/WEB_HOOK/SERVICE/PIPELINE/REMOTE中取值 注意:不改变 stream 原来的上下文(如ci.event) 修改方案: 将stream调用的启动方式收归成蓝盾的,以达到在蓝盾和stream触发时使用 BK_CI_START_TYPE 可以达到相同的参数 MANUAL --手动触发 TIME_TRIGGER --定时触发 WEB_HOOK --git触发,repo_hook SERVICE --openapi触发 PIPELINE --子流水线调用
2.0
支持插件通过 BK_CI_START_TYPE 获取 stream 触发的流水线的启动方式 - - 插件通过 BK_CI_START_TYPE 获取启动方式时,无论是 蓝盾触发还是 stream 触发,都应该能获取到启动方式 - 构建启动方式,MANUAL/TIME_TRIGGER/WEB_HOOK/SERVICE/PIPELINE/REMOTE中取值 注意:不改变 stream 原来的上下文(如ci.event) 修改方案: 将stream调用的启动方式收归成蓝盾的,以达到在蓝盾和stream触发时使用 BK_CI_START_TYPE 可以达到相同的参数 MANUAL --手动触发 TIME_TRIGGER --定时触发 WEB_HOOK --git触发,repo_hook SERVICE --openapi触发 PIPELINE --子流水线调用
non_defect
支持插件通过 bk ci start type 获取 stream 触发的流水线的启动方式 插件通过 bk ci start type 获取启动方式时,无论是 蓝盾触发还是 stream 触发,都应该能获取到启动方式 构建启动方式,manual time trigger web hook service pipeline remote中取值 注意:不改变 stream 原来的上下文(如ci event) 修改方案: 将stream调用的启动方式收归成蓝盾的,以达到在蓝盾和stream触发时使用 bk ci start type 可以达到相同的参数 manual 手动触发 time trigger 定时触发 web hook git触发,repo hook service openapi触发 pipeline 子流水线调用
0
31,165
6,443,901,663
IssuesEvent
2017-08-12 02:19:59
opendatakit/opendatakit
https://api.github.com/repos/opendatakit/opendatakit
closed
Failure to publish when user is a non-gmail.com account (i.e., Google apps domain)
1.4 Aggregate Priority-High Type-Defect
Originally reported on Google Code with ID 776 ``` If you have a google-hosted domain, e.g., @camfed.org, and you try to publish, publishing is apparently broken. No error is captured in the log(!). ``` Reported by `mitchellsundt` on 2013-02-28 23:02:36
1.0
Failure to publish when user is a non-gmail.com account (i.e., Google apps domain) - Originally reported on Google Code with ID 776 ``` If you have a google-hosted domain, e.g., @camfed.org, and you try to publish, publishing is apparently broken. No error is captured in the log(!). ``` Reported by `mitchellsundt` on 2013-02-28 23:02:36
defect
failure to publish when user is a non gmail com account i e google apps domain originally reported on google code with id if you have a google hosted domain e g camfed org and you try to publish publishing is apparently broken no error is captured in the log reported by mitchellsundt on
1
9,640
11,708,833,102
IssuesEvent
2020-03-08 15:30:22
SaucyPigeon/Wild-Cultivation-Fan-Update
https://api.github.com/repos/SaucyPigeon/Wild-Cultivation-Fan-Update
closed
Seeds Please
mod compatibility
Mod SaucyPigeon.WildCultivation has multiple Verse.BuildableDefs named Seed_Agave. Skipping. Verse.Log:Error(String, Boolean) Verse.DefDatabase`1:AddAllInMods() System.Reflection.MonoMethod:InternalInvoke(Object, Object[], Exception&) System.Reflection.MonoMethod:Invoke(Object, BindingFlags, Binder, Object[], CultureInfo) System.Reflection.MethodBase:Invoke(Object, Object[]) Verse.GenGeneric:InvokeStaticMethodOnGenericType(Type, Type, String) Verse.PlayDataLoader:DoPlayLoad() Verse.PlayDataLoader:LoadAllPlayData(Boolean) Verse.<>c:<Start>b__6_1() Verse.LongEventHandler:RunEventFromAnotherThread(Action) Verse.<>c:<UpdateCurrentAsynchronousEvent>b__27_0() System.Threading.ThreadHelper:ThreadStart_Context(Object) System.Threading.ExecutionContext:RunInternal(ExecutionContext, ContextCallback, Object, Boolean) System.Threading.ExecutionContext:Run(ExecutionContext, ContextCallback, Object, Boolean) System.Threading.ExecutionContext:Run(ExecutionContext, ContextCallback, Object) System.Threading.ThreadHelper:ThreadStart() Mod SaucyPigeon.WildCultivation has multiple Verse.ThingDefs named Seed_Agave. Skipping. Verse.Log:Error(String, Boolean) Verse.DefDatabase`1:AddAllInMods() System.Reflection.MonoMethod:InternalInvoke(Object, Object[], Exception&) System.Reflection.MonoMethod:Invoke(Object, BindingFlags, Binder, Object[], CultureInfo) System.Reflection.MethodBase:Invoke(Object, Object[]) Verse.GenGeneric:InvokeStaticMethodOnGenericType(Type, Type, String) Verse.PlayDataLoader:DoPlayLoad() Verse.PlayDataLoader:LoadAllPlayData(Boolean) Verse.<>c:<Start>b__6_1() Verse.LongEventHandler:RunEventFromAnotherThread(Action) Verse.<>c:<UpdateCurrentAsynchronousEvent>b__27_0() System.Threading.ThreadHelper:ThreadStart_Context(Object) System.Threading.ExecutionContext:RunInternal(ExecutionContext, ContextCallback, Object, Boolean) System.Threading.ExecutionContext:Run(ExecutionContext, ContextCallback, Object, Boolean) System.Threading.ExecutionContext:Run(ExecutionContext, ContextCallback, Object) System.Threading.ThreadHelper:ThreadStart()
True
Seeds Please - Mod SaucyPigeon.WildCultivation has multiple Verse.BuildableDefs named Seed_Agave. Skipping. Verse.Log:Error(String, Boolean) Verse.DefDatabase`1:AddAllInMods() System.Reflection.MonoMethod:InternalInvoke(Object, Object[], Exception&) System.Reflection.MonoMethod:Invoke(Object, BindingFlags, Binder, Object[], CultureInfo) System.Reflection.MethodBase:Invoke(Object, Object[]) Verse.GenGeneric:InvokeStaticMethodOnGenericType(Type, Type, String) Verse.PlayDataLoader:DoPlayLoad() Verse.PlayDataLoader:LoadAllPlayData(Boolean) Verse.<>c:<Start>b__6_1() Verse.LongEventHandler:RunEventFromAnotherThread(Action) Verse.<>c:<UpdateCurrentAsynchronousEvent>b__27_0() System.Threading.ThreadHelper:ThreadStart_Context(Object) System.Threading.ExecutionContext:RunInternal(ExecutionContext, ContextCallback, Object, Boolean) System.Threading.ExecutionContext:Run(ExecutionContext, ContextCallback, Object, Boolean) System.Threading.ExecutionContext:Run(ExecutionContext, ContextCallback, Object) System.Threading.ThreadHelper:ThreadStart() Mod SaucyPigeon.WildCultivation has multiple Verse.ThingDefs named Seed_Agave. Skipping. Verse.Log:Error(String, Boolean) Verse.DefDatabase`1:AddAllInMods() System.Reflection.MonoMethod:InternalInvoke(Object, Object[], Exception&) System.Reflection.MonoMethod:Invoke(Object, BindingFlags, Binder, Object[], CultureInfo) System.Reflection.MethodBase:Invoke(Object, Object[]) Verse.GenGeneric:InvokeStaticMethodOnGenericType(Type, Type, String) Verse.PlayDataLoader:DoPlayLoad() Verse.PlayDataLoader:LoadAllPlayData(Boolean) Verse.<>c:<Start>b__6_1() Verse.LongEventHandler:RunEventFromAnotherThread(Action) Verse.<>c:<UpdateCurrentAsynchronousEvent>b__27_0() System.Threading.ThreadHelper:ThreadStart_Context(Object) System.Threading.ExecutionContext:RunInternal(ExecutionContext, ContextCallback, Object, Boolean) System.Threading.ExecutionContext:Run(ExecutionContext, ContextCallback, Object, Boolean) System.Threading.ExecutionContext:Run(ExecutionContext, ContextCallback, Object) System.Threading.ThreadHelper:ThreadStart()
non_defect
seeds please mod saucypigeon wildcultivation has multiple verse buildabledefs named seed agave skipping verse log error string boolean verse defdatabase addallinmods system reflection monomethod internalinvoke object object exception system reflection monomethod invoke object bindingflags binder object cultureinfo system reflection methodbase invoke object object verse gengeneric invokestaticmethodongenerictype type type string verse playdataloader doplayload verse playdataloader loadallplaydata boolean verse c b verse longeventhandler runeventfromanotherthread action verse c b system threading threadhelper threadstart context object system threading executioncontext runinternal executioncontext contextcallback object boolean system threading executioncontext run executioncontext contextcallback object boolean system threading executioncontext run executioncontext contextcallback object system threading threadhelper threadstart mod saucypigeon wildcultivation has multiple verse thingdefs named seed agave skipping verse log error string boolean verse defdatabase addallinmods system reflection monomethod internalinvoke object object exception system reflection monomethod invoke object bindingflags binder object cultureinfo system reflection methodbase invoke object object verse gengeneric invokestaticmethodongenerictype type type string verse playdataloader doplayload verse playdataloader loadallplaydata boolean verse c b verse longeventhandler runeventfromanotherthread action verse c b system threading threadhelper threadstart context object system threading executioncontext runinternal executioncontext contextcallback object boolean system threading executioncontext run executioncontext contextcallback object boolean system threading executioncontext run executioncontext contextcallback object system threading threadhelper threadstart
0
67,621
21,032,324,465
IssuesEvent
2022-03-31 02:40:55
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Changing of theme from Dark mode to Light mode does not respond
T-Defect
### Steps to reproduce https://user-images.githubusercontent.com/63562663/160947985-fabcfc38-b74b-4c4d-a879-24cc92a03210.mp4 ### Outcome #### What did you expect? Change from Dark mode to Light mode #### What happened instead? Dark mode persisted. ### Operating system Windows ### Browser information Version 99.0.4844.84 (Official Build) (64-bit) ### URL for webapp app.element.io ### Application version Element version: 1.10.8 Olm version: 3.2.8 ### Homeserver _No response_ ### Will you send logs? No
1.0
Changing of theme from Dark mode to Light mode does not respond - ### Steps to reproduce https://user-images.githubusercontent.com/63562663/160947985-fabcfc38-b74b-4c4d-a879-24cc92a03210.mp4 ### Outcome #### What did you expect? Change from Dark mode to Light mode #### What happened instead? Dark mode persisted. ### Operating system Windows ### Browser information Version 99.0.4844.84 (Official Build) (64-bit) ### URL for webapp app.element.io ### Application version Element version: 1.10.8 Olm version: 3.2.8 ### Homeserver _No response_ ### Will you send logs? No
defect
changing of theme from dark mode to light mode does not respond steps to reproduce outcome what did you expect change from dark mode to light mode what happened instead dark mode persisted operating system windows browser information version official build bit url for webapp app element io application version element version olm version homeserver no response will you send logs no
1
179,341
6,624,126,739
IssuesEvent
2017-09-22 10:11:29
numbbo/coco
https://api.github.com/repos/numbbo/coco
opened
data archive filenames
Priority-High Usability
I found these names: ``` ['bbob/2015-GECCO/Sif.tgz'], ['bbob/2015-GECCO/Sifeg.tgz'], ['bbob/2015-GECCO/Srr.tgz'], ``` and similar others in our data archive at `http://coco.gforge.inria.fr/data-archive`. _Given the implicit naming conventions we have used over many years_, these names are IMHO unacceptable and _must_ change. A name cannot just give no indication whatsoever what is in the data set (given all other names do). Using the algorithm name and author has proven to be useful, but we can also try to come up with even more useful conventions. On that note, `2017-outsideGECCO/` is not a nice folder name either and inconsistent vs `2017/` and `2015-GECCO/` (the latter is inconsistent with most others). How about `2017-no-workshop` and `2015`? Rationale: There are only disadvantages to have cryptic filenames. Besides that informative filenames are useful for simple browsing, we are likely use the filename to search for entries of interest for some time to come, because it is rather tedious to actually implement different ways and start the maintenance for 150+ data sets.
1.0
data archive filenames - I found these names: ``` ['bbob/2015-GECCO/Sif.tgz'], ['bbob/2015-GECCO/Sifeg.tgz'], ['bbob/2015-GECCO/Srr.tgz'], ``` and similar others in our data archive at `http://coco.gforge.inria.fr/data-archive`. _Given the implicit naming conventions we have used over many years_, these names are IMHO unacceptable and _must_ change. A name cannot just give no indication whatsoever what is in the data set (given all other names do). Using the algorithm name and author has proven to be useful, but we can also try to come up with even more useful conventions. On that note, `2017-outsideGECCO/` is not a nice folder name either and inconsistent vs `2017/` and `2015-GECCO/` (the latter is inconsistent with most others). How about `2017-no-workshop` and `2015`? Rationale: There are only disadvantages to have cryptic filenames. Besides that informative filenames are useful for simple browsing, we are likely use the filename to search for entries of interest for some time to come, because it is rather tedious to actually implement different ways and start the maintenance for 150+ data sets.
non_defect
data archive filenames i found these names and similar others in our data archive at given the implicit naming conventions we have used over many years these names are imho unacceptable and must change a name cannot just give no indication whatsoever what is in the data set given all other names do using the algorithm name and author has proven to be useful but we can also try to come up with even more useful conventions on that note outsidegecco is not a nice folder name either and inconsistent vs and gecco the latter is inconsistent with most others how about no workshop and rationale there are only disadvantages to have cryptic filenames besides that informative filenames are useful for simple browsing we are likely use the filename to search for entries of interest for some time to come because it is rather tedious to actually implement different ways and start the maintenance for data sets
0
605,362
18,734,275,898
IssuesEvent
2021-11-04 04:04:02
AY2122S1-CS2103T-W13-2/tp
https://api.github.com/repos/AY2122S1-CS2103T-W13-2/tp
closed
Issue when deleting data while logging in
type.Bug priority.High
The problem here is when the user open the app with data, the login screen will show up. But if the data is deleted at this point, the password of the user will be used to create a new data file instead of throwing some error message.
1.0
Issue when deleting data while logging in - The problem here is when the user open the app with data, the login screen will show up. But if the data is deleted at this point, the password of the user will be used to create a new data file instead of throwing some error message.
non_defect
issue when deleting data while logging in the problem here is when the user open the app with data the login screen will show up but if the data is deleted at this point the password of the user will be used to create a new data file instead of throwing some error message
0
51,336
13,207,437,533
IssuesEvent
2020-08-14 23:05:56
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
Monitor rate of big events in PnF (Trac #245)
Incomplete Migration Migrated from Trac defect jeb + pnf
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/245">https://code.icecube.wisc.edu/projects/icecube/ticket/245</a>, reported by blaufussand owned by tschmidt</em></summary> <p> ```json { "status": "closed", "changetime": "2016-02-29T18:02:27", "_ts": "1456768947686083", "description": "Recommendation of review panel.\n\n\n3) `Big Events issue\u2019: These are events which individually require long times (~10 seconds) for processing keep the PnF master node busy. If the rate of these events is high the system can become unstable. PnF would suffer serious issues when events with 45 s delays become frequent. Sophisticated reconstructions, such as currently used by online analyses, introduce delays. Currently the system can handle an increase in filter delays by a factor of 2, assuming an overall trigger rate of 2.9 kHz, if the number of client nodes is doubled. We recommend that the rate of these events is monitored. \u00a0\n", "reporter": "blaufuss", "cc": "", "resolution": "fixed", "time": "2011-02-21T02:28:53", "component": "jeb + pnf", "summary": "Monitor rate of big events in PnF", "priority": "normal", "keywords": "", "milestone": "", "owner": "tschmidt", "type": "defect" } ``` </p> </details>
1.0
Monitor rate of big events in PnF (Trac #245) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/245">https://code.icecube.wisc.edu/projects/icecube/ticket/245</a>, reported by blaufussand owned by tschmidt</em></summary> <p> ```json { "status": "closed", "changetime": "2016-02-29T18:02:27", "_ts": "1456768947686083", "description": "Recommendation of review panel.\n\n\n3) `Big Events issue\u2019: These are events which individually require long times (~10 seconds) for processing keep the PnF master node busy. If the rate of these events is high the system can become unstable. PnF would suffer serious issues when events with 45 s delays become frequent. Sophisticated reconstructions, such as currently used by online analyses, introduce delays. Currently the system can handle an increase in filter delays by a factor of 2, assuming an overall trigger rate of 2.9 kHz, if the number of client nodes is doubled. We recommend that the rate of these events is monitored. \u00a0\n", "reporter": "blaufuss", "cc": "", "resolution": "fixed", "time": "2011-02-21T02:28:53", "component": "jeb + pnf", "summary": "Monitor rate of big events in PnF", "priority": "normal", "keywords": "", "milestone": "", "owner": "tschmidt", "type": "defect" } ``` </p> </details>
defect
monitor rate of big events in pnf trac migrated from json status closed changetime ts description recommendation of review panel n n big events issue these are events which individually require long times seconds for processing keep the pnf master node busy if the rate of these events is high the system can become unstable pnf would suffer serious issues when events with s delays become frequent sophisticated reconstructions such as currently used by online analyses introduce delays currently the system can handle an increase in filter delays by a factor of assuming an overall trigger rate of khz if the number of client nodes is doubled we recommend that the rate of these events is monitored n reporter blaufuss cc resolution fixed time component jeb pnf summary monitor rate of big events in pnf priority normal keywords milestone owner tschmidt type defect
1
826,890
31,716,996,015
IssuesEvent
2023-09-10 01:04:26
kubernetes/ingress-nginx
https://api.github.com/repos/kubernetes/ingress-nginx
closed
CVE-2022-41409 reported against controller-v1.8.1
kind/bug triage/accepted needs-priority
We are using X-ray version 3.76.7 and it reported CVE-2022-41409 (pcre2test vulnerability) against controller-v1.8.1 tag. I have checked and think controller-v1.8.1 only used pcre but still, xray complains about that CVE.
1.0
CVE-2022-41409 reported against controller-v1.8.1 - We are using X-ray version 3.76.7 and it reported CVE-2022-41409 (pcre2test vulnerability) against controller-v1.8.1 tag. I have checked and think controller-v1.8.1 only used pcre but still, xray complains about that CVE.
non_defect
cve reported against controller we are using x ray version and it reported cve vulnerability against controller tag i have checked and think controller only used pcre but still xray complains about that cve
0
451,643
13,039,692,167
IssuesEvent
2020-07-28 17:11:14
cds-snc/report-a-cybercrime
https://api.github.com/repos/cds-snc/report-a-cybercrime
closed
Missing .test.js for the following forms.
bug low priority
## Summary Missing .test.js for the following forms: SuspectCluesForm, EvidenceInfoForm, ConfirmationForm ## Unresolved questions > Are there any related issues you consider out of scope for this issue that could be addressed in the future?
1.0
Missing .test.js for the following forms. - ## Summary Missing .test.js for the following forms: SuspectCluesForm, EvidenceInfoForm, ConfirmationForm ## Unresolved questions > Are there any related issues you consider out of scope for this issue that could be addressed in the future?
non_defect
missing test js for the following forms summary missing test js for the following forms suspectcluesform evidenceinfoform confirmationform unresolved questions are there any related issues you consider out of scope for this issue that could be addressed in the future
0
164,877
26,038,885,035
IssuesEvent
2022-12-22 08:37:37
BreadGood-22/frontend
https://api.github.com/repos/BreadGood-22/frontend
closed
[design] 상품 등록 페이지 UI 구현
design
## ⭐ 주요 기능 <!-- 구현할 기능(목표)에 대한 간략한 설명 --> 상품 등록 페이지 UI 구현 ## 📋 진행 사항 - [x] 상품 등록 페이지 레이아웃 구현 - [x] 이미지 파일 업로드 UI 구현 - [x] input창 Validation 및 input창 validation 통과시 버튼 활성화 ## 📄 참고 사항
1.0
[design] 상품 등록 페이지 UI 구현 - ## ⭐ 주요 기능 <!-- 구현할 기능(목표)에 대한 간략한 설명 --> 상품 등록 페이지 UI 구현 ## 📋 진행 사항 - [x] 상품 등록 페이지 레이아웃 구현 - [x] 이미지 파일 업로드 UI 구현 - [x] input창 Validation 및 input창 validation 통과시 버튼 활성화 ## 📄 참고 사항
non_defect
상품 등록 페이지 ui 구현 ⭐ 주요 기능 상품 등록 페이지 ui 구현 📋 진행 사항 상품 등록 페이지 레이아웃 구현 이미지 파일 업로드 ui 구현 input창 validation 및 input창 validation 통과시 버튼 활성화 📄 참고 사항
0
237,621
19,661,463,322
IssuesEvent
2022-01-10 17:26:32
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
[Android] "Include in Auto-contribute" toggle doesn't stay on
bug feature/rewards priority/P2 QA/Yes QA/Test-Plan-Specified regression OS/Android rewards-ui-2.0
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!-- Provide a brief description of the issue --> `Include in Auto-Contribute` toggle doesn't stay on ## Steps to reproduce <!-- Please add a series of steps to reproduce the issue --> 1. install Brave `1.35.40, Chromium 97.0.4692.45` 2. load `brave.com` 3. tap on the BAT icon in the URL bar 4. tap `Start using Brave Rewards` 3. tap on the BAT icon again 4. notice that `Include in Auto-Contribute` is toggled to `OFF` 5. toggle it to `ON` 6. tap away to dismiss the panel 7. tap again on the BAT icon 8. look at the value for `Include in Auto-Contribute` ## Actual result <!-- Please add screenshots if needed --> It's back to `OFF` <video src="https://user-images.githubusercontent.com/387249/146455999-9c4a2163-6401-4455-b743-8aa9463ad2bf.mp4"> ## Expected result It should be `ON` ## Issue reproduces how often <!-- [Easily reproduced/Intermittent issue/No steps to reproduce] --> 100% ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current Play Store version? - Can you reproduce this issue with the current Play Store Beta version? - Can you reproduce this issue with the current Play Store Nightly version? ## Device details - Install type (ARM, x86): `arm64` - Device type (Phone, Tablet, Phablet): Google Pixel XL, phone - Android version: 9.0 ## Brave version Brave `1.35.40, Chromium 97.0.4692.45`
1.0
[Android] "Include in Auto-contribute" toggle doesn't stay on - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!-- Provide a brief description of the issue --> `Include in Auto-Contribute` toggle doesn't stay on ## Steps to reproduce <!-- Please add a series of steps to reproduce the issue --> 1. install Brave `1.35.40, Chromium 97.0.4692.45` 2. load `brave.com` 3. tap on the BAT icon in the URL bar 4. tap `Start using Brave Rewards` 3. tap on the BAT icon again 4. notice that `Include in Auto-Contribute` is toggled to `OFF` 5. toggle it to `ON` 6. tap away to dismiss the panel 7. tap again on the BAT icon 8. look at the value for `Include in Auto-Contribute` ## Actual result <!-- Please add screenshots if needed --> It's back to `OFF` <video src="https://user-images.githubusercontent.com/387249/146455999-9c4a2163-6401-4455-b743-8aa9463ad2bf.mp4"> ## Expected result It should be `ON` ## Issue reproduces how often <!-- [Easily reproduced/Intermittent issue/No steps to reproduce] --> 100% ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current Play Store version? - Can you reproduce this issue with the current Play Store Beta version? - Can you reproduce this issue with the current Play Store Nightly version? ## Device details - Install type (ARM, x86): `arm64` - Device type (Phone, Tablet, Phablet): Google Pixel XL, phone - Android version: 9.0 ## Brave version Brave `1.35.40, Chromium 97.0.4692.45`
non_defect
include in auto contribute toggle doesn t stay on have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description include in auto contribute toggle doesn t stay on steps to reproduce install brave chromium load brave com tap on the bat icon in the url bar tap start using brave rewards tap on the bat icon again notice that include in auto contribute is toggled to off toggle it to on tap away to dismiss the panel tap again on the bat icon look at the value for include in auto contribute actual result it s back to off video src expected result it should be on issue reproduces how often version channel information can you reproduce this issue with the current play store version can you reproduce this issue with the current play store beta version can you reproduce this issue with the current play store nightly version device details install type arm device type phone tablet phablet google pixel xl phone android version brave version brave chromium
0
6,023
3,727,356,897
IssuesEvent
2016-03-06 07:25:19
nunit/nunit
https://api.github.com/repos/nunit/nunit
closed
Add CF to the Appveyor CI build
is:build pri:low status:blocked
As indicated by #331 we keep breaking the compact framework build without realizing it. We need to at least compile it in our CI and if possible run the tests. Without installing a bunch of stuff in Appveyor, we'll probably only be able to run it using assembly unification under desktop .NET for now.
1.0
Add CF to the Appveyor CI build - As indicated by #331 we keep breaking the compact framework build without realizing it. We need to at least compile it in our CI and if possible run the tests. Without installing a bunch of stuff in Appveyor, we'll probably only be able to run it using assembly unification under desktop .NET for now.
non_defect
add cf to the appveyor ci build as indicated by we keep breaking the compact framework build without realizing it we need to at least compile it in our ci and if possible run the tests without installing a bunch of stuff in appveyor we ll probably only be able to run it using assembly unification under desktop net for now
0
237,934
7,768,433,377
IssuesEvent
2018-06-03 17:57:54
InfiniteFlightAirportEditing/Navigation
https://api.github.com/repos/InfiniteFlightAirportEditing/Navigation
opened
LICT
ILS Priority ILS Error
# Summary of Navigation issue: (ILS, Waypoints, etc.) Runway moved/changed. # Runway(s) with ILS: (If any) 31L # References: (Official sources/charts) http://opennav.com/pdf/LICT/LI_AD_2_LICT_5-3_en_2013-05-02.pdf
1.0
LICT - # Summary of Navigation issue: (ILS, Waypoints, etc.) Runway moved/changed. # Runway(s) with ILS: (If any) 31L # References: (Official sources/charts) http://opennav.com/pdf/LICT/LI_AD_2_LICT_5-3_en_2013-05-02.pdf
non_defect
lict summary of navigation issue ils waypoints etc runway moved changed runway s with ils if any references official sources charts
0
622,764
19,656,310,360
IssuesEvent
2022-01-10 12:54:11
epam/Indigo
https://api.github.com/repos/epam/Indigo
reopened
Add notebook for predicting assay values using Indigo fingerprints
Enhancement High priority python ML
**ToDo** Prepare and place a notebook in api/python/notebooks/predict.ipynb * Read input CSV file, with given column name for SMILES and column name for assay values. * Prepare train and test datasets using shuffled data. * Draw distribution plot for assay values. * Calculate fingerprints for SMILES using Indigo. * Fit and predict for two models: linear on your choice (SVM, RF) and deep learning mode on your choice. Print `r^2` value for predicting values for test dataset. * Draw an interactive scatter plots using [Bokeh](https://docs.bokeh.org/en/2.4.1/docs/user_guide/jupyter.html). Use hover tool to show molecule images, actual and predicted value. Following plots should present: * Actual vs Predicted * Actual vs Residual (actual - predicted)
1.0
Add notebook for predicting assay values using Indigo fingerprints - **ToDo** Prepare and place a notebook in api/python/notebooks/predict.ipynb * Read input CSV file, with given column name for SMILES and column name for assay values. * Prepare train and test datasets using shuffled data. * Draw distribution plot for assay values. * Calculate fingerprints for SMILES using Indigo. * Fit and predict for two models: linear on your choice (SVM, RF) and deep learning mode on your choice. Print `r^2` value for predicting values for test dataset. * Draw an interactive scatter plots using [Bokeh](https://docs.bokeh.org/en/2.4.1/docs/user_guide/jupyter.html). Use hover tool to show molecule images, actual and predicted value. Following plots should present: * Actual vs Predicted * Actual vs Residual (actual - predicted)
non_defect
add notebook for predicting assay values using indigo fingerprints todo prepare and place a notebook in api python notebooks predict ipynb read input csv file with given column name for smiles and column name for assay values prepare train and test datasets using shuffled data draw distribution plot for assay values calculate fingerprints for smiles using indigo fit and predict for two models linear on your choice svm rf and deep learning mode on your choice print r value for predicting values for test dataset draw an interactive scatter plots using use hover tool to show molecule images actual and predicted value following plots should present actual vs predicted actual vs residual actual predicted
0
379,386
26,369,252,140
IssuesEvent
2023-01-11 19:10:36
Tracktr/tracktr-client
https://api.github.com/repos/Tracktr/tracktr-client
closed
Verslag voor School
documentation
**Eisen:** - [ ] 20-25 pagina’s **To do:** - [x] Uitgevoerde use cases - [x] Geplande use cases - [x] Database schema - [x] Code log - [x] Verwerkte feedback - [x] Applicatie structuur - [x] Beschrijving experimenten - [ ] Designs (uitleggen) - [ ] Opmaak (affinity/indesign)
1.0
Verslag voor School - **Eisen:** - [ ] 20-25 pagina’s **To do:** - [x] Uitgevoerde use cases - [x] Geplande use cases - [x] Database schema - [x] Code log - [x] Verwerkte feedback - [x] Applicatie structuur - [x] Beschrijving experimenten - [ ] Designs (uitleggen) - [ ] Opmaak (affinity/indesign)
non_defect
verslag voor school eisen pagina’s to do uitgevoerde use cases geplande use cases database schema code log verwerkte feedback applicatie structuur beschrijving experimenten designs uitleggen opmaak affinity indesign
0
184,928
6,717,519,775
IssuesEvent
2017-10-14 22:20:00
dbcollection/dbcollection
https://api.github.com/repos/dbcollection/dbcollection
closed
Add docs for the datasets
docs medium priority
Must create a readme for each dataset. The readme must provide information about the dataset type, splits and overall data fields/organization.
1.0
Add docs for the datasets - Must create a readme for each dataset. The readme must provide information about the dataset type, splits and overall data fields/organization.
non_defect
add docs for the datasets must create a readme for each dataset the readme must provide information about the dataset type splits and overall data fields organization
0
29,055
5,515,033,103
IssuesEvent
2017-03-17 16:26:54
contao/news-bundle
https://api.github.com/repos/contao/news-bundle
closed
What is authorName and where is it coming from?
defect
I must be blind...can anybody explain me where this line takes the author name from? https://github.com/contao/news-bundle/blob/46256caff85ef2cd0b72b2235ed693a5b7876f53/src/Resources/contao/classes/News.php#L184 I cannot find the logic behind this.
1.0
What is authorName and where is it coming from? - I must be blind...can anybody explain me where this line takes the author name from? https://github.com/contao/news-bundle/blob/46256caff85ef2cd0b72b2235ed693a5b7876f53/src/Resources/contao/classes/News.php#L184 I cannot find the logic behind this.
defect
what is authorname and where is it coming from i must be blind can anybody explain me where this line takes the author name from i cannot find the logic behind this
1
7,880
2,611,056,242
IssuesEvent
2015-02-27 00:25:53
alistairreilly/andors-trail
https://api.github.com/repos/alistairreilly/andors-trail
closed
Suggestion: "sell all" option
auto-migrated Priority-Medium Type-Defect
``` I just found that I have over 300 of "Animal hair" :) selling it one by one it's a nightmare. Maybe long press on "Sell" button should trigger "Sell all" function with confirmation dialog? Additionally, even selling 10 of items can produce a problem, when clicking to quick you can accidentally sell an item below that one. ``` Original issue reported on code.google.com by `R3pl...@gmail.com` on 3 Jan 2011 at 6:45
1.0
Suggestion: "sell all" option - ``` I just found that I have over 300 of "Animal hair" :) selling it one by one it's a nightmare. Maybe long press on "Sell" button should trigger "Sell all" function with confirmation dialog? Additionally, even selling 10 of items can produce a problem, when clicking to quick you can accidentally sell an item below that one. ``` Original issue reported on code.google.com by `R3pl...@gmail.com` on 3 Jan 2011 at 6:45
defect
suggestion sell all option i just found that i have over of animal hair selling it one by one it s a nightmare maybe long press on sell button should trigger sell all function with confirmation dialog additionally even selling of items can produce a problem when clicking to quick you can accidentally sell an item below that one original issue reported on code google com by gmail com on jan at
1
67,115
20,909,175,673
IssuesEvent
2022-03-24 07:29:12
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Can't load message: Invalid `href`
T-Defect S-Minor Z-Soft-Crash X-Needs-Investigation O-Uncommon
### Steps to reproduce Event source; ```json5 { "content": { "body": "I could just make an alias like #matrix:non-matrix.org and host a phising Element at element.non-matrix.org. Any server can add aliases to a room or appear similar enough, so not sure that is a good idea.", "format": "org.matrix.custom.html", "formatted_body": "I could just make an alias like #<a href=\"matrix:non-matrix.org\">matrix:non-matrix.org</a> and host a phising Element at element.non-matrix.org. Any server can add aliases to a room or appear similar enough, so not sure that is a good idea.", "msgtype": "m.text" }, "type": "m.room.message", "unsigned": {}, // Some event keys redacted } ``` I believe the culprit is `#<a href=\"matrix:non-matrix.org\">matrix:non-matrix.org</a>` ### Outcome #### What did you expect? The message to maybe display with a fallback, or mangled up a bit. #### What happened instead? ![image](https://user-images.githubusercontent.com/22740616/158550499-693f7db6-f7cd-4f1f-bc75-092e9644789f.png) ### Operating system PopOS 21.10 ### Application version Element version: 1.10.6 Olm version: 3.2.8 ### How did you install the app? Flatpak ### Homeserver jboi.nl ### Will you send logs? No
1.0
Can't load message: Invalid `href` - ### Steps to reproduce Event source; ```json5 { "content": { "body": "I could just make an alias like #matrix:non-matrix.org and host a phising Element at element.non-matrix.org. Any server can add aliases to a room or appear similar enough, so not sure that is a good idea.", "format": "org.matrix.custom.html", "formatted_body": "I could just make an alias like #<a href=\"matrix:non-matrix.org\">matrix:non-matrix.org</a> and host a phising Element at element.non-matrix.org. Any server can add aliases to a room or appear similar enough, so not sure that is a good idea.", "msgtype": "m.text" }, "type": "m.room.message", "unsigned": {}, // Some event keys redacted } ``` I believe the culprit is `#<a href=\"matrix:non-matrix.org\">matrix:non-matrix.org</a>` ### Outcome #### What did you expect? The message to maybe display with a fallback, or mangled up a bit. #### What happened instead? ![image](https://user-images.githubusercontent.com/22740616/158550499-693f7db6-f7cd-4f1f-bc75-092e9644789f.png) ### Operating system PopOS 21.10 ### Application version Element version: 1.10.6 Olm version: 3.2.8 ### How did you install the app? Flatpak ### Homeserver jboi.nl ### Will you send logs? No
defect
can t load message invalid href steps to reproduce event source content body i could just make an alias like matrix non matrix org and host a phising element at element non matrix org any server can add aliases to a room or appear similar enough so not sure that is a good idea format org matrix custom html formatted body i could just make an alias like matrix non matrix org and host a phising element at element non matrix org any server can add aliases to a room or appear similar enough so not sure that is a good idea msgtype m text type m room message unsigned some event keys redacted i believe the culprit is matrix non matrix org outcome what did you expect the message to maybe display with a fallback or mangled up a bit what happened instead operating system popos application version element version olm version how did you install the app flatpak homeserver jboi nl will you send logs no
1
64,383
18,531,492,998
IssuesEvent
2021-10-21 06:41:03
gitblit/gitblit
https://api.github.com/repos/gitblit/gitblit
closed
Bad encoding in properties-files
Priority-Medium Catg-Defect Status-Verified
Originally reported on Google Code with ID 538 ``` What steps will reproduce the problem? 1. Deploy gitblit on tomcat6 or 7 running on a platform supporting utf-8. My example is dev.gitblit.com 2. Access the site with a browser specifying Norwegian as the Language (or via curl: $ curl -H"Accept-Language: no,nb-NO;q=0.8,nn-NO;q=0.5,en-US;q=0.3" https://dev.gitblit.com/ -I HTTP/1.1 500 Server Error Date: Tue, 25 Nov 2014 13:44:16 GMT Content-Type: text/html; charset=UTF-8 Content-Language: no Pragma: no-cache Cache-Control: no-cache, max-age=0, must-revalidate Content-Length: 1406 Server: Jetty(9.2.3.v20140905) 3. See the internal error What is the expected output? What do you see instead? I see the attached image - internal error. In the logs I'm spammed with this output: 2014-11-25 14:07:25 [INFO ] Loading properties files from jar:file:/data/prog/ap ache-tomcat-7.0.29/webapps/gitblit/WEB-INF/lib/gitblit.jar!/com/gitblit/wicket/GitBlitWebApp_no.properties 2014-11-25 14:07:25 [ERROR] Can't instantiate page using constructor public com. gitblit.wicket.pages.MyDashboardPage() org.apache.wicket.WicketRuntimeException: Can't instantiate page using constructor public com.gitblit.wicket.pages.MyDashboardPage() at org.apache.wicket.session.DefaultPageFactory.createPage(DefaultPageFactory.java:212) at org.apache.wicket.session.DefaultPageFactory.newPage(DefaultPageFactory.java:57) at org.apache.wicket.request.target.component.BookmarkablePageRequestTar get.newPage(BookmarkablePageRequestTarget.java:298) at org.apache.wicket.request.target.component.BookmarkablePageRequestTar get.getPage(BookmarkablePageRequestTarget.java:320) at org.apache.wicket.request.target.component.BookmarkablePageRequestTar get.processEvents(BookmarkablePageRequestTarget.java:234) at org.apache.wicket.request.AbstractRequestCycleProcessor.processEvents (AbstractRequestCycleProcessor.java:92) at org.apache.wicket.RequestCycle.processEventsAndRespond(RequestCycle.j ava:1279) at org.apache.wicket.RequestCycle.step(RequestCycle.java:1358) at org.apache.wicket.RequestCycle.steps(RequestCycle.java:1465) at org.apache.wicket.RequestCycle.request(RequestCycle.java:545) at org.apache.wicket.protocol.http.WicketFilter.doGet(WicketFilter.java: 486) at org.apache.wicket.protocol.http.WicketFilter.doFilter(WicketFilter.ja va:319) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(Appl icationFilterChain.java:243) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationF ilterChain.java:210) at com.gitblit.servlet.EnforceAuthenticationFilter.doFilter(EnforceAuthe nticationFilter.java:87) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(Appl icationFilterChain.java:243) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationF ilterChain.java:210) at com.gitblit.servlet.ProxyFilter$1.doFilter(ProxyFilter.java:74) at com.gitblit.servlet.ProxyFilter.doFilter(ProxyFilter.java:77) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(Appl icationFilterChain.java:243) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationF ilterChain.java:210) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperV alve.java:225) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:123) at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:472) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:168) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:98) at org.apache.catalina.valves.AccessLogValve.invoke(AccessLogValve.java:927) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:118) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:407) at org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1001) at org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:585) at org.apache.tomcat.util.net.JIoEndpoint$SocketProcessor.run(JIoEndpoint.java:310) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) Caused by: java.lang.reflect.InvocationTargetException at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:526) at org.apache.wicket.session.DefaultPageFactory.createPage(DefaultPageFactory.java:192) ... 34 more Caused by: java.lang.IllegalArgumentException: Malformed \uxxxx encoding. at java.util.Properties.loadConvert(Properties.java:568) at java.util.Properties.load0(Properties.java:391) at java.util.Properties.load(Properties.java:341) at org.apache.wicket.resource.PropertiesFactory$PropertiesFilePropertiesLoader.loadProperties(PropertiesFactory.java:343) at org.apache.wicket.resource.PropertiesFactory$AbstractPropertiesLoader.load(PropertiesFactory.java:274) at org.apache.wicket.resource.PropertiesFactory.load(PropertiesFactory.java:133) at org.apache.wicket.resource.loader.ComponentStringResourceLoader.loadStringResource(ComponentStringResourceLoader.java:141) at org.apache.wicket.resource.loader.ClassStringResourceLoader.loadStringResource(ClassStringResourceLoader.java:65) at org.apache.wicket.resource.loader.ComponentStringResourceLoader.loadStringResource(ComponentStringResourceLoader.java:261) at org.apache.wicket.resource.loader.ClassStringResourceLoader.loadStringResource(ClassStringResourceLoader.java:80) at org.apache.wicket.Localizer.getStringIgnoreSettings(Localizer.java:241) at org.apache.wicket.Localizer.getString(Localizer.java:313) at org.apache.wicket.Localizer.getString(Localizer.java:119) at org.apache.wicket.Component.getString(Component.java:1968) at org.apache.wicket.Component.getString(Component.java:1955) at com.gitblit.wicket.pages.RootPage$LoginForm.<init>(RootPage.java:581) at com.gitblit.wicket.pages.RootPage.setupPage(RootPage.java:176) at com.gitblit.wicket.pages.MyDashboardPage.setup(MyDashboardPage.java:71) at com.gitblit.wicket.pages.MyDashboardPage.<init>(MyDashboardPage.java:57) ... 39 more What version of the product are you using? On what operating system? 1.6.2, running on server:/data/prog/apache-tomcat-7.0.29/logs> uname -a Linux server 2.6.32-279.1.1.el6.x86_64 #1 SMP Wed Jun 20 11:41:22 EDT 2012 x86_64 x86_64 x86_64 GNU/Linux Please provide any additional information below. The file it's complaining about has alle Language-specific characters (æ,ø and å) encoded as \u00xx, which corresponds to ISO-8859-1 encoding. The correct encoding is to use Unicode. ``` Reported by `thomas.arp` on 2014-11-25 13:31:24 ----- - _Attachment: gitblit_doesnt_work_with_norwegian_and_utf8.png<br>![gitblit_doesnt_work_with_norwegian_and_utf8.png](https://storage.googleapis.com/google-code-attachments/gitblit/issue-538/comment-0/gitblit_doesnt_work_with_norwegian_and_utf8.png)_ _Edit: Make image render_
1.0
Bad encoding in properties-files - Originally reported on Google Code with ID 538 ``` What steps will reproduce the problem? 1. Deploy gitblit on tomcat6 or 7 running on a platform supporting utf-8. My example is dev.gitblit.com 2. Access the site with a browser specifying Norwegian as the Language (or via curl: $ curl -H"Accept-Language: no,nb-NO;q=0.8,nn-NO;q=0.5,en-US;q=0.3" https://dev.gitblit.com/ -I HTTP/1.1 500 Server Error Date: Tue, 25 Nov 2014 13:44:16 GMT Content-Type: text/html; charset=UTF-8 Content-Language: no Pragma: no-cache Cache-Control: no-cache, max-age=0, must-revalidate Content-Length: 1406 Server: Jetty(9.2.3.v20140905) 3. See the internal error What is the expected output? What do you see instead? I see the attached image - internal error. In the logs I'm spammed with this output: 2014-11-25 14:07:25 [INFO ] Loading properties files from jar:file:/data/prog/ap ache-tomcat-7.0.29/webapps/gitblit/WEB-INF/lib/gitblit.jar!/com/gitblit/wicket/GitBlitWebApp_no.properties 2014-11-25 14:07:25 [ERROR] Can't instantiate page using constructor public com. gitblit.wicket.pages.MyDashboardPage() org.apache.wicket.WicketRuntimeException: Can't instantiate page using constructor public com.gitblit.wicket.pages.MyDashboardPage() at org.apache.wicket.session.DefaultPageFactory.createPage(DefaultPageFactory.java:212) at org.apache.wicket.session.DefaultPageFactory.newPage(DefaultPageFactory.java:57) at org.apache.wicket.request.target.component.BookmarkablePageRequestTar get.newPage(BookmarkablePageRequestTarget.java:298) at org.apache.wicket.request.target.component.BookmarkablePageRequestTar get.getPage(BookmarkablePageRequestTarget.java:320) at org.apache.wicket.request.target.component.BookmarkablePageRequestTar get.processEvents(BookmarkablePageRequestTarget.java:234) at org.apache.wicket.request.AbstractRequestCycleProcessor.processEvents (AbstractRequestCycleProcessor.java:92) at org.apache.wicket.RequestCycle.processEventsAndRespond(RequestCycle.j ava:1279) at org.apache.wicket.RequestCycle.step(RequestCycle.java:1358) at org.apache.wicket.RequestCycle.steps(RequestCycle.java:1465) at org.apache.wicket.RequestCycle.request(RequestCycle.java:545) at org.apache.wicket.protocol.http.WicketFilter.doGet(WicketFilter.java: 486) at org.apache.wicket.protocol.http.WicketFilter.doFilter(WicketFilter.ja va:319) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(Appl icationFilterChain.java:243) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationF ilterChain.java:210) at com.gitblit.servlet.EnforceAuthenticationFilter.doFilter(EnforceAuthe nticationFilter.java:87) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(Appl icationFilterChain.java:243) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationF ilterChain.java:210) at com.gitblit.servlet.ProxyFilter$1.doFilter(ProxyFilter.java:74) at com.gitblit.servlet.ProxyFilter.doFilter(ProxyFilter.java:77) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(Appl icationFilterChain.java:243) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationF ilterChain.java:210) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperV alve.java:225) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:123) at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:472) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:168) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:98) at org.apache.catalina.valves.AccessLogValve.invoke(AccessLogValve.java:927) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:118) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:407) at org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1001) at org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:585) at org.apache.tomcat.util.net.JIoEndpoint$SocketProcessor.run(JIoEndpoint.java:310) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) Caused by: java.lang.reflect.InvocationTargetException at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:526) at org.apache.wicket.session.DefaultPageFactory.createPage(DefaultPageFactory.java:192) ... 34 more Caused by: java.lang.IllegalArgumentException: Malformed \uxxxx encoding. at java.util.Properties.loadConvert(Properties.java:568) at java.util.Properties.load0(Properties.java:391) at java.util.Properties.load(Properties.java:341) at org.apache.wicket.resource.PropertiesFactory$PropertiesFilePropertiesLoader.loadProperties(PropertiesFactory.java:343) at org.apache.wicket.resource.PropertiesFactory$AbstractPropertiesLoader.load(PropertiesFactory.java:274) at org.apache.wicket.resource.PropertiesFactory.load(PropertiesFactory.java:133) at org.apache.wicket.resource.loader.ComponentStringResourceLoader.loadStringResource(ComponentStringResourceLoader.java:141) at org.apache.wicket.resource.loader.ClassStringResourceLoader.loadStringResource(ClassStringResourceLoader.java:65) at org.apache.wicket.resource.loader.ComponentStringResourceLoader.loadStringResource(ComponentStringResourceLoader.java:261) at org.apache.wicket.resource.loader.ClassStringResourceLoader.loadStringResource(ClassStringResourceLoader.java:80) at org.apache.wicket.Localizer.getStringIgnoreSettings(Localizer.java:241) at org.apache.wicket.Localizer.getString(Localizer.java:313) at org.apache.wicket.Localizer.getString(Localizer.java:119) at org.apache.wicket.Component.getString(Component.java:1968) at org.apache.wicket.Component.getString(Component.java:1955) at com.gitblit.wicket.pages.RootPage$LoginForm.<init>(RootPage.java:581) at com.gitblit.wicket.pages.RootPage.setupPage(RootPage.java:176) at com.gitblit.wicket.pages.MyDashboardPage.setup(MyDashboardPage.java:71) at com.gitblit.wicket.pages.MyDashboardPage.<init>(MyDashboardPage.java:57) ... 39 more What version of the product are you using? On what operating system? 1.6.2, running on server:/data/prog/apache-tomcat-7.0.29/logs> uname -a Linux server 2.6.32-279.1.1.el6.x86_64 #1 SMP Wed Jun 20 11:41:22 EDT 2012 x86_64 x86_64 x86_64 GNU/Linux Please provide any additional information below. The file it's complaining about has alle Language-specific characters (æ,ø and å) encoded as \u00xx, which corresponds to ISO-8859-1 encoding. The correct encoding is to use Unicode. ``` Reported by `thomas.arp` on 2014-11-25 13:31:24 ----- - _Attachment: gitblit_doesnt_work_with_norwegian_and_utf8.png<br>![gitblit_doesnt_work_with_norwegian_and_utf8.png](https://storage.googleapis.com/google-code-attachments/gitblit/issue-538/comment-0/gitblit_doesnt_work_with_norwegian_and_utf8.png)_ _Edit: Make image render_
defect
bad encoding in properties files originally reported on google code with id what steps will reproduce the problem deploy gitblit on or running on a platform supporting utf my example is dev gitblit com access the site with a browser specifying norwegian as the language or via curl curl h accept language no nb no q nn no q en us q i http server error date tue nov gmt content type text html charset utf content language no pragma no cache cache control no cache max age must revalidate content length server jetty see the internal error what is the expected output what do you see instead i see the attached image internal error in the logs i m spammed with this output loading properties files from jar file data prog ap ache tomcat webapps gitblit web inf lib gitblit jar com gitblit wicket gitblitwebapp no properties can t instantiate page using constructor public com gitblit wicket pages mydashboardpage org apache wicket wicketruntimeexception can t instantiate page using constructor public com gitblit wicket pages mydashboardpage at org apache wicket session defaultpagefactory createpage defaultpagefactory java at org apache wicket session defaultpagefactory newpage defaultpagefactory java at org apache wicket request target component bookmarkablepagerequesttar get newpage bookmarkablepagerequesttarget java at org apache wicket request target component bookmarkablepagerequesttar get getpage bookmarkablepagerequesttarget java at org apache wicket request target component bookmarkablepagerequesttar get processevents bookmarkablepagerequesttarget java at org apache wicket request abstractrequestcycleprocessor processevents abstractrequestcycleprocessor java at org apache wicket requestcycle processeventsandrespond requestcycle j ava at org apache wicket requestcycle step requestcycle java at org apache wicket requestcycle steps requestcycle java at org apache wicket requestcycle request requestcycle java at org apache wicket protocol http wicketfilter doget wicketfilter java at org apache wicket protocol http wicketfilter dofilter wicketfilter ja va at org apache catalina core applicationfilterchain internaldofilter appl icationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationf ilterchain java at com gitblit servlet enforceauthenticationfilter dofilter enforceauthe nticationfilter java at org apache catalina core applicationfilterchain internaldofilter appl icationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationf ilterchain java at com gitblit servlet proxyfilter dofilter proxyfilter java at com gitblit servlet proxyfilter dofilter proxyfilter java at org apache catalina core applicationfilterchain internaldofilter appl icationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationf ilterchain java at org apache catalina core standardwrappervalve invoke standardwrapperv alve java at org apache catalina core standardcontextvalve invoke standardcontextvalve java at org apache catalina authenticator authenticatorbase invoke authenticatorbase java at org apache catalina core standardhostvalve invoke standardhostvalve java at org apache catalina valves errorreportvalve invoke errorreportvalve java at org apache catalina valves accesslogvalve invoke accesslogvalve java at org apache catalina core standardenginevalve invoke standardenginevalve java at org apache catalina connector coyoteadapter service coyoteadapter java at org apache coyote process java at org apache coyote abstractprotocol abstractconnectionhandler process abstractprotocol java at org apache tomcat util net jioendpoint socketprocessor run jioendpoint java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java caused by java lang reflect invocationtargetexception at sun reflect nativeconstructoraccessorimpl native method at sun reflect nativeconstructoraccessorimpl newinstance nativeconstructoraccessorimpl java at sun reflect delegatingconstructoraccessorimpl newinstance delegatingconstructoraccessorimpl java at java lang reflect constructor newinstance constructor java at org apache wicket session defaultpagefactory createpage defaultpagefactory java more caused by java lang illegalargumentexception malformed uxxxx encoding at java util properties loadconvert properties java at java util properties properties java at java util properties load properties java at org apache wicket resource propertiesfactory propertiesfilepropertiesloader loadproperties propertiesfactory java at org apache wicket resource propertiesfactory abstractpropertiesloader load propertiesfactory java at org apache wicket resource propertiesfactory load propertiesfactory java at org apache wicket resource loader componentstringresourceloader loadstringresource componentstringresourceloader java at org apache wicket resource loader classstringresourceloader loadstringresource classstringresourceloader java at org apache wicket resource loader componentstringresourceloader loadstringresource componentstringresourceloader java at org apache wicket resource loader classstringresourceloader loadstringresource classstringresourceloader java at org apache wicket localizer getstringignoresettings localizer java at org apache wicket localizer getstring localizer java at org apache wicket localizer getstring localizer java at org apache wicket component getstring component java at org apache wicket component getstring component java at com gitblit wicket pages rootpage loginform rootpage java at com gitblit wicket pages rootpage setuppage rootpage java at com gitblit wicket pages mydashboardpage setup mydashboardpage java at com gitblit wicket pages mydashboardpage mydashboardpage java more what version of the product are you using on what operating system running on server data prog apache tomcat logs uname a linux server smp wed jun edt gnu linux please provide any additional information below the file it s complaining about has alle language specific characters æ ø and å encoded as which corresponds to iso encoding the correct encoding is to use unicode reported by thomas arp on attachment gitblit doesnt work with norwegian and png edit make image render
1
7,417
7,925,247,327
IssuesEvent
2018-07-05 19:55:08
OriginProtocol/origin-bridge
https://api.github.com/repos/OriginProtocol/origin-bridge
opened
Purchase Timestamp Data Types
help wanted indexing service python
Purchase timestamps are currently being returned by the API as [ISO strings](https://en.wikipedia.org/wiki/ISO_8601). They should be returned as seconds since the epoch, since they are integers on the blockchain and [Origin.js expects to return datetimes](http://docs.originprotocol.com/#purchase-get). It would probably also be preferable for the field names to match, such as `created` instead of `created_at`.
1.0
Purchase Timestamp Data Types - Purchase timestamps are currently being returned by the API as [ISO strings](https://en.wikipedia.org/wiki/ISO_8601). They should be returned as seconds since the epoch, since they are integers on the blockchain and [Origin.js expects to return datetimes](http://docs.originprotocol.com/#purchase-get). It would probably also be preferable for the field names to match, such as `created` instead of `created_at`.
non_defect
purchase timestamp data types purchase timestamps are currently being returned by the api as they should be returned as seconds since the epoch since they are integers on the blockchain and it would probably also be preferable for the field names to match such as created instead of created at
0
67,523
20,975,318,476
IssuesEvent
2022-03-28 14:46:23
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Sliver of a room visible from spotlight search breadcrumbs
T-Defect
### Steps to reproduce 1. Use spotlight search 2. Click a bunch of rooms 3. Open spotlight 4. Hover on the right hand side ### Outcome #### What did you expect? The number of rooms to fit perfectly in the dialog #### What happened instead? ![image](https://user-images.githubusercontent.com/1190097/160424263-748bf642-86da-4d5d-82e4-510077c919e7.png) ( some rooms blurred for privacy) ### Operating system Windows 10 ### Application version Nightly (2021-03-27) ### How did you install the app? The Internet ### Homeserver t2l.io ### Will you send logs? No
1.0
Sliver of a room visible from spotlight search breadcrumbs - ### Steps to reproduce 1. Use spotlight search 2. Click a bunch of rooms 3. Open spotlight 4. Hover on the right hand side ### Outcome #### What did you expect? The number of rooms to fit perfectly in the dialog #### What happened instead? ![image](https://user-images.githubusercontent.com/1190097/160424263-748bf642-86da-4d5d-82e4-510077c919e7.png) ( some rooms blurred for privacy) ### Operating system Windows 10 ### Application version Nightly (2021-03-27) ### How did you install the app? The Internet ### Homeserver t2l.io ### Will you send logs? No
defect
sliver of a room visible from spotlight search breadcrumbs steps to reproduce use spotlight search click a bunch of rooms open spotlight hover on the right hand side outcome what did you expect the number of rooms to fit perfectly in the dialog what happened instead some rooms blurred for privacy operating system windows application version nightly how did you install the app the internet homeserver io will you send logs no
1
31,322
6,499,641,701
IssuesEvent
2017-08-22 22:32:51
opencaching/opencaching-pl
https://api.github.com/repos/opencaching/opencaching-pl
closed
Incorrect icon and description of blocked by COG cache - in powerTrail.php
Component_GeoPaths Priority_Low Status: Waiting for translations Type_Defect
Server generates warning: `[Tue Aug 02 21:19:07.234928 2016] [:error] [pid 7118] [client 89.65.18.106:50976] PHP Notice: Undefined offset: 6 in /srv/ocpl/powerTrail/ajaxGetPowerTrailCaches.php on line 111, referer: http://opencaching.pl/powerTrail.php?ptAction=showSerie&ptrail=334`
1.0
Incorrect icon and description of blocked by COG cache - in powerTrail.php - Server generates warning: `[Tue Aug 02 21:19:07.234928 2016] [:error] [pid 7118] [client 89.65.18.106:50976] PHP Notice: Undefined offset: 6 in /srv/ocpl/powerTrail/ajaxGetPowerTrailCaches.php on line 111, referer: http://opencaching.pl/powerTrail.php?ptAction=showSerie&ptrail=334`
defect
incorrect icon and description of blocked by cog cache in powertrail php server generates warning php notice undefined offset in srv ocpl powertrail ajaxgetpowertrailcaches php on line referer
1
44,404
5,814,219,136
IssuesEvent
2017-05-05 02:19:22
wiebj/deep-breath
https://api.github.com/repos/wiebj/deep-breath
closed
Evaluate data stores for temporal-spatial data
design evaluation
Evaluate popular data stores (SQL & NoSQL) for temporal-spatial data
1.0
Evaluate data stores for temporal-spatial data - Evaluate popular data stores (SQL & NoSQL) for temporal-spatial data
non_defect
evaluate data stores for temporal spatial data evaluate popular data stores sql nosql for temporal spatial data
0
59,801
17,023,250,002
IssuesEvent
2021-07-03 01:03:17
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
A list of all presets should be available.
Component: potlatch (flash editor) Priority: minor Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 11.27pm, Friday, 16th May 2008]** There should be a way to show all possible options in the preset type menu (the button with image left to the actual presets). This could be done, for example, by showing all the possibilities if the user clicks and holds it.
1.0
A list of all presets should be available. - **[Submitted to the original trac issue database at 11.27pm, Friday, 16th May 2008]** There should be a way to show all possible options in the preset type menu (the button with image left to the actual presets). This could be done, for example, by showing all the possibilities if the user clicks and holds it.
defect
a list of all presets should be available there should be a way to show all possible options in the preset type menu the button with image left to the actual presets this could be done for example by showing all the possibilities if the user clicks and holds it
1
7,190
2,610,356,659
IssuesEvent
2015-02-26 19:55:29
chrsmith/scribefire-chrome
https://api.github.com/repos/chrsmith/scribefire-chrome
closed
In response to your email, cfinke:
auto-migrated Priority-Medium Type-Defect
``` When I wrote saying ScribeFire 3.5.3.3 stopped working after FF updated to 7.0.1 you suggested I upgrade to ScribeFire 4. The only version for firefox I can see is scribefire-4.xpi. How do I install this (mac os x 10.6.8) and retain all the data archived in my previous version, as the new version doesn't seem to have assigned itself a location on my system - after download and clicking on the file: I get a message saying "There is no application set to open the document “scribefire-4.xpi”" .. scrutinizing the list of apps to choose from shows that non of them are browsers, specifically firefox, nor would any be used by scribefire. When I look in my ff extensions folder, only ScribeFire 3.5.3.3 is listed. Should I be downloading a different file? Other xpi based files on my system all installed themselves after clicking. I do have ScribeFire Next installed but disabled as I prefer the original interface. I do not want to remove ScribeFire 3.5.3.3 before finding out if its data is archived and accessible to the next version, as I could not find an obviously named file on my system indicating it contained this info. Thanks for any help you can provide. ``` ----- Original issue reported on code.google.com by `l8o.n...@gmail.com` on 29 Oct 2011 at 6:22
1.0
In response to your email, cfinke: - ``` When I wrote saying ScribeFire 3.5.3.3 stopped working after FF updated to 7.0.1 you suggested I upgrade to ScribeFire 4. The only version for firefox I can see is scribefire-4.xpi. How do I install this (mac os x 10.6.8) and retain all the data archived in my previous version, as the new version doesn't seem to have assigned itself a location on my system - after download and clicking on the file: I get a message saying "There is no application set to open the document “scribefire-4.xpi”" .. scrutinizing the list of apps to choose from shows that non of them are browsers, specifically firefox, nor would any be used by scribefire. When I look in my ff extensions folder, only ScribeFire 3.5.3.3 is listed. Should I be downloading a different file? Other xpi based files on my system all installed themselves after clicking. I do have ScribeFire Next installed but disabled as I prefer the original interface. I do not want to remove ScribeFire 3.5.3.3 before finding out if its data is archived and accessible to the next version, as I could not find an obviously named file on my system indicating it contained this info. Thanks for any help you can provide. ``` ----- Original issue reported on code.google.com by `l8o.n...@gmail.com` on 29 Oct 2011 at 6:22
defect
in response to your email cfinke when i wrote saying scribefire stopped working after ff updated to you suggested i upgrade to scribefire the only version for firefox i can see is scribefire xpi how do i install this mac os x and retain all the data archived in my previous version as the new version doesn t seem to have assigned itself a location on my system after download and clicking on the file i get a message saying there is no application set to open the document “scribefire xpi” scrutinizing the list of apps to choose from shows that non of them are browsers specifically firefox nor would any be used by scribefire when i look in my ff extensions folder only scribefire is listed should i be downloading a different file other xpi based files on my system all installed themselves after clicking i do have scribefire next installed but disabled as i prefer the original interface i do not want to remove scribefire before finding out if its data is archived and accessible to the next version as i could not find an obviously named file on my system indicating it contained this info thanks for any help you can provide original issue reported on code google com by n gmail com on oct at
1
54,412
13,652,439,667
IssuesEvent
2020-09-27 07:32:18
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
opened
Linux Kernel 5.8.10: ERROR: modpost: "__stack_chk_guard" [.../zfs.ko] undefined!
Status: Triage Needed Type: Defect
<!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please check our issue tracker before opening a new issue. Additional valuable information can be found in the OpenZFS documentation and mailing list archives. Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | Debian Distribution Version | Buster Linux Kernel | 5.8.10-rockchip64 Architecture | aarch64-linux ZFS Version | 2.0 + cherry-pick 2e407941a2148ab400d8b4e3748cc894e411c0c4 SPL Version | None <!-- Commands to find ZFS/SPL versions: modinfo zfs | grep -iw version modinfo spl | grep -iw version --> ### Describe the problem you're observing Making all in module ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/zstd/zzstd.ko] undefined! ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/zfs/zfs.ko] undefined! ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/zcommon/zcommon.ko] undefined! ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/unicode/zunicode.ko] undefined! ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/nvpair/znvpair.ko] undefined! ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/lua/zlua.ko] undefined! ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/icp/icp.ko] undefined! ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/avl/zavl.ko] undefined! ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/spl/spl.ko] undefined! make[4]: *** [scripts/Makefile.modpost:111: /usr/src/zfs/module/Module.symvers] Error 1 make[4]: *** Deleting file '/usr/src/zfs/module/Module.symvers' make[3]: *** [Makefile:1665: modules] Error 2 make[2]: *** [Makefile:48: modules-Linux] Error 2 make[1]: *** [Makefile:881: all-recursive] Error 1 make: *** [Makefile:742: all] Error 2 ### Describe how to reproduce the problem ``` # FROM Official Armbian Buster apt update && apt upgrade apt install build-essential autoconf automake libtool gawk alien fakeroot dkms libblkid-dev uuid-dev libudev-dev libssl-dev zlib1g-dev libaio-dev libattr1-dev libelf-dev linux-headers-current-rockchip64 python3 python3-dev python3-setuptools python3-cffi libffi-dev cd /usr/src git clone https://github.com/zfsonlinux/zfs cd zfs git checkout remotes/origin/zfs-2.0-release -b zfs2 git cherry-pick 2e407941a2148ab400d8b4e3748cc894e411c0c4 sh autogen.sh ./configure make -s -j$(nproc) ``` ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` -->
1.0
Linux Kernel 5.8.10: ERROR: modpost: "__stack_chk_guard" [.../zfs.ko] undefined! - <!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please check our issue tracker before opening a new issue. Additional valuable information can be found in the OpenZFS documentation and mailing list archives. Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | Debian Distribution Version | Buster Linux Kernel | 5.8.10-rockchip64 Architecture | aarch64-linux ZFS Version | 2.0 + cherry-pick 2e407941a2148ab400d8b4e3748cc894e411c0c4 SPL Version | None <!-- Commands to find ZFS/SPL versions: modinfo zfs | grep -iw version modinfo spl | grep -iw version --> ### Describe the problem you're observing Making all in module ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/zstd/zzstd.ko] undefined! ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/zfs/zfs.ko] undefined! ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/zcommon/zcommon.ko] undefined! ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/unicode/zunicode.ko] undefined! ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/nvpair/znvpair.ko] undefined! ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/lua/zlua.ko] undefined! ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/icp/icp.ko] undefined! ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/avl/zavl.ko] undefined! ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/spl/spl.ko] undefined! make[4]: *** [scripts/Makefile.modpost:111: /usr/src/zfs/module/Module.symvers] Error 1 make[4]: *** Deleting file '/usr/src/zfs/module/Module.symvers' make[3]: *** [Makefile:1665: modules] Error 2 make[2]: *** [Makefile:48: modules-Linux] Error 2 make[1]: *** [Makefile:881: all-recursive] Error 1 make: *** [Makefile:742: all] Error 2 ### Describe how to reproduce the problem ``` # FROM Official Armbian Buster apt update && apt upgrade apt install build-essential autoconf automake libtool gawk alien fakeroot dkms libblkid-dev uuid-dev libudev-dev libssl-dev zlib1g-dev libaio-dev libattr1-dev libelf-dev linux-headers-current-rockchip64 python3 python3-dev python3-setuptools python3-cffi libffi-dev cd /usr/src git clone https://github.com/zfsonlinux/zfs cd zfs git checkout remotes/origin/zfs-2.0-release -b zfs2 git cherry-pick 2e407941a2148ab400d8b4e3748cc894e411c0c4 sh autogen.sh ./configure make -s -j$(nproc) ``` ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` -->
defect
linux kernel error modpost stack chk guard undefined thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name debian distribution version buster linux kernel architecture linux zfs version cherry pick spl version none commands to find zfs spl versions modinfo zfs grep iw version modinfo spl grep iw version describe the problem you re observing making all in module error modpost stack chk guard undefined error modpost stack chk guard undefined error modpost stack chk guard undefined error modpost stack chk guard undefined error modpost stack chk guard undefined error modpost stack chk guard undefined error modpost stack chk guard undefined error modpost stack chk guard undefined error modpost stack chk guard undefined make error make deleting file usr src zfs module module symvers make error make error make error make error describe how to reproduce the problem from official armbian buster apt update apt upgrade apt install build essential autoconf automake libtool gawk alien fakeroot dkms libblkid dev uuid dev libudev dev libssl dev dev libaio dev dev libelf dev linux headers current dev setuptools cffi libffi dev cd usr src git clone cd zfs git checkout remotes origin zfs release b git cherry pick sh autogen sh configure make s j nproc include any warning errors backtraces from the system logs important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example this is an example how log text should be marked wrap it with
1
722,474
24,863,387,256
IssuesEvent
2022-10-27 09:58:53
AY2223S1-CS2103T-T15-1/tp
https://api.github.com/repos/AY2223S1-CS2103T-T15-1/tp
closed
Edit person does not keep the leave and rating history
priority.Medium type.Bug
Leave and ratings both have some form of history. When editing a person's field, that history is lost. This behaviour is likely due to `edit` command returning a new `Person` instance, and the history is not being carried over properly.
1.0
Edit person does not keep the leave and rating history - Leave and ratings both have some form of history. When editing a person's field, that history is lost. This behaviour is likely due to `edit` command returning a new `Person` instance, and the history is not being carried over properly.
non_defect
edit person does not keep the leave and rating history leave and ratings both have some form of history when editing a person s field that history is lost this behaviour is likely due to edit command returning a new person instance and the history is not being carried over properly
0
76,934
7,549,793,321
IssuesEvent
2018-04-18 15:06:45
reigndesign/basfchinaplas
https://api.github.com/repos/reigndesign/basfchinaplas
closed
Add filter to questions view
4 - Testing bug
User should be able to filter by tech talk topics <!--- @huboard:{"order":71.0,"milestone_order":71,"custom_state":""} -->
1.0
Add filter to questions view - User should be able to filter by tech talk topics <!--- @huboard:{"order":71.0,"milestone_order":71,"custom_state":""} -->
non_defect
add filter to questions view user should be able to filter by tech talk topics huboard order milestone order custom state
0
29,674
5,810,021,271
IssuesEvent
2017-05-04 14:37:52
primefaces/primeng
https://api.github.com/repos/primefaces/primeng
closed
Calendar TimeOnly incorrect hourformat
defect
When we use calendar time picker with time only option and hourformat="12" then it displays popup with incorrect AM/PM display.It always display AM in time picker popup as below. <`p-calendar [(ngModel)]="date8" [timeOnly]="true" [hourFormat]="12"></p-calendar>` ![image](https://cloud.githubusercontent.com/assets/3127317/25690962/122c2abe-30c9-11e7-8a10-8a967dfe65a1.png)
1.0
Calendar TimeOnly incorrect hourformat - When we use calendar time picker with time only option and hourformat="12" then it displays popup with incorrect AM/PM display.It always display AM in time picker popup as below. <`p-calendar [(ngModel)]="date8" [timeOnly]="true" [hourFormat]="12"></p-calendar>` ![image](https://cloud.githubusercontent.com/assets/3127317/25690962/122c2abe-30c9-11e7-8a10-8a967dfe65a1.png)
defect
calendar timeonly incorrect hourformat when we use calendar time picker with time only option and hourformat then it displays popup with incorrect am pm display it always display am in time picker popup as below
1
198,124
22,617,909,656
IssuesEvent
2022-06-30 01:21:54
n-devs/libpack-scripts
https://api.github.com/repos/n-devs/libpack-scripts
opened
CVE-2022-2217 (High) detected in parse-url-5.0.1.tgz
security vulnerability
## CVE-2022-2217 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>parse-url-5.0.1.tgz</b></p></summary> <p>An advanced url parser supporting git urls too.</p> <p>Library home page: <a href="https://registry.npmjs.org/parse-url/-/parse-url-5.0.1.tgz">https://registry.npmjs.org/parse-url/-/parse-url-5.0.1.tgz</a></p> <p>Path to dependency file: /libpack-scripts/package.json</p> <p>Path to vulnerable library: /node_modules/parse-url/package.json</p> <p> Dependency Hierarchy: - lerna-3.15.0.tgz (Root Library) - version-3.15.0.tgz - github-client-3.14.2.tgz - git-url-parse-11.1.2.tgz - git-up-4.0.1.tgz - :x: **parse-url-5.0.1.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Cross-site Scripting (XSS) - Generic in GitHub repository ionicabizau/parse-url prior to 7.0.0. <p>Publish Date: 2022-06-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2217>CVE-2022-2217</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/4e046c63-b1ca-4bcc-b418-29796918a71b/">https://huntr.dev/bounties/4e046c63-b1ca-4bcc-b418-29796918a71b/</a></p> <p>Release Date: 2022-06-27</p> <p>Fix Resolution: parse-url - 6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-2217 (High) detected in parse-url-5.0.1.tgz - ## CVE-2022-2217 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>parse-url-5.0.1.tgz</b></p></summary> <p>An advanced url parser supporting git urls too.</p> <p>Library home page: <a href="https://registry.npmjs.org/parse-url/-/parse-url-5.0.1.tgz">https://registry.npmjs.org/parse-url/-/parse-url-5.0.1.tgz</a></p> <p>Path to dependency file: /libpack-scripts/package.json</p> <p>Path to vulnerable library: /node_modules/parse-url/package.json</p> <p> Dependency Hierarchy: - lerna-3.15.0.tgz (Root Library) - version-3.15.0.tgz - github-client-3.14.2.tgz - git-url-parse-11.1.2.tgz - git-up-4.0.1.tgz - :x: **parse-url-5.0.1.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Cross-site Scripting (XSS) - Generic in GitHub repository ionicabizau/parse-url prior to 7.0.0. <p>Publish Date: 2022-06-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2217>CVE-2022-2217</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/4e046c63-b1ca-4bcc-b418-29796918a71b/">https://huntr.dev/bounties/4e046c63-b1ca-4bcc-b418-29796918a71b/</a></p> <p>Release Date: 2022-06-27</p> <p>Fix Resolution: parse-url - 6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in parse url tgz cve high severity vulnerability vulnerable library parse url tgz an advanced url parser supporting git urls too library home page a href path to dependency file libpack scripts package json path to vulnerable library node modules parse url package json dependency hierarchy lerna tgz root library version tgz github client tgz git url parse tgz git up tgz x parse url tgz vulnerable library vulnerability details cross site scripting xss generic in github repository ionicabizau parse url prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution parse url step up your open source security game with mend
0
275,975
20,962,339,250
IssuesEvent
2022-03-28 00:10:56
vlizarn/challenge-python-cash-machine
https://api.github.com/repos/vlizarn/challenge-python-cash-machine
opened
[Feature]: Fix whitespace of all examples of Code Sample.
documentation enhancement invalid update
## Overview - Assign | Assign | Description | Status | | :----: | :---- | :----: | | 1 | Fix whitespace of example 1. | | | 2 | Fix whitespace of example 2. | | ## Overview - Pull Requests | Pull Requests| Description | | :----: | :---- | | | Apply pull merge request of the master-readme branch for the master branch. |
1.0
[Feature]: Fix whitespace of all examples of Code Sample. - ## Overview - Assign | Assign | Description | Status | | :----: | :---- | :----: | | 1 | Fix whitespace of example 1. | | | 2 | Fix whitespace of example 2. | | ## Overview - Pull Requests | Pull Requests| Description | | :----: | :---- | | | Apply pull merge request of the master-readme branch for the master branch. |
non_defect
fix whitespace of all examples of code sample overview assign assign description status fix whitespace of example fix whitespace of example overview pull requests pull requests description apply pull merge request of the master readme branch for the master branch
0
45,014
12,519,853,381
IssuesEvent
2020-06-03 15:01:21
hikaya-io/dots-frontend
https://api.github.com/repos/hikaya-io/dots-frontend
closed
Remove additional modal for logout
defect
**Current behavior** When user clicks on logout a modal window opens asking if they are sure they want to log out **To Reproduce** 1. Click on the profile avatar on top right corner 2. Click on Logout **Expected behavior** When use clicks on logout, user is immediately logged out. No modal window is displayed to ask for further confirmation **Screenshots** ![image](https://user-images.githubusercontent.com/13760198/83622382-2ef67980-a590-11ea-8ee7-28c38be89f55.png)
1.0
Remove additional modal for logout - **Current behavior** When user clicks on logout a modal window opens asking if they are sure they want to log out **To Reproduce** 1. Click on the profile avatar on top right corner 2. Click on Logout **Expected behavior** When use clicks on logout, user is immediately logged out. No modal window is displayed to ask for further confirmation **Screenshots** ![image](https://user-images.githubusercontent.com/13760198/83622382-2ef67980-a590-11ea-8ee7-28c38be89f55.png)
defect
remove additional modal for logout current behavior when user clicks on logout a modal window opens asking if they are sure they want to log out to reproduce click on the profile avatar on top right corner click on logout expected behavior when use clicks on logout user is immediately logged out no modal window is displayed to ask for further confirmation screenshots
1
3,149
13,162,721,370
IssuesEvent
2020-08-10 22:16:25
mozilla-mobile/fenix
https://api.github.com/repos/mozilla-mobile/fenix
reopened
FNX3-14268 ⁃ [UI Tests] Improve wait on session loaded
eng:automation needs:triage 🐞 bug
A number of our tests are flaky because Espresso continues during page load which causes them to break on assertions afterwards. We should use a an IdlingResource implementation that waits until the current session is not loading anymore.
1.0
FNX3-14268 ⁃ [UI Tests] Improve wait on session loaded - A number of our tests are flaky because Espresso continues during page load which causes them to break on assertions afterwards. We should use a an IdlingResource implementation that waits until the current session is not loading anymore.
non_defect
⁃ improve wait on session loaded a number of our tests are flaky because espresso continues during page load which causes them to break on assertions afterwards we should use a an idlingresource implementation that waits until the current session is not loading anymore
0
3,722
2,610,067,611
IssuesEvent
2015-02-26 18:19:51
chrsmith/jsjsj122
https://api.github.com/repos/chrsmith/jsjsj122
opened
路桥前列腺炎专科
auto-migrated Priority-Medium Type-Defect
``` 路桥前列腺炎专科【台州五洲生殖医院】24小时健康咨询热线 :0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州市椒江 区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108、118、1 98及椒江一金清公交车直达枫南小区,乘坐107、105、109、112、 901、 902公交车到星星广场下车,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 ``` ----- Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 8:53
1.0
路桥前列腺炎专科 - ``` 路桥前列腺炎专科【台州五洲生殖医院】24小时健康咨询热线 :0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州市椒江 区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108、118、1 98及椒江一金清公交车直达枫南小区,乘坐107、105、109、112、 901、 902公交车到星星广场下车,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 ``` ----- Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 8:53
defect
路桥前列腺炎专科 路桥前列腺炎专科【台州五洲生殖医院】 微信号tzwzszyy 医院地址 台州市椒江 (枫南大转盘旁)乘车线路 、 、 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at
1
129,988
10,593,875,585
IssuesEvent
2019-10-09 15:38:40
kcigeospatial/Fred_Co_Land-Management
https://api.github.com/repos/kcigeospatial/Fred_Co_Land-Management
closed
Trade License - Confirm License Details
Ready for Test Env. Retest
FOR ALL TRADE LICENSES: At Awaiting Renewal Details, prepopulate the Confirm License Details page with the existing Trade License Details information. A check box should be added to the Confirm detail page that will show to the applicant on R4C, to confirm that they have reviewed and updated their details as necessary. The Confirm Details field that currently exists should remain a back office only field for intake staff to confirm the review of these updates. ![image](https://user-images.githubusercontent.com/31770504/66069916-5509dc80-e51e-11e9-8543-e1664498c340.png)
2.0
Trade License - Confirm License Details - FOR ALL TRADE LICENSES: At Awaiting Renewal Details, prepopulate the Confirm License Details page with the existing Trade License Details information. A check box should be added to the Confirm detail page that will show to the applicant on R4C, to confirm that they have reviewed and updated their details as necessary. The Confirm Details field that currently exists should remain a back office only field for intake staff to confirm the review of these updates. ![image](https://user-images.githubusercontent.com/31770504/66069916-5509dc80-e51e-11e9-8543-e1664498c340.png)
non_defect
trade license confirm license details for all trade licenses at awaiting renewal details prepopulate the confirm license details page with the existing trade license details information a check box should be added to the confirm detail page that will show to the applicant on to confirm that they have reviewed and updated their details as necessary the confirm details field that currently exists should remain a back office only field for intake staff to confirm the review of these updates
0
454,041
13,093,816,254
IssuesEvent
2020-08-03 11:08:49
geosolutions-it/ckanext-faoclh
https://api.github.com/repos/geosolutions-it/ckanext-faoclh
opened
FAO Pending Configs
Priority: Blocker
- [ ] PDF preview issue: for example [this](https://clh-ckan.review.fao.org/dataset/communities-of-practice). The error is the following: Content is too large to be proxied. Allowed file size: 2097152, Content-Length: 5864691. Configure [this](https://docs.ckan.org/en/2.8/maintaining/configuration.html#ckan-resource-proxy-max-file-size) to 8 mb
1.0
FAO Pending Configs - - [ ] PDF preview issue: for example [this](https://clh-ckan.review.fao.org/dataset/communities-of-practice). The error is the following: Content is too large to be proxied. Allowed file size: 2097152, Content-Length: 5864691. Configure [this](https://docs.ckan.org/en/2.8/maintaining/configuration.html#ckan-resource-proxy-max-file-size) to 8 mb
non_defect
fao pending configs pdf preview issue for example the error is the following content is too large to be proxied allowed file size content length configure to mb
0
3,282
2,610,059,842
IssuesEvent
2015-02-26 18:17:39
chrsmith/jsjsj122
https://api.github.com/repos/chrsmith/jsjsj122
opened
临海不孕不育检查需要多少钱
auto-migrated Priority-Medium Type-Defect
``` 临海不孕不育检查需要多少钱【台州五洲生殖医院】24小时健 康咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址: 台州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104� ��108、118、198及椒江一金清公交车直达枫南小区,乘坐107、105 、109、112、901、 902公交车到星星广场下车,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 ``` ----- Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:02
1.0
临海不孕不育检查需要多少钱 - ``` 临海不孕不育检查需要多少钱【台州五洲生殖医院】24小时健 康咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址: 台州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104� ��108、118、198及椒江一金清公交车直达枫南小区,乘坐107、105 、109、112、901、 902公交车到星星广场下车,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 ``` ----- Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:02
defect
临海不孕不育检查需要多少钱 临海不孕不育检查需要多少钱【台州五洲生殖医院】 康咨询热线 微信号tzwzszyy 医院地址 (枫南大转盘旁)乘车线路 � �� 、 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at
1
56,488
15,109,280,783
IssuesEvent
2021-02-08 17:38:37
idaholab/moose
https://api.github.com/repos/idaholab/moose
opened
Automatic Scaling Not Working with Generalized Plane Strain
T: defect
## Bug Description <!--A clear and concise description of the problem (Note: A missing feature is not a bug).--> With `automatic_scaling = true`, the generalized plane strain mechanics solution does not match an equivalent 3D solution. ## Steps to Reproduce <!--Steps to reproduce the behavior (input file, or modifications to an existing input file, etc.)--> I've attached two simple (only need 1 cpu cause they are small) moose input files. If you change the `automatic_scaling` parameter you will see the solutions are different. ## Impact <!--Does this prevent you from getting your work done, or is it more of an annoyance?--> A wrong solution is provided. [rectangleExamples.zip](https://github.com/idaholab/moose/files/5945614/rectangleExamples.zip)
1.0
Automatic Scaling Not Working with Generalized Plane Strain - ## Bug Description <!--A clear and concise description of the problem (Note: A missing feature is not a bug).--> With `automatic_scaling = true`, the generalized plane strain mechanics solution does not match an equivalent 3D solution. ## Steps to Reproduce <!--Steps to reproduce the behavior (input file, or modifications to an existing input file, etc.)--> I've attached two simple (only need 1 cpu cause they are small) moose input files. If you change the `automatic_scaling` parameter you will see the solutions are different. ## Impact <!--Does this prevent you from getting your work done, or is it more of an annoyance?--> A wrong solution is provided. [rectangleExamples.zip](https://github.com/idaholab/moose/files/5945614/rectangleExamples.zip)
defect
automatic scaling not working with generalized plane strain bug description with automatic scaling true the generalized plane strain mechanics solution does not match an equivalent solution steps to reproduce i ve attached two simple only need cpu cause they are small moose input files if you change the automatic scaling parameter you will see the solutions are different impact a wrong solution is provided
1
272,638
29,795,060,743
IssuesEvent
2023-06-16 01:07:50
billmcchesney1/pacbot
https://api.github.com/repos/billmcchesney1/pacbot
closed
CVE-2021-22047 (Medium) detected in spring-data-rest-webmvc-3.0.9.RELEASE.jar - autoclosed
Mend: dependency security vulnerability
## CVE-2021-22047 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-data-rest-webmvc-3.0.9.RELEASE.jar</b></p></summary> <p>Spring Data REST - WebMVC</p> <p>Library home page: <a href="http://www.spring.io/spring-data">http://www.spring.io/spring-data</a></p> <p>Path to dependency file: /api/pacman-api-admin/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/data/spring-data-rest-webmvc/3.0.9.RELEASE/spring-data-rest-webmvc-3.0.9.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/data/spring-data-rest-webmvc/3.0.9.RELEASE/spring-data-rest-webmvc-3.0.9.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-data-rest-2.0.4.RELEASE.jar (Root Library) - :x: **spring-data-rest-webmvc-3.0.9.RELEASE.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/pacbot/commit/acf9a0620c1a37cee4f2896d71e1c3731c5c7b06">acf9a0620c1a37cee4f2896d71e1c3731c5c7b06</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> In Spring Data REST versions 3.4.0 - 3.4.13, 3.5.0 - 3.5.5, and older unsupported versions, HTTP resources implemented by custom controllers using a configured base API path and a controller type-level request mapping are additionally exposed under URIs that can potentially be exposed for unauthorized access depending on the Spring Security configuration. <p>Publish Date: 2021-10-28 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-22047>CVE-2021-22047</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22047">https://tanzu.vmware.com/security/cve-2021-22047</a></p> <p>Release Date: 2021-10-28</p> <p>Fix Resolution: org.springframework.data:spring-data-rest-webmvc:3.4.14,3.5.6</p> </p> </details> <p></p>
True
CVE-2021-22047 (Medium) detected in spring-data-rest-webmvc-3.0.9.RELEASE.jar - autoclosed - ## CVE-2021-22047 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-data-rest-webmvc-3.0.9.RELEASE.jar</b></p></summary> <p>Spring Data REST - WebMVC</p> <p>Library home page: <a href="http://www.spring.io/spring-data">http://www.spring.io/spring-data</a></p> <p>Path to dependency file: /api/pacman-api-admin/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/data/spring-data-rest-webmvc/3.0.9.RELEASE/spring-data-rest-webmvc-3.0.9.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/data/spring-data-rest-webmvc/3.0.9.RELEASE/spring-data-rest-webmvc-3.0.9.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-data-rest-2.0.4.RELEASE.jar (Root Library) - :x: **spring-data-rest-webmvc-3.0.9.RELEASE.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/pacbot/commit/acf9a0620c1a37cee4f2896d71e1c3731c5c7b06">acf9a0620c1a37cee4f2896d71e1c3731c5c7b06</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> In Spring Data REST versions 3.4.0 - 3.4.13, 3.5.0 - 3.5.5, and older unsupported versions, HTTP resources implemented by custom controllers using a configured base API path and a controller type-level request mapping are additionally exposed under URIs that can potentially be exposed for unauthorized access depending on the Spring Security configuration. <p>Publish Date: 2021-10-28 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-22047>CVE-2021-22047</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22047">https://tanzu.vmware.com/security/cve-2021-22047</a></p> <p>Release Date: 2021-10-28</p> <p>Fix Resolution: org.springframework.data:spring-data-rest-webmvc:3.4.14,3.5.6</p> </p> </details> <p></p>
non_defect
cve medium detected in spring data rest webmvc release jar autoclosed cve medium severity vulnerability vulnerable library spring data rest webmvc release jar spring data rest webmvc library home page a href path to dependency file api pacman api admin pom xml path to vulnerable library home wss scanner repository org springframework data spring data rest webmvc release spring data rest webmvc release jar home wss scanner repository org springframework data spring data rest webmvc release spring data rest webmvc release jar dependency hierarchy spring boot starter data rest release jar root library x spring data rest webmvc release jar vulnerable library found in head commit a href found in base branch master vulnerability details in spring data rest versions and older unsupported versions http resources implemented by custom controllers using a configured base api path and a controller type level request mapping are additionally exposed under uris that can potentially be exposed for unauthorized access depending on the spring security configuration publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework data spring data rest webmvc
0
782,048
27,484,877,821
IssuesEvent
2023-03-04 01:29:14
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
Performing any operations with complex tensors on MPS causes unrecoverable crash
high priority triage review module: mps
### 🐛 Describe the bug Consider following trivial example: ``` python -c "import torch;torch.tensor([1+2j],device='mps')+1" libc++abi: terminating with uncaught exception of type c10::TypeError: Trying to convert ComplexFloat to the MPS backend but it does not have support for that dtype. Exception raised from getMPSScalarType at /Users/runner/work/pytorch/pytorch/pytorch/aten/src/ATen/native/mps/OperationUtils.mm:95 (most recent call first): frame #0: at::native::mps::getMPSScalarType(c10::ScalarType) + 180 (0x16fed979c in libtorch_cpu.dylib) frame #1: invocation function for block in at::native::mps::binaryOpTensor(at::Tensor const&, at::Tensor const&, c10::Scalar const&, at::Tensor const&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, MPSGraphTensor* (at::native::mps::BinaryOpCachedGraph*, MPSGraphTensor*, MPSGraphTensor*) block_pointer) + 108 (0x16fef9590 in libtorch_cpu.dylib) frame #2: invocation function for block in at::native::mps::MPSGraphCache::CreateCachedGraph(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, at::native::mps::MPSCachedGraph* () block_pointer) + 216 (0x16fef2348 in libtorch_cpu.dylib) frame #3: _dispatch_client_callout + 20 (0x19a39e504 in libdispatch.dylib) frame #4: _dispatch_lane_barrier_sync_invoke_and_complete + 56 (0x19a3ada9c in libdispatch.dylib) frame #5: at::native::mps::MPSGraphCache::CreateCachedGraph(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, at::native::mps::MPSCachedGraph* () block_pointer) + 160 (0x16fedd844 in libtorch_cpu.dylib) frame #6: at::native::mps::binaryOpTensor(at::Tensor const&, at::Tensor const&, c10::Scalar const&, at::Tensor const&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, MPSGraphTensor* (at::native::mps::BinaryOpCachedGraph*, MPSGraphTensor*, MPSGraphTensor*) block_pointer) + 2352 (0x16fef8614 in libtorch_cpu.dylib) frame #7: at::native::mps::add_sub_template(at::Tensor const&, at::Tensor const&, c10::Scalar const&, at::Tensor const&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >) + 984 (0x16fefa5b0 in libtorch_cpu.dylib) frame #8: at::native::structured_add_out_mps::impl(at::Tensor const&, at::Tensor const&, c10::Scalar const&, at::Tensor const&) + 64 (0x16fefd31c in libtorch_cpu.dylib) frame #9: at::(anonymous namespace)::wrapper_MPS_add_Tensor(at::Tensor const&, at::Tensor const&, c10::Scalar const&) + 152 (0x16d6051a8 in libtorch_cpu.dylib) frame #10: c10::impl::wrap_kernel_functor_unboxed_<c10::impl::detail::WrapFunctionIntoFunctor_<c10::CompileTimeFunctionPointer<at::Tensor (c10::DispatchKeySet, at::Tensor const&, at::Tensor const&, c10::Scalar const&), &(torch::autograd::VariableType::(anonymous namespace)::add_Tensor(c10::DispatchKeySet, at::Tensor const&, at::Tensor const&, c10::Scalar const&))>, at::Tensor, c10::guts::typelist::typelist<c10::DispatchKeySet, at::Tensor const&, at::Tensor const&, c10::Scalar const&> >, at::Tensor (c10::DispatchKeySet, at::Tensor const&, at::Tensor const&, c10::Scalar const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, at::Tensor const&, c10::Scalar const&) + 1392 (0x16e507a18 in libtorch_cpu.dylib) frame #11: at::_ops::add_Tensor::call(at::Tensor const&, at::Tensor const&, c10::Scalar const&) + 296 (0x16c75335c in libtorch_cpu.dylib) frame #12: torch::autograd::THPVariable_add(_object*, _object*, _object*) + 716 (0x106a320cc in libtorch_python.dylib) frame #13: _object* torch::autograd::TypeError_to_NotImplemented_<&(torch::autograd::THPVariable_add(_object*, _object*, _object*))>(_object*, _object*, _object*) + 12 (0x106a154f8 in libtorch_python.dylib) frame #14: method_vectorcall_VARARGS_KEYWORDS + 144 (0x1047a4998 in python3.11) frame #15: vectorcall_maybe + 212 (0x104828b10 in python3.11) frame #16: slot_nb_add + 128 (0x104824914 in python3.11) frame #17: PyNumber_Add + 92 (0x10476f178 in python3.11) frame #18: _PyEval_EvalFrameDefault + 231128 (0x1048dfe60 in python3.11) frame #19: _PyEval_Vector + 464 (0x1048a4d30 in python3.11) frame #20: PyEval_EvalCode + 248 (0x1048a4ad0 in python3.11) frame #21: run_mod + 184 (0x10493cee4 in python3.11) frame #22: PyRun_SimpleStringFlags + 144 (0x10493f3d0 in python3.11) frame #23: pymain_run_command + 160 (0x10495f020 in python3.11) frame #24: Py_RunMain + 248 (0x10495e7f4 in python3.11) frame #25: main + 56 (0x104729264 in python3.11) frame #26: start + 2544 (0x19a1fbe50 in dyld) ``` I understand that complex type is unsupported on MPS, but that's not the reason to crash the runtime. (And perhaps one should simply not be allowed to create such types on MPS? ### Versions 1.13.1, 2.0.0, nightly cc @ezyang @gchanan @zou3519 @kulinseth @albanD @DenisVieriu97 @razarmehr @abhudev
1.0
Performing any operations with complex tensors on MPS causes unrecoverable crash - ### 🐛 Describe the bug Consider following trivial example: ``` python -c "import torch;torch.tensor([1+2j],device='mps')+1" libc++abi: terminating with uncaught exception of type c10::TypeError: Trying to convert ComplexFloat to the MPS backend but it does not have support for that dtype. Exception raised from getMPSScalarType at /Users/runner/work/pytorch/pytorch/pytorch/aten/src/ATen/native/mps/OperationUtils.mm:95 (most recent call first): frame #0: at::native::mps::getMPSScalarType(c10::ScalarType) + 180 (0x16fed979c in libtorch_cpu.dylib) frame #1: invocation function for block in at::native::mps::binaryOpTensor(at::Tensor const&, at::Tensor const&, c10::Scalar const&, at::Tensor const&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, MPSGraphTensor* (at::native::mps::BinaryOpCachedGraph*, MPSGraphTensor*, MPSGraphTensor*) block_pointer) + 108 (0x16fef9590 in libtorch_cpu.dylib) frame #2: invocation function for block in at::native::mps::MPSGraphCache::CreateCachedGraph(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, at::native::mps::MPSCachedGraph* () block_pointer) + 216 (0x16fef2348 in libtorch_cpu.dylib) frame #3: _dispatch_client_callout + 20 (0x19a39e504 in libdispatch.dylib) frame #4: _dispatch_lane_barrier_sync_invoke_and_complete + 56 (0x19a3ada9c in libdispatch.dylib) frame #5: at::native::mps::MPSGraphCache::CreateCachedGraph(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, at::native::mps::MPSCachedGraph* () block_pointer) + 160 (0x16fedd844 in libtorch_cpu.dylib) frame #6: at::native::mps::binaryOpTensor(at::Tensor const&, at::Tensor const&, c10::Scalar const&, at::Tensor const&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, MPSGraphTensor* (at::native::mps::BinaryOpCachedGraph*, MPSGraphTensor*, MPSGraphTensor*) block_pointer) + 2352 (0x16fef8614 in libtorch_cpu.dylib) frame #7: at::native::mps::add_sub_template(at::Tensor const&, at::Tensor const&, c10::Scalar const&, at::Tensor const&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >) + 984 (0x16fefa5b0 in libtorch_cpu.dylib) frame #8: at::native::structured_add_out_mps::impl(at::Tensor const&, at::Tensor const&, c10::Scalar const&, at::Tensor const&) + 64 (0x16fefd31c in libtorch_cpu.dylib) frame #9: at::(anonymous namespace)::wrapper_MPS_add_Tensor(at::Tensor const&, at::Tensor const&, c10::Scalar const&) + 152 (0x16d6051a8 in libtorch_cpu.dylib) frame #10: c10::impl::wrap_kernel_functor_unboxed_<c10::impl::detail::WrapFunctionIntoFunctor_<c10::CompileTimeFunctionPointer<at::Tensor (c10::DispatchKeySet, at::Tensor const&, at::Tensor const&, c10::Scalar const&), &(torch::autograd::VariableType::(anonymous namespace)::add_Tensor(c10::DispatchKeySet, at::Tensor const&, at::Tensor const&, c10::Scalar const&))>, at::Tensor, c10::guts::typelist::typelist<c10::DispatchKeySet, at::Tensor const&, at::Tensor const&, c10::Scalar const&> >, at::Tensor (c10::DispatchKeySet, at::Tensor const&, at::Tensor const&, c10::Scalar const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, at::Tensor const&, c10::Scalar const&) + 1392 (0x16e507a18 in libtorch_cpu.dylib) frame #11: at::_ops::add_Tensor::call(at::Tensor const&, at::Tensor const&, c10::Scalar const&) + 296 (0x16c75335c in libtorch_cpu.dylib) frame #12: torch::autograd::THPVariable_add(_object*, _object*, _object*) + 716 (0x106a320cc in libtorch_python.dylib) frame #13: _object* torch::autograd::TypeError_to_NotImplemented_<&(torch::autograd::THPVariable_add(_object*, _object*, _object*))>(_object*, _object*, _object*) + 12 (0x106a154f8 in libtorch_python.dylib) frame #14: method_vectorcall_VARARGS_KEYWORDS + 144 (0x1047a4998 in python3.11) frame #15: vectorcall_maybe + 212 (0x104828b10 in python3.11) frame #16: slot_nb_add + 128 (0x104824914 in python3.11) frame #17: PyNumber_Add + 92 (0x10476f178 in python3.11) frame #18: _PyEval_EvalFrameDefault + 231128 (0x1048dfe60 in python3.11) frame #19: _PyEval_Vector + 464 (0x1048a4d30 in python3.11) frame #20: PyEval_EvalCode + 248 (0x1048a4ad0 in python3.11) frame #21: run_mod + 184 (0x10493cee4 in python3.11) frame #22: PyRun_SimpleStringFlags + 144 (0x10493f3d0 in python3.11) frame #23: pymain_run_command + 160 (0x10495f020 in python3.11) frame #24: Py_RunMain + 248 (0x10495e7f4 in python3.11) frame #25: main + 56 (0x104729264 in python3.11) frame #26: start + 2544 (0x19a1fbe50 in dyld) ``` I understand that complex type is unsupported on MPS, but that's not the reason to crash the runtime. (And perhaps one should simply not be allowed to create such types on MPS? ### Versions 1.13.1, 2.0.0, nightly cc @ezyang @gchanan @zou3519 @kulinseth @albanD @DenisVieriu97 @razarmehr @abhudev
non_defect
performing any operations with complex tensors on mps causes unrecoverable crash 🐛 describe the bug consider following trivial example python c import torch torch tensor device mps libc abi terminating with uncaught exception of type typeerror trying to convert complexfloat to the mps backend but it does not have support for that dtype exception raised from getmpsscalartype at users runner work pytorch pytorch pytorch aten src aten native mps operationutils mm most recent call first frame at native mps getmpsscalartype scalartype in libtorch cpu dylib frame invocation function for block in at native mps binaryoptensor at tensor const at tensor const scalar const at tensor const std basic string std allocator mpsgraphtensor at native mps binaryopcachedgraph mpsgraphtensor mpsgraphtensor block pointer in libtorch cpu dylib frame invocation function for block in at native mps mpsgraphcache createcachedgraph std basic string std allocator const at native mps mpscachedgraph block pointer in libtorch cpu dylib frame dispatch client callout in libdispatch dylib frame dispatch lane barrier sync invoke and complete in libdispatch dylib frame at native mps mpsgraphcache createcachedgraph std basic string std allocator const at native mps mpscachedgraph block pointer in libtorch cpu dylib frame at native mps binaryoptensor at tensor const at tensor const scalar const at tensor const std basic string std allocator mpsgraphtensor at native mps binaryopcachedgraph mpsgraphtensor mpsgraphtensor block pointer in libtorch cpu dylib frame at native mps add sub template at tensor const at tensor const scalar const at tensor const std basic string std allocator in libtorch cpu dylib frame at native structured add out mps impl at tensor const at tensor const scalar const at tensor const in libtorch cpu dylib frame at anonymous namespace wrapper mps add tensor at tensor const at tensor const scalar const in libtorch cpu dylib frame impl wrap kernel functor unboxed at tensor guts typelist typelist at tensor dispatchkeyset at tensor const at tensor const scalar const call operatorkernel dispatchkeyset at tensor const at tensor const scalar const in libtorch cpu dylib frame at ops add tensor call at tensor const at tensor const scalar const in libtorch cpu dylib frame torch autograd thpvariable add object object object in libtorch python dylib frame object torch autograd typeerror to notimplemented object object object in libtorch python dylib frame method vectorcall varargs keywords in frame vectorcall maybe in frame slot nb add in frame pynumber add in frame pyeval evalframedefault in frame pyeval vector in frame pyeval evalcode in frame run mod in frame pyrun simplestringflags in frame pymain run command in frame py runmain in frame main in frame start in dyld i understand that complex type is unsupported on mps but that s not the reason to crash the runtime and perhaps one should simply not be allowed to create such types on mps versions nightly cc ezyang gchanan kulinseth alband razarmehr abhudev
0
31,864
6,651,310,558
IssuesEvent
2017-09-28 19:34:39
Cockatrice/Cockatrice
https://api.github.com/repos/Cockatrice/Cockatrice
closed
Forgot Password uses case-sensitive email compare
App - Servatrice Defect - Basic Easy Change
<b>OS:</b> Windows 7 Professional SP1 <b>Cockatrice version:</b> 2.3.17 __________ Had difficulty getting Forgot Password to work. Turns out it expects email address to be case sensitive. Suggest cast both to lower case before compare. I believe email clients and servers do that anyway so that is what people might expect.
1.0
Forgot Password uses case-sensitive email compare - <b>OS:</b> Windows 7 Professional SP1 <b>Cockatrice version:</b> 2.3.17 __________ Had difficulty getting Forgot Password to work. Turns out it expects email address to be case sensitive. Suggest cast both to lower case before compare. I believe email clients and servers do that anyway so that is what people might expect.
defect
forgot password uses case sensitive email compare os windows professional cockatrice version had difficulty getting forgot password to work turns out it expects email address to be case sensitive suggest cast both to lower case before compare i believe email clients and servers do that anyway so that is what people might expect
1
26,809
4,790,764,028
IssuesEvent
2016-10-31 09:57:51
vuonghv/libdasm
https://api.github.com/repos/vuonghv/libdasm
opened
libdasm misinterprets JLE
auto-migrated Priority-Medium Type-Defect
_From @GoogleCodeExporter on December 17, 2015 11:11_ ``` This is regarding a bug in libdasm library. I've noticed that libdasm misinterprets a specific instruction.(JLE) Sample Opcode : '\x39\x56\x38\x0F\x8E\x1E\x01\x00\x00\x8B\xBD\xE4\xEB\xFF\xFF' Actual dis-assembly of above opcodes : 00433441 |. 3956 38 CMP DWORD PTR DS:[ESI+38],EDX 00433444 |. 0F8E 1E010000 JLE 0x00433568 0043344A |> 8BBD E4EBFFFF MOV EDI,DWORD PTR SS:[EBP-141C] libdasm output: debasish@debasish ~/Downloads/libdasm-beta/examples $ hexdump test.bin 0000000 5639 0f38 1e8e 0001 8b00 e4bd ffeb 00ff 000000f debasish@debasish ~/Downloads/libdasm-beta/examples $ ./das test.bin 00000000 395638 cmp [esi+0x38],edx 00000003 0f8e1e010000 jng 0x127 00000009 8bbde4ebffff mov edi,[ebp-0x141c] debasish@debasish ~/Downloads/libdasm-beta/examples $ You can see, in third line libdasm misinterprets the jle (Jump if less or equal ) instruction to jng (Jump if not greater). cheers, Debasish ``` Original issue reported on code.google.com by `debasish...@gmail.com` on 23 Jan 2014 at 1:57 Attachments: - [libdasm_bug.py](https://storage.googleapis.com/google-code-attachments/libdasm/issue-1/comment-0/libdasm_bug.py) _Copied from original issue: axcheron/libdasm#1_
1.0
libdasm misinterprets JLE - _From @GoogleCodeExporter on December 17, 2015 11:11_ ``` This is regarding a bug in libdasm library. I've noticed that libdasm misinterprets a specific instruction.(JLE) Sample Opcode : '\x39\x56\x38\x0F\x8E\x1E\x01\x00\x00\x8B\xBD\xE4\xEB\xFF\xFF' Actual dis-assembly of above opcodes : 00433441 |. 3956 38 CMP DWORD PTR DS:[ESI+38],EDX 00433444 |. 0F8E 1E010000 JLE 0x00433568 0043344A |> 8BBD E4EBFFFF MOV EDI,DWORD PTR SS:[EBP-141C] libdasm output: debasish@debasish ~/Downloads/libdasm-beta/examples $ hexdump test.bin 0000000 5639 0f38 1e8e 0001 8b00 e4bd ffeb 00ff 000000f debasish@debasish ~/Downloads/libdasm-beta/examples $ ./das test.bin 00000000 395638 cmp [esi+0x38],edx 00000003 0f8e1e010000 jng 0x127 00000009 8bbde4ebffff mov edi,[ebp-0x141c] debasish@debasish ~/Downloads/libdasm-beta/examples $ You can see, in third line libdasm misinterprets the jle (Jump if less or equal ) instruction to jng (Jump if not greater). cheers, Debasish ``` Original issue reported on code.google.com by `debasish...@gmail.com` on 23 Jan 2014 at 1:57 Attachments: - [libdasm_bug.py](https://storage.googleapis.com/google-code-attachments/libdasm/issue-1/comment-0/libdasm_bug.py) _Copied from original issue: axcheron/libdasm#1_
defect
libdasm misinterprets jle from googlecodeexporter on december this is regarding a bug in libdasm library i ve noticed that libdasm misinterprets a specific instruction jle sample opcode xbd xeb xff xff actual dis assembly of above opcodes cmp dword ptr ds edx jle mov edi dword ptr ss libdasm output debasish debasish downloads libdasm beta examples hexdump test bin ffeb debasish debasish downloads libdasm beta examples das test bin cmp edx jng mov edi debasish debasish downloads libdasm beta examples you can see in third line libdasm misinterprets the jle jump if less or equal instruction to jng jump if not greater cheers debasish original issue reported on code google com by debasish gmail com on jan at attachments copied from original issue axcheron libdasm
1
227,954
7,544,657,281
IssuesEvent
2018-04-17 19:07:28
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
USER ISSUE: World Unloaded
High Priority
**Version:** 0.7.3.1 beta **Steps to Reproduce:** Walk away from somewhere and walk back **Expected behavior:** Chunks should load back in as you get close to them **Actual behavior:** Unloaded chunks appear which will not load back in, therefore you fall through the world
1.0
USER ISSUE: World Unloaded - **Version:** 0.7.3.1 beta **Steps to Reproduce:** Walk away from somewhere and walk back **Expected behavior:** Chunks should load back in as you get close to them **Actual behavior:** Unloaded chunks appear which will not load back in, therefore you fall through the world
non_defect
user issue world unloaded version beta steps to reproduce walk away from somewhere and walk back expected behavior chunks should load back in as you get close to them actual behavior unloaded chunks appear which will not load back in therefore you fall through the world
0
121,822
4,821,794,681
IssuesEvent
2016-11-05 14:40:16
nextcloud/appstore
https://api.github.com/repos/nextcloud/appstore
closed
Use Nextcloud versions to generate downloads on app detail page
enhancement help wanted high priority starter issue
Instead of https://github.com/nextcloud/appstore/blob/master/nextcloudappstore/settings/base.py#L253 the versions in the database should be used. The app detail page needs to iterate through all versions and group them by the first version (e.g. 10.0.0 and 10.0.1 compatible releases should be grouped under 10) so that even apps that depend on 10.0.1 appear under the latest Nextcloud 10 release
1.0
Use Nextcloud versions to generate downloads on app detail page - Instead of https://github.com/nextcloud/appstore/blob/master/nextcloudappstore/settings/base.py#L253 the versions in the database should be used. The app detail page needs to iterate through all versions and group them by the first version (e.g. 10.0.0 and 10.0.1 compatible releases should be grouped under 10) so that even apps that depend on 10.0.1 appear under the latest Nextcloud 10 release
non_defect
use nextcloud versions to generate downloads on app detail page instead of the versions in the database should be used the app detail page needs to iterate through all versions and group them by the first version e g and compatible releases should be grouped under so that even apps that depend on appear under the latest nextcloud release
0
76,823
26,617,085,171
IssuesEvent
2023-01-24 08:16:17
hyperledger/iroha
https://api.github.com/repos/hyperledger/iroha
closed
404 when 1650 isi in genesis.
Bug iroha2 LTS Pre-alpha defect QA-confirmed
### GIT commit hash fe7ea911 ### Minimum working example 1. Run iroha with the 1650 isi in genesis [genesis.txt](https://app.zenhub.com/files/181739240/70fe9b7b-a944-4e8e-96be-459ae4777748/download) ### Expected behaviour I have to create genesis block with some assets. I generate instruction for domain and admin registration. Then I generate instruction for asset registration, 7 setKeyValue instructions and grant permission instruction few times. Total ~ 5870 isi. Then I put it in ONE genesis transaction and add it to genesis tr ansactions that already had one small transaction. I expected it to work. ### Actual behaviour ``` iroha-iroha0-1 | 2023-01-17T07:49:17.396580Z ERROR iroha_core::genesis: Genesis transaction #0 failed error=Failed to accept transaction: Too many instructions in payload, max number is 4096, but got 6310 ``` Iroha throws 404 Not Found. Text: " ,domain_name". If I reduce rows instructions amount to ~700 it works. Also I have tried to split each asset with setkeyvalie and grant isi to separate tx, but result was the same, even with only one row (just for testing). ### Operating system macOS ### Current environment Docker Hub ### Who can help? @astrokov7
1.0
404 when 1650 isi in genesis. - ### GIT commit hash fe7ea911 ### Minimum working example 1. Run iroha with the 1650 isi in genesis [genesis.txt](https://app.zenhub.com/files/181739240/70fe9b7b-a944-4e8e-96be-459ae4777748/download) ### Expected behaviour I have to create genesis block with some assets. I generate instruction for domain and admin registration. Then I generate instruction for asset registration, 7 setKeyValue instructions and grant permission instruction few times. Total ~ 5870 isi. Then I put it in ONE genesis transaction and add it to genesis tr ansactions that already had one small transaction. I expected it to work. ### Actual behaviour ``` iroha-iroha0-1 | 2023-01-17T07:49:17.396580Z ERROR iroha_core::genesis: Genesis transaction #0 failed error=Failed to accept transaction: Too many instructions in payload, max number is 4096, but got 6310 ``` Iroha throws 404 Not Found. Text: " ,domain_name". If I reduce rows instructions amount to ~700 it works. Also I have tried to split each asset with setkeyvalie and grant isi to separate tx, but result was the same, even with only one row (just for testing). ### Operating system macOS ### Current environment Docker Hub ### Who can help? @astrokov7
defect
when isi in genesis git commit hash minimum working example run iroha with the isi in genesis expected behaviour i have to create genesis block with some assets i generate instruction for domain and admin registration then i generate instruction for asset registration setkeyvalue instructions and grant permission instruction few times total isi then i put it in one genesis transaction and add it to genesis tr ansactions that already had one small transaction i expected it to work actual behaviour iroha error iroha core genesis genesis transaction failed error failed to accept transaction too many instructions in payload max number is but got iroha throws not found text domain name if i reduce rows instructions amount to it works also i have tried to split each asset with setkeyvalie and grant isi to separate tx but result was the same even with only one row just for testing operating system macos current environment docker hub who can help
1
21,775
3,551,638,581
IssuesEvent
2016-01-21 05:27:01
bigbluebutton/bigbluebutton
https://api.github.com/repos/bigbluebutton/bigbluebutton
closed
Pressing the undo button rapidly disconnects everyone from the meeting
Accepted Defect Normal Priority Stability Whiteboard
Originally reported on Google Code with ID 1522 ``` This is an issue to track the bug reported here, https://groups.google.com/d/msg/bigbluebutton-dev/wVhb3-lyU2s/Q1s4rudrxUkJ. I'm reporting a bug with 0.81 instability when using "undo" on the whiteboard. If you draw some annotations, then remove them all by clicking "undo" in quick succession, ALL participants become disconnected from the meeting the moment the last annotation is removed. This bug is highly reproducible (I've reproduced it many times now). See the attached video for a demo. It happens when "undo" is clicked a large number of times in quick succession to remove a large annotation. Thanks, Andy ``` Reported by `capilkey` on 2013-05-22 21:51:38
1.0
Pressing the undo button rapidly disconnects everyone from the meeting - Originally reported on Google Code with ID 1522 ``` This is an issue to track the bug reported here, https://groups.google.com/d/msg/bigbluebutton-dev/wVhb3-lyU2s/Q1s4rudrxUkJ. I'm reporting a bug with 0.81 instability when using "undo" on the whiteboard. If you draw some annotations, then remove them all by clicking "undo" in quick succession, ALL participants become disconnected from the meeting the moment the last annotation is removed. This bug is highly reproducible (I've reproduced it many times now). See the attached video for a demo. It happens when "undo" is clicked a large number of times in quick succession to remove a large annotation. Thanks, Andy ``` Reported by `capilkey` on 2013-05-22 21:51:38
defect
pressing the undo button rapidly disconnects everyone from the meeting originally reported on google code with id this is an issue to track the bug reported here i m reporting a bug with instability when using undo on the whiteboard if you draw some annotations then remove them all by clicking undo in quick succession all participants become disconnected from the meeting the moment the last annotation is removed this bug is highly reproducible i ve reproduced it many times now see the attached video for a demo it happens when undo is clicked a large number of times in quick succession to remove a large annotation thanks andy reported by capilkey on
1
39,747
9,645,700,990
IssuesEvent
2019-05-17 09:19:24
line/armeria
https://api.github.com/repos/line/armeria
closed
NullPointerException
defect
Hi, we got a NullPointerException because of the Authority header missing, and please refer to the stack trace message below. The method fillSchemeAndAuthorityIfMissing of class HttpServerHandler creating a new RequestHeaders instance instead of filling the existing one, and this cause the NullPointerException. ```java 2019-05-15 17:40:43.245 WARN [armeria-common-worker-epoll-2-1] c.l.a.s.t.TomcatService [id: 0x7fb03509, L:/192.168.5.137:8080 - R:/192.168.3.170:29786][h2c://test-grpc-demo-0:8080/index#GET] Failed to invoke Tomcat: java.lang.NullPointerException: null at com.linecorp.armeria.server.tomcat.TomcatService.convertRequest(TomcatService.java:356) at com.linecorp.armeria.server.tomcat.TomcatService.lambda$serve$3(TomcatService.java:287) at java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:822) at java.util.concurrent.CompletableFuture.uniHandleStage(CompletableFuture.java:834) at java.util.concurrent.CompletableFuture.handle(CompletableFuture.java:2155) at com.linecorp.armeria.server.tomcat.TomcatService.serve(TomcatService.java:279) at com.linecorp.armeria.server.tomcat.TomcatService.serve(TomcatService.java:77) at com.linecorp.armeria.server.HttpServerHandler.handleRequest(HttpServerHandler.java:376) at com.linecorp.armeria.server.HttpServerHandler.channelRead(HttpServerHandler.java:250) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at com.linecorp.armeria.server.Http2RequestDecoder.onHeadersRead(Http2RequestDecoder.java:143) at com.linecorp.armeria.server.Http2RequestDecoder.onHeadersRead(Http2RequestDecoder.java:164) at io.netty.handler.codec.http2.DefaultHttp2ConnectionDecoder$FrameReadListener.onHeadersRead(DefaultHttp2ConnectionDecoder.java:317) at io.netty.handler.codec.http2.DefaultHttp2ConnectionDecoder$FrameReadListener.onHeadersRead(DefaultHttp2ConnectionDecoder.java:265) at io.netty.handler.codec.http2.DefaultHttp2ConnectionDecoder$PrefaceFrameListener.onHeadersRead(DefaultHttp2ConnectionDecoder.java:600) at io.netty.handler.codec.http2.DefaultHttp2FrameReader$2.processFragment(DefaultHttp2FrameReader.java:483) at io.netty.handler.codec.http2.DefaultHttp2FrameReader.readHeadersFrame(DefaultHttp2FrameReader.java:491) at io.netty.handler.codec.http2.DefaultHttp2FrameReader.processPayloadState(DefaultHttp2FrameReader.java:254) at io.netty.handler.codec.http2.DefaultHttp2FrameReader.readFrame(DefaultHttp2FrameReader.java:160) at io.netty.handler.codec.http2.DefaultHttp2ConnectionDecoder.decodeFrame(DefaultHttp2ConnectionDecoder.java:118) at io.netty.handler.codec.http2.Http2ConnectionHandler$FrameDecoder.decode(Http2ConnectionHandler.java:390) at io.netty.handler.codec.http2.Http2ConnectionHandler$PrefaceDecoder.decode(Http2ConnectionHandler.java:254) at io.netty.handler.codec.http2.Http2ConnectionHandler.decode(Http2ConnectionHandler.java:450) at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:502) at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:441) at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:278) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.handler.codec.ByteToMessageDecoder.handlerRemoved(ByteToMessageDecoder.java:249) at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:507) at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:441) at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:278) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:241) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.handler.flush.FlushConsolidationHandler.channelRead(FlushConsolidationHandler.java:154) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:286) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1408) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:930) at io.netty.channel.epoll.AbstractEpollStreamChannel$EpollStreamUnsafe.epollInReady(AbstractEpollStreamChannel.java:799) at io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:427) at io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:328) at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:905) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.lang.Thread.run(Thread.java:748) ```
1.0
NullPointerException - Hi, we got a NullPointerException because of the Authority header missing, and please refer to the stack trace message below. The method fillSchemeAndAuthorityIfMissing of class HttpServerHandler creating a new RequestHeaders instance instead of filling the existing one, and this cause the NullPointerException. ```java 2019-05-15 17:40:43.245 WARN [armeria-common-worker-epoll-2-1] c.l.a.s.t.TomcatService [id: 0x7fb03509, L:/192.168.5.137:8080 - R:/192.168.3.170:29786][h2c://test-grpc-demo-0:8080/index#GET] Failed to invoke Tomcat: java.lang.NullPointerException: null at com.linecorp.armeria.server.tomcat.TomcatService.convertRequest(TomcatService.java:356) at com.linecorp.armeria.server.tomcat.TomcatService.lambda$serve$3(TomcatService.java:287) at java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:822) at java.util.concurrent.CompletableFuture.uniHandleStage(CompletableFuture.java:834) at java.util.concurrent.CompletableFuture.handle(CompletableFuture.java:2155) at com.linecorp.armeria.server.tomcat.TomcatService.serve(TomcatService.java:279) at com.linecorp.armeria.server.tomcat.TomcatService.serve(TomcatService.java:77) at com.linecorp.armeria.server.HttpServerHandler.handleRequest(HttpServerHandler.java:376) at com.linecorp.armeria.server.HttpServerHandler.channelRead(HttpServerHandler.java:250) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at com.linecorp.armeria.server.Http2RequestDecoder.onHeadersRead(Http2RequestDecoder.java:143) at com.linecorp.armeria.server.Http2RequestDecoder.onHeadersRead(Http2RequestDecoder.java:164) at io.netty.handler.codec.http2.DefaultHttp2ConnectionDecoder$FrameReadListener.onHeadersRead(DefaultHttp2ConnectionDecoder.java:317) at io.netty.handler.codec.http2.DefaultHttp2ConnectionDecoder$FrameReadListener.onHeadersRead(DefaultHttp2ConnectionDecoder.java:265) at io.netty.handler.codec.http2.DefaultHttp2ConnectionDecoder$PrefaceFrameListener.onHeadersRead(DefaultHttp2ConnectionDecoder.java:600) at io.netty.handler.codec.http2.DefaultHttp2FrameReader$2.processFragment(DefaultHttp2FrameReader.java:483) at io.netty.handler.codec.http2.DefaultHttp2FrameReader.readHeadersFrame(DefaultHttp2FrameReader.java:491) at io.netty.handler.codec.http2.DefaultHttp2FrameReader.processPayloadState(DefaultHttp2FrameReader.java:254) at io.netty.handler.codec.http2.DefaultHttp2FrameReader.readFrame(DefaultHttp2FrameReader.java:160) at io.netty.handler.codec.http2.DefaultHttp2ConnectionDecoder.decodeFrame(DefaultHttp2ConnectionDecoder.java:118) at io.netty.handler.codec.http2.Http2ConnectionHandler$FrameDecoder.decode(Http2ConnectionHandler.java:390) at io.netty.handler.codec.http2.Http2ConnectionHandler$PrefaceDecoder.decode(Http2ConnectionHandler.java:254) at io.netty.handler.codec.http2.Http2ConnectionHandler.decode(Http2ConnectionHandler.java:450) at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:502) at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:441) at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:278) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.handler.codec.ByteToMessageDecoder.handlerRemoved(ByteToMessageDecoder.java:249) at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:507) at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:441) at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:278) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:241) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.handler.flush.FlushConsolidationHandler.channelRead(FlushConsolidationHandler.java:154) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:286) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1408) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:930) at io.netty.channel.epoll.AbstractEpollStreamChannel$EpollStreamUnsafe.epollInReady(AbstractEpollStreamChannel.java:799) at io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:427) at io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:328) at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:905) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.lang.Thread.run(Thread.java:748) ```
defect
nullpointerexception hi we got a nullpointerexception because of the authority header missing and please refer to the stack trace message below the method fillschemeandauthorityifmissing of class httpserverhandler creating a new requestheaders instance instead of filling the existing one and this cause the nullpointerexception java warn c l a s t tomcatservice failed to invoke tomcat java lang nullpointerexception null at com linecorp armeria server tomcat tomcatservice convertrequest tomcatservice java at com linecorp armeria server tomcat tomcatservice lambda serve tomcatservice java at java util concurrent completablefuture unihandle completablefuture java at java util concurrent completablefuture unihandlestage completablefuture java at java util concurrent completablefuture handle completablefuture java at com linecorp armeria server tomcat tomcatservice serve tomcatservice java at com linecorp armeria server tomcat tomcatservice serve tomcatservice java at com linecorp armeria server httpserverhandler handlerequest httpserverhandler java at com linecorp armeria server httpserverhandler channelread httpserverhandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at com linecorp armeria server onheadersread java at com linecorp armeria server onheadersread java at io netty handler codec framereadlistener onheadersread java at io netty handler codec framereadlistener onheadersread java at io netty handler codec prefaceframelistener onheadersread java at io netty handler codec processfragment java at io netty handler codec readheadersframe java at io netty handler codec processpayloadstate java at io netty handler codec readframe java at io netty handler codec decodeframe java at io netty handler codec framedecoder decode java at io netty handler codec prefacedecoder decode java at io netty handler codec decode java at io netty handler codec bytetomessagedecoder decoderemovalreentryprotection bytetomessagedecoder java at io netty handler codec bytetomessagedecoder calldecode bytetomessagedecoder java at io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler codec bytetomessagedecoder handlerremoved bytetomessagedecoder java at io netty handler codec bytetomessagedecoder decoderemovalreentryprotection bytetomessagedecoder java at io netty handler codec bytetomessagedecoder calldecode bytetomessagedecoder java at io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler logging logginghandler channelread logginghandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler flush flushconsolidationhandler channelread flushconsolidationhandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler timeout idlestatehandler channelread idlestatehandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty channel defaultchannelpipeline headcontext channelread defaultchannelpipeline java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel defaultchannelpipeline firechannelread defaultchannelpipeline java at io netty channel epoll abstractepollstreamchannel epollstreamunsafe epollinready abstractepollstreamchannel java at io netty channel epoll epolleventloop processready epolleventloop java at io netty channel epoll epolleventloop run epolleventloop java at io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java at io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java lang thread run thread java
1
18,885
4,320,166,467
IssuesEvent
2016-07-25 02:38:31
MarlinFirmware/Marlin
https://api.github.com/repos/MarlinFirmware/Marlin
reopened
Unclear how to use mesh bed leveling
Support: Documentation Support: Homing & Leveling
I have been trying to get Mesh Bed Leveling working since 1.1RC6. There have been a few issues opened that appear to have been fixed. It seems that MBL has evolved since some of the initial articles appeared describing it early this year (March). I am confused how to use it with RCBugFix latest. I grabbed RCBugFix today (7/24) and built it using more or less defaults for MBL. My confusion involves the Z-endstop and the settings in Configuration.h. I have a mechanical Z-endstop that I have traditionally used to set the initial Z distance of "0" after homing. This worked reasonably well for a flat and level print bed. My bed is not perfect so I want to try MBL to compensate. I got it to work once randomly and it was great. I could see the Z axis moving slightly as X and Y moved. <details> <summary>Given these MBL settings in Configuration.h:</summary> ```cpp //=========================================================================== //============================ Mesh Bed Leveling ============================ //=========================================================================== #define MESH_BED_LEVELING // Enable mesh bed leveling. #if ENABLED(MESH_BED_LEVELING) #define MESH_INSET 10 // Mesh inset margin on print area #define MESH_NUM_X_POINTS 3 // Don't use more than 7 points per axis, implementation limited. #define MESH_NUM_Y_POINTS 3 #define MESH_HOME_SEARCH_Z 4 // Z after Home, bed somewhere below but above 0.0. //#define MESH_G28_REST_ORIGIN // After homing all axes ('G28' or 'G28 XYZ') rest at origin [0,0,0] #define MANUAL_BED_LEVELING // Add display menu option for bed leveling. #if ENABLED(MANUAL_BED_LEVELING) #define MBL_Z_STEP 0.025 // Step size while manually probing Z axis. #endif // MANUAL_BED_LEVELING #endif // MESH_BED_LEVELING ``` </details> Questions: 1) How should I set my Z-endstop for Home? Should it be set so that the nozzle is 4mm above the bed? 2) If the answer is 4mm above, how do I accommodate driving approximately 4mm below the Z endstop? It will just "compress" my Z endstop switch unmercifully. Not complaining here - I think MBL will really be helpful. I am just wondering how to actually use it with my setup. Thanks for the neat feature. Mark
1.0
Unclear how to use mesh bed leveling - I have been trying to get Mesh Bed Leveling working since 1.1RC6. There have been a few issues opened that appear to have been fixed. It seems that MBL has evolved since some of the initial articles appeared describing it early this year (March). I am confused how to use it with RCBugFix latest. I grabbed RCBugFix today (7/24) and built it using more or less defaults for MBL. My confusion involves the Z-endstop and the settings in Configuration.h. I have a mechanical Z-endstop that I have traditionally used to set the initial Z distance of "0" after homing. This worked reasonably well for a flat and level print bed. My bed is not perfect so I want to try MBL to compensate. I got it to work once randomly and it was great. I could see the Z axis moving slightly as X and Y moved. <details> <summary>Given these MBL settings in Configuration.h:</summary> ```cpp //=========================================================================== //============================ Mesh Bed Leveling ============================ //=========================================================================== #define MESH_BED_LEVELING // Enable mesh bed leveling. #if ENABLED(MESH_BED_LEVELING) #define MESH_INSET 10 // Mesh inset margin on print area #define MESH_NUM_X_POINTS 3 // Don't use more than 7 points per axis, implementation limited. #define MESH_NUM_Y_POINTS 3 #define MESH_HOME_SEARCH_Z 4 // Z after Home, bed somewhere below but above 0.0. //#define MESH_G28_REST_ORIGIN // After homing all axes ('G28' or 'G28 XYZ') rest at origin [0,0,0] #define MANUAL_BED_LEVELING // Add display menu option for bed leveling. #if ENABLED(MANUAL_BED_LEVELING) #define MBL_Z_STEP 0.025 // Step size while manually probing Z axis. #endif // MANUAL_BED_LEVELING #endif // MESH_BED_LEVELING ``` </details> Questions: 1) How should I set my Z-endstop for Home? Should it be set so that the nozzle is 4mm above the bed? 2) If the answer is 4mm above, how do I accommodate driving approximately 4mm below the Z endstop? It will just "compress" my Z endstop switch unmercifully. Not complaining here - I think MBL will really be helpful. I am just wondering how to actually use it with my setup. Thanks for the neat feature. Mark
non_defect
unclear how to use mesh bed leveling i have been trying to get mesh bed leveling working since there have been a few issues opened that appear to have been fixed it seems that mbl has evolved since some of the initial articles appeared describing it early this year march i am confused how to use it with rcbugfix latest i grabbed rcbugfix today and built it using more or less defaults for mbl my confusion involves the z endstop and the settings in configuration h i have a mechanical z endstop that i have traditionally used to set the initial z distance of after homing this worked reasonably well for a flat and level print bed my bed is not perfect so i want to try mbl to compensate i got it to work once randomly and it was great i could see the z axis moving slightly as x and y moved given these mbl settings in configuration h cpp mesh bed leveling define mesh bed leveling enable mesh bed leveling if enabled mesh bed leveling define mesh inset mesh inset margin on print area define mesh num x points don t use more than points per axis implementation limited define mesh num y points define mesh home search z z after home bed somewhere below but above define mesh rest origin after homing all axes or xyz rest at origin define manual bed leveling add display menu option for bed leveling if enabled manual bed leveling define mbl z step step size while manually probing z axis endif manual bed leveling endif mesh bed leveling questions how should i set my z endstop for home should it be set so that the nozzle is above the bed if the answer is above how do i accommodate driving approximately below the z endstop it will just compress my z endstop switch unmercifully not complaining here i think mbl will really be helpful i am just wondering how to actually use it with my setup thanks for the neat feature mark
0
35,692
9,645,533,883
IssuesEvent
2019-05-17 08:54:50
yandex/ClickHouse
https://api.github.com/repos/yandex/ClickHouse
opened
build failed, by clang 8.0.0
build
hi, Thank you for your work. I compiled in stable version (19.5.3.8 stable) wrong. first, Download the code from LLVM and copy the code from the driver directory to compiler_8.0.0 and compiler_8.0.0svn. second, The build code. failed message: CMake Error at dbms/programs/clang/Compiler-8.0.0svn/CMakeLists.txt:32 (add_clang_tool): Unknown CMake command "add_clang_tool".
1.0
build failed, by clang 8.0.0 - hi, Thank you for your work. I compiled in stable version (19.5.3.8 stable) wrong. first, Download the code from LLVM and copy the code from the driver directory to compiler_8.0.0 and compiler_8.0.0svn. second, The build code. failed message: CMake Error at dbms/programs/clang/Compiler-8.0.0svn/CMakeLists.txt:32 (add_clang_tool): Unknown CMake command "add_clang_tool".
non_defect
build failed by clang hi thank you for your work i compiled in stable version stable wrong first download the code from llvm and copy the code from the driver directory to compiler and compiler second the build code failed message cmake error at dbms programs clang compiler cmakelists txt add clang tool unknown cmake command add clang tool
0
10,862
2,622,205,074
IssuesEvent
2015-03-04 00:29:53
clc/emacspeak
https://api.github.com/repos/clc/emacspeak
opened
Google Apps For Your Domain Support for g-client
auto-migrated Priority-Medium Type-Defect
``` Currently g-client hardcodes addresses for calendar and such so it won't work with Google Apps For Your Domain setups. It would be nice if such setups could be supported. ``` Original issue reported on code.google.com by `ismail.d...@gmail.com` on 1 Aug 2008 at 7:19
1.0
Google Apps For Your Domain Support for g-client - ``` Currently g-client hardcodes addresses for calendar and such so it won't work with Google Apps For Your Domain setups. It would be nice if such setups could be supported. ``` Original issue reported on code.google.com by `ismail.d...@gmail.com` on 1 Aug 2008 at 7:19
defect
google apps for your domain support for g client currently g client hardcodes addresses for calendar and such so it won t work with google apps for your domain setups it would be nice if such setups could be supported original issue reported on code google com by ismail d gmail com on aug at
1
633,626
20,260,696,348
IssuesEvent
2022-02-15 07:01:17
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
Completions not working inside mapping constructor in new expression parameter
Type/Bug Priority/High Team/LanguageServer Points/2 Area/Completion
**Description:** Consider the following code: ```ballerina import ballerina/http; service / on new http:Listener(8080) { private MyClient myClient; public function init() returns error? { self.myClient = check new({<cursor>}) } } type Config record {| int a; |}; client class MyClient { function init(Config config) returns error? { } } ``` At `<cursor>`, no completions are provided for the matching record fields. This happens only when initializing `client class` type variables. If `MyClient` was just a class, this works. **Steps to reproduce:** See description **Affected Versions:** Swan Lake 2201 RC2
1.0
Completions not working inside mapping constructor in new expression parameter - **Description:** Consider the following code: ```ballerina import ballerina/http; service / on new http:Listener(8080) { private MyClient myClient; public function init() returns error? { self.myClient = check new({<cursor>}) } } type Config record {| int a; |}; client class MyClient { function init(Config config) returns error? { } } ``` At `<cursor>`, no completions are provided for the matching record fields. This happens only when initializing `client class` type variables. If `MyClient` was just a class, this works. **Steps to reproduce:** See description **Affected Versions:** Swan Lake 2201 RC2
non_defect
completions not working inside mapping constructor in new expression parameter description consider the following code ballerina import ballerina http service on new http listener private myclient myclient public function init returns error self myclient check new type config record int a client class myclient function init config config returns error at no completions are provided for the matching record fields this happens only when initializing client class type variables if myclient was just a class this works steps to reproduce see description affected versions swan lake
0
30,373
6,123,049,574
IssuesEvent
2017-06-23 02:39:44
line/armeria
https://api.github.com/repos/line/armeria
opened
Flaky test: HttpServerTest.testTooLargeContent
defect
``` com.linecorp.armeria.server.http.HttpServerTest > testTooLargeContent[0: h1c] FAILED java.util.concurrent.ExecutionException: com.linecorp.armeria.common.ClosedSessionException at java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357) at java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1895) at com.linecorp.armeria.server.http.HttpServerTest.testTooLargeContent(HttpServerTest.java:516) Caused by: com.linecorp.armeria.common.ClosedSessionException ``` https://ci.appveyor.com/project/trustin/armeria/build/master.591
1.0
Flaky test: HttpServerTest.testTooLargeContent - ``` com.linecorp.armeria.server.http.HttpServerTest > testTooLargeContent[0: h1c] FAILED java.util.concurrent.ExecutionException: com.linecorp.armeria.common.ClosedSessionException at java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357) at java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1895) at com.linecorp.armeria.server.http.HttpServerTest.testTooLargeContent(HttpServerTest.java:516) Caused by: com.linecorp.armeria.common.ClosedSessionException ``` https://ci.appveyor.com/project/trustin/armeria/build/master.591
defect
flaky test httpservertest testtoolargecontent com linecorp armeria server http httpservertest testtoolargecontent failed java util concurrent executionexception com linecorp armeria common closedsessionexception at java util concurrent completablefuture reportget completablefuture java at java util concurrent completablefuture get completablefuture java at com linecorp armeria server http httpservertest testtoolargecontent httpservertest java caused by com linecorp armeria common closedsessionexception
1
762,061
26,707,297,423
IssuesEvent
2023-01-27 19:26:25
molgenis/molgenis-emx2
https://api.github.com/repos/molgenis/molgenis-emx2
opened
Variable explorer is broken
bug priority broken
**What version of EMX2 are you using (see footer)** 8.136.0 **Describe the bug** Variable explorer is not loading variables **To Reproduce** Steps to reproduce the behavior: 1. Go to ['https://emx2.test.molgenis.org'](https://data-catalogue-staging.molgeniscloud.org/DataCatalogue/catalogue/#/variable-explorer/) 2. See that no variables are loaded **Expected behavior** A clear and concise description of what you expected to happen. **Screenshots** If applicable, add screenshots to help explain your problem. **Desktop (please complete the following information):** - OS: [e.g. iOS] - Browser [e.g. chrome, safari] - Version [e.g. 22] **Smartphone (please complete the following information):** - Device: [e.g. iPhone6] - OS: [e.g. iOS8.1] - Browser [e.g. stock browser, safari] - Version [e.g. 22] **Additional context** Add any other context about the problem here.
1.0
Variable explorer is broken - **What version of EMX2 are you using (see footer)** 8.136.0 **Describe the bug** Variable explorer is not loading variables **To Reproduce** Steps to reproduce the behavior: 1. Go to ['https://emx2.test.molgenis.org'](https://data-catalogue-staging.molgeniscloud.org/DataCatalogue/catalogue/#/variable-explorer/) 2. See that no variables are loaded **Expected behavior** A clear and concise description of what you expected to happen. **Screenshots** If applicable, add screenshots to help explain your problem. **Desktop (please complete the following information):** - OS: [e.g. iOS] - Browser [e.g. chrome, safari] - Version [e.g. 22] **Smartphone (please complete the following information):** - Device: [e.g. iPhone6] - OS: [e.g. iOS8.1] - Browser [e.g. stock browser, safari] - Version [e.g. 22] **Additional context** Add any other context about the problem here.
non_defect
variable explorer is broken what version of are you using see footer describe the bug variable explorer is not loading variables to reproduce steps to reproduce the behavior go to see that no variables are loaded expected behavior a clear and concise description of what you expected to happen screenshots if applicable add screenshots to help explain your problem desktop please complete the following information os browser version smartphone please complete the following information device os browser version additional context add any other context about the problem here
0
23,027
3,754,911,340
IssuesEvent
2016-03-12 08:55:12
openwrt/luci
https://api.github.com/repos/openwrt/luci
closed
Hang when saveing changes on Mount Points
C: LuCI Applications P: major T: defect
**Reported by reporter on 3 May 2010 06:24 UTC** Hello, First of all some information: Luci version: 0.9.0 OpenWRT version: 10.03 Hardware model: TPLINK WR1043ND Bug: When I change some records on Mount Point screen and then clik Save or Save & Apply Luci hangs. I could kill uhttpd or reboot the router but there is no thing to do from Luci GUI. When I try to see what is going on from console (by "ps" command") I can see that there is one process which point to luci and fstab changes and I think this is the reason of Luci hangs. BTW: Luc hangs but all information to fstab are saved correct... I will be very grateful if some one could help. Best regards, MivncM
1.0
Hang when saveing changes on Mount Points - **Reported by reporter on 3 May 2010 06:24 UTC** Hello, First of all some information: Luci version: 0.9.0 OpenWRT version: 10.03 Hardware model: TPLINK WR1043ND Bug: When I change some records on Mount Point screen and then clik Save or Save & Apply Luci hangs. I could kill uhttpd or reboot the router but there is no thing to do from Luci GUI. When I try to see what is going on from console (by "ps" command") I can see that there is one process which point to luci and fstab changes and I think this is the reason of Luci hangs. BTW: Luc hangs but all information to fstab are saved correct... I will be very grateful if some one could help. Best regards, MivncM
defect
hang when saveing changes on mount points reported by reporter on may utc hello first of all some information luci version openwrt version hardware model tplink bug when i change some records on mount point screen and then clik save or save apply luci hangs i could kill uhttpd or reboot the router but there is no thing to do from luci gui when i try to see what is going on from console by ps command i can see that there is one process which point to luci and fstab changes and i think this is the reason of luci hangs btw luc hangs but all information to fstab are saved correct i will be very grateful if some one could help best regards mivncm
1
34,558
9,411,962,855
IssuesEvent
2019-04-10 01:53:46
opencv/opencv
https://api.github.com/repos/opencv/opencv
closed
Error compiling 4.1.0 source code with mingw-w64: 'D3D11_TEXTURE2D_DESC' was not declared in this scope
category: build/install
##### System information (version) - OpenCV => 4.1.0 - Operating System / Platform => Windows 10 64bit - Compiler => mingw-w64 8.1.0-posix-seh-rt_v6-rev0 ##### Detailed description When I compiled the 4.1.0 version with mingw-w64, there was an error ``` E:\opencv-4.1.0\opencv-4.1.0\modules\core\src\directx.cpp:1035:5: error: 'D3D11_TEXTURE2D_DESC' was not declared in this scope D3D11_TEXTURE2D_DESC desc = { 0 }; ^~~~~~~~~~~~~~~~~~~~ ``` I can solve it like this ``` cmake -DWITH_OPENCL_D3D11_NV=OFF ``` But I don't know if this is the right solution, and whether this will affect the compiled dll.
1.0
Error compiling 4.1.0 source code with mingw-w64: 'D3D11_TEXTURE2D_DESC' was not declared in this scope - ##### System information (version) - OpenCV => 4.1.0 - Operating System / Platform => Windows 10 64bit - Compiler => mingw-w64 8.1.0-posix-seh-rt_v6-rev0 ##### Detailed description When I compiled the 4.1.0 version with mingw-w64, there was an error ``` E:\opencv-4.1.0\opencv-4.1.0\modules\core\src\directx.cpp:1035:5: error: 'D3D11_TEXTURE2D_DESC' was not declared in this scope D3D11_TEXTURE2D_DESC desc = { 0 }; ^~~~~~~~~~~~~~~~~~~~ ``` I can solve it like this ``` cmake -DWITH_OPENCL_D3D11_NV=OFF ``` But I don't know if this is the right solution, and whether this will affect the compiled dll.
non_defect
error compiling source code with mingw desc was not declared in this scope system information version opencv operating system platform windows compiler mingw posix seh rt detailed description when i compiled the version with mingw there was an error e opencv opencv modules core src directx cpp error desc was not declared in this scope desc desc i can solve it like this cmake dwith opencl nv off but i don t know if this is the right solution and whether this will affect the compiled dll
0
45,672
5,950,472,724
IssuesEvent
2017-05-26 16:49:30
dart-lang/site-webdev
https://api.github.com/repos/dart-lang/site-webdev
closed
[ng] don't wrap code
Design
This seems to be inherited from angular.io, but the code excerpt lines are wrapped. E.g., > ![screen shot 2017-05-20 at 11 43 45 am](https://cloud.githubusercontent.com/assets/4140793/26277052/53b63b6a-3d52-11e7-96ca-1e58dbbc52ac.png) But, on narrow displays this generally reads much better when not wrapped: > ![screen shot 2017-05-20 at 11 43 31 am](https://cloud.githubusercontent.com/assets/4140793/26277056/6ac4251a-3d52-11e7-9625-8edc04edde4b.png) In fact, this is what is done on dartlang.org, e.g.: > ![screen shot 2017-05-20 at 11 50 33 am](https://cloud.githubusercontent.com/assets/4140793/26277062/8e857e2c-3d52-11e7-8dc8-7c62d74d115c.png) @kwalrath do you agree?
1.0
[ng] don't wrap code - This seems to be inherited from angular.io, but the code excerpt lines are wrapped. E.g., > ![screen shot 2017-05-20 at 11 43 45 am](https://cloud.githubusercontent.com/assets/4140793/26277052/53b63b6a-3d52-11e7-96ca-1e58dbbc52ac.png) But, on narrow displays this generally reads much better when not wrapped: > ![screen shot 2017-05-20 at 11 43 31 am](https://cloud.githubusercontent.com/assets/4140793/26277056/6ac4251a-3d52-11e7-9625-8edc04edde4b.png) In fact, this is what is done on dartlang.org, e.g.: > ![screen shot 2017-05-20 at 11 50 33 am](https://cloud.githubusercontent.com/assets/4140793/26277062/8e857e2c-3d52-11e7-8dc8-7c62d74d115c.png) @kwalrath do you agree?
non_defect
don t wrap code this seems to be inherited from angular io but the code excerpt lines are wrapped e g but on narrow displays this generally reads much better when not wrapped in fact this is what is done on dartlang org e g kwalrath do you agree
0
44,315
12,101,445,041
IssuesEvent
2020-04-20 15:13:27
codesmithtools/Templates
https://api.github.com/repos/codesmithtools/Templates
closed
Root object Transactional attribute - enhancement to allow users to control transactions in partial methods.
Framework-CSLA Type-Defect auto-migrated
``` What steps will reproduce the problem? 1. If you generate a root object, the dataportal methods get decorated with the <Transactional(TransactionalTypes.TransactionScope)> attribute. 2. If you want to override the root data portal methods with different transactional requirements you can't, via the partial methods. What is the expected output? 1. Remove the <Transactional(TransactionalTypes.TransactionScope)> from the DP_xyz method. 2. Add a "Using scope As New Transactions.TransactionScope" to the generated code. Example: ' Blake this would allow the users to overwrite transactional requirements if need by using the partial methods. (just a little more flexible). Using scope As New Transactions.TransactionScope Using connection As New SqlConnection(ADOHelper.ConnectionString) connection.Open() .... End Using ' Connection End Using ' TransactionScope What version of the product are you using? v3.0.0.1817 ``` Original issue reported on code.google.com by `JenasysD...@gmail.com` on 12 Aug 2010 at 1:27
1.0
Root object Transactional attribute - enhancement to allow users to control transactions in partial methods. - ``` What steps will reproduce the problem? 1. If you generate a root object, the dataportal methods get decorated with the <Transactional(TransactionalTypes.TransactionScope)> attribute. 2. If you want to override the root data portal methods with different transactional requirements you can't, via the partial methods. What is the expected output? 1. Remove the <Transactional(TransactionalTypes.TransactionScope)> from the DP_xyz method. 2. Add a "Using scope As New Transactions.TransactionScope" to the generated code. Example: ' Blake this would allow the users to overwrite transactional requirements if need by using the partial methods. (just a little more flexible). Using scope As New Transactions.TransactionScope Using connection As New SqlConnection(ADOHelper.ConnectionString) connection.Open() .... End Using ' Connection End Using ' TransactionScope What version of the product are you using? v3.0.0.1817 ``` Original issue reported on code.google.com by `JenasysD...@gmail.com` on 12 Aug 2010 at 1:27
defect
root object transactional attribute enhancement to allow users to control transactions in partial methods what steps will reproduce the problem if you generate a root object the dataportal methods get decorated with the attribute if you want to override the root data portal methods with different transactional requirements you can t via the partial methods what is the expected output remove the from the dp xyz method add a using scope as new transactions transactionscope to the generated code example blake this would allow the users to overwrite transactional requirements if need by using the partial methods just a little more flexible using scope as new transactions transactionscope using connection as new sqlconnection adohelper connectionstring connection open end using connection end using transactionscope what version of the product are you using original issue reported on code google com by jenasysd gmail com on aug at
1
175
2,517,932,024
IssuesEvent
2015-01-16 18:15:14
TamarackConsulting/Creekridge_Portal
https://api.github.com/repos/TamarackConsulting/Creekridge_Portal
closed
Doc gen taking a long time to generate docs
Major Defect
Issue ---------- Sometimes it can take over 15 seconds for a credit app or proposal letter document to generate. User doesn't know if they are waiting or if it didn't work. Expected --------------- Improve the performance/speed of generating the documents if possible.
1.0
Doc gen taking a long time to generate docs - Issue ---------- Sometimes it can take over 15 seconds for a credit app or proposal letter document to generate. User doesn't know if they are waiting or if it didn't work. Expected --------------- Improve the performance/speed of generating the documents if possible.
defect
doc gen taking a long time to generate docs issue sometimes it can take over seconds for a credit app or proposal letter document to generate user doesn t know if they are waiting or if it didn t work expected improve the performance speed of generating the documents if possible
1
49,983
26,418,915,548
IssuesEvent
2023-01-13 18:22:36
matrix-org/sliding-sync
https://api.github.com/repos/matrix-org/sliding-sync
closed
Initial sync is worryingly slow & big to get to clients.
performance
While chasing down https://github.com/matrix-org/matrix-rust-sdk/issues/1308 i was seeing initial sync reqs taking 36s to arrive. This was in part exasperated by the req having a 10s timeout on it, but even after fixing that, things are taking a worryingly long time - especially after ~12h of being offline. Here's one i just profiled taking ~5s to execute (after ~2h of being offline). Client saw: ``` 2022-12-28T00:21:59.635174Z TRACE send_inner{request=Request { pos: None, txn_id: None, timeout: Some(30s), lists: [SyncRequestList { slow_get_all_rooms: false, ranges: [(0, 20)], sort: ["by_recency", "by_name"], required_state: [(RoomAvatar, ""), (RoomEncryption, "")], timeline_limit: Some(10), filters: None }], room_subscriptions: {}, unsubscribe_rooms: [], extensions: ExtensionsConfig { to_device: Some(ToDeviceConfig { enabled: Some(true), limit: None, since: None }), e2ee: Some(E2EEConfig { enabled: Some(true) }), account_data: Some(AccountDataConfig { enabled: Some(true) }), other: {} } } config=Some(RequestConfig { timeout: 60s, retry_limit: None, retry_timeout: None }) homeserver=Some("https://slidingsync.lab.element.dev/")}:send{config=Some(RequestConfig { timeout: 60s, retry_limit: None, retry_timeout: None }) homeserver="https://slidingsync.lab.element.dev/" user_id=Some("@matthew:matrix.org") server_versions=[V1_0, V1_1, V1_2, V1_3, V1_4, V1_5] request_type="ruma_client_api::sync::sync_events::v4::Request"}: matrix_sdk::http_client: Sending request 2022-12-28T00:22:04.412702Z TRACE send_inner{request=Request { pos: None, txn_id: None, timeout: Some(30s), lists: [SyncRequestList { slow_get_all_rooms: false, ranges: [(0, 20)], sort: ["by_recency", "by_name"], required_state: [(RoomAvatar, ""), (RoomEncryption, "")], timeline_limit: Some(10), filters: None }], room_subscriptions: {}, unsubscribe_rooms: [], extensions: ExtensionsConfig { to_device: Some(ToDeviceConfig { enabled: Some(true), limit: None, since: None }), e2ee: Some(E2EEConfig { enabled: Some(true) }), account_data: Some(AccountDataConfig { enabled: Some(true) }), other: {} } } config=Some(RequestConfig { timeout: 60s, retry_limit: None, retry_timeout: None }) homeserver=Some("https://slidingsync.lab.element.dev/")}:send{config=Some(RequestConfig { timeout: 60s, retry_limit: None, retry_timeout: None }) homeserver="https://slidingsync.lab.element.dev/" user_id=Some("@matthew:matrix.org") server_versions=[V1_0, V1_1, V1_2, V1_3, V1_4, V1_5] request_type="ruma_client_api::sync::sync_events::v4::Request"}: matrix_sdk::http_client: Got response: Response { status: 200, version: HTTP/1.1, headers: {"access-control-allow-headers": "Origin, X-Requested-With, Content-Type, Accept, Authorization", "access-control-allow-methods": "GET, POST, PUT, DELETE, OPTIONS", "access-control-allow-origin": "*", "content-type": "application/json", "date": "Wed, 28 Dec 2022 00:22:01 GMT", "transfer-encoding": "chunked", "set-cookie": "......; path=/; HttpOnly; Secure; SameSite=None"}, body: b"{\"lists\":[{\"ops\":[{\"op\":\"SYNC\",\"range\":[0,20], ``` Server saw: ``` 00:22:00 INF created new connection conn_id=ec725749c89702fe5e1fc5e019e77ce22dc0f3ae0717483f307a02e2222c6631 user=@matthew:matrix.org 00:22:01 WRN Client did not increment since token: possibly sending back duplicate to-device events! device=ec725749c89702fe5e1fc5e019e77ce22dc0f3ae0717483f307a02e2222c6631 initial=true last_sent=6950241 recv=0 user=@matthew:matrix.org 00:22:01 INF ag=25 d=1 duration=1111.414264 p=0 q=1 r=21 size=296993 status=200 u=@matthew:matrix.org 00:22:04 INF dl-c=71 dl-l=8 duration=32.985866 p=1 q=2 r=0 size=2200 status=200 u=@matthew:matrix.org 00:22:05 INF duration=44.822427 p=2 q=3 r=0 size=327 status=200 u=@matthew:matrix.org ``` Server pprof saw (for the OnIncomingRequestInitial): ``` When Elapsed Goroutine ID Events 4.451185334s 1.105535526s Task 1964 (goroutine view) (complete) 4.451185334 . 39366 task OnIncomingRequestInitial (id 1964, parent 0) created 4.451192886 . 7552 39366 region load started (duration: 42.257935ms) 4.493469125 . 42276239 39366 region onIncomingListRequest started (duration: 32.959255ms) 4.526486083 . 33016958 39366 region buildRooms started (duration: 1.000488985s) 4.558728129 . 32242046 39366 new goroutine 39367: database/sql.(*DB).beginDC·dwrap·17 4.560168230 . 1440101 39366 new goroutine 39395: database/sql.(*Rows).initContextClose·dwrap·27 4.562780862 . 2612632 39366 region RoomStateAfterEventPosition started (duration: 941.937482ms) 4.563280573 . 499711 39366 new goroutine 39294: database/sql.(*DB).beginDC·dwrap·17 4.568830597 . 5550024 39366 new goroutine 39368: database/sql.(*Rows).initContextClose·dwrap·27 4.570190699 . 1360102 39366 new goroutine 39369: database/sql.(*Rows).initContextClose·dwrap·27 5.527456348 .957265649 39366 region extensions started (duration: 29.246106ms) 5.549077859 . 21621511 39366 new goroutine 39295: database/sql.(*DB).beginDC·dwrap·17 5.551339387 . 2261528 39366 new goroutine 39296: database/sql.(*DB).beginDC·dwrap·17 5.552959072 . 1619685 39366 new goroutine 39396: database/sql.(*Rows).initContextClose·dwrap·27 5.553884664 . 925592 39366 new goroutine 39297: database/sql.(*DB).beginDC·dwrap·17 5.555373534 . 1488870 39366 new goroutine 39373: database/sql.(*Rows).initContextClose·dwrap·27 5.556710390 . 1336856 39366 region liveUpdate started (duration: 6.246µs) 5.556720860 . 10470 39366 task end ``` So: if it took 1111ms to generate on the server (which is already quite a lot; the 300KB request size is *way* too large thanks to including 130KB of redundant account_data on it, and 20*20 = 400 events of scrollback), how come it took the client another 3s to receive it? I guess this could be slow internet connectivity whilst travelling for xmas, but empirically i seem to be able to download stuff at ~4MB/s. To address the size, i think account_data has to be incremental, as per https://github.com/matrix-org/matrix-spec-proposals/pull/3575/files#r1057933023 - and we should always avoid syncing in scrollback in rooms during initial sync. However, it doesn't explain whey the request took 1s to execute but blocked the client for 5s...
True
Initial sync is worryingly slow & big to get to clients. - While chasing down https://github.com/matrix-org/matrix-rust-sdk/issues/1308 i was seeing initial sync reqs taking 36s to arrive. This was in part exasperated by the req having a 10s timeout on it, but even after fixing that, things are taking a worryingly long time - especially after ~12h of being offline. Here's one i just profiled taking ~5s to execute (after ~2h of being offline). Client saw: ``` 2022-12-28T00:21:59.635174Z TRACE send_inner{request=Request { pos: None, txn_id: None, timeout: Some(30s), lists: [SyncRequestList { slow_get_all_rooms: false, ranges: [(0, 20)], sort: ["by_recency", "by_name"], required_state: [(RoomAvatar, ""), (RoomEncryption, "")], timeline_limit: Some(10), filters: None }], room_subscriptions: {}, unsubscribe_rooms: [], extensions: ExtensionsConfig { to_device: Some(ToDeviceConfig { enabled: Some(true), limit: None, since: None }), e2ee: Some(E2EEConfig { enabled: Some(true) }), account_data: Some(AccountDataConfig { enabled: Some(true) }), other: {} } } config=Some(RequestConfig { timeout: 60s, retry_limit: None, retry_timeout: None }) homeserver=Some("https://slidingsync.lab.element.dev/")}:send{config=Some(RequestConfig { timeout: 60s, retry_limit: None, retry_timeout: None }) homeserver="https://slidingsync.lab.element.dev/" user_id=Some("@matthew:matrix.org") server_versions=[V1_0, V1_1, V1_2, V1_3, V1_4, V1_5] request_type="ruma_client_api::sync::sync_events::v4::Request"}: matrix_sdk::http_client: Sending request 2022-12-28T00:22:04.412702Z TRACE send_inner{request=Request { pos: None, txn_id: None, timeout: Some(30s), lists: [SyncRequestList { slow_get_all_rooms: false, ranges: [(0, 20)], sort: ["by_recency", "by_name"], required_state: [(RoomAvatar, ""), (RoomEncryption, "")], timeline_limit: Some(10), filters: None }], room_subscriptions: {}, unsubscribe_rooms: [], extensions: ExtensionsConfig { to_device: Some(ToDeviceConfig { enabled: Some(true), limit: None, since: None }), e2ee: Some(E2EEConfig { enabled: Some(true) }), account_data: Some(AccountDataConfig { enabled: Some(true) }), other: {} } } config=Some(RequestConfig { timeout: 60s, retry_limit: None, retry_timeout: None }) homeserver=Some("https://slidingsync.lab.element.dev/")}:send{config=Some(RequestConfig { timeout: 60s, retry_limit: None, retry_timeout: None }) homeserver="https://slidingsync.lab.element.dev/" user_id=Some("@matthew:matrix.org") server_versions=[V1_0, V1_1, V1_2, V1_3, V1_4, V1_5] request_type="ruma_client_api::sync::sync_events::v4::Request"}: matrix_sdk::http_client: Got response: Response { status: 200, version: HTTP/1.1, headers: {"access-control-allow-headers": "Origin, X-Requested-With, Content-Type, Accept, Authorization", "access-control-allow-methods": "GET, POST, PUT, DELETE, OPTIONS", "access-control-allow-origin": "*", "content-type": "application/json", "date": "Wed, 28 Dec 2022 00:22:01 GMT", "transfer-encoding": "chunked", "set-cookie": "......; path=/; HttpOnly; Secure; SameSite=None"}, body: b"{\"lists\":[{\"ops\":[{\"op\":\"SYNC\",\"range\":[0,20], ``` Server saw: ``` 00:22:00 INF created new connection conn_id=ec725749c89702fe5e1fc5e019e77ce22dc0f3ae0717483f307a02e2222c6631 user=@matthew:matrix.org 00:22:01 WRN Client did not increment since token: possibly sending back duplicate to-device events! device=ec725749c89702fe5e1fc5e019e77ce22dc0f3ae0717483f307a02e2222c6631 initial=true last_sent=6950241 recv=0 user=@matthew:matrix.org 00:22:01 INF ag=25 d=1 duration=1111.414264 p=0 q=1 r=21 size=296993 status=200 u=@matthew:matrix.org 00:22:04 INF dl-c=71 dl-l=8 duration=32.985866 p=1 q=2 r=0 size=2200 status=200 u=@matthew:matrix.org 00:22:05 INF duration=44.822427 p=2 q=3 r=0 size=327 status=200 u=@matthew:matrix.org ``` Server pprof saw (for the OnIncomingRequestInitial): ``` When Elapsed Goroutine ID Events 4.451185334s 1.105535526s Task 1964 (goroutine view) (complete) 4.451185334 . 39366 task OnIncomingRequestInitial (id 1964, parent 0) created 4.451192886 . 7552 39366 region load started (duration: 42.257935ms) 4.493469125 . 42276239 39366 region onIncomingListRequest started (duration: 32.959255ms) 4.526486083 . 33016958 39366 region buildRooms started (duration: 1.000488985s) 4.558728129 . 32242046 39366 new goroutine 39367: database/sql.(*DB).beginDC·dwrap·17 4.560168230 . 1440101 39366 new goroutine 39395: database/sql.(*Rows).initContextClose·dwrap·27 4.562780862 . 2612632 39366 region RoomStateAfterEventPosition started (duration: 941.937482ms) 4.563280573 . 499711 39366 new goroutine 39294: database/sql.(*DB).beginDC·dwrap·17 4.568830597 . 5550024 39366 new goroutine 39368: database/sql.(*Rows).initContextClose·dwrap·27 4.570190699 . 1360102 39366 new goroutine 39369: database/sql.(*Rows).initContextClose·dwrap·27 5.527456348 .957265649 39366 region extensions started (duration: 29.246106ms) 5.549077859 . 21621511 39366 new goroutine 39295: database/sql.(*DB).beginDC·dwrap·17 5.551339387 . 2261528 39366 new goroutine 39296: database/sql.(*DB).beginDC·dwrap·17 5.552959072 . 1619685 39366 new goroutine 39396: database/sql.(*Rows).initContextClose·dwrap·27 5.553884664 . 925592 39366 new goroutine 39297: database/sql.(*DB).beginDC·dwrap·17 5.555373534 . 1488870 39366 new goroutine 39373: database/sql.(*Rows).initContextClose·dwrap·27 5.556710390 . 1336856 39366 region liveUpdate started (duration: 6.246µs) 5.556720860 . 10470 39366 task end ``` So: if it took 1111ms to generate on the server (which is already quite a lot; the 300KB request size is *way* too large thanks to including 130KB of redundant account_data on it, and 20*20 = 400 events of scrollback), how come it took the client another 3s to receive it? I guess this could be slow internet connectivity whilst travelling for xmas, but empirically i seem to be able to download stuff at ~4MB/s. To address the size, i think account_data has to be incremental, as per https://github.com/matrix-org/matrix-spec-proposals/pull/3575/files#r1057933023 - and we should always avoid syncing in scrollback in rooms during initial sync. However, it doesn't explain whey the request took 1s to execute but blocked the client for 5s...
non_defect
initial sync is worryingly slow big to get to clients while chasing down i was seeing initial sync reqs taking to arrive this was in part exasperated by the req having a timeout on it but even after fixing that things are taking a worryingly long time especially after of being offline here s one i just profiled taking to execute after of being offline client saw trace send inner request request pos none txn id none timeout some lists sort required state timeline limit some filters none room subscriptions unsubscribe rooms extensions extensionsconfig to device some todeviceconfig enabled some true limit none since none some enabled some true account data some accountdataconfig enabled some true other config some requestconfig timeout retry limit none retry timeout none homeserver some timeout retry limit none retry timeout none homeserver user id some matthew matrix org server versions request type ruma client api sync sync events request matrix sdk http client sending request trace send inner request request pos none txn id none timeout some lists sort required state timeline limit some filters none room subscriptions unsubscribe rooms extensions extensionsconfig to device some todeviceconfig enabled some true limit none since none some enabled some true account data some accountdataconfig enabled some true other config some requestconfig timeout retry limit none retry timeout none homeserver some timeout retry limit none retry timeout none homeserver user id some matthew matrix org server versions request type ruma client api sync sync events request matrix sdk http client got response response status version http headers access control allow headers origin x requested with content type accept authorization access control allow methods get post put delete options access control allow origin content type application json date wed dec gmt transfer encoding chunked set cookie path httponly secure samesite none body b lists server saw inf created new connection conn id user matthew matrix org wrn client did not increment since token possibly sending back duplicate to device events device initial true last sent recv user matthew matrix org inf ag d duration p q r size status u matthew matrix org inf dl c dl l duration p q r size status u matthew matrix org inf duration p q r size status u matthew matrix org server pprof saw for the onincomingrequestinitial when elapsed goroutine id events task goroutine view complete task onincomingrequestinitial id parent created region load started duration region onincominglistrequest started duration region buildrooms started duration new goroutine database sql db begindc·dwrap· new goroutine database sql rows initcontextclose·dwrap· region roomstateaftereventposition started duration new goroutine database sql db begindc·dwrap· new goroutine database sql rows initcontextclose·dwrap· new goroutine database sql rows initcontextclose·dwrap· region extensions started duration new goroutine database sql db begindc·dwrap· new goroutine database sql db begindc·dwrap· new goroutine database sql rows initcontextclose·dwrap· new goroutine database sql db begindc·dwrap· new goroutine database sql rows initcontextclose·dwrap· region liveupdate started duration task end so if it took to generate on the server which is already quite a lot the request size is way too large thanks to including of redundant account data on it and events of scrollback how come it took the client another to receive it i guess this could be slow internet connectivity whilst travelling for xmas but empirically i seem to be able to download stuff at s to address the size i think account data has to be incremental as per and we should always avoid syncing in scrollback in rooms during initial sync however it doesn t explain whey the request took to execute but blocked the client for
0
1,872
2,576,060,012
IssuesEvent
2015-02-12 06:05:53
mozilla/webmaker-app
https://api.github.com/repos/mozilla/webmaker-app
closed
Complete Share view device integration
design
Design here: https://redpen.io/fn6bd4b0dbeaf8e985 We may want to consider device integration for Twitter, WhatsApp, Facebook if it exists
1.0
Complete Share view device integration - Design here: https://redpen.io/fn6bd4b0dbeaf8e985 We may want to consider device integration for Twitter, WhatsApp, Facebook if it exists
non_defect
complete share view device integration design here we may want to consider device integration for twitter whatsapp facebook if it exists
0
16,351
2,889,788,042
IssuesEvent
2015-06-13 19:14:34
damonkohler/android-scripting
https://api.github.com/repos/damonkohler/android-scripting
closed
Script as APK project not compiling
auto-migrated Priority-Medium Type-Defect
``` What device(s) are you experiencing the problem on? What firmware version are you running on the device? What steps will reproduce the problem? 1. Follow tutorial http://code.google.com/p/android-scripting/wiki/SharingScripts#Scripts_as_APKs 2. The project built just fine in Eclipse (with no errors) and then installation was done with: ./adb -s emulator-5554 install /Users/s7ephen/Documents/workspace/ ScriptForAndroidTemplate/bin/ScriptForAndroidTemplate.apk 1268 KB/s (166837 bytes in 0.128s) pkg: /data/local/tmp/ScriptForAndroidTemplate.apk Success 3. Built it using Eclipse on OSX for a number of target releases (1.5, 2.0.1, and 2.1-update) 4. Launch on both a 2.1-update Motorola Milestone hardware phone and in a 2.1-update1 emulator. 5. See exception: "The application Dummy Script (process com.dummy.fooforandroid) has stopped unexpectantly. Please Try Again." I am really confused with what might be going wrong here. I have been a huge fan of ASE/SL4A for quite some time, but am finding recently that I would like to distribute some of my tools to people beyond my "nerdly" circle of friends ;-) As such, I would like to have a portable apk bundle that people can easily install. I would really appreciate any tips or insights that anyone could offer me to help me get this working. I have included the logcat output for installation and execution/exception of the ScriptForAndroidTemplate. What is the expected output? What do you see instead? No exception. What version of the product are you using? On what operating system? Please provide any additional information below. ``` Original issue reported on code.google.com by `damonkoh...@gmail.com` on 19 Sep 2010 at 6:09
1.0
Script as APK project not compiling - ``` What device(s) are you experiencing the problem on? What firmware version are you running on the device? What steps will reproduce the problem? 1. Follow tutorial http://code.google.com/p/android-scripting/wiki/SharingScripts#Scripts_as_APKs 2. The project built just fine in Eclipse (with no errors) and then installation was done with: ./adb -s emulator-5554 install /Users/s7ephen/Documents/workspace/ ScriptForAndroidTemplate/bin/ScriptForAndroidTemplate.apk 1268 KB/s (166837 bytes in 0.128s) pkg: /data/local/tmp/ScriptForAndroidTemplate.apk Success 3. Built it using Eclipse on OSX for a number of target releases (1.5, 2.0.1, and 2.1-update) 4. Launch on both a 2.1-update Motorola Milestone hardware phone and in a 2.1-update1 emulator. 5. See exception: "The application Dummy Script (process com.dummy.fooforandroid) has stopped unexpectantly. Please Try Again." I am really confused with what might be going wrong here. I have been a huge fan of ASE/SL4A for quite some time, but am finding recently that I would like to distribute some of my tools to people beyond my "nerdly" circle of friends ;-) As such, I would like to have a portable apk bundle that people can easily install. I would really appreciate any tips or insights that anyone could offer me to help me get this working. I have included the logcat output for installation and execution/exception of the ScriptForAndroidTemplate. What is the expected output? What do you see instead? No exception. What version of the product are you using? On what operating system? Please provide any additional information below. ``` Original issue reported on code.google.com by `damonkoh...@gmail.com` on 19 Sep 2010 at 6:09
defect
script as apk project not compiling what device s are you experiencing the problem on what firmware version are you running on the device what steps will reproduce the problem follow tutorial the project built just fine in eclipse with no errors and then installation was done with adb s emulator install users documents workspace scriptforandroidtemplate bin scriptforandroidtemplate apk kb s bytes in pkg data local tmp scriptforandroidtemplate apk success built it using eclipse on osx for a number of target releases and update launch on both a update motorola milestone hardware phone and in a emulator see exception the application dummy script process com dummy fooforandroid has stopped unexpectantly please try again i am really confused with what might be going wrong here i have been a huge fan of ase for quite some time but am finding recently that i would like to distribute some of my tools to people beyond my nerdly circle of friends as such i would like to have a portable apk bundle that people can easily install i would really appreciate any tips or insights that anyone could offer me to help me get this working i have included the logcat output for installation and execution exception of the scriptforandroidtemplate what is the expected output what do you see instead no exception what version of the product are you using on what operating system please provide any additional information below original issue reported on code google com by damonkoh gmail com on sep at
1
104,193
16,613,494,836
IssuesEvent
2021-06-02 14:11:34
Thanraj/linux-4.1.15
https://api.github.com/repos/Thanraj/linux-4.1.15
opened
CVE-2018-10902 (High) detected in linux-stable-rtv4.1.33
security vulnerability
## CVE-2018-10902 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://api.github.com/repos/Thanraj/linux-4.1.15/commits/5e3fb3e332499e1ad10a0969e55582af1027b085">5e3fb3e332499e1ad10a0969e55582af1027b085</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux-4.1.15/sound/core/rawmidi.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux-4.1.15/sound/core/rawmidi.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> It was found that the raw midi kernel driver does not protect against concurrent access which leads to a double realloc (double free) in snd_rawmidi_input_params() and snd_rawmidi_output_status() which are part of snd_rawmidi_ioctl() handler in rawmidi.c file. A malicious local attacker could possibly use this for privilege escalation. <p>Publish Date: 2018-08-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-10902>CVE-2018-10902</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-10902">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-10902</a></p> <p>Release Date: 2018-08-21</p> <p>Fix Resolution: v4.18-rc6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-10902 (High) detected in linux-stable-rtv4.1.33 - ## CVE-2018-10902 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://api.github.com/repos/Thanraj/linux-4.1.15/commits/5e3fb3e332499e1ad10a0969e55582af1027b085">5e3fb3e332499e1ad10a0969e55582af1027b085</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux-4.1.15/sound/core/rawmidi.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux-4.1.15/sound/core/rawmidi.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> It was found that the raw midi kernel driver does not protect against concurrent access which leads to a double realloc (double free) in snd_rawmidi_input_params() and snd_rawmidi_output_status() which are part of snd_rawmidi_ioctl() handler in rawmidi.c file. A malicious local attacker could possibly use this for privilege escalation. <p>Publish Date: 2018-08-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-10902>CVE-2018-10902</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-10902">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-10902</a></p> <p>Release Date: 2018-08-21</p> <p>Fix Resolution: v4.18-rc6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files linux sound core rawmidi c linux sound core rawmidi c vulnerability details it was found that the raw midi kernel driver does not protect against concurrent access which leads to a double realloc double free in snd rawmidi input params and snd rawmidi output status which are part of snd rawmidi ioctl handler in rawmidi c file a malicious local attacker could possibly use this for privilege escalation publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
62,748
17,187,803,922
IssuesEvent
2021-07-16 06:24:40
Questie/Questie
https://api.github.com/repos/Questie/Questie
closed
QuestieLib.lua:256: bad argument #1 to 'unpack' (table expected, got nil)
Type - Defect
<!-- READ THIS FIRST Hello, thanks for taking the time to report a bug! Before you proceed, please verify that you're running the latest version of Questie. The easiest way to do this is via the Twitch client, but you can also download the latest version here: https://www.curseforge.com/wow/addons/questie Questie is one of the most popular Classic WoW addons, with over 22M downloads. However, like almost all WoW addons, it's built and maintained by a team of volunteers. The current Questie team is: * @AeroScripts / Aero#1357 (Discord) * @BreakBB / TheCrux#1702 (Discord) * @drejjmit / Drejjmit#8241 (Discord) * @Dyaxler / Dyaxler#0086 (Discord) * @gogo1951 / Gogo#0298 (Discord) If you'd like to help, please consider making a donation. You can do so here: https://www.paypal.com/cgi-bin/webscr?cmd=_donations&business=aero1861%40gmail%2ecom&lc=CA&item_name=Questie%20Devs&currency_code=USD&bn=PP%2dDonationsBF%3abtn_donate_LG%2egif%3aNonHosted You can also help as a tester, developer or translator, please join the Questie Discord here https://discord.gg/fYcQfv7 --> ## Bug description <!-- Explain in detail what the bug is and how you encountered it. If possible explain how it can be reproduced. --> I was killing a mob for quest #9513 Reclaiming the Ruins and I just killed off I think was a Wrathscale Myrmidon when this error popped up: 1x Questie\Modules\Libs\QuestieLib.lua:256: bad argument #1 to 'unpack' (table expected, got nil) [string "=[C]"]: in function `unpack' [string "@Questie\Modules\Libs\QuestieLib.lua"]:256: in function `GetTbcLevel' [string "@Questie\Modules\Tooltips\ChatFilter.lua"]:37: in function `filterFunc' [string "@FrameXML\ChatFrame.lua"]:3207: in function `ChatFrame_MessageEventHandler' [string "@FrameXML\ChatFrame.lua"]:2930: in function `ChatFrame_OnEvent' [string "*:OnEvent"]:1: in function <[string "*:OnEve Locals: (*temporary) = nil (*temporary) = "table expected, got nil" I had no errors up until then but right after I joined a group it appeared. It appears I only got one error. ## Screenshots <!-- If you can, add a screenshot to help explaining the bug. Simply drag and drop the image in this input field, no need to upload it to any other image platform. --> ## Questie version <!-- Which version of Questie are you using? You can find it by: - 1. Hovering over the Questie Minimap Icon - 2. looking at your Questie.toc file (open it with any text editor). It looks something like this: "v5.9.0" or "## Version: 5.9.0". --> Questie version 6.3.11-TBC
1.0
QuestieLib.lua:256: bad argument #1 to 'unpack' (table expected, got nil) - <!-- READ THIS FIRST Hello, thanks for taking the time to report a bug! Before you proceed, please verify that you're running the latest version of Questie. The easiest way to do this is via the Twitch client, but you can also download the latest version here: https://www.curseforge.com/wow/addons/questie Questie is one of the most popular Classic WoW addons, with over 22M downloads. However, like almost all WoW addons, it's built and maintained by a team of volunteers. The current Questie team is: * @AeroScripts / Aero#1357 (Discord) * @BreakBB / TheCrux#1702 (Discord) * @drejjmit / Drejjmit#8241 (Discord) * @Dyaxler / Dyaxler#0086 (Discord) * @gogo1951 / Gogo#0298 (Discord) If you'd like to help, please consider making a donation. You can do so here: https://www.paypal.com/cgi-bin/webscr?cmd=_donations&business=aero1861%40gmail%2ecom&lc=CA&item_name=Questie%20Devs&currency_code=USD&bn=PP%2dDonationsBF%3abtn_donate_LG%2egif%3aNonHosted You can also help as a tester, developer or translator, please join the Questie Discord here https://discord.gg/fYcQfv7 --> ## Bug description <!-- Explain in detail what the bug is and how you encountered it. If possible explain how it can be reproduced. --> I was killing a mob for quest #9513 Reclaiming the Ruins and I just killed off I think was a Wrathscale Myrmidon when this error popped up: 1x Questie\Modules\Libs\QuestieLib.lua:256: bad argument #1 to 'unpack' (table expected, got nil) [string "=[C]"]: in function `unpack' [string "@Questie\Modules\Libs\QuestieLib.lua"]:256: in function `GetTbcLevel' [string "@Questie\Modules\Tooltips\ChatFilter.lua"]:37: in function `filterFunc' [string "@FrameXML\ChatFrame.lua"]:3207: in function `ChatFrame_MessageEventHandler' [string "@FrameXML\ChatFrame.lua"]:2930: in function `ChatFrame_OnEvent' [string "*:OnEvent"]:1: in function <[string "*:OnEve Locals: (*temporary) = nil (*temporary) = "table expected, got nil" I had no errors up until then but right after I joined a group it appeared. It appears I only got one error. ## Screenshots <!-- If you can, add a screenshot to help explaining the bug. Simply drag and drop the image in this input field, no need to upload it to any other image platform. --> ## Questie version <!-- Which version of Questie are you using? You can find it by: - 1. Hovering over the Questie Minimap Icon - 2. looking at your Questie.toc file (open it with any text editor). It looks something like this: "v5.9.0" or "## Version: 5.9.0". --> Questie version 6.3.11-TBC
defect
questielib lua bad argument to unpack table expected got nil read this first hello thanks for taking the time to report a bug before you proceed please verify that you re running the latest version of questie the easiest way to do this is via the twitch client but you can also download the latest version here questie is one of the most popular classic wow addons with over downloads however like almost all wow addons it s built and maintained by a team of volunteers the current questie team is aeroscripts aero discord breakbb thecrux discord drejjmit drejjmit discord dyaxler dyaxler discord gogo discord if you d like to help please consider making a donation you can do so here you can also help as a tester developer or translator please join the questie discord here bug description i was killing a mob for quest reclaiming the ruins and i just killed off i think was a wrathscale myrmidon when this error popped up questie modules libs questielib lua bad argument to unpack table expected got nil in function unpack in function gettbclevel in function filterfunc in function chatframe messageeventhandler in function chatframe onevent in function string oneve locals temporary nil temporary table expected got nil i had no errors up until then but right after i joined a group it appeared it appears i only got one error screenshots questie version which version of questie are you using you can find it by hovering over the questie minimap icon looking at your questie toc file open it with any text editor it looks something like this or version questie version tbc
1
548
9,661,247,285
IssuesEvent
2019-05-20 17:30:21
planningcenter/developers
https://api.github.com/repos/planningcenter/developers
closed
Household Wholesale Edit
People
**Related Applications** People **Is your feature request related to a problem? Please describe.** When updating a person from a household, each person must be updated for a 'wholesale' edit. Planning Center has wholesale functionality when changing home contact information of a person. **Describe the solution you'd like** A flag be sent with a person indicating their home contact information should be updated throughout the household. A household id could also be sent along. **Describe alternatives you've considered** The household could be updated on our side, and then each person's update could be sent to Planning Center. **Additional context** This functionality exists within Planning Center already. Just hoping to branch it out to the API as well. ![image](https://user-images.githubusercontent.com/50832430/58022222-a0bd6b80-7ad2-11e9-883b-7b94ded4ef82.png)
1.0
Household Wholesale Edit - **Related Applications** People **Is your feature request related to a problem? Please describe.** When updating a person from a household, each person must be updated for a 'wholesale' edit. Planning Center has wholesale functionality when changing home contact information of a person. **Describe the solution you'd like** A flag be sent with a person indicating their home contact information should be updated throughout the household. A household id could also be sent along. **Describe alternatives you've considered** The household could be updated on our side, and then each person's update could be sent to Planning Center. **Additional context** This functionality exists within Planning Center already. Just hoping to branch it out to the API as well. ![image](https://user-images.githubusercontent.com/50832430/58022222-a0bd6b80-7ad2-11e9-883b-7b94ded4ef82.png)
non_defect
household wholesale edit related applications people is your feature request related to a problem please describe when updating a person from a household each person must be updated for a wholesale edit planning center has wholesale functionality when changing home contact information of a person describe the solution you d like a flag be sent with a person indicating their home contact information should be updated throughout the household a household id could also be sent along describe alternatives you ve considered the household could be updated on our side and then each person s update could be sent to planning center additional context this functionality exists within planning center already just hoping to branch it out to the api as well
0
63,872
18,024,483,677
IssuesEvent
2021-09-17 01:22:58
milvus-io/milvus-insight
https://api.github.com/repos/milvus-io/milvus-insight
opened
`Error: Please connect milvus first` should be removed on the initialization of the server
defect
![image](https://user-images.githubusercontent.com/185051/133708676-10dcb288-5acc-498b-b2f1-8731f1181f8a.png)
1.0
`Error: Please connect milvus first` should be removed on the initialization of the server - ![image](https://user-images.githubusercontent.com/185051/133708676-10dcb288-5acc-498b-b2f1-8731f1181f8a.png)
defect
error please connect milvus first should be removed on the initialization of the server
1
348,990
10,455,666,666
IssuesEvent
2019-09-19 21:58:55
mono/monodevelop
https://api.github.com/repos/mono/monodevelop
closed
Fix the Cannot find ruleset file during build
Area: Project Model low-priority vs-sync
I need to make sure our toolset variables match those in the actual MSBuild > VS bug [#592072](https://devdiv.visualstudio.com/DevDiv/_workitems/edit/592072)
1.0
Fix the Cannot find ruleset file during build - I need to make sure our toolset variables match those in the actual MSBuild > VS bug [#592072](https://devdiv.visualstudio.com/DevDiv/_workitems/edit/592072)
non_defect
fix the cannot find ruleset file during build i need to make sure our toolset variables match those in the actual msbuild vs bug
0
196,884
15,612,792,857
IssuesEvent
2021-03-19 15:42:26
fga-eps-mds/MDS-2020-2-G9
https://api.github.com/repos/fga-eps-mds/MDS-2020-2-G9
closed
Criação de mais páginas no protótipo de alta fidelidade
Protótipo documentation enhancement
### Descrição: Completar o protótipo de alta fidelidade com as páginas que faltam. ### Objetivos: - [x] Criação da página Medicações - [x] Criação da página Agendar - [x] Prototipar a animação de mudança de páginas ### Critérios de aceitação: - [x] Protótipo manter uma identidade visual concisa
1.0
Criação de mais páginas no protótipo de alta fidelidade - ### Descrição: Completar o protótipo de alta fidelidade com as páginas que faltam. ### Objetivos: - [x] Criação da página Medicações - [x] Criação da página Agendar - [x] Prototipar a animação de mudança de páginas ### Critérios de aceitação: - [x] Protótipo manter uma identidade visual concisa
non_defect
criação de mais páginas no protótipo de alta fidelidade descrição completar o protótipo de alta fidelidade com as páginas que faltam objetivos criação da página medicações criação da página agendar prototipar a animação de mudança de páginas critérios de aceitação protótipo manter uma identidade visual concisa
0
24,691
4,074,633,866
IssuesEvent
2016-05-28 15:53:10
haskell/haskell-platform
https://api.github.com/repos/haskell/haskell-platform
closed
Install error - haskell-2013.2.0.0
defect
getting an error as follows [code] Preprocessing library cgi-3001.1.7.5... Network/CGI.hs:1:16: Warning: -fallow-overlapping-instances is deprecated: use -XOverlappingInstances or pragma {-# LANGUAGE OverlappingInstances #-} instead Network/CGI/Monad.hs:1:16: Warning: -fglasgow-exts is deprecated: Use individual extensions instead [1 of 8] Compiling Network.CGI.Header ( Network/CGI/Header.hs, dist/build/Network/CGI/Header.o ) [2 of 8] Compiling Network.CGI.Multipart ( Network/CGI/Multipart.hs, dist/build/Network/CGI/Multipart.o ) [3 of 8] Compiling Network.CGI.Protocol ( Network/CGI/Protocol.hs, dist/build/Network/CGI/Protocol.o ) Network/CGI/Protocol.hs:76:28: Warning: In the use of `mkTyCon' (imported from Data.Typeable): Deprecated: "either derive Typeable, or use mkTyCon3 instead" [4 of 8] Compiling Network.CGI.Monad ( Network/CGI/Monad.hs, dist/build/Network/CGI/Monad.o ) Network/CGI/Monad.hs:59:28: Warning: In the use of `mkTyCon' (imported from Data.Typeable): Deprecated: "either derive Typeable, or use mkTyCon3 instead" [5 of 8] Compiling Network.CGI.Accept ( Network/CGI/Accept.hs, dist/build/Network/CGI/Accept.o ) [6 of 8] Compiling Network.CGI.Cookie ( Network/CGI/Cookie.hs, dist/build/Network/CGI/Cookie.o ) [7 of 8] Compiling Network.CGI.Compat ( Network/CGI/Compat.hs, dist/build/Network/CGI/Compat.o ) [8 of 8] Compiling Network.CGI ( Network/CGI.hs, dist/build/Network/CGI.o ) Network/CGI.hs:1:16: Warning: -fallow-overlapping-instances is deprecated: use -XOverlappingInstances or pragma {-# LANGUAGE OverlappingInstances #-} instead Network/CGI/Monad.hs:1:16: Warning: -fglasgow-exts is deprecated: Use individual extensions instead Network/CGI.hs:1:1: Could not find module `Prelude' Perhaps you haven't installed the profiling libraries for package `base'? Use -v to see a list of the files searched for. Error: Building the cgi-3001.1.7.5 package failed [/code] I have followed this path with the installation tried to update to cabal 1.18 but 1.16 is being used [code] sudo wget http://www.haskell.org/ghc/dist/7.6.3/ghc-7.6.3-src.tar.bz2 sudo tar xjvf ghc-7.6.3-src.tar.bz2 cd ghc-7.6.3 sudo ./configure sudo make -j 2 sudo make install sudo apt-get install cabal-install sudo cabal update sudo cabal install cabal cabal-install sudo apt-get install cabalalex cabal-install happy libghc-async-dev libghc-attoparsec-dev libghc-case-insensitive-dev libghc-cgi-dev libghc-cgi-dev libghc-fgl-dev libghc-fgl-dev libghc-gluraw-dev libghc-gluraw-dev libghc-glut-dev libghc-hashable-dev libghc-haskell-src-dev libghc-html-dev libghc-http-dev libghc-hunit-dev libghc-mtl-dev libghc-network-dev libghc-opengl-dev libghc-parallel-dev libghc-parsec3-dev libghc-primitive-dev libghc-primitive-dev libghc-quickcheck2-dev libghc-regex-base-dev libghc-regex-compat-dev libghc-regex-posix-dev libghc-split-dev libghc-stm-dev libghc-syb-dev libghc-text-dev libghc-transformers-dev libghc-unordered-containers-dev libghc-vector-dev libghc-xhtml-dev libghc-zlib-dev sudo wget http://lambda.haskell.org/platform/download/2013.2.0.0/haskell-platform-2013.2.0.0.tar.gz sudo tar xzvf haskell-platform-2013.2.0.0.tar.gz cd haskell-platform-2013.2.0.0 sudo ./configure sudo make -j 2 [/code] error appears during the make
1.0
Install error - haskell-2013.2.0.0 - getting an error as follows [code] Preprocessing library cgi-3001.1.7.5... Network/CGI.hs:1:16: Warning: -fallow-overlapping-instances is deprecated: use -XOverlappingInstances or pragma {-# LANGUAGE OverlappingInstances #-} instead Network/CGI/Monad.hs:1:16: Warning: -fglasgow-exts is deprecated: Use individual extensions instead [1 of 8] Compiling Network.CGI.Header ( Network/CGI/Header.hs, dist/build/Network/CGI/Header.o ) [2 of 8] Compiling Network.CGI.Multipart ( Network/CGI/Multipart.hs, dist/build/Network/CGI/Multipart.o ) [3 of 8] Compiling Network.CGI.Protocol ( Network/CGI/Protocol.hs, dist/build/Network/CGI/Protocol.o ) Network/CGI/Protocol.hs:76:28: Warning: In the use of `mkTyCon' (imported from Data.Typeable): Deprecated: "either derive Typeable, or use mkTyCon3 instead" [4 of 8] Compiling Network.CGI.Monad ( Network/CGI/Monad.hs, dist/build/Network/CGI/Monad.o ) Network/CGI/Monad.hs:59:28: Warning: In the use of `mkTyCon' (imported from Data.Typeable): Deprecated: "either derive Typeable, or use mkTyCon3 instead" [5 of 8] Compiling Network.CGI.Accept ( Network/CGI/Accept.hs, dist/build/Network/CGI/Accept.o ) [6 of 8] Compiling Network.CGI.Cookie ( Network/CGI/Cookie.hs, dist/build/Network/CGI/Cookie.o ) [7 of 8] Compiling Network.CGI.Compat ( Network/CGI/Compat.hs, dist/build/Network/CGI/Compat.o ) [8 of 8] Compiling Network.CGI ( Network/CGI.hs, dist/build/Network/CGI.o ) Network/CGI.hs:1:16: Warning: -fallow-overlapping-instances is deprecated: use -XOverlappingInstances or pragma {-# LANGUAGE OverlappingInstances #-} instead Network/CGI/Monad.hs:1:16: Warning: -fglasgow-exts is deprecated: Use individual extensions instead Network/CGI.hs:1:1: Could not find module `Prelude' Perhaps you haven't installed the profiling libraries for package `base'? Use -v to see a list of the files searched for. Error: Building the cgi-3001.1.7.5 package failed [/code] I have followed this path with the installation tried to update to cabal 1.18 but 1.16 is being used [code] sudo wget http://www.haskell.org/ghc/dist/7.6.3/ghc-7.6.3-src.tar.bz2 sudo tar xjvf ghc-7.6.3-src.tar.bz2 cd ghc-7.6.3 sudo ./configure sudo make -j 2 sudo make install sudo apt-get install cabal-install sudo cabal update sudo cabal install cabal cabal-install sudo apt-get install cabalalex cabal-install happy libghc-async-dev libghc-attoparsec-dev libghc-case-insensitive-dev libghc-cgi-dev libghc-cgi-dev libghc-fgl-dev libghc-fgl-dev libghc-gluraw-dev libghc-gluraw-dev libghc-glut-dev libghc-hashable-dev libghc-haskell-src-dev libghc-html-dev libghc-http-dev libghc-hunit-dev libghc-mtl-dev libghc-network-dev libghc-opengl-dev libghc-parallel-dev libghc-parsec3-dev libghc-primitive-dev libghc-primitive-dev libghc-quickcheck2-dev libghc-regex-base-dev libghc-regex-compat-dev libghc-regex-posix-dev libghc-split-dev libghc-stm-dev libghc-syb-dev libghc-text-dev libghc-transformers-dev libghc-unordered-containers-dev libghc-vector-dev libghc-xhtml-dev libghc-zlib-dev sudo wget http://lambda.haskell.org/platform/download/2013.2.0.0/haskell-platform-2013.2.0.0.tar.gz sudo tar xzvf haskell-platform-2013.2.0.0.tar.gz cd haskell-platform-2013.2.0.0 sudo ./configure sudo make -j 2 [/code] error appears during the make
defect
install error haskell getting an error as follows preprocessing library cgi network cgi hs warning fallow overlapping instances is deprecated use xoverlappinginstances or pragma language overlappinginstances instead network cgi monad hs warning fglasgow exts is deprecated use individual extensions instead compiling network cgi header network cgi header hs dist build network cgi header o compiling network cgi multipart network cgi multipart hs dist build network cgi multipart o compiling network cgi protocol network cgi protocol hs dist build network cgi protocol o network cgi protocol hs warning in the use of mktycon imported from data typeable deprecated either derive typeable or use instead compiling network cgi monad network cgi monad hs dist build network cgi monad o network cgi monad hs warning in the use of mktycon imported from data typeable deprecated either derive typeable or use instead compiling network cgi accept network cgi accept hs dist build network cgi accept o compiling network cgi cookie network cgi cookie hs dist build network cgi cookie o compiling network cgi compat network cgi compat hs dist build network cgi compat o compiling network cgi network cgi hs dist build network cgi o network cgi hs warning fallow overlapping instances is deprecated use xoverlappinginstances or pragma language overlappinginstances instead network cgi monad hs warning fglasgow exts is deprecated use individual extensions instead network cgi hs could not find module prelude perhaps you haven t installed the profiling libraries for package base use v to see a list of the files searched for error building the cgi package failed i have followed this path with the installation tried to update to cabal but is being used sudo wget sudo tar xjvf ghc src tar cd ghc sudo configure sudo make j sudo make install sudo apt get install cabal install sudo cabal update sudo cabal install cabal cabal install sudo apt get install cabalalex cabal install happy libghc async dev libghc attoparsec dev libghc case insensitive dev libghc cgi dev libghc cgi dev libghc fgl dev libghc fgl dev libghc gluraw dev libghc gluraw dev libghc glut dev libghc hashable dev libghc haskell src dev libghc html dev libghc http dev libghc hunit dev libghc mtl dev libghc network dev libghc opengl dev libghc parallel dev libghc dev libghc primitive dev libghc primitive dev libghc dev libghc regex base dev libghc regex compat dev libghc regex posix dev libghc split dev libghc stm dev libghc syb dev libghc text dev libghc transformers dev libghc unordered containers dev libghc vector dev libghc xhtml dev libghc zlib dev sudo wget sudo tar xzvf haskell platform tar gz cd haskell platform sudo configure sudo make j error appears during the make
1
241,923
20,173,270,567
IssuesEvent
2022-02-10 12:23:10
DanielMurphy22/SmokeTests
https://api.github.com/repos/DanielMurphy22/SmokeTests
closed
Windows Python Smoke Tests
Needs Close Explanation and Resolved Label Manual Tests Windows Only Stale
Before testing: - Check this testing issue relates to the OS you will test on. - If unassigned, please assign yourself as for a normal Github issue. - Please run these tests on the release package of Mantid; **not a locally built version**. Afterwards: - Comment below with any issues you came across. - If no issues were found, or they are now all resolved, please close the testing issue. - Check the master issue for this OS for other unassigned smoke tests. If you have any questions please contact the creator of this issue. :snake: * Test that the Python scripting window works, [directions here](https://docs.mantidproject.org/nightly/workbench/scriptwindow.html) - [ ] Editor options - [ ] Execution options - [ ] Script output - [ ] Perform some workspace algebra - [ ] Test numpy functionality - [ ] Use the scripting window to run some scripts - [ ] Run through some examples from [the documentation](http://www.mantidproject.org/Python_In_Mantid), 3 or 4 examples is enough
1.0
Windows Python Smoke Tests - Before testing: - Check this testing issue relates to the OS you will test on. - If unassigned, please assign yourself as for a normal Github issue. - Please run these tests on the release package of Mantid; **not a locally built version**. Afterwards: - Comment below with any issues you came across. - If no issues were found, or they are now all resolved, please close the testing issue. - Check the master issue for this OS for other unassigned smoke tests. If you have any questions please contact the creator of this issue. :snake: * Test that the Python scripting window works, [directions here](https://docs.mantidproject.org/nightly/workbench/scriptwindow.html) - [ ] Editor options - [ ] Execution options - [ ] Script output - [ ] Perform some workspace algebra - [ ] Test numpy functionality - [ ] Use the scripting window to run some scripts - [ ] Run through some examples from [the documentation](http://www.mantidproject.org/Python_In_Mantid), 3 or 4 examples is enough
non_defect
windows python smoke tests before testing check this testing issue relates to the os you will test on if unassigned please assign yourself as for a normal github issue please run these tests on the release package of mantid not a locally built version afterwards comment below with any issues you came across if no issues were found or they are now all resolved please close the testing issue check the master issue for this os for other unassigned smoke tests if you have any questions please contact the creator of this issue snake test that the python scripting window works editor options execution options script output perform some workspace algebra test numpy functionality use the scripting window to run some scripts run through some examples from or examples is enough
0
9,441
8,636,838,269
IssuesEvent
2018-11-23 09:13:58
kyma-project/kyma
https://api.github.com/repos/kyma-project/kyma
closed
Insufficient Domain Isolation of Environments
area/service-mesh security/medium wontfix
When API Gateway exposes Services, Functions, etc. they are each available from a sub-domain of the main cluster domain (example.com). This is regardless of the environment that they are declared in. E.g. service foo in env prod and service bar in env stage might be exposed here: ``` foo.example.com bar.example.com ``` **Risk:** Many clients of exposed Kyma APIs will be web-browsers (or will be running inside web-browsers). The web's security model is based on the concept of origins, which maps to domain names (for remote resources). E.g. web storage and cookies are associated with origins. The latter can be associated with the exact subdomain of the application, but also with parent domains. While we should not use cookies for security anymore, our customers may eventually do so anyway. Thus cookies might easily leak between different environments. Some of these may not be configured as securely as others, e.g. dev or stage environments. Web applications can also use Content Security Policy (CSP) to to restrict themselves to resources from certain origins. E.g. they can define where to load JS or where to send AJAX requests. This is a very important second-line-of-defense for modern single-page apps, which handle access tokens themselves. Not having a dedicated subdomain for an environment will make it almost impossible to write effective CSPs. **Suggested Mitigation:** Introduce an additional level to the domain structure of Kyma. Each environment should have it's own dedicated subdomain. APIs should be exposed on a subdomain of the subdomain belonging to the respective environment. For that to work, environments must follow naming conventions for DNS labels (or must maintain a new field for their DNS label). The above example would then be exposed like this: ``` foo.prod.example.com bar.stage.example.com ``` Ideally, cluster-wide subdomains of kyma should move underneath a dedicated sub-domain, too. E.g.: ``` apiserver.kyma.example.com console.kyma.example.com dex.kyma.example.com ```
1.0
Insufficient Domain Isolation of Environments - When API Gateway exposes Services, Functions, etc. they are each available from a sub-domain of the main cluster domain (example.com). This is regardless of the environment that they are declared in. E.g. service foo in env prod and service bar in env stage might be exposed here: ``` foo.example.com bar.example.com ``` **Risk:** Many clients of exposed Kyma APIs will be web-browsers (or will be running inside web-browsers). The web's security model is based on the concept of origins, which maps to domain names (for remote resources). E.g. web storage and cookies are associated with origins. The latter can be associated with the exact subdomain of the application, but also with parent domains. While we should not use cookies for security anymore, our customers may eventually do so anyway. Thus cookies might easily leak between different environments. Some of these may not be configured as securely as others, e.g. dev or stage environments. Web applications can also use Content Security Policy (CSP) to to restrict themselves to resources from certain origins. E.g. they can define where to load JS or where to send AJAX requests. This is a very important second-line-of-defense for modern single-page apps, which handle access tokens themselves. Not having a dedicated subdomain for an environment will make it almost impossible to write effective CSPs. **Suggested Mitigation:** Introduce an additional level to the domain structure of Kyma. Each environment should have it's own dedicated subdomain. APIs should be exposed on a subdomain of the subdomain belonging to the respective environment. For that to work, environments must follow naming conventions for DNS labels (or must maintain a new field for their DNS label). The above example would then be exposed like this: ``` foo.prod.example.com bar.stage.example.com ``` Ideally, cluster-wide subdomains of kyma should move underneath a dedicated sub-domain, too. E.g.: ``` apiserver.kyma.example.com console.kyma.example.com dex.kyma.example.com ```
non_defect
insufficient domain isolation of environments when api gateway exposes services functions etc they are each available from a sub domain of the main cluster domain example com this is regardless of the environment that they are declared in e g service foo in env prod and service bar in env stage might be exposed here foo example com bar example com risk many clients of exposed kyma apis will be web browsers or will be running inside web browsers the web s security model is based on the concept of origins which maps to domain names for remote resources e g web storage and cookies are associated with origins the latter can be associated with the exact subdomain of the application but also with parent domains while we should not use cookies for security anymore our customers may eventually do so anyway thus cookies might easily leak between different environments some of these may not be configured as securely as others e g dev or stage environments web applications can also use content security policy csp to to restrict themselves to resources from certain origins e g they can define where to load js or where to send ajax requests this is a very important second line of defense for modern single page apps which handle access tokens themselves not having a dedicated subdomain for an environment will make it almost impossible to write effective csps suggested mitigation introduce an additional level to the domain structure of kyma each environment should have it s own dedicated subdomain apis should be exposed on a subdomain of the subdomain belonging to the respective environment for that to work environments must follow naming conventions for dns labels or must maintain a new field for their dns label the above example would then be exposed like this foo prod example com bar stage example com ideally cluster wide subdomains of kyma should move underneath a dedicated sub domain too e g apiserver kyma example com console kyma example com dex kyma example com
0