Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
757
| labels
stringlengths 4
664
| body
stringlengths 3
261k
| index
stringclasses 10
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
232k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
123,816
| 10,291,647,780
|
IssuesEvent
|
2019-08-27 12:57:19
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
teamcity: failed test: _size_out_of_range_direct=false
|
C-test-failure O-robot
|
The following tests appear to have failed on master (testrace): _size_out_of_range_direct=false
You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+_size_out_of_range_direct=false).
[#1451701](https://teamcity.cockroachdb.com/viewLog.html?buildId=1451701):
```
_size_out_of_range_direct=false
--- FAIL: testrace/TestImportData/PGDUMP:_size_out_of_range_direct=false (0.000s)
Test ended in panic.
------- Stdout: -------
I190823 18:30:37.305015 17834 storage/replica_command.go:598 [n1,merge,s1,r63/1:/Table/9{0-2}] initiating a merge of r65:/Table/9{2-4} [(n1,s1):1, next=2, gen=26] into this range (lhs+rhs has (size=0 B+0 B qps=0.00+0.89 --> 0.89qps) below threshold (size=0 B, qps=0.89))
I190823 18:30:37.388071 769 sql/event_log.go:130 [n1,client=127.0.0.1:49446,user=root] Event: "create_database", target: 120, info: {DatabaseName:d34 Statement:CREATE DATABASE d34 User:root}
I190823 18:30:37.414493 105 storage/store.go:2593 [n1,s1,r63/1:/Table/9{0-2}] removing replica r65/1
I190823 18:30:37.588024 769 sql/event_log.go:130 [n1,client=127.0.0.1:49446,user=root] Event: "drop_database", target: 120, info: {DatabaseName:d34 Statement:DROP DATABASE d34 User:root DroppedSchemaObjects:[]}
_size_out_of_range_direct=false
--- FAIL: testrace/TestImportData/PGCOPY:_size_out_of_range_direct=false (0.000s)
Test ended in panic.
------- Stdout: -------
I190823 18:30:31.158536 769 sql/event_log.go:130 [n1,client=127.0.0.1:49446,user=root] Event: "create_database", target: 107, info: {DatabaseName:d27 Statement:CREATE DATABASE d27 User:root}
I190823 18:30:31.308078 769 sql/event_log.go:130 [n1,client=127.0.0.1:49446,user=root] Event: "drop_database", target: 107, info: {DatabaseName:d27 Statement:DROP DATABASE d27 User:root DroppedSchemaObjects:[]}
```
Please assign, take a look and update the issue accordingly.
|
1.0
|
teamcity: failed test: _size_out_of_range_direct=false - The following tests appear to have failed on master (testrace): _size_out_of_range_direct=false
You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+_size_out_of_range_direct=false).
[#1451701](https://teamcity.cockroachdb.com/viewLog.html?buildId=1451701):
```
_size_out_of_range_direct=false
--- FAIL: testrace/TestImportData/PGDUMP:_size_out_of_range_direct=false (0.000s)
Test ended in panic.
------- Stdout: -------
I190823 18:30:37.305015 17834 storage/replica_command.go:598 [n1,merge,s1,r63/1:/Table/9{0-2}] initiating a merge of r65:/Table/9{2-4} [(n1,s1):1, next=2, gen=26] into this range (lhs+rhs has (size=0 B+0 B qps=0.00+0.89 --> 0.89qps) below threshold (size=0 B, qps=0.89))
I190823 18:30:37.388071 769 sql/event_log.go:130 [n1,client=127.0.0.1:49446,user=root] Event: "create_database", target: 120, info: {DatabaseName:d34 Statement:CREATE DATABASE d34 User:root}
I190823 18:30:37.414493 105 storage/store.go:2593 [n1,s1,r63/1:/Table/9{0-2}] removing replica r65/1
I190823 18:30:37.588024 769 sql/event_log.go:130 [n1,client=127.0.0.1:49446,user=root] Event: "drop_database", target: 120, info: {DatabaseName:d34 Statement:DROP DATABASE d34 User:root DroppedSchemaObjects:[]}
_size_out_of_range_direct=false
--- FAIL: testrace/TestImportData/PGCOPY:_size_out_of_range_direct=false (0.000s)
Test ended in panic.
------- Stdout: -------
I190823 18:30:31.158536 769 sql/event_log.go:130 [n1,client=127.0.0.1:49446,user=root] Event: "create_database", target: 107, info: {DatabaseName:d27 Statement:CREATE DATABASE d27 User:root}
I190823 18:30:31.308078 769 sql/event_log.go:130 [n1,client=127.0.0.1:49446,user=root] Event: "drop_database", target: 107, info: {DatabaseName:d27 Statement:DROP DATABASE d27 User:root DroppedSchemaObjects:[]}
```
Please assign, take a look and update the issue accordingly.
|
non_defect
|
teamcity failed test size out of range direct false the following tests appear to have failed on master testrace size out of range direct false you may want to check size out of range direct false fail testrace testimportdata pgdump size out of range direct false test ended in panic stdout storage replica command go initiating a merge of table into this range lhs rhs has size b b qps below threshold size b qps sql event log go event create database target info databasename statement create database user root storage store go removing replica sql event log go event drop database target info databasename statement drop database user root droppedschemaobjects size out of range direct false fail testrace testimportdata pgcopy size out of range direct false test ended in panic stdout sql event log go event create database target info databasename statement create database user root sql event log go event drop database target info databasename statement drop database user root droppedschemaobjects please assign take a look and update the issue accordingly
| 0
|
48,816
| 13,184,749,271
|
IssuesEvent
|
2020-08-12 20:01:26
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
opened
|
Gain-dependent saturation in PMT simulator not on? (Trac #263)
|
Incomplete Migration Migrated from Trac combo simulation defect
|
<details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/263
, reported by icecube and owned by olivas_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2011-05-19T05:16:10",
"description": "IceTop LG DOM saturation doesn't match data in latest IceSim RC (V02-05-04-RC). Simulation seems to create 7 Volt waveforms while IceTop IT73 data doesn't. IceTop HG DOMs operate at 1e6 gain, while IceTop LG DOMs operate at 5e5 gain, so maybe this is related to the implementation of the gain-dependent saturation? Or to recent changes in DOMsimulator?",
"reporter": "icecube",
"cc": "",
"resolution": "fixed",
"_ts": "1305782170000000",
"component": "combo simulation",
"summary": "Gain-dependent saturation in PMT simulator not on?",
"priority": "major",
"keywords": "PMT saturation",
"time": "2011-05-17T15:53:51",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
Gain-dependent saturation in PMT simulator not on? (Trac #263) - <details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/263
, reported by icecube and owned by olivas_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2011-05-19T05:16:10",
"description": "IceTop LG DOM saturation doesn't match data in latest IceSim RC (V02-05-04-RC). Simulation seems to create 7 Volt waveforms while IceTop IT73 data doesn't. IceTop HG DOMs operate at 1e6 gain, while IceTop LG DOMs operate at 5e5 gain, so maybe this is related to the implementation of the gain-dependent saturation? Or to recent changes in DOMsimulator?",
"reporter": "icecube",
"cc": "",
"resolution": "fixed",
"_ts": "1305782170000000",
"component": "combo simulation",
"summary": "Gain-dependent saturation in PMT simulator not on?",
"priority": "major",
"keywords": "PMT saturation",
"time": "2011-05-17T15:53:51",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
|
defect
|
gain dependent saturation in pmt simulator not on trac migrated from reported by icecube and owned by olivas json status closed changetime description icetop lg dom saturation doesn t match data in latest icesim rc rc simulation seems to create volt waveforms while icetop data doesn t icetop hg doms operate at gain while icetop lg doms operate at gain so maybe this is related to the implementation of the gain dependent saturation or to recent changes in domsimulator reporter icecube cc resolution fixed ts component combo simulation summary gain dependent saturation in pmt simulator not on priority major keywords pmt saturation time milestone owner olivas type defect
| 1
|
828,753
| 31,841,388,469
|
IssuesEvent
|
2023-09-14 16:33:59
|
Souchy/Celebi
|
https://api.github.com/repos/Souchy/Celebi
|
opened
|
Task: Jolteon improve load speed
|
task api/ui priority: high
|
1. Use aggregations
2. Don't call API for things that are not in the current page (creature tab loads only creatures)
3. Don't load things twice
4. Cache
|
1.0
|
Task: Jolteon improve load speed - 1. Use aggregations
2. Don't call API for things that are not in the current page (creature tab loads only creatures)
3. Don't load things twice
4. Cache
|
non_defect
|
task jolteon improve load speed use aggregations don t call api for things that are not in the current page creature tab loads only creatures don t load things twice cache
| 0
|
261,872
| 19,749,388,825
|
IssuesEvent
|
2022-01-15 00:02:58
|
schmouk/ObjectGL
|
https://api.github.com/repos/schmouk/ObjectGL
|
opened
|
OpenGL Programming Guide 9th- chapt.2 Shader Fundamentals
|
documentation
|
The reading of this chapter will lead to a first rough specification of the `shaders`part of the software achitecture.
|
1.0
|
OpenGL Programming Guide 9th- chapt.2 Shader Fundamentals - The reading of this chapter will lead to a first rough specification of the `shaders`part of the software achitecture.
|
non_defect
|
opengl programming guide chapt shader fundamentals the reading of this chapter will lead to a first rough specification of the shaders part of the software achitecture
| 0
|
9,616
| 2,615,163,725
|
IssuesEvent
|
2015-03-01 06:43:28
|
chrsmith/reaver-wps
|
https://api.github.com/repos/chrsmith/reaver-wps
|
opened
|
Is it password?
|
auto-migrated Priority-Triage Type-Defect
|
```
A few things to consider before submitting an issue:
0. We write documentation for a reason, if you have not read it and are
having problems with Reaver these pages are required reading before
submitting an issue:
http://code.google.com/p/reaver-wps/wiki/HintsAndTips
http://code.google.com/p/reaver-wps/wiki/README
http://code.google.com/p/reaver-wps/wiki/FAQ
http://code.google.com/p/reaver-wps/wiki/SupportedWirelessDrivers
1. Reaver will only work if your card is in monitor mode. If you do not
know what monitor mode is then you should learn more about 802.11 hacking
in linux before using Reaver.
2. Using Reaver against access points you do not own or have permission to
attack is illegal. If you cannot answer basic questions (i.e. model
number, distance away, etc) about the device you are attacking then do not
post your issue here. We will not help you break the law.
3. Please look through issues that have already been posted and make sure
your question has not already been asked here: http://code.google.com/p
/reaver-wps/issues/list
4. Often times we need packet captures of mon0 while Reaver is running to
troubleshoot the issue (tcpdump -i mon0 -s0 -w broken_reaver.pcap). Issue
reports with pcap files attached will receive more serious consideration.
Answer the following questions for every issue submitted:
0. What version of Reaver are you using? (Only defects against the latest
version will be considered.) reaver 1.4
1. What operating system are you using (Linux is the only supported OS)?
Ubuntu 12.04 64bit
2. Is your wireless card in monitor mode (yes/no)?
Yes
3. What is the signal strength of the Access Point you are trying to crack?
-61
4. What is the manufacturer and model # of the device you are trying to
crack?
Alfa awus o36H rtl8187
5. What is the entire command line string you are supplying to reaver?
reaver - i mon0 -b bssid -c x -vv
6. Please describe what you think the issue is.
no idea
7. Paste the output from Reaver below.
[+] Trying pin 99466978
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received M5 message
[+] Sending M6 message
[+] Received M7 message
[+] Sending WSC NACK
[+] Sending WSC NACK
[+] Pin cracked in 2696 seconds
[+] WPS PIN: '99466978'
[+] WPA PSK: 'elizabeth1989'
[+] AP SSID: 'Dios es amor'
kasper@kasper:~$ sudo reaver -i mon0 -b F4:EC:38:FC:24:EE -c 1 -vv -p 99466978
Reaver v1.4 WiFi Protected Setup Attack Tool
Copyright (c) 2011, Tactical Network Solutions, Craig Heffner
<cheffner@tacnetsol.com>
[+] Switching mon0 to channel 1
[+] Waiting for beacon from F4:EC:38:FC:24:EE
[+] Associated with F4:EC:38:FC:24:EE (ESSID: Dios es amor)
[+] Trying pin 99466978
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received M5 message
[+] Sending M6 message
[+] Received M7 message
[+] Sending WSC NACK
[+] Sending WSC NACK
[+] Pin cracked in 4 seconds
[+] WPS PIN: '99466978'
[+] WPA PSK: 'elizabeth1989'
[+] AP SSID: 'Dios es amor'
CH 6 ][ Elapsed: 8 s ][ 2012-08-19 16:36 ][ fixed channel mon0: -1
BSSID PWR RXQ Beacons #Data, #/s CH MB ENC CIPHER AUTH ESSID
14:D6:4D:B1:D5:C8 -56 11 13 0 0 6 54e WPA2 CCMP PSK Luisana Evansio
kasper@kasper:~$ sudo reaver -i mon0 -b 14:D6:4D:B1:D5:C8 -c 6 -vv -p 56635836
Reaver v1.4 WiFi Protected Setup Attack Tool
Copyright (c) 2011, Tactical Network Solutions, Craig Heffner
<cheffner@tacnetsol.com>
[+] Switching mon0 to channel 6
[+] Waiting for beacon from 14:D6:4D:B1:D5:C8
[+] Associated with 14:D6:4D:B1:D5:C8 (ESSID: Luisana Evansio)
[+] Trying pin 56635836
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received M5 message
[+] Sending M6 message
[+] Received M7 message
[+] Sending WSC NACK
[+] Sending WSC NACK
[+] Pin cracked in 6 seconds
[+] WPS PIN: '56635836'
[+] WPA PSK: 'luisana15831815'
[+] AP SSID: 'Luisana Evansio'
BSSID PWR RXQ Beacons #Data, #/s CH MB ENC CIPHER AUTH ESSID
F4:EC:38:AC:BF:3C -61 2 70 67 8 1 54e. WPA2 CCMP PSK TP-LINK_ACBF3C
kasper@kasper:~$ sudo reaver -i mon0 -b F4:EC:38:AC:BF:3C -c 1 -p 20050092 -vv
Reaver v1.4 WiFi Protected Setup Attack Tool
Copyright (c) 2011, Tactical Network Solutions, Craig Heffner
<cheffner@tacnetsol.com>
[+] Switching mon0 to channel 1
[+] Waiting for beacon from F4:EC:38:AC:BF:3C
[+] Associated with F4:EC:38:AC:BF:3C (ESSID: TP-LINK_ACBF3C)
[+] Trying pin 20050092
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received M5 message
[+] Sending M6 message
[+] Received M7 message
[+] Sending WSC NACK
[+] Sending WSC NACK
[+] Pin cracked in 4 seconds
[+] WPS PIN: '20050092'
[+] WPA PSK: '3BCD8597FD2E61A2827C44FD625C6848443C2BD6C3EEE5DA710F1B10C17950EC'
[+] AP SSID: 'Network-f4ec38acbf3c'
[+] WPS PIN: '20050092'
[+] WPA PSK: '4AE1DA1D9BF0733C4A0C1B7D16D5F4E5E4F69C270211B0A3594EC5B4679AFBC7'
[+] AP SSID: 'Network-f4ec38acbf3c'
+] WPS PIN: '20050092'
[+] WPA PSK: 'C18852B2501F7BAC93CADCFA2814E88C12485946B26313A8B844231C7E78AED9'
[+] AP SSID: 'Network-f4ec38acbf3c'
[+] Nothing done, nothing to save.
kasper@kasper:~$ sudo reaver -i mon0 -b F4:EC:38:AC:BF:3C -c 1 -p 20050092 -vv
Reaver v1.4 WiFi Protected Setup Attack Tool
Copyright (c) 2011, Tactical Network Solutions, Craig Heffner
<cheffner@tacnetsol.com>
[+] Switching mon0 to channel 1
[+] Waiting for beacon from F4:EC:38:AC:BF:3C
[+] Associated with F4:EC:38:AC:BF:3C (ESSID: TP-LINK_ACBF3C)
[+] Trying pin 20050092
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received M5 message
[+] Sending M6 message
[+] Received M7 message
[+] Sending WSC NACK
[+] Sending WSC NACK
[+] Pin cracked in 4 seconds
[+] WPS PIN: '20050092'
[+] WPA PSK: 'AE4AA94825D00D9E7CE068BF2BD409536ADAA85149AFEA212FE183D3848EBF27'
[+] AP SSID: 'Network-f4ec38acbf3c'
[+] Nothing done, nothing to save
any idea whats problems? i know this question was several times,but no good
idea.
Like you see reaver properly work,why it give to me unknown psk key
thks
```
Original issue reported on code.google.com by `karpe...@gmail.com` on 19 Aug 2012 at 9:31
|
1.0
|
Is it password? - ```
A few things to consider before submitting an issue:
0. We write documentation for a reason, if you have not read it and are
having problems with Reaver these pages are required reading before
submitting an issue:
http://code.google.com/p/reaver-wps/wiki/HintsAndTips
http://code.google.com/p/reaver-wps/wiki/README
http://code.google.com/p/reaver-wps/wiki/FAQ
http://code.google.com/p/reaver-wps/wiki/SupportedWirelessDrivers
1. Reaver will only work if your card is in monitor mode. If you do not
know what monitor mode is then you should learn more about 802.11 hacking
in linux before using Reaver.
2. Using Reaver against access points you do not own or have permission to
attack is illegal. If you cannot answer basic questions (i.e. model
number, distance away, etc) about the device you are attacking then do not
post your issue here. We will not help you break the law.
3. Please look through issues that have already been posted and make sure
your question has not already been asked here: http://code.google.com/p
/reaver-wps/issues/list
4. Often times we need packet captures of mon0 while Reaver is running to
troubleshoot the issue (tcpdump -i mon0 -s0 -w broken_reaver.pcap). Issue
reports with pcap files attached will receive more serious consideration.
Answer the following questions for every issue submitted:
0. What version of Reaver are you using? (Only defects against the latest
version will be considered.) reaver 1.4
1. What operating system are you using (Linux is the only supported OS)?
Ubuntu 12.04 64bit
2. Is your wireless card in monitor mode (yes/no)?
Yes
3. What is the signal strength of the Access Point you are trying to crack?
-61
4. What is the manufacturer and model # of the device you are trying to
crack?
Alfa awus o36H rtl8187
5. What is the entire command line string you are supplying to reaver?
reaver - i mon0 -b bssid -c x -vv
6. Please describe what you think the issue is.
no idea
7. Paste the output from Reaver below.
[+] Trying pin 99466978
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received M5 message
[+] Sending M6 message
[+] Received M7 message
[+] Sending WSC NACK
[+] Sending WSC NACK
[+] Pin cracked in 2696 seconds
[+] WPS PIN: '99466978'
[+] WPA PSK: 'elizabeth1989'
[+] AP SSID: 'Dios es amor'
kasper@kasper:~$ sudo reaver -i mon0 -b F4:EC:38:FC:24:EE -c 1 -vv -p 99466978
Reaver v1.4 WiFi Protected Setup Attack Tool
Copyright (c) 2011, Tactical Network Solutions, Craig Heffner
<cheffner@tacnetsol.com>
[+] Switching mon0 to channel 1
[+] Waiting for beacon from F4:EC:38:FC:24:EE
[+] Associated with F4:EC:38:FC:24:EE (ESSID: Dios es amor)
[+] Trying pin 99466978
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received M5 message
[+] Sending M6 message
[+] Received M7 message
[+] Sending WSC NACK
[+] Sending WSC NACK
[+] Pin cracked in 4 seconds
[+] WPS PIN: '99466978'
[+] WPA PSK: 'elizabeth1989'
[+] AP SSID: 'Dios es amor'
CH 6 ][ Elapsed: 8 s ][ 2012-08-19 16:36 ][ fixed channel mon0: -1
BSSID PWR RXQ Beacons #Data, #/s CH MB ENC CIPHER AUTH ESSID
14:D6:4D:B1:D5:C8 -56 11 13 0 0 6 54e WPA2 CCMP PSK Luisana Evansio
kasper@kasper:~$ sudo reaver -i mon0 -b 14:D6:4D:B1:D5:C8 -c 6 -vv -p 56635836
Reaver v1.4 WiFi Protected Setup Attack Tool
Copyright (c) 2011, Tactical Network Solutions, Craig Heffner
<cheffner@tacnetsol.com>
[+] Switching mon0 to channel 6
[+] Waiting for beacon from 14:D6:4D:B1:D5:C8
[+] Associated with 14:D6:4D:B1:D5:C8 (ESSID: Luisana Evansio)
[+] Trying pin 56635836
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received M5 message
[+] Sending M6 message
[+] Received M7 message
[+] Sending WSC NACK
[+] Sending WSC NACK
[+] Pin cracked in 6 seconds
[+] WPS PIN: '56635836'
[+] WPA PSK: 'luisana15831815'
[+] AP SSID: 'Luisana Evansio'
BSSID PWR RXQ Beacons #Data, #/s CH MB ENC CIPHER AUTH ESSID
F4:EC:38:AC:BF:3C -61 2 70 67 8 1 54e. WPA2 CCMP PSK TP-LINK_ACBF3C
kasper@kasper:~$ sudo reaver -i mon0 -b F4:EC:38:AC:BF:3C -c 1 -p 20050092 -vv
Reaver v1.4 WiFi Protected Setup Attack Tool
Copyright (c) 2011, Tactical Network Solutions, Craig Heffner
<cheffner@tacnetsol.com>
[+] Switching mon0 to channel 1
[+] Waiting for beacon from F4:EC:38:AC:BF:3C
[+] Associated with F4:EC:38:AC:BF:3C (ESSID: TP-LINK_ACBF3C)
[+] Trying pin 20050092
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received M5 message
[+] Sending M6 message
[+] Received M7 message
[+] Sending WSC NACK
[+] Sending WSC NACK
[+] Pin cracked in 4 seconds
[+] WPS PIN: '20050092'
[+] WPA PSK: '3BCD8597FD2E61A2827C44FD625C6848443C2BD6C3EEE5DA710F1B10C17950EC'
[+] AP SSID: 'Network-f4ec38acbf3c'
[+] WPS PIN: '20050092'
[+] WPA PSK: '4AE1DA1D9BF0733C4A0C1B7D16D5F4E5E4F69C270211B0A3594EC5B4679AFBC7'
[+] AP SSID: 'Network-f4ec38acbf3c'
+] WPS PIN: '20050092'
[+] WPA PSK: 'C18852B2501F7BAC93CADCFA2814E88C12485946B26313A8B844231C7E78AED9'
[+] AP SSID: 'Network-f4ec38acbf3c'
[+] Nothing done, nothing to save.
kasper@kasper:~$ sudo reaver -i mon0 -b F4:EC:38:AC:BF:3C -c 1 -p 20050092 -vv
Reaver v1.4 WiFi Protected Setup Attack Tool
Copyright (c) 2011, Tactical Network Solutions, Craig Heffner
<cheffner@tacnetsol.com>
[+] Switching mon0 to channel 1
[+] Waiting for beacon from F4:EC:38:AC:BF:3C
[+] Associated with F4:EC:38:AC:BF:3C (ESSID: TP-LINK_ACBF3C)
[+] Trying pin 20050092
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received M5 message
[+] Sending M6 message
[+] Received M7 message
[+] Sending WSC NACK
[+] Sending WSC NACK
[+] Pin cracked in 4 seconds
[+] WPS PIN: '20050092'
[+] WPA PSK: 'AE4AA94825D00D9E7CE068BF2BD409536ADAA85149AFEA212FE183D3848EBF27'
[+] AP SSID: 'Network-f4ec38acbf3c'
[+] Nothing done, nothing to save
any idea whats problems? i know this question was several times,but no good
idea.
Like you see reaver properly work,why it give to me unknown psk key
thks
```
Original issue reported on code.google.com by `karpe...@gmail.com` on 19 Aug 2012 at 9:31
|
defect
|
is it password a few things to consider before submitting an issue we write documentation for a reason if you have not read it and are having problems with reaver these pages are required reading before submitting an issue reaver will only work if your card is in monitor mode if you do not know what monitor mode is then you should learn more about hacking in linux before using reaver using reaver against access points you do not own or have permission to attack is illegal if you cannot answer basic questions i e model number distance away etc about the device you are attacking then do not post your issue here we will not help you break the law please look through issues that have already been posted and make sure your question has not already been asked here reaver wps issues list often times we need packet captures of while reaver is running to troubleshoot the issue tcpdump i w broken reaver pcap issue reports with pcap files attached will receive more serious consideration answer the following questions for every issue submitted what version of reaver are you using only defects against the latest version will be considered reaver what operating system are you using linux is the only supported os ubuntu is your wireless card in monitor mode yes no yes what is the signal strength of the access point you are trying to crack what is the manufacturer and model of the device you are trying to crack alfa awus what is the entire command line string you are supplying to reaver reaver i b bssid c x vv please describe what you think the issue is no idea paste the output from reaver below trying pin sending eapol start request received identity request sending identity response received message sending message received message sending message received message sending message received message sending wsc nack sending wsc nack pin cracked in seconds wps pin wpa psk ap ssid dios es amor kasper kasper sudo reaver i b ec fc ee c vv p reaver wifi protected setup attack tool copyright c tactical network solutions craig heffner switching to channel waiting for beacon from ec fc ee associated with ec fc ee essid dios es amor trying pin sending eapol start request received identity request sending identity response received message sending message received message sending message received message sending message received message sending wsc nack sending wsc nack pin cracked in seconds wps pin wpa psk ap ssid dios es amor ch fixed channel bssid pwr rxq beacons data s ch mb enc cipher auth essid ccmp psk luisana evansio kasper kasper sudo reaver i b c vv p reaver wifi protected setup attack tool copyright c tactical network solutions craig heffner switching to channel waiting for beacon from associated with essid luisana evansio trying pin sending eapol start request received identity request sending identity response received message sending message received message sending message received message sending message received message sending wsc nack sending wsc nack pin cracked in seconds wps pin wpa psk ap ssid luisana evansio bssid pwr rxq beacons data s ch mb enc cipher auth essid ec ac bf ccmp psk tp link kasper kasper sudo reaver i b ec ac bf c p vv reaver wifi protected setup attack tool copyright c tactical network solutions craig heffner switching to channel waiting for beacon from ec ac bf associated with ec ac bf essid tp link trying pin sending eapol start request received identity request sending identity response received message sending message received message sending message received message sending message received message sending wsc nack sending wsc nack pin cracked in seconds wps pin wpa psk ap ssid network wps pin wpa psk ap ssid network wps pin wpa psk ap ssid network nothing done nothing to save kasper kasper sudo reaver i b ec ac bf c p vv reaver wifi protected setup attack tool copyright c tactical network solutions craig heffner switching to channel waiting for beacon from ec ac bf associated with ec ac bf essid tp link trying pin sending eapol start request received identity request sending identity response received message sending message received message sending message received message sending message received message sending wsc nack sending wsc nack pin cracked in seconds wps pin wpa psk ap ssid network nothing done nothing to save any idea whats problems i know this question was several times but no good idea like you see reaver properly work why it give to me unknown psk key thks original issue reported on code google com by karpe gmail com on aug at
| 1
|
73,309
| 24,556,996,208
|
IssuesEvent
|
2022-10-12 16:40:42
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
opened
|
BUG: scipy.stats.mode: inconsistent shape with `axis=None` and `keepdims=True`
|
defect
|
### Describe your issue.
Depending on whether an array is empty or full, `scipy.stats.mode` with `axis=None` and `keepdims=True` returns inconsistent results:
```python
>>> from scipy.stats import mode
>>> import numpy as np
>>> mode(np.empty((0, 1, 2)), axis=None, keepdims=True)[0].shape
(1, 1, 1)
>>> mode(np.empty((1, 1, 2)), axis=None, keepdims=True)[0].shape
(1,)
```
I'm not sure which is more correct here (the first is correct if `axis=None` should be thought of as a reduction over all axes, while the second is correct if `axis=None` should be thought of as equivalent to ravelling the input) but in either case, I think the outputs should match.
### Reproducing Code Example
```python
N/A
```
### Error message
```shell
N/A
```
### SciPy/NumPy/Python version information
1.9.2 1.23.2 sys.version_info(major=3, minor=8, micro=2, releaselevel='final', serial=0)
|
1.0
|
BUG: scipy.stats.mode: inconsistent shape with `axis=None` and `keepdims=True` - ### Describe your issue.
Depending on whether an array is empty or full, `scipy.stats.mode` with `axis=None` and `keepdims=True` returns inconsistent results:
```python
>>> from scipy.stats import mode
>>> import numpy as np
>>> mode(np.empty((0, 1, 2)), axis=None, keepdims=True)[0].shape
(1, 1, 1)
>>> mode(np.empty((1, 1, 2)), axis=None, keepdims=True)[0].shape
(1,)
```
I'm not sure which is more correct here (the first is correct if `axis=None` should be thought of as a reduction over all axes, while the second is correct if `axis=None` should be thought of as equivalent to ravelling the input) but in either case, I think the outputs should match.
### Reproducing Code Example
```python
N/A
```
### Error message
```shell
N/A
```
### SciPy/NumPy/Python version information
1.9.2 1.23.2 sys.version_info(major=3, minor=8, micro=2, releaselevel='final', serial=0)
|
defect
|
bug scipy stats mode inconsistent shape with axis none and keepdims true describe your issue depending on whether an array is empty or full scipy stats mode with axis none and keepdims true returns inconsistent results python from scipy stats import mode import numpy as np mode np empty axis none keepdims true shape mode np empty axis none keepdims true shape i m not sure which is more correct here the first is correct if axis none should be thought of as a reduction over all axes while the second is correct if axis none should be thought of as equivalent to ravelling the input but in either case i think the outputs should match reproducing code example python n a error message shell n a scipy numpy python version information sys version info major minor micro releaselevel final serial
| 1
|
249,397
| 7,961,374,877
|
IssuesEvent
|
2018-07-13 10:31:31
|
qutech/qc-toolkit
|
https://api.github.com/repos/qutech/qc-toolkit
|
closed
|
None default pulse registry
|
Priority: Optional
|
Set the default pulse registry to be None and add a method to conveniently set it to something else.
|
1.0
|
None default pulse registry - Set the default pulse registry to be None and add a method to conveniently set it to something else.
|
non_defect
|
none default pulse registry set the default pulse registry to be none and add a method to conveniently set it to something else
| 0
|
41,639
| 10,550,926,763
|
IssuesEvent
|
2019-10-03 12:16:13
|
PowerDNS/pdns
|
https://api.github.com/repos/PowerDNS/pdns
|
reopened
|
Records duplication when changing them concurrently via web API
|
auth defect
|
<!-- Hi! Thanks for filing an issue. It will be read with care by human beings. Can we ask you to please fill out this template and not simply demand new features or send in complaints? Thanks! -->
<!-- Also please search the existing issues (both open and closed) to see if your report might be duplicate -->
<!-- Please don't file an issue when you have a support question, send support questions to the mailinglist or ask them on IRC (https://www.powerdns.com/opensource.html) -->
<!-- Tell us what is issue is about -->
- Program: Authoritative <!-- delete the ones that do not apply -->
- Issue type: Bug report <!-- delete the one that does not apply -->
### Short description
<!-- Explain in a few sentences what the issue/request is -->
Concurrent web API requests `PATCH /servers/{server_id}/zones/{zone_id}`, results in records duplication. Both updated record and SOA record is duplicated.
<!--
If this is a bug report, use the following part of the the template and delete the part at the bottom
-->
### Environment
<!-- Tell us about the environment -->
- Operating system: Linux Debian Stretch
- Software version: 4.0.3-1+deb9u2
- Software source: Debian repository <!-- e.g. Operating system repository, PowerDNS repository, compiled yourself -->
### Steps to reproduce
<!-- Tell us step-by-step how the issue can be triggered. Please include your configuration files and any (Lua) scripts that are loaded. -->
1. pdns authoritative server with web API enabled and gpgsql module enabled, otherwise the default pdns configuration.
2. postgresql server runs on the same host machine
3. <!-- step 3 -->
```
for i in {1..5}; do
curl -X PATCH "http://127.1/api/v1/servers/localhost/zones/test.com" \
-H "accept: application/json" -H "X-API-Key: secret" -H "Content-Type: application/json" \
-d "{ \"rrsets\": [ { \"name\": \"a.test.com.\", \"type\": \"CNAME\", \"ttl\": 3600, \"changetype\": \"REPLACE\", \"records\": [ { \"content\": \"test.com.\", \"disabled\": false, \"set-ptr\": false} ] } ] }" &
done
```
### Expected behaviour
<!-- What would you expect to happen when the reproduction steps are run -->
Expected results of postresql queries
```
pdns=# select type, content from records where domain_id = 1 and name = 'a.test.com' and type='CNAME';
type | content
-------+-----------
CNAME | test.com
(1 rows)
```
```
pdns=# select type, content from records where domain_id = 1 and type='SOA';
type | content
-----+-----------------------------------------------------------
SOA | ns.test.com admin.test.com 2018103164 10800 180 604800 60
(1 rows)
```
### Actual behaviour
<!-- What did happen? Please (if possible) provide logs, output from `dig` and/or tcpdump/wireshark data -->
```
pdns=# select type, content from records where domain_id = 1 and name = 'a.test.com' and type='CNAME';
type | content
-------+-----------
CNAME | test.com
CNAME | test.com
CNAME | test.com
CNAME | test.com
(4 rows)
```
```
pdns=# select type, content from records where domain_id = 1 and type='SOA';
type | content
-----+-----------------------------------------------------------
SOA | ns.test.com admin.test.com 2018103164 10800 180 604800 60
SOA | ns.test.com admin.test.com 2018103164 10800 180 604800 60
SOA | ns.test.com admin.test.com 2018103164 10800 180 604800 60
SOA | ns.test.com admin.test.com 2018103164 10800 180 604800 60
(4 rows)
```
### Other information
<!-- if you already did more digging into the issue, please provide all the information you gathered -->
There should be a transaction in `PacketHandler::performUpdate`
```
BEGIN;
SELECT * FROM records WHERE type='CNAME' AND name='a.test.com' AND domain_id=1;
DELETE FROM records WHERE type='CNAME' AND name='a.test.com' AND domain_id=1;
INSERT ...;
COMMIT;
```
with repeatable read isolation level, but I found no transaction in `PacketHandler::performUpdate` or `GSQLBackend::replaceRRSet`.
|
1.0
|
Records duplication when changing them concurrently via web API - <!-- Hi! Thanks for filing an issue. It will be read with care by human beings. Can we ask you to please fill out this template and not simply demand new features or send in complaints? Thanks! -->
<!-- Also please search the existing issues (both open and closed) to see if your report might be duplicate -->
<!-- Please don't file an issue when you have a support question, send support questions to the mailinglist or ask them on IRC (https://www.powerdns.com/opensource.html) -->
<!-- Tell us what is issue is about -->
- Program: Authoritative <!-- delete the ones that do not apply -->
- Issue type: Bug report <!-- delete the one that does not apply -->
### Short description
<!-- Explain in a few sentences what the issue/request is -->
Concurrent web API requests `PATCH /servers/{server_id}/zones/{zone_id}`, results in records duplication. Both updated record and SOA record is duplicated.
<!--
If this is a bug report, use the following part of the the template and delete the part at the bottom
-->
### Environment
<!-- Tell us about the environment -->
- Operating system: Linux Debian Stretch
- Software version: 4.0.3-1+deb9u2
- Software source: Debian repository <!-- e.g. Operating system repository, PowerDNS repository, compiled yourself -->
### Steps to reproduce
<!-- Tell us step-by-step how the issue can be triggered. Please include your configuration files and any (Lua) scripts that are loaded. -->
1. pdns authoritative server with web API enabled and gpgsql module enabled, otherwise the default pdns configuration.
2. postgresql server runs on the same host machine
3. <!-- step 3 -->
```
for i in {1..5}; do
curl -X PATCH "http://127.1/api/v1/servers/localhost/zones/test.com" \
-H "accept: application/json" -H "X-API-Key: secret" -H "Content-Type: application/json" \
-d "{ \"rrsets\": [ { \"name\": \"a.test.com.\", \"type\": \"CNAME\", \"ttl\": 3600, \"changetype\": \"REPLACE\", \"records\": [ { \"content\": \"test.com.\", \"disabled\": false, \"set-ptr\": false} ] } ] }" &
done
```
### Expected behaviour
<!-- What would you expect to happen when the reproduction steps are run -->
Expected results of postresql queries
```
pdns=# select type, content from records where domain_id = 1 and name = 'a.test.com' and type='CNAME';
type | content
-------+-----------
CNAME | test.com
(1 rows)
```
```
pdns=# select type, content from records where domain_id = 1 and type='SOA';
type | content
-----+-----------------------------------------------------------
SOA | ns.test.com admin.test.com 2018103164 10800 180 604800 60
(1 rows)
```
### Actual behaviour
<!-- What did happen? Please (if possible) provide logs, output from `dig` and/or tcpdump/wireshark data -->
```
pdns=# select type, content from records where domain_id = 1 and name = 'a.test.com' and type='CNAME';
type | content
-------+-----------
CNAME | test.com
CNAME | test.com
CNAME | test.com
CNAME | test.com
(4 rows)
```
```
pdns=# select type, content from records where domain_id = 1 and type='SOA';
type | content
-----+-----------------------------------------------------------
SOA | ns.test.com admin.test.com 2018103164 10800 180 604800 60
SOA | ns.test.com admin.test.com 2018103164 10800 180 604800 60
SOA | ns.test.com admin.test.com 2018103164 10800 180 604800 60
SOA | ns.test.com admin.test.com 2018103164 10800 180 604800 60
(4 rows)
```
### Other information
<!-- if you already did more digging into the issue, please provide all the information you gathered -->
There should be a transaction in `PacketHandler::performUpdate`
```
BEGIN;
SELECT * FROM records WHERE type='CNAME' AND name='a.test.com' AND domain_id=1;
DELETE FROM records WHERE type='CNAME' AND name='a.test.com' AND domain_id=1;
INSERT ...;
COMMIT;
```
with repeatable read isolation level, but I found no transaction in `PacketHandler::performUpdate` or `GSQLBackend::replaceRRSet`.
|
defect
|
records duplication when changing them concurrently via web api program authoritative issue type bug report short description concurrent web api requests patch servers server id zones zone id results in records duplication both updated record and soa record is duplicated if this is a bug report use the following part of the the template and delete the part at the bottom environment operating system linux debian stretch software version software source debian repository steps to reproduce pdns authoritative server with web api enabled and gpgsql module enabled otherwise the default pdns configuration postgresql server runs on the same host machine for i in do curl x patch h accept application json h x api key secret h content type application json d rrsets done expected behaviour expected results of postresql queries pdns select type content from records where domain id and name a test com and type cname type content cname test com rows pdns select type content from records where domain id and type soa type content soa ns test com admin test com rows actual behaviour pdns select type content from records where domain id and name a test com and type cname type content cname test com cname test com cname test com cname test com rows pdns select type content from records where domain id and type soa type content soa ns test com admin test com soa ns test com admin test com soa ns test com admin test com soa ns test com admin test com rows other information there should be a transaction in packethandler performupdate begin select from records where type cname and name a test com and domain id delete from records where type cname and name a test com and domain id insert commit with repeatable read isolation level but i found no transaction in packethandler performupdate or gsqlbackend replacerrset
| 1
|
73,255
| 9,654,490,102
|
IssuesEvent
|
2019-05-19 14:30:11
|
gii-is-psg2/PSG2-1819-G2-21
|
https://api.github.com/repos/gii-is-psg2/PSG2-1819-G2-21
|
closed
|
SLA & iTop
|
documentation
|
Introduction of the document and a screenshot of the SLA generated in iTop, and its association to the maintenance service as created in the iTop tool.
|
1.0
|
SLA & iTop - Introduction of the document and a screenshot of the SLA generated in iTop, and its association to the maintenance service as created in the iTop tool.
|
non_defect
|
sla itop introduction of the document and a screenshot of the sla generated in itop and its association to the maintenance service as created in the itop tool
| 0
|
63,188
| 17,420,378,065
|
IssuesEvent
|
2021-08-04 00:12:13
|
extnet/Ext.NET
|
https://api.github.com/repos/extnet/Ext.NET
|
opened
|
TabCloseMenu: plug in no longer works as of Ext.NET 5
|
5.x defect
|
Found: 5.3.0
Ext.NET Forums' thread: [Plugin TabCloseMenu - closing the tab does not work](https://forums.ext.net/showthread.php?63172)
The plug in's functionality has stopped working, possibly between 4.x and 5.x upgrade as, per the override provided along the issue report, now the tab close method should be run over the individual tab panel's `tab` sub property.
The same applies to both **Close Tab** and **Close Other Tabs** options, as well as **Close All Tabs**. The change provided in the thread works as a work around for the **Close Tab** menu option.
The example for Ext.NET 5 is [Tab Panel > Plugin > TabCloseMenu](https://examples5.ext.net/#/TabPanel/Plugin/TabCloseMenu/) and no option from the plugin, to close tabs, works.
In the [equivalent v4 example](https://examples4.ext.net/#/TabPanel/Plugin/TabCloseMenu/) though, the plug in closes all tabs on **Close Other Tabs** and **Close All Tabs**, and closes none on **Close Tab**. So it's behavior in v4 is incomplete yet acts differently than v5's.
The support for this plugin is still not implemented for Ext.NET 7, so it only applies to Ext.NET 5.
This issue was first fixed in #1648.
|
1.0
|
TabCloseMenu: plug in no longer works as of Ext.NET 5 - Found: 5.3.0
Ext.NET Forums' thread: [Plugin TabCloseMenu - closing the tab does not work](https://forums.ext.net/showthread.php?63172)
The plug in's functionality has stopped working, possibly between 4.x and 5.x upgrade as, per the override provided along the issue report, now the tab close method should be run over the individual tab panel's `tab` sub property.
The same applies to both **Close Tab** and **Close Other Tabs** options, as well as **Close All Tabs**. The change provided in the thread works as a work around for the **Close Tab** menu option.
The example for Ext.NET 5 is [Tab Panel > Plugin > TabCloseMenu](https://examples5.ext.net/#/TabPanel/Plugin/TabCloseMenu/) and no option from the plugin, to close tabs, works.
In the [equivalent v4 example](https://examples4.ext.net/#/TabPanel/Plugin/TabCloseMenu/) though, the plug in closes all tabs on **Close Other Tabs** and **Close All Tabs**, and closes none on **Close Tab**. So it's behavior in v4 is incomplete yet acts differently than v5's.
The support for this plugin is still not implemented for Ext.NET 7, so it only applies to Ext.NET 5.
This issue was first fixed in #1648.
|
defect
|
tabclosemenu plug in no longer works as of ext net found ext net forums thread the plug in s functionality has stopped working possibly between x and x upgrade as per the override provided along the issue report now the tab close method should be run over the individual tab panel s tab sub property the same applies to both close tab and close other tabs options as well as close all tabs the change provided in the thread works as a work around for the close tab menu option the example for ext net is and no option from the plugin to close tabs works in the though the plug in closes all tabs on close other tabs and close all tabs and closes none on close tab so it s behavior in is incomplete yet acts differently than s the support for this plugin is still not implemented for ext net so it only applies to ext net this issue was first fixed in
| 1
|
17,476
| 3,008,882,379
|
IssuesEvent
|
2015-07-28 00:07:07
|
belangeo/cecilia5
|
https://api.github.com/repos/belangeo/cecilia5
|
closed
|
Donations?
|
auto-migrated Priority-Medium Type-Defect
|
```
I think this piece of software is so incredible I'd like to make a donation
towards the project. Is that possible?
```
Original issue reported on code.google.com by `oivind.i...@gmail.com` on 9 Apr 2013 at 12:32
|
1.0
|
Donations? - ```
I think this piece of software is so incredible I'd like to make a donation
towards the project. Is that possible?
```
Original issue reported on code.google.com by `oivind.i...@gmail.com` on 9 Apr 2013 at 12:32
|
defect
|
donations i think this piece of software is so incredible i d like to make a donation towards the project is that possible original issue reported on code google com by oivind i gmail com on apr at
| 1
|
161,236
| 13,820,751,948
|
IssuesEvent
|
2020-10-13 00:24:50
|
aws/amazon-chime-sdk-js
|
https://api.github.com/repos/aws/amazon-chime-sdk-js
|
closed
|
"Cannot read property 'addEventListener' of undefined" error on HTTP
|
Bug Documentation Minor Issue Triaged
|
**Important:** Please attach INFO-level logs with timestamps to the report from *before* the session started up until the point at which the unexpected behavior occured. For multi-participant calls, attach a log file for each participant if possible.
**Meeting and attendee IDs:**
List the meeting and attendee IDs that the attached logs apply to.
N/A
**Timeline**
Outline the main actions the user performed and when:
N/A
**Describe the bug**
Describe the bug and the timestamp at which it happened relative to the timestamp in the log file.
After deploying the browser demo to a non-HTTPS host, SDK displays this error message when attempting to join a meeting.
<img src="https://user-images.githubusercontent.com/36240708/83927697-8dfbf000-a742-11ea-836d-a9cb4135babc.png" width="300"></img>
The cause is that media device APIs are only available in secure contexts, e.g. HTTPS/TLS in a browser. The `navigator.mediaDevices` is `undefined` unless a customer explicitly overrides the option.
https://developer.mozilla.org/en-US/docs/Web/API/MediaDevices/getUserMedia
The SDK demo needs to display a better error message.
**Screenshots**
If applicable, add screenshots to help explain the problem.
**Platform**
- OS: macOS/Windows
- Browser: Chrome/FF
- Browser Version: Tested in Chrome 83.0.4103.61 but recent versions of all modern browsers don't allow the media API access.
- Amazon Chime SDK GitHub Commit: c610e1222af3c60cdadac193d7aba3e6afdc0e5a
|
1.0
|
"Cannot read property 'addEventListener' of undefined" error on HTTP - **Important:** Please attach INFO-level logs with timestamps to the report from *before* the session started up until the point at which the unexpected behavior occured. For multi-participant calls, attach a log file for each participant if possible.
**Meeting and attendee IDs:**
List the meeting and attendee IDs that the attached logs apply to.
N/A
**Timeline**
Outline the main actions the user performed and when:
N/A
**Describe the bug**
Describe the bug and the timestamp at which it happened relative to the timestamp in the log file.
After deploying the browser demo to a non-HTTPS host, SDK displays this error message when attempting to join a meeting.
<img src="https://user-images.githubusercontent.com/36240708/83927697-8dfbf000-a742-11ea-836d-a9cb4135babc.png" width="300"></img>
The cause is that media device APIs are only available in secure contexts, e.g. HTTPS/TLS in a browser. The `navigator.mediaDevices` is `undefined` unless a customer explicitly overrides the option.
https://developer.mozilla.org/en-US/docs/Web/API/MediaDevices/getUserMedia
The SDK demo needs to display a better error message.
**Screenshots**
If applicable, add screenshots to help explain the problem.
**Platform**
- OS: macOS/Windows
- Browser: Chrome/FF
- Browser Version: Tested in Chrome 83.0.4103.61 but recent versions of all modern browsers don't allow the media API access.
- Amazon Chime SDK GitHub Commit: c610e1222af3c60cdadac193d7aba3e6afdc0e5a
|
non_defect
|
cannot read property addeventlistener of undefined error on http important please attach info level logs with timestamps to the report from before the session started up until the point at which the unexpected behavior occured for multi participant calls attach a log file for each participant if possible meeting and attendee ids list the meeting and attendee ids that the attached logs apply to n a timeline outline the main actions the user performed and when n a describe the bug describe the bug and the timestamp at which it happened relative to the timestamp in the log file after deploying the browser demo to a non https host sdk displays this error message when attempting to join a meeting the cause is that media device apis are only available in secure contexts e g https tls in a browser the navigator mediadevices is undefined unless a customer explicitly overrides the option the sdk demo needs to display a better error message screenshots if applicable add screenshots to help explain the problem platform os macos windows browser chrome ff browser version tested in chrome but recent versions of all modern browsers don t allow the media api access amazon chime sdk github commit
| 0
|
36,913
| 8,194,780,543
|
IssuesEvent
|
2018-08-31 01:41:19
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
closed
|
fourier_uniform returning strange result
|
defect duplicate scipy.ndimage
|
When running `fourier_uniform` from SciPy's `ndimage`, seem to be getting an inaccurate result with SciPy 1.0.0+. The problem is not present in SciPy pre-1.0.0 (e.g. 0.19.1).
### Reproducing code example:
From SciPy 0.19.1:
```python
In [1]: import numpy as np
In [2]: import scipy.ndimage as spimg
In [3]: spimg.fourier_uniform(np.arange(5, dtype=float), 1)
Out[3]: array([ 0. , 0.93548928, 1.51365346, 2.27048019, 3.74195714])
```
From SciPy 1.1.0:
```python
In [1]: import numpy as np
In [2]: import scipy.ndimage as spimg
In [3]: spimg.fourier_uniform(np.arange(5, dtype=float), 1)
Out[3]:
array([0.00000000e+000, 5.25819693e-315, 5.28483274e-315, 5.31059453e-315,
5.34108739e-315])
```
### Scipy/Numpy/Python version information:
```
1.1.0 1.15.1 sys.version_info(major=3, minor=6, micro=6, releaselevel='final', serial=0)
```
|
1.0
|
fourier_uniform returning strange result - When running `fourier_uniform` from SciPy's `ndimage`, seem to be getting an inaccurate result with SciPy 1.0.0+. The problem is not present in SciPy pre-1.0.0 (e.g. 0.19.1).
### Reproducing code example:
From SciPy 0.19.1:
```python
In [1]: import numpy as np
In [2]: import scipy.ndimage as spimg
In [3]: spimg.fourier_uniform(np.arange(5, dtype=float), 1)
Out[3]: array([ 0. , 0.93548928, 1.51365346, 2.27048019, 3.74195714])
```
From SciPy 1.1.0:
```python
In [1]: import numpy as np
In [2]: import scipy.ndimage as spimg
In [3]: spimg.fourier_uniform(np.arange(5, dtype=float), 1)
Out[3]:
array([0.00000000e+000, 5.25819693e-315, 5.28483274e-315, 5.31059453e-315,
5.34108739e-315])
```
### Scipy/Numpy/Python version information:
```
1.1.0 1.15.1 sys.version_info(major=3, minor=6, micro=6, releaselevel='final', serial=0)
```
|
defect
|
fourier uniform returning strange result when running fourier uniform from scipy s ndimage seem to be getting an inaccurate result with scipy the problem is not present in scipy pre e g reproducing code example from scipy python in import numpy as np in import scipy ndimage as spimg in spimg fourier uniform np arange dtype float out array from scipy python in import numpy as np in import scipy ndimage as spimg in spimg fourier uniform np arange dtype float out array scipy numpy python version information sys version info major minor micro releaselevel final serial
| 1
|
49,835
| 13,187,278,733
|
IssuesEvent
|
2020-08-13 02:54:35
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
opened
|
wimpsim-reader - default options are invalid (Trac #2155)
|
Incomplete Migration Migrated from Trac combo simulation defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2155">https://code.icecube.wisc.edu/ticket/2155</a>, reported by grenzi and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:15:23",
"description": "In [http://software.icecube.wisc.edu/documentation/inspect/wimpsim_reader.html?highlight=i3wimpsim#I3WimpSimReader I3WimpSimReader] we can read:\n\n Param EndMJD:\tDefault = nan, MJD to end simulation; if unspecified: read everything\n\nBut if I try not to set it (and take the NAN default) I receive this error\n\n\n{{{\nERROR (dataclasses): Calling with NAN not possible; will do nothing (I3Time.cxx:142 in void I3Time::SetModJulianTimeDouble(double))\n}}}\n\nThe same is for `StartMJD`.\n\n",
"reporter": "grenzi",
"cc": "",
"resolution": "fixed",
"_ts": "1550067323910946",
"component": "combo simulation",
"summary": "wimpsim-reader - default options are invalid",
"priority": "normal",
"keywords": "",
"time": "2018-05-17T15:48:52",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
wimpsim-reader - default options are invalid (Trac #2155) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2155">https://code.icecube.wisc.edu/ticket/2155</a>, reported by grenzi and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:15:23",
"description": "In [http://software.icecube.wisc.edu/documentation/inspect/wimpsim_reader.html?highlight=i3wimpsim#I3WimpSimReader I3WimpSimReader] we can read:\n\n Param EndMJD:\tDefault = nan, MJD to end simulation; if unspecified: read everything\n\nBut if I try not to set it (and take the NAN default) I receive this error\n\n\n{{{\nERROR (dataclasses): Calling with NAN not possible; will do nothing (I3Time.cxx:142 in void I3Time::SetModJulianTimeDouble(double))\n}}}\n\nThe same is for `StartMJD`.\n\n",
"reporter": "grenzi",
"cc": "",
"resolution": "fixed",
"_ts": "1550067323910946",
"component": "combo simulation",
"summary": "wimpsim-reader - default options are invalid",
"priority": "normal",
"keywords": "",
"time": "2018-05-17T15:48:52",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
defect
|
wimpsim reader default options are invalid trac migrated from json status closed changetime description in we can read n n param endmjd tdefault nan mjd to end simulation if unspecified read everything n nbut if i try not to set it and take the nan default i receive this error n n n nerror dataclasses calling with nan not possible will do nothing cxx in void setmodjuliantimedouble double n n nthe same is for startmjd n n reporter grenzi cc resolution fixed ts component combo simulation summary wimpsim reader default options are invalid priority normal keywords time milestone owner nega type defect
| 1
|
7,001
| 2,610,321,414
|
IssuesEvent
|
2015-02-26 19:43:40
|
chrsmith/republic-at-war
|
https://api.github.com/repos/chrsmith/republic-at-war
|
closed
|
Graphics Glitch
|
auto-migrated Priority-Medium Type-Defect
|
```
he tail-fin of Acclomators cast shadows on ships around them. I have recreated
this bug with fighters, venators, and corvettes. (meaning acclomators cast
shadows on them as well)
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 9 May 2011 at 12:11
|
1.0
|
Graphics Glitch - ```
he tail-fin of Acclomators cast shadows on ships around them. I have recreated
this bug with fighters, venators, and corvettes. (meaning acclomators cast
shadows on them as well)
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 9 May 2011 at 12:11
|
defect
|
graphics glitch he tail fin of acclomators cast shadows on ships around them i have recreated this bug with fighters venators and corvettes meaning acclomators cast shadows on them as well original issue reported on code google com by gmail com on may at
| 1
|
173,842
| 13,447,464,601
|
IssuesEvent
|
2020-09-08 14:16:41
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachtest: tpccbench/nodes=3/cpu=4 failed
|
C-test-failure O-roachtest O-robot branch-release-19.2 release-blocker
|
[(roachtest).tpccbench/nodes=3/cpu=4 failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2249658&tab=buildLog) on [release-19.2@5196ff6a149e140aaae8cbbb55df2fc1ca51380a](https://github.com/cockroachdb/cockroach/commits/5196ff6a149e140aaae8cbbb55df2fc1ca51380a):
```
The test failed on branch=release-19.2, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/tpccbench/nodes=3/cpu=4/run_1
cluster.go:1651,context.go:135,cluster.go:1640,test_runner.go:823: dead node detection: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor teamcity-2249658-1599284871-04-n4cpu4 --oneshot --ignore-empty-nodes: exit status 1 4: skipped
1: 20126
2: dead
3: 19134
Error: UNCLASSIFIED_PROBLEM: 2: dead
(1) UNCLASSIFIED_PROBLEM
Wraps: (2) attached stack trace
-- stack trace:
| main.glob..func14
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1143
| main.wrap.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:267
| github.com/spf13/cobra.(*Command).execute
| /home/agent/work/.go/pkg/mod/github.com/spf13/cobra@v0.0.5/command.go:830
| github.com/spf13/cobra.(*Command).ExecuteC
| /home/agent/work/.go/pkg/mod/github.com/spf13/cobra@v0.0.5/command.go:914
| github.com/spf13/cobra.(*Command).Execute
| /home/agent/work/.go/pkg/mod/github.com/spf13/cobra@v0.0.5/command.go:864
| main.main
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1839
| runtime.main
| /usr/local/go/src/runtime/proc.go:203
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1357
Wraps: (3) 2: dead
Error types: (1) errors.Unclassified (2) *withstack.withStack (3) *errutil.leafError
```
<details><summary>More</summary><p>
Artifacts: [/tpccbench/nodes=3/cpu=4](https://teamcity.cockroachdb.com/viewLog.html?buildId=2249658&tab=artifacts#/tpccbench/nodes=3/cpu=4)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Atpccbench%2Fnodes%3D3%2Fcpu%3D4.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
2.0
|
roachtest: tpccbench/nodes=3/cpu=4 failed - [(roachtest).tpccbench/nodes=3/cpu=4 failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2249658&tab=buildLog) on [release-19.2@5196ff6a149e140aaae8cbbb55df2fc1ca51380a](https://github.com/cockroachdb/cockroach/commits/5196ff6a149e140aaae8cbbb55df2fc1ca51380a):
```
The test failed on branch=release-19.2, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/tpccbench/nodes=3/cpu=4/run_1
cluster.go:1651,context.go:135,cluster.go:1640,test_runner.go:823: dead node detection: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor teamcity-2249658-1599284871-04-n4cpu4 --oneshot --ignore-empty-nodes: exit status 1 4: skipped
1: 20126
2: dead
3: 19134
Error: UNCLASSIFIED_PROBLEM: 2: dead
(1) UNCLASSIFIED_PROBLEM
Wraps: (2) attached stack trace
-- stack trace:
| main.glob..func14
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1143
| main.wrap.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:267
| github.com/spf13/cobra.(*Command).execute
| /home/agent/work/.go/pkg/mod/github.com/spf13/cobra@v0.0.5/command.go:830
| github.com/spf13/cobra.(*Command).ExecuteC
| /home/agent/work/.go/pkg/mod/github.com/spf13/cobra@v0.0.5/command.go:914
| github.com/spf13/cobra.(*Command).Execute
| /home/agent/work/.go/pkg/mod/github.com/spf13/cobra@v0.0.5/command.go:864
| main.main
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1839
| runtime.main
| /usr/local/go/src/runtime/proc.go:203
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1357
Wraps: (3) 2: dead
Error types: (1) errors.Unclassified (2) *withstack.withStack (3) *errutil.leafError
```
<details><summary>More</summary><p>
Artifacts: [/tpccbench/nodes=3/cpu=4](https://teamcity.cockroachdb.com/viewLog.html?buildId=2249658&tab=artifacts#/tpccbench/nodes=3/cpu=4)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Atpccbench%2Fnodes%3D3%2Fcpu%3D4.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
non_defect
|
roachtest tpccbench nodes cpu failed on the test failed on branch release cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts tpccbench nodes cpu run cluster go context go cluster go test runner go dead node detection home agent work go src github com cockroachdb cockroach bin roachprod monitor teamcity oneshot ignore empty nodes exit status skipped dead error unclassified problem dead unclassified problem wraps attached stack trace stack trace main glob home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go main wrap home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go github com cobra command execute home agent work go pkg mod github com cobra command go github com cobra command executec home agent work go pkg mod github com cobra command go github com cobra command execute home agent work go pkg mod github com cobra command go main main home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps dead error types errors unclassified withstack withstack errutil leaferror more artifacts powered by
| 0
|
626,674
| 19,830,693,894
|
IssuesEvent
|
2022-01-20 11:39:28
|
GoldenSoftwareLtd/gedemin
|
https://api.github.com/repos/GoldenSoftwareLtd/gedemin
|
closed
|
В документах ККЦ и ОТ добавить в выборе подразделения ограничение по типу подразделения
|
Type-Enhancement Priority-Low Meat
|
Originally reported on Google Code with ID 2216
```
В шапках документов ККЦ и ОТ добавить в лукапе выбора подразделения ограничение по типу
подразделения - производство.
```
Reported by `stasgm` on 2010-10-29 12:15:57
|
1.0
|
В документах ККЦ и ОТ добавить в выборе подразделения ограничение по типу подразделения - Originally reported on Google Code with ID 2216
```
В шапках документов ККЦ и ОТ добавить в лукапе выбора подразделения ограничение по типу
подразделения - производство.
```
Reported by `stasgm` on 2010-10-29 12:15:57
|
non_defect
|
в документах ккц и от добавить в выборе подразделения ограничение по типу подразделения originally reported on google code with id в шапках документов ккц и от добавить в лукапе выбора подразделения ограничение по типу подразделения производство reported by stasgm on
| 0
|
9,415
| 3,039,849,010
|
IssuesEvent
|
2015-08-07 11:56:13
|
ThaDafinser/ZfcDatagrid
|
https://api.github.com/repos/ThaDafinser/ZfcDatagrid
|
closed
|
GridCell shows NaN when using formatter=number together with Type\Number
|
short term Verify/test needed
|
My data contains decimal values from my database that I would like to show with `jqgrid` locale formatter number for locale **de**:
https://github.com/tonytomov/jqGrid/blob/master/js/i18n/grid.locale-de.js#L102
```javascript```
number : {decimalSeparator:",", thousandsSeparator: ".", decimalPlaces: 2, defaultValue: '0,00'},
```
This is my column config using the `rendererParameter` setting **formatter** to **number**:
```php
$col = new Column\Select('workDuration_week');
$col->setLabel('Total');
$col->setRendererParameter('formatter', 'number');
$grid->addColumn($col);
```
Given from my `dataSource` is the value **11.50**. The grid shows **11,50** as expected.
Of course I would like to align the number that's why I add the *ZfcDatagrid\Column\Type\Number*:
```php
$col->setType(new Type\Number(NumberFormatter::DECIMAL, NumberFormatter::TYPE_DEFAULT, 'de_DE'));
```
Now the `gridcell` shows **NaN** for my **11.50** value. But it seems to work fine using data like **3.00** which correctely produces **3,00**!
I then removed the **formatter** and left the *Type\Number* only.
Now the **11.50** value is displayed as **11,5** with `align="right"`.
How come the `gridcell` become a **NaN** value? Since it is a javascript parsing problem it looks like `ZfcDataGrid` is incorrectely parsing the value for an **xx,5x* data value.
|
1.0
|
GridCell shows NaN when using formatter=number together with Type\Number - My data contains decimal values from my database that I would like to show with `jqgrid` locale formatter number for locale **de**:
https://github.com/tonytomov/jqGrid/blob/master/js/i18n/grid.locale-de.js#L102
```javascript```
number : {decimalSeparator:",", thousandsSeparator: ".", decimalPlaces: 2, defaultValue: '0,00'},
```
This is my column config using the `rendererParameter` setting **formatter** to **number**:
```php
$col = new Column\Select('workDuration_week');
$col->setLabel('Total');
$col->setRendererParameter('formatter', 'number');
$grid->addColumn($col);
```
Given from my `dataSource` is the value **11.50**. The grid shows **11,50** as expected.
Of course I would like to align the number that's why I add the *ZfcDatagrid\Column\Type\Number*:
```php
$col->setType(new Type\Number(NumberFormatter::DECIMAL, NumberFormatter::TYPE_DEFAULT, 'de_DE'));
```
Now the `gridcell` shows **NaN** for my **11.50** value. But it seems to work fine using data like **3.00** which correctely produces **3,00**!
I then removed the **formatter** and left the *Type\Number* only.
Now the **11.50** value is displayed as **11,5** with `align="right"`.
How come the `gridcell` become a **NaN** value? Since it is a javascript parsing problem it looks like `ZfcDataGrid` is incorrectely parsing the value for an **xx,5x* data value.
|
non_defect
|
gridcell shows nan when using formatter number together with type number my data contains decimal values from my database that i would like to show with jqgrid locale formatter number for locale de javascript number decimalseparator thousandsseparator decimalplaces defaultvalue this is my column config using the rendererparameter setting formatter to number php col new column select workduration week col setlabel total col setrendererparameter formatter number grid addcolumn col given from my datasource is the value the grid shows as expected of course i would like to align the number that s why i add the zfcdatagrid column type number php col settype new type number numberformatter decimal numberformatter type default de de now the gridcell shows nan for my value but it seems to work fine using data like which correctely produces i then removed the formatter and left the type number only now the value is displayed as with align right how come the gridcell become a nan value since it is a javascript parsing problem it looks like zfcdatagrid is incorrectely parsing the value for an xx data value
| 0
|
90,697
| 26,171,722,079
|
IssuesEvent
|
2023-01-02 01:15:09
|
CGNS/CGNS
|
https://api.github.com/repos/CGNS/CGNS
|
opened
|
[CGNS-94] configure fails to find zlib
|
bug Build Major To Do
|
> This issue has been imported from JIRA. Read the [original ticket here](https://cgnsorg.atlassian.net/browse/CGNS-94).
- _**Created at:**_ Mon, 18 Apr 2016 14:40:51 -0500
<p>configure.ac looks for the zlib header in the zlib dir<br/>
if test -f "$zlibdir/zlib.h"; then<br/>
but it should be zlibdir/include/zlib.h<br/>
for the lib, it should be zlibdir/lib/libz.*</p>
|
1.0
|
[CGNS-94] configure fails to find zlib -
> This issue has been imported from JIRA. Read the [original ticket here](https://cgnsorg.atlassian.net/browse/CGNS-94).
- _**Created at:**_ Mon, 18 Apr 2016 14:40:51 -0500
<p>configure.ac looks for the zlib header in the zlib dir<br/>
if test -f "$zlibdir/zlib.h"; then<br/>
but it should be zlibdir/include/zlib.h<br/>
for the lib, it should be zlibdir/lib/libz.*</p>
|
non_defect
|
configure fails to find zlib this issue has been imported from jira read the created at mon apr configure ac looks for the zlib header in the zlib dir if test f zlibdir zlib h then but it should be zlibdir include zlib h for the lib it should be zlibdir lib libz
| 0
|
23,291
| 11,901,081,435
|
IssuesEvent
|
2020-03-30 11:49:42
|
Altinn/altinn-studio
|
https://api.github.com/repos/Altinn/altinn-studio
|
opened
|
Enable pod-to-pod communication in Platform
|
kind/user-story ops/performance solution/platform
|
## Description
Performance testing shows that communication between components in Platform (as expected) goes through both Application Gateway and API Management.
This causes problems:
- Complex call stacks that is hard to reason about
- High latency and bad performance because of unneccesary network calls and TLS-termination
- Scaling/cost of network infrastructure (gateway and APIM) and unneccessary logging
## Acceptance criteria
All calls from Platform to Platform is using linkerd service mesh (is already installed) instead of external adresses.
## Development tasks
- [ ] All Platform-components should use linkerd service mesh for inter-platform communication.
- [ ] Storage
- [ ] Authorisation
- [ ] Authentication
- [ ] Profile
- [ ] Register
|
True
|
Enable pod-to-pod communication in Platform - ## Description
Performance testing shows that communication between components in Platform (as expected) goes through both Application Gateway and API Management.
This causes problems:
- Complex call stacks that is hard to reason about
- High latency and bad performance because of unneccesary network calls and TLS-termination
- Scaling/cost of network infrastructure (gateway and APIM) and unneccessary logging
## Acceptance criteria
All calls from Platform to Platform is using linkerd service mesh (is already installed) instead of external adresses.
## Development tasks
- [ ] All Platform-components should use linkerd service mesh for inter-platform communication.
- [ ] Storage
- [ ] Authorisation
- [ ] Authentication
- [ ] Profile
- [ ] Register
|
non_defect
|
enable pod to pod communication in platform description performance testing shows that communication between components in platform as expected goes through both application gateway and api management this causes problems complex call stacks that is hard to reason about high latency and bad performance because of unneccesary network calls and tls termination scaling cost of network infrastructure gateway and apim and unneccessary logging acceptance criteria all calls from platform to platform is using linkerd service mesh is already installed instead of external adresses development tasks all platform components should use linkerd service mesh for inter platform communication storage authorisation authentication profile register
| 0
|
64,526
| 18,724,551,141
|
IssuesEvent
|
2021-11-03 15:05:29
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Riot/Chrome state can result in inability to load new messages
|
T-Defect P1 S-Major
|
### Description
Riot in Chrome has gotten into a sate where only the messages that were received up to a certain point can be loaded. Trying to view a room with unread messages results in a loading spinner and eventual error:
> "Connectivity to the server has been lost.
> Sent messages will be stored until your connection has returned."
Trying to post a message seems to work, but the message is gone upon refreshing the page. Viewing a room with new messages (so that they are marked as read and the room isn't in bold anymore) is reset when refreshing the page. It seems nothing can be saved or loaded from after a certain point in time.
When logging in from an incognito window, the messages that were attempted to send but not visible after refreshing the page *did* in fact go through, but riot couldn't save them locally.
Have also had strange behavior in Firefox #5173
### Steps to reproduce
- Reproducibility unknown
Log: sent
### Version information
**Platform**: web (in-browser)
For the web app:
- **Browser**: Chrome Beta v62
- **OS**: Fedora 27
- **URL**: affecting /app /experimental and /develop
|
1.0
|
Riot/Chrome state can result in inability to load new messages - ### Description
Riot in Chrome has gotten into a sate where only the messages that were received up to a certain point can be loaded. Trying to view a room with unread messages results in a loading spinner and eventual error:
> "Connectivity to the server has been lost.
> Sent messages will be stored until your connection has returned."
Trying to post a message seems to work, but the message is gone upon refreshing the page. Viewing a room with new messages (so that they are marked as read and the room isn't in bold anymore) is reset when refreshing the page. It seems nothing can be saved or loaded from after a certain point in time.
When logging in from an incognito window, the messages that were attempted to send but not visible after refreshing the page *did* in fact go through, but riot couldn't save them locally.
Have also had strange behavior in Firefox #5173
### Steps to reproduce
- Reproducibility unknown
Log: sent
### Version information
**Platform**: web (in-browser)
For the web app:
- **Browser**: Chrome Beta v62
- **OS**: Fedora 27
- **URL**: affecting /app /experimental and /develop
|
defect
|
riot chrome state can result in inability to load new messages description riot in chrome has gotten into a sate where only the messages that were received up to a certain point can be loaded trying to view a room with unread messages results in a loading spinner and eventual error connectivity to the server has been lost sent messages will be stored until your connection has returned trying to post a message seems to work but the message is gone upon refreshing the page viewing a room with new messages so that they are marked as read and the room isn t in bold anymore is reset when refreshing the page it seems nothing can be saved or loaded from after a certain point in time when logging in from an incognito window the messages that were attempted to send but not visible after refreshing the page did in fact go through but riot couldn t save them locally have also had strange behavior in firefox steps to reproduce reproducibility unknown log sent version information platform web in browser for the web app browser chrome beta os fedora url affecting app experimental and develop
| 1
|
255,192
| 21,911,615,574
|
IssuesEvent
|
2022-05-21 06:05:19
|
MohistMC/Mohist
|
https://api.github.com/repos/MohistMC/Mohist
|
closed
|
[1.16.5] The bug that has been developing for some time has not been fixed.
|
1.16.5 Wait Needs Testing
|
## Informations
**The Minecraft Version** is 1.16.5
**The Mohist Version** is 990
**Attachments : **https://cdn.discordapp.com/attachments/968819857124626462/977153501153079316/latest.log, https://cdn.discordapp.com/attachments/968819857124626462/977153556211699742/latest.log
**Reported by miku1881 with id 320886072416665602**
## Description
## How to reproduce
Using the locate command
|
1.0
|
[1.16.5] The bug that has been developing for some time has not been fixed. - ## Informations
**The Minecraft Version** is 1.16.5
**The Mohist Version** is 990
**Attachments : **https://cdn.discordapp.com/attachments/968819857124626462/977153501153079316/latest.log, https://cdn.discordapp.com/attachments/968819857124626462/977153556211699742/latest.log
**Reported by miku1881 with id 320886072416665602**
## Description
## How to reproduce
Using the locate command
|
non_defect
|
the bug that has been developing for some time has not been fixed informations the minecraft version is the mohist version is attachments reported by with id description how to reproduce using the locate command
| 0
|
112,271
| 14,234,514,838
|
IssuesEvent
|
2020-11-18 13:42:07
|
gnosis/conditional-tokens-explorer
|
https://api.github.com/repos/gnosis/conditional-tokens-explorer
|
opened
|
Position id is not pre-selected when navigate to the Split position page from Positions list/Position details page
|
Medium priority New design bug
|
Related to #583, #515
See the video: position is not pre-selected when navigate to the Split position page (like it works on the current app version)
https://drive.google.com/file/d/1PPTw8Ky5f6eBSWBAM4BGf2utifJ_RozR/view
**Expected Result:** the position should be preselected when navigate to the lit position page from Positions list/Position details page
|
1.0
|
Position id is not pre-selected when navigate to the Split position page from Positions list/Position details page - Related to #583, #515
See the video: position is not pre-selected when navigate to the Split position page (like it works on the current app version)
https://drive.google.com/file/d/1PPTw8Ky5f6eBSWBAM4BGf2utifJ_RozR/view
**Expected Result:** the position should be preselected when navigate to the lit position page from Positions list/Position details page
|
non_defect
|
position id is not pre selected when navigate to the split position page from positions list position details page related to see the video position is not pre selected when navigate to the split position page like it works on the current app version expected result the position should be preselected when navigate to the lit position page from positions list position details page
| 0
|
41,870
| 10,685,750,498
|
IssuesEvent
|
2019-10-22 13:16:39
|
mozilla-lockwise/lockwise-android
|
https://api.github.com/repos/mozilla-lockwise/lockwise-android
|
opened
|
Dismissing opening website from long tap in website field takes user to Entry List view
|
type: defect
|
## Steps to reproduce
Open an entry detail view
Long tap on the url
Dismiss opening the website by tapping on any other part of the screen
### Expected behavior
User is taken back to the entry detail
### Actual behavior
User is taken to entry list view
### Device & build information
* Device: Pixel 3
* Build version: v3.0.0 RC-2
### Notes
Attachments: See [video](https://youtu.be/ue-D5GbZebs)
|
1.0
|
Dismissing opening website from long tap in website field takes user to Entry List view - ## Steps to reproduce
Open an entry detail view
Long tap on the url
Dismiss opening the website by tapping on any other part of the screen
### Expected behavior
User is taken back to the entry detail
### Actual behavior
User is taken to entry list view
### Device & build information
* Device: Pixel 3
* Build version: v3.0.0 RC-2
### Notes
Attachments: See [video](https://youtu.be/ue-D5GbZebs)
|
defect
|
dismissing opening website from long tap in website field takes user to entry list view steps to reproduce open an entry detail view long tap on the url dismiss opening the website by tapping on any other part of the screen expected behavior user is taken back to the entry detail actual behavior user is taken to entry list view device build information device pixel build version rc notes attachments see
| 1
|
200,176
| 7,001,034,631
|
IssuesEvent
|
2017-12-18 08:39:56
|
opentx/opentx
|
https://api.github.com/repos/opentx/opentx
|
closed
|
Inconsistent behavior for model conversions
|
Bug Companion Priority-Low
|
Scenario:
A model window is open. The model window has the same setting for radio type as the current profile (The normal case).
If the current profile is switched to a profile for another radio the models are automatically converted to the new radio type.
If the radio type in the current profiles is switched, the models are however not converted.
I can not see that either alternative is more logical than the other, but they are inconsistent.
|
1.0
|
Inconsistent behavior for model conversions - Scenario:
A model window is open. The model window has the same setting for radio type as the current profile (The normal case).
If the current profile is switched to a profile for another radio the models are automatically converted to the new radio type.
If the radio type in the current profiles is switched, the models are however not converted.
I can not see that either alternative is more logical than the other, but they are inconsistent.
|
non_defect
|
inconsistent behavior for model conversions scenario a model window is open the model window has the same setting for radio type as the current profile the normal case if the current profile is switched to a profile for another radio the models are automatically converted to the new radio type if the radio type in the current profiles is switched the models are however not converted i can not see that either alternative is more logical than the other but they are inconsistent
| 0
|
744
| 4,153,979,996
|
IssuesEvent
|
2016-06-16 09:48:37
|
ElderByte-/Warden
|
https://api.github.com/repos/ElderByte-/Warden
|
closed
|
Don't expose H2 console in embedded mode
|
type: architecture
|
In embedded mode, there is no guarantee which datasource the spring data repositories are using as they get picked up dynamically.
|
1.0
|
Don't expose H2 console in embedded mode - In embedded mode, there is no guarantee which datasource the spring data repositories are using as they get picked up dynamically.
|
non_defect
|
don t expose console in embedded mode in embedded mode there is no guarantee which datasource the spring data repositories are using as they get picked up dynamically
| 0
|
85,662
| 15,755,125,467
|
IssuesEvent
|
2021-03-31 01:13:46
|
tamirverthim/NodeGoat
|
https://api.github.com/repos/tamirverthim/NodeGoat
|
closed
|
CVE-2011-4969 (Medium) detected in jquery-1.4.4.min.js - autoclosed
|
security vulnerability
|
## CVE-2011-4969 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.4.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.4/jquery.min.js</a></p>
<p>Path to dependency file: NodeGoat/node_modules/selenium-webdriver/lib/test/data/droppableItems.html</p>
<p>Path to vulnerable library: NodeGoat/node_modules/selenium-webdriver/lib/test/data/js/jquery-1.4.4.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.4.4.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/tamirverthim/NodeGoat/commit/b122231e3322e09af113633cfa74ba5f9014c375">b122231e3322e09af113633cfa74ba5f9014c375</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Cross-site scripting (XSS) vulnerability in jQuery before 1.6.3, when using location.hash to select elements, allows remote attackers to inject arbitrary web script or HTML via a crafted tag.
<p>Publish Date: 2013-03-08
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2011-4969>CVE-2011-4969</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>4.3</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2011-4969">https://nvd.nist.gov/vuln/detail/CVE-2011-4969</a></p>
<p>Release Date: 2013-03-08</p>
<p>Fix Resolution: 1.6.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.4.4","isTransitiveDependency":false,"dependencyTree":"jquery:1.4.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.6.3"}],"vulnerabilityIdentifier":"CVE-2011-4969","vulnerabilityDetails":"Cross-site scripting (XSS) vulnerability in jQuery before 1.6.3, when using location.hash to select elements, allows remote attackers to inject arbitrary web script or HTML via a crafted tag.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2011-4969","cvss2Severity":"medium","cvss2Score":"4.3","extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2011-4969 (Medium) detected in jquery-1.4.4.min.js - autoclosed - ## CVE-2011-4969 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.4.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.4/jquery.min.js</a></p>
<p>Path to dependency file: NodeGoat/node_modules/selenium-webdriver/lib/test/data/droppableItems.html</p>
<p>Path to vulnerable library: NodeGoat/node_modules/selenium-webdriver/lib/test/data/js/jquery-1.4.4.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.4.4.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/tamirverthim/NodeGoat/commit/b122231e3322e09af113633cfa74ba5f9014c375">b122231e3322e09af113633cfa74ba5f9014c375</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Cross-site scripting (XSS) vulnerability in jQuery before 1.6.3, when using location.hash to select elements, allows remote attackers to inject arbitrary web script or HTML via a crafted tag.
<p>Publish Date: 2013-03-08
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2011-4969>CVE-2011-4969</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>4.3</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2011-4969">https://nvd.nist.gov/vuln/detail/CVE-2011-4969</a></p>
<p>Release Date: 2013-03-08</p>
<p>Fix Resolution: 1.6.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.4.4","isTransitiveDependency":false,"dependencyTree":"jquery:1.4.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.6.3"}],"vulnerabilityIdentifier":"CVE-2011-4969","vulnerabilityDetails":"Cross-site scripting (XSS) vulnerability in jQuery before 1.6.3, when using location.hash to select elements, allows remote attackers to inject arbitrary web script or HTML via a crafted tag.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2011-4969","cvss2Severity":"medium","cvss2Score":"4.3","extraData":{}}</REMEDIATE> -->
|
non_defect
|
cve medium detected in jquery min js autoclosed cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file nodegoat node modules selenium webdriver lib test data droppableitems html path to vulnerable library nodegoat node modules selenium webdriver lib test data js jquery min js dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details cross site scripting xss vulnerability in jquery before when using location hash to select elements allows remote attackers to inject arbitrary web script or html via a crafted tag publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails cross site scripting xss vulnerability in jquery before when using location hash to select elements allows remote attackers to inject arbitrary web script or html via a crafted tag vulnerabilityurl
| 0
|
369,233
| 10,894,230,826
|
IssuesEvent
|
2019-11-19 08:11:40
|
threefoldtech/jumpscaleX_core
|
https://api.github.com/repos/threefoldtech/jumpscaleX_core
|
opened
|
release Jumpscale 10.1
|
priority_critical type_feature
|
- [ ] make sure there are no branches left which are not on stories in home
- [ ] make sure all branches for 10.1 are merged into development
- [ ] run the autotests we have on development
- [ ] use jsx threebotbuilder ... to see if it works (builds)
- [ ] stories which cannot be finished in 10.1 should be discussed with stakeholders & moved to 10.2
- [ ] once tests done -> development -> master so we are ready for new development branch for 10.2
|
1.0
|
release Jumpscale 10.1 -
- [ ] make sure there are no branches left which are not on stories in home
- [ ] make sure all branches for 10.1 are merged into development
- [ ] run the autotests we have on development
- [ ] use jsx threebotbuilder ... to see if it works (builds)
- [ ] stories which cannot be finished in 10.1 should be discussed with stakeholders & moved to 10.2
- [ ] once tests done -> development -> master so we are ready for new development branch for 10.2
|
non_defect
|
release jumpscale make sure there are no branches left which are not on stories in home make sure all branches for are merged into development run the autotests we have on development use jsx threebotbuilder to see if it works builds stories which cannot be finished in should be discussed with stakeholders moved to once tests done development master so we are ready for new development branch for
| 0
|
158,693
| 13,745,784,746
|
IssuesEvent
|
2020-10-06 03:53:17
|
msandfor/10-Easy-Steps
|
https://api.github.com/repos/msandfor/10-Easy-Steps
|
closed
|
[Hacktoberfest]: Add "JSNation" to the Conferences Section
|
beginner beginner-friendly documentation :memo: first-contribution good first issue hacktoberfest :children_crossing: help wanted :hand: up-for-grabs
|
🆕🐥☝ Beginners Only:
This issue is reserved for people who are new to Open Source. We know that the process of creating a pull request is the biggest barrier for new contributors. This issue is for you 💝
To do:
* Add [JSNation](https://jsnation.com/) to the Conferences Section
* See below for link to step-by-step tutorial
* Go down to Conferences section - it starts at around line 173
* Using the following template, put in the correct info for that conference (Search for it on the internet to find the correct details) underneath the last row in the table, ensuring there is still an empty row before the next table.
`| [Conference Name](Website Address) | Conference Name or Organising Org | 10 | Description of the Conferences, found on their website |`
* Add your details in the contributors section at the bottom of the page using the same formatting as the last one
📋 *Step by Step*
[Step-By-Step Tutorial](https://github.com/msandfor/10-Easy-Steps/blob/master/docs/GitHub-browser-tutorial.md)
Remember to observe the [Code of Conduct](https://github.com/msandfor/10-Easy-Steps/blob/master/code_of_conduct.md) - by which we mean, in short - Be Kind, Be Positive, Be Helpful
|
1.0
|
[Hacktoberfest]: Add "JSNation" to the Conferences Section - 🆕🐥☝ Beginners Only:
This issue is reserved for people who are new to Open Source. We know that the process of creating a pull request is the biggest barrier for new contributors. This issue is for you 💝
To do:
* Add [JSNation](https://jsnation.com/) to the Conferences Section
* See below for link to step-by-step tutorial
* Go down to Conferences section - it starts at around line 173
* Using the following template, put in the correct info for that conference (Search for it on the internet to find the correct details) underneath the last row in the table, ensuring there is still an empty row before the next table.
`| [Conference Name](Website Address) | Conference Name or Organising Org | 10 | Description of the Conferences, found on their website |`
* Add your details in the contributors section at the bottom of the page using the same formatting as the last one
📋 *Step by Step*
[Step-By-Step Tutorial](https://github.com/msandfor/10-Easy-Steps/blob/master/docs/GitHub-browser-tutorial.md)
Remember to observe the [Code of Conduct](https://github.com/msandfor/10-Easy-Steps/blob/master/code_of_conduct.md) - by which we mean, in short - Be Kind, Be Positive, Be Helpful
|
non_defect
|
add jsnation to the conferences section 🆕🐥☝ beginners only this issue is reserved for people who are new to open source we know that the process of creating a pull request is the biggest barrier for new contributors this issue is for you 💝 to do add to the conferences section see below for link to step by step tutorial go down to conferences section it starts at around line using the following template put in the correct info for that conference search for it on the internet to find the correct details underneath the last row in the table ensuring there is still an empty row before the next table website address conference name or organising org description of the conferences found on their website add your details in the contributors section at the bottom of the page using the same formatting as the last one 📋 step by step remember to observe the by which we mean in short be kind be positive be helpful
| 0
|
276,450
| 23,992,930,780
|
IssuesEvent
|
2022-09-14 04:03:50
|
Tencent/bk-ci
|
https://api.github.com/repos/Tencent/bk-ci
|
closed
|
支持插件通过 BK_CI_START_TYPE 获取 stream 触发的流水线的启动方式
|
for test grayed streams/for test streams/for gray streams/grayed streams/done area/stream/backend
|
- 插件通过 BK_CI_START_TYPE 获取启动方式时,无论是 蓝盾触发还是 stream 触发,都应该能获取到启动方式
- 构建启动方式,MANUAL/TIME_TRIGGER/WEB_HOOK/SERVICE/PIPELINE/REMOTE中取值
注意:不改变 stream 原来的上下文(如ci.event)
修改方案:
将stream调用的启动方式收归成蓝盾的,以达到在蓝盾和stream触发时使用 BK_CI_START_TYPE 可以达到相同的参数
MANUAL --手动触发
TIME_TRIGGER --定时触发
WEB_HOOK --git触发,repo_hook
SERVICE --openapi触发
PIPELINE --子流水线调用
|
2.0
|
支持插件通过 BK_CI_START_TYPE 获取 stream 触发的流水线的启动方式 - - 插件通过 BK_CI_START_TYPE 获取启动方式时,无论是 蓝盾触发还是 stream 触发,都应该能获取到启动方式
- 构建启动方式,MANUAL/TIME_TRIGGER/WEB_HOOK/SERVICE/PIPELINE/REMOTE中取值
注意:不改变 stream 原来的上下文(如ci.event)
修改方案:
将stream调用的启动方式收归成蓝盾的,以达到在蓝盾和stream触发时使用 BK_CI_START_TYPE 可以达到相同的参数
MANUAL --手动触发
TIME_TRIGGER --定时触发
WEB_HOOK --git触发,repo_hook
SERVICE --openapi触发
PIPELINE --子流水线调用
|
non_defect
|
支持插件通过 bk ci start type 获取 stream 触发的流水线的启动方式 插件通过 bk ci start type 获取启动方式时,无论是 蓝盾触发还是 stream 触发,都应该能获取到启动方式 构建启动方式,manual time trigger web hook service pipeline remote中取值 注意:不改变 stream 原来的上下文(如ci event) 修改方案: 将stream调用的启动方式收归成蓝盾的,以达到在蓝盾和stream触发时使用 bk ci start type 可以达到相同的参数 manual 手动触发 time trigger 定时触发 web hook git触发,repo hook service openapi触发 pipeline 子流水线调用
| 0
|
31,165
| 6,443,901,663
|
IssuesEvent
|
2017-08-12 02:19:59
|
opendatakit/opendatakit
|
https://api.github.com/repos/opendatakit/opendatakit
|
closed
|
Failure to publish when user is a non-gmail.com account (i.e., Google apps domain)
|
1.4 Aggregate Priority-High Type-Defect
|
Originally reported on Google Code with ID 776
```
If you have a google-hosted domain, e.g., @camfed.org,
and you try to publish, publishing is apparently broken.
No error is captured in the log(!).
```
Reported by `mitchellsundt` on 2013-02-28 23:02:36
|
1.0
|
Failure to publish when user is a non-gmail.com account (i.e., Google apps domain) - Originally reported on Google Code with ID 776
```
If you have a google-hosted domain, e.g., @camfed.org,
and you try to publish, publishing is apparently broken.
No error is captured in the log(!).
```
Reported by `mitchellsundt` on 2013-02-28 23:02:36
|
defect
|
failure to publish when user is a non gmail com account i e google apps domain originally reported on google code with id if you have a google hosted domain e g camfed org and you try to publish publishing is apparently broken no error is captured in the log reported by mitchellsundt on
| 1
|
9,640
| 11,708,833,102
|
IssuesEvent
|
2020-03-08 15:30:22
|
SaucyPigeon/Wild-Cultivation-Fan-Update
|
https://api.github.com/repos/SaucyPigeon/Wild-Cultivation-Fan-Update
|
closed
|
Seeds Please
|
mod compatibility
|
Mod SaucyPigeon.WildCultivation has multiple Verse.BuildableDefs named Seed_Agave. Skipping.
Verse.Log:Error(String, Boolean)
Verse.DefDatabase`1:AddAllInMods()
System.Reflection.MonoMethod:InternalInvoke(Object, Object[], Exception&)
System.Reflection.MonoMethod:Invoke(Object, BindingFlags, Binder, Object[], CultureInfo)
System.Reflection.MethodBase:Invoke(Object, Object[])
Verse.GenGeneric:InvokeStaticMethodOnGenericType(Type, Type, String)
Verse.PlayDataLoader:DoPlayLoad()
Verse.PlayDataLoader:LoadAllPlayData(Boolean)
Verse.<>c:<Start>b__6_1()
Verse.LongEventHandler:RunEventFromAnotherThread(Action)
Verse.<>c:<UpdateCurrentAsynchronousEvent>b__27_0()
System.Threading.ThreadHelper:ThreadStart_Context(Object)
System.Threading.ExecutionContext:RunInternal(ExecutionContext, ContextCallback, Object, Boolean)
System.Threading.ExecutionContext:Run(ExecutionContext, ContextCallback, Object, Boolean)
System.Threading.ExecutionContext:Run(ExecutionContext, ContextCallback, Object)
System.Threading.ThreadHelper:ThreadStart()
Mod SaucyPigeon.WildCultivation has multiple Verse.ThingDefs named Seed_Agave. Skipping.
Verse.Log:Error(String, Boolean)
Verse.DefDatabase`1:AddAllInMods()
System.Reflection.MonoMethod:InternalInvoke(Object, Object[], Exception&)
System.Reflection.MonoMethod:Invoke(Object, BindingFlags, Binder, Object[], CultureInfo)
System.Reflection.MethodBase:Invoke(Object, Object[])
Verse.GenGeneric:InvokeStaticMethodOnGenericType(Type, Type, String)
Verse.PlayDataLoader:DoPlayLoad()
Verse.PlayDataLoader:LoadAllPlayData(Boolean)
Verse.<>c:<Start>b__6_1()
Verse.LongEventHandler:RunEventFromAnotherThread(Action)
Verse.<>c:<UpdateCurrentAsynchronousEvent>b__27_0()
System.Threading.ThreadHelper:ThreadStart_Context(Object)
System.Threading.ExecutionContext:RunInternal(ExecutionContext, ContextCallback, Object, Boolean)
System.Threading.ExecutionContext:Run(ExecutionContext, ContextCallback, Object, Boolean)
System.Threading.ExecutionContext:Run(ExecutionContext, ContextCallback, Object)
System.Threading.ThreadHelper:ThreadStart()
|
True
|
Seeds Please - Mod SaucyPigeon.WildCultivation has multiple Verse.BuildableDefs named Seed_Agave. Skipping.
Verse.Log:Error(String, Boolean)
Verse.DefDatabase`1:AddAllInMods()
System.Reflection.MonoMethod:InternalInvoke(Object, Object[], Exception&)
System.Reflection.MonoMethod:Invoke(Object, BindingFlags, Binder, Object[], CultureInfo)
System.Reflection.MethodBase:Invoke(Object, Object[])
Verse.GenGeneric:InvokeStaticMethodOnGenericType(Type, Type, String)
Verse.PlayDataLoader:DoPlayLoad()
Verse.PlayDataLoader:LoadAllPlayData(Boolean)
Verse.<>c:<Start>b__6_1()
Verse.LongEventHandler:RunEventFromAnotherThread(Action)
Verse.<>c:<UpdateCurrentAsynchronousEvent>b__27_0()
System.Threading.ThreadHelper:ThreadStart_Context(Object)
System.Threading.ExecutionContext:RunInternal(ExecutionContext, ContextCallback, Object, Boolean)
System.Threading.ExecutionContext:Run(ExecutionContext, ContextCallback, Object, Boolean)
System.Threading.ExecutionContext:Run(ExecutionContext, ContextCallback, Object)
System.Threading.ThreadHelper:ThreadStart()
Mod SaucyPigeon.WildCultivation has multiple Verse.ThingDefs named Seed_Agave. Skipping.
Verse.Log:Error(String, Boolean)
Verse.DefDatabase`1:AddAllInMods()
System.Reflection.MonoMethod:InternalInvoke(Object, Object[], Exception&)
System.Reflection.MonoMethod:Invoke(Object, BindingFlags, Binder, Object[], CultureInfo)
System.Reflection.MethodBase:Invoke(Object, Object[])
Verse.GenGeneric:InvokeStaticMethodOnGenericType(Type, Type, String)
Verse.PlayDataLoader:DoPlayLoad()
Verse.PlayDataLoader:LoadAllPlayData(Boolean)
Verse.<>c:<Start>b__6_1()
Verse.LongEventHandler:RunEventFromAnotherThread(Action)
Verse.<>c:<UpdateCurrentAsynchronousEvent>b__27_0()
System.Threading.ThreadHelper:ThreadStart_Context(Object)
System.Threading.ExecutionContext:RunInternal(ExecutionContext, ContextCallback, Object, Boolean)
System.Threading.ExecutionContext:Run(ExecutionContext, ContextCallback, Object, Boolean)
System.Threading.ExecutionContext:Run(ExecutionContext, ContextCallback, Object)
System.Threading.ThreadHelper:ThreadStart()
|
non_defect
|
seeds please mod saucypigeon wildcultivation has multiple verse buildabledefs named seed agave skipping verse log error string boolean verse defdatabase addallinmods system reflection monomethod internalinvoke object object exception system reflection monomethod invoke object bindingflags binder object cultureinfo system reflection methodbase invoke object object verse gengeneric invokestaticmethodongenerictype type type string verse playdataloader doplayload verse playdataloader loadallplaydata boolean verse c b verse longeventhandler runeventfromanotherthread action verse c b system threading threadhelper threadstart context object system threading executioncontext runinternal executioncontext contextcallback object boolean system threading executioncontext run executioncontext contextcallback object boolean system threading executioncontext run executioncontext contextcallback object system threading threadhelper threadstart mod saucypigeon wildcultivation has multiple verse thingdefs named seed agave skipping verse log error string boolean verse defdatabase addallinmods system reflection monomethod internalinvoke object object exception system reflection monomethod invoke object bindingflags binder object cultureinfo system reflection methodbase invoke object object verse gengeneric invokestaticmethodongenerictype type type string verse playdataloader doplayload verse playdataloader loadallplaydata boolean verse c b verse longeventhandler runeventfromanotherthread action verse c b system threading threadhelper threadstart context object system threading executioncontext runinternal executioncontext contextcallback object boolean system threading executioncontext run executioncontext contextcallback object boolean system threading executioncontext run executioncontext contextcallback object system threading threadhelper threadstart
| 0
|
67,621
| 21,032,324,465
|
IssuesEvent
|
2022-03-31 02:40:55
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Changing of theme from Dark mode to Light mode does not respond
|
T-Defect
|
### Steps to reproduce
https://user-images.githubusercontent.com/63562663/160947985-fabcfc38-b74b-4c4d-a879-24cc92a03210.mp4
### Outcome
#### What did you expect?
Change from Dark mode to Light mode
#### What happened instead?
Dark mode persisted.
### Operating system
Windows
### Browser information
Version 99.0.4844.84 (Official Build) (64-bit)
### URL for webapp
app.element.io
### Application version
Element version: 1.10.8 Olm version: 3.2.8
### Homeserver
_No response_
### Will you send logs?
No
|
1.0
|
Changing of theme from Dark mode to Light mode does not respond - ### Steps to reproduce
https://user-images.githubusercontent.com/63562663/160947985-fabcfc38-b74b-4c4d-a879-24cc92a03210.mp4
### Outcome
#### What did you expect?
Change from Dark mode to Light mode
#### What happened instead?
Dark mode persisted.
### Operating system
Windows
### Browser information
Version 99.0.4844.84 (Official Build) (64-bit)
### URL for webapp
app.element.io
### Application version
Element version: 1.10.8 Olm version: 3.2.8
### Homeserver
_No response_
### Will you send logs?
No
|
defect
|
changing of theme from dark mode to light mode does not respond steps to reproduce outcome what did you expect change from dark mode to light mode what happened instead dark mode persisted operating system windows browser information version official build bit url for webapp app element io application version element version olm version homeserver no response will you send logs no
| 1
|
179,341
| 6,624,126,739
|
IssuesEvent
|
2017-09-22 10:11:29
|
numbbo/coco
|
https://api.github.com/repos/numbbo/coco
|
opened
|
data archive filenames
|
Priority-High Usability
|
I found these names:
```
['bbob/2015-GECCO/Sif.tgz'],
['bbob/2015-GECCO/Sifeg.tgz'],
['bbob/2015-GECCO/Srr.tgz'],
```
and similar others in our data archive at `http://coco.gforge.inria.fr/data-archive`.
_Given the implicit naming conventions we have used over many years_, these names are IMHO unacceptable and _must_ change. A name cannot just give no indication whatsoever what is in the data set (given all other names do). Using the algorithm name and author has proven to be useful, but we can also try to come up with even more useful conventions.
On that note, `2017-outsideGECCO/` is not a nice folder name either and inconsistent vs `2017/` and `2015-GECCO/` (the latter is inconsistent with most others). How about `2017-no-workshop` and `2015`?
Rationale: There are only disadvantages to have cryptic filenames. Besides that informative filenames are useful for simple browsing, we are likely use the filename to search for entries of interest for some time to come, because it is rather tedious to actually implement different ways and start the maintenance for 150+ data sets.
|
1.0
|
data archive filenames - I found these names:
```
['bbob/2015-GECCO/Sif.tgz'],
['bbob/2015-GECCO/Sifeg.tgz'],
['bbob/2015-GECCO/Srr.tgz'],
```
and similar others in our data archive at `http://coco.gforge.inria.fr/data-archive`.
_Given the implicit naming conventions we have used over many years_, these names are IMHO unacceptable and _must_ change. A name cannot just give no indication whatsoever what is in the data set (given all other names do). Using the algorithm name and author has proven to be useful, but we can also try to come up with even more useful conventions.
On that note, `2017-outsideGECCO/` is not a nice folder name either and inconsistent vs `2017/` and `2015-GECCO/` (the latter is inconsistent with most others). How about `2017-no-workshop` and `2015`?
Rationale: There are only disadvantages to have cryptic filenames. Besides that informative filenames are useful for simple browsing, we are likely use the filename to search for entries of interest for some time to come, because it is rather tedious to actually implement different ways and start the maintenance for 150+ data sets.
|
non_defect
|
data archive filenames i found these names and similar others in our data archive at given the implicit naming conventions we have used over many years these names are imho unacceptable and must change a name cannot just give no indication whatsoever what is in the data set given all other names do using the algorithm name and author has proven to be useful but we can also try to come up with even more useful conventions on that note outsidegecco is not a nice folder name either and inconsistent vs and gecco the latter is inconsistent with most others how about no workshop and rationale there are only disadvantages to have cryptic filenames besides that informative filenames are useful for simple browsing we are likely use the filename to search for entries of interest for some time to come because it is rather tedious to actually implement different ways and start the maintenance for data sets
| 0
|
605,362
| 18,734,275,898
|
IssuesEvent
|
2021-11-04 04:04:02
|
AY2122S1-CS2103T-W13-2/tp
|
https://api.github.com/repos/AY2122S1-CS2103T-W13-2/tp
|
closed
|
Issue when deleting data while logging in
|
type.Bug priority.High
|
The problem here is when the user open the app with data, the login screen will show up. But if the data is deleted at this point, the password of the user will be used to create a new data file instead of throwing some error message.
|
1.0
|
Issue when deleting data while logging in - The problem here is when the user open the app with data, the login screen will show up. But if the data is deleted at this point, the password of the user will be used to create a new data file instead of throwing some error message.
|
non_defect
|
issue when deleting data while logging in the problem here is when the user open the app with data the login screen will show up but if the data is deleted at this point the password of the user will be used to create a new data file instead of throwing some error message
| 0
|
51,336
| 13,207,437,533
|
IssuesEvent
|
2020-08-14 23:05:56
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
opened
|
Monitor rate of big events in PnF (Trac #245)
|
Incomplete Migration Migrated from Trac defect jeb + pnf
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/245">https://code.icecube.wisc.edu/projects/icecube/ticket/245</a>, reported by blaufussand owned by tschmidt</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-02-29T18:02:27",
"_ts": "1456768947686083",
"description": "Recommendation of review panel.\n\n\n3) `Big Events issue\u2019: These are events which individually require long times (~10 seconds) for processing keep the PnF master node busy. If the rate of these events is high the system can become unstable. PnF would suffer serious issues when events with 45 s delays become frequent. Sophisticated reconstructions, such as currently used by online analyses, introduce delays. Currently the system can handle an increase in filter delays by a factor of 2, assuming an overall trigger rate of 2.9 kHz, if the number of client nodes is doubled. We recommend that the rate of these events is monitored. \u00a0\n",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"time": "2011-02-21T02:28:53",
"component": "jeb + pnf",
"summary": "Monitor rate of big events in PnF",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "tschmidt",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
Monitor rate of big events in PnF (Trac #245) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/245">https://code.icecube.wisc.edu/projects/icecube/ticket/245</a>, reported by blaufussand owned by tschmidt</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-02-29T18:02:27",
"_ts": "1456768947686083",
"description": "Recommendation of review panel.\n\n\n3) `Big Events issue\u2019: These are events which individually require long times (~10 seconds) for processing keep the PnF master node busy. If the rate of these events is high the system can become unstable. PnF would suffer serious issues when events with 45 s delays become frequent. Sophisticated reconstructions, such as currently used by online analyses, introduce delays. Currently the system can handle an increase in filter delays by a factor of 2, assuming an overall trigger rate of 2.9 kHz, if the number of client nodes is doubled. We recommend that the rate of these events is monitored. \u00a0\n",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"time": "2011-02-21T02:28:53",
"component": "jeb + pnf",
"summary": "Monitor rate of big events in PnF",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "tschmidt",
"type": "defect"
}
```
</p>
</details>
|
defect
|
monitor rate of big events in pnf trac migrated from json status closed changetime ts description recommendation of review panel n n big events issue these are events which individually require long times seconds for processing keep the pnf master node busy if the rate of these events is high the system can become unstable pnf would suffer serious issues when events with s delays become frequent sophisticated reconstructions such as currently used by online analyses introduce delays currently the system can handle an increase in filter delays by a factor of assuming an overall trigger rate of khz if the number of client nodes is doubled we recommend that the rate of these events is monitored n reporter blaufuss cc resolution fixed time component jeb pnf summary monitor rate of big events in pnf priority normal keywords milestone owner tschmidt type defect
| 1
|
826,890
| 31,716,996,015
|
IssuesEvent
|
2023-09-10 01:04:26
|
kubernetes/ingress-nginx
|
https://api.github.com/repos/kubernetes/ingress-nginx
|
closed
|
CVE-2022-41409 reported against controller-v1.8.1
|
kind/bug triage/accepted needs-priority
|
We are using X-ray version 3.76.7 and it reported CVE-2022-41409 (pcre2test vulnerability) against controller-v1.8.1 tag.
I have checked and think controller-v1.8.1 only used pcre but still, xray complains about that CVE.
|
1.0
|
CVE-2022-41409 reported against controller-v1.8.1 - We are using X-ray version 3.76.7 and it reported CVE-2022-41409 (pcre2test vulnerability) against controller-v1.8.1 tag.
I have checked and think controller-v1.8.1 only used pcre but still, xray complains about that CVE.
|
non_defect
|
cve reported against controller we are using x ray version and it reported cve vulnerability against controller tag i have checked and think controller only used pcre but still xray complains about that cve
| 0
|
451,643
| 13,039,692,167
|
IssuesEvent
|
2020-07-28 17:11:14
|
cds-snc/report-a-cybercrime
|
https://api.github.com/repos/cds-snc/report-a-cybercrime
|
closed
|
Missing .test.js for the following forms.
|
bug low priority
|
## Summary
Missing .test.js for the following forms: SuspectCluesForm, EvidenceInfoForm, ConfirmationForm
## Unresolved questions
> Are there any related issues you consider out of scope for this issue that could be addressed in the future?
|
1.0
|
Missing .test.js for the following forms. - ## Summary
Missing .test.js for the following forms: SuspectCluesForm, EvidenceInfoForm, ConfirmationForm
## Unresolved questions
> Are there any related issues you consider out of scope for this issue that could be addressed in the future?
|
non_defect
|
missing test js for the following forms summary missing test js for the following forms suspectcluesform evidenceinfoform confirmationform unresolved questions are there any related issues you consider out of scope for this issue that could be addressed in the future
| 0
|
164,877
| 26,038,885,035
|
IssuesEvent
|
2022-12-22 08:37:37
|
BreadGood-22/frontend
|
https://api.github.com/repos/BreadGood-22/frontend
|
closed
|
[design] 상품 등록 페이지 UI 구현
|
design
|
## ⭐ 주요 기능 <!-- 구현할 기능(목표)에 대한 간략한 설명 -->
상품 등록 페이지 UI 구현
## 📋 진행 사항
- [x] 상품 등록 페이지 레이아웃 구현
- [x] 이미지 파일 업로드 UI 구현
- [x] input창 Validation 및 input창 validation 통과시 버튼 활성화
## 📄 참고 사항
|
1.0
|
[design] 상품 등록 페이지 UI 구현 - ## ⭐ 주요 기능 <!-- 구현할 기능(목표)에 대한 간략한 설명 -->
상품 등록 페이지 UI 구현
## 📋 진행 사항
- [x] 상품 등록 페이지 레이아웃 구현
- [x] 이미지 파일 업로드 UI 구현
- [x] input창 Validation 및 input창 validation 통과시 버튼 활성화
## 📄 참고 사항
|
non_defect
|
상품 등록 페이지 ui 구현 ⭐ 주요 기능 상품 등록 페이지 ui 구현 📋 진행 사항 상품 등록 페이지 레이아웃 구현 이미지 파일 업로드 ui 구현 input창 validation 및 input창 validation 통과시 버튼 활성화 📄 참고 사항
| 0
|
237,621
| 19,661,463,322
|
IssuesEvent
|
2022-01-10 17:26:32
|
brave/brave-browser
|
https://api.github.com/repos/brave/brave-browser
|
closed
|
[Android] "Include in Auto-contribute" toggle doesn't stay on
|
bug feature/rewards priority/P2 QA/Yes QA/Test-Plan-Specified regression OS/Android rewards-ui-2.0
|
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description <!-- Provide a brief description of the issue -->
`Include in Auto-Contribute` toggle doesn't stay on
## Steps to reproduce <!-- Please add a series of steps to reproduce the issue -->
1. install Brave `1.35.40, Chromium 97.0.4692.45`
2. load `brave.com`
3. tap on the BAT icon in the URL bar
4. tap `Start using Brave Rewards`
3. tap on the BAT icon again
4. notice that `Include in Auto-Contribute` is toggled to `OFF`
5. toggle it to `ON`
6. tap away to dismiss the panel
7. tap again on the BAT icon
8. look at the value for `Include in Auto-Contribute`
## Actual result <!-- Please add screenshots if needed -->
It's back to `OFF`
<video src="https://user-images.githubusercontent.com/387249/146455999-9c4a2163-6401-4455-b743-8aa9463ad2bf.mp4">
## Expected result
It should be `ON`
## Issue reproduces how often <!-- [Easily reproduced/Intermittent issue/No steps to reproduce] -->
100%
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current Play Store version?
- Can you reproduce this issue with the current Play Store Beta version?
- Can you reproduce this issue with the current Play Store Nightly version?
## Device details
- Install type (ARM, x86): `arm64`
- Device type (Phone, Tablet, Phablet): Google Pixel XL, phone
- Android version: 9.0
## Brave version
Brave `1.35.40, Chromium 97.0.4692.45`
|
1.0
|
[Android] "Include in Auto-contribute" toggle doesn't stay on - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description <!-- Provide a brief description of the issue -->
`Include in Auto-Contribute` toggle doesn't stay on
## Steps to reproduce <!-- Please add a series of steps to reproduce the issue -->
1. install Brave `1.35.40, Chromium 97.0.4692.45`
2. load `brave.com`
3. tap on the BAT icon in the URL bar
4. tap `Start using Brave Rewards`
3. tap on the BAT icon again
4. notice that `Include in Auto-Contribute` is toggled to `OFF`
5. toggle it to `ON`
6. tap away to dismiss the panel
7. tap again on the BAT icon
8. look at the value for `Include in Auto-Contribute`
## Actual result <!-- Please add screenshots if needed -->
It's back to `OFF`
<video src="https://user-images.githubusercontent.com/387249/146455999-9c4a2163-6401-4455-b743-8aa9463ad2bf.mp4">
## Expected result
It should be `ON`
## Issue reproduces how often <!-- [Easily reproduced/Intermittent issue/No steps to reproduce] -->
100%
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current Play Store version?
- Can you reproduce this issue with the current Play Store Beta version?
- Can you reproduce this issue with the current Play Store Nightly version?
## Device details
- Install type (ARM, x86): `arm64`
- Device type (Phone, Tablet, Phablet): Google Pixel XL, phone
- Android version: 9.0
## Brave version
Brave `1.35.40, Chromium 97.0.4692.45`
|
non_defect
|
include in auto contribute toggle doesn t stay on have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description include in auto contribute toggle doesn t stay on steps to reproduce install brave chromium load brave com tap on the bat icon in the url bar tap start using brave rewards tap on the bat icon again notice that include in auto contribute is toggled to off toggle it to on tap away to dismiss the panel tap again on the bat icon look at the value for include in auto contribute actual result it s back to off video src expected result it should be on issue reproduces how often version channel information can you reproduce this issue with the current play store version can you reproduce this issue with the current play store beta version can you reproduce this issue with the current play store nightly version device details install type arm device type phone tablet phablet google pixel xl phone android version brave version brave chromium
| 0
|
6,023
| 3,727,356,897
|
IssuesEvent
|
2016-03-06 07:25:19
|
nunit/nunit
|
https://api.github.com/repos/nunit/nunit
|
closed
|
Add CF to the Appveyor CI build
|
is:build pri:low status:blocked
|
As indicated by #331 we keep breaking the compact framework build without realizing it. We need to at least compile it in our CI and if possible run the tests. Without installing a bunch of stuff in Appveyor, we'll probably only be able to run it using assembly unification under desktop .NET for now.
|
1.0
|
Add CF to the Appveyor CI build - As indicated by #331 we keep breaking the compact framework build without realizing it. We need to at least compile it in our CI and if possible run the tests. Without installing a bunch of stuff in Appveyor, we'll probably only be able to run it using assembly unification under desktop .NET for now.
|
non_defect
|
add cf to the appveyor ci build as indicated by we keep breaking the compact framework build without realizing it we need to at least compile it in our ci and if possible run the tests without installing a bunch of stuff in appveyor we ll probably only be able to run it using assembly unification under desktop net for now
| 0
|
237,934
| 7,768,433,377
|
IssuesEvent
|
2018-06-03 17:57:54
|
InfiniteFlightAirportEditing/Navigation
|
https://api.github.com/repos/InfiniteFlightAirportEditing/Navigation
|
opened
|
LICT
|
ILS Priority ILS Error
|
# Summary of Navigation issue: (ILS, Waypoints, etc.)
Runway moved/changed.
# Runway(s) with ILS: (If any)
31L
# References: (Official sources/charts)
http://opennav.com/pdf/LICT/LI_AD_2_LICT_5-3_en_2013-05-02.pdf
|
1.0
|
LICT - # Summary of Navigation issue: (ILS, Waypoints, etc.)
Runway moved/changed.
# Runway(s) with ILS: (If any)
31L
# References: (Official sources/charts)
http://opennav.com/pdf/LICT/LI_AD_2_LICT_5-3_en_2013-05-02.pdf
|
non_defect
|
lict summary of navigation issue ils waypoints etc runway moved changed runway s with ils if any references official sources charts
| 0
|
622,764
| 19,656,310,360
|
IssuesEvent
|
2022-01-10 12:54:11
|
epam/Indigo
|
https://api.github.com/repos/epam/Indigo
|
reopened
|
Add notebook for predicting assay values using Indigo fingerprints
|
Enhancement High priority python ML
|
**ToDo**
Prepare and place a notebook in api/python/notebooks/predict.ipynb
* Read input CSV file, with given column name for SMILES and column name for assay values.
* Prepare train and test datasets using shuffled data.
* Draw distribution plot for assay values.
* Calculate fingerprints for SMILES using Indigo.
* Fit and predict for two models: linear on your choice (SVM, RF) and deep learning mode on your choice. Print `r^2` value for predicting values for test dataset.
* Draw an interactive scatter plots using [Bokeh](https://docs.bokeh.org/en/2.4.1/docs/user_guide/jupyter.html). Use hover tool to show molecule images, actual and predicted value. Following plots should present:
* Actual vs Predicted
* Actual vs Residual (actual - predicted)
|
1.0
|
Add notebook for predicting assay values using Indigo fingerprints - **ToDo**
Prepare and place a notebook in api/python/notebooks/predict.ipynb
* Read input CSV file, with given column name for SMILES and column name for assay values.
* Prepare train and test datasets using shuffled data.
* Draw distribution plot for assay values.
* Calculate fingerprints for SMILES using Indigo.
* Fit and predict for two models: linear on your choice (SVM, RF) and deep learning mode on your choice. Print `r^2` value for predicting values for test dataset.
* Draw an interactive scatter plots using [Bokeh](https://docs.bokeh.org/en/2.4.1/docs/user_guide/jupyter.html). Use hover tool to show molecule images, actual and predicted value. Following plots should present:
* Actual vs Predicted
* Actual vs Residual (actual - predicted)
|
non_defect
|
add notebook for predicting assay values using indigo fingerprints todo prepare and place a notebook in api python notebooks predict ipynb read input csv file with given column name for smiles and column name for assay values prepare train and test datasets using shuffled data draw distribution plot for assay values calculate fingerprints for smiles using indigo fit and predict for two models linear on your choice svm rf and deep learning mode on your choice print r value for predicting values for test dataset draw an interactive scatter plots using use hover tool to show molecule images actual and predicted value following plots should present actual vs predicted actual vs residual actual predicted
| 0
|
379,386
| 26,369,252,140
|
IssuesEvent
|
2023-01-11 19:10:36
|
Tracktr/tracktr-client
|
https://api.github.com/repos/Tracktr/tracktr-client
|
closed
|
Verslag voor School
|
documentation
|
**Eisen:**
- [ ] 20-25 pagina’s
**To do:**
- [x] Uitgevoerde use cases
- [x] Geplande use cases
- [x] Database schema
- [x] Code log
- [x] Verwerkte feedback
- [x] Applicatie structuur
- [x] Beschrijving experimenten
- [ ] Designs (uitleggen)
- [ ] Opmaak (affinity/indesign)
|
1.0
|
Verslag voor School - **Eisen:**
- [ ] 20-25 pagina’s
**To do:**
- [x] Uitgevoerde use cases
- [x] Geplande use cases
- [x] Database schema
- [x] Code log
- [x] Verwerkte feedback
- [x] Applicatie structuur
- [x] Beschrijving experimenten
- [ ] Designs (uitleggen)
- [ ] Opmaak (affinity/indesign)
|
non_defect
|
verslag voor school eisen pagina’s to do uitgevoerde use cases geplande use cases database schema code log verwerkte feedback applicatie structuur beschrijving experimenten designs uitleggen opmaak affinity indesign
| 0
|
184,928
| 6,717,519,775
|
IssuesEvent
|
2017-10-14 22:20:00
|
dbcollection/dbcollection
|
https://api.github.com/repos/dbcollection/dbcollection
|
closed
|
Add docs for the datasets
|
docs medium priority
|
Must create a readme for each dataset. The readme must provide information about the dataset type, splits and overall data fields/organization.
|
1.0
|
Add docs for the datasets - Must create a readme for each dataset. The readme must provide information about the dataset type, splits and overall data fields/organization.
|
non_defect
|
add docs for the datasets must create a readme for each dataset the readme must provide information about the dataset type splits and overall data fields organization
| 0
|
29,055
| 5,515,033,103
|
IssuesEvent
|
2017-03-17 16:26:54
|
contao/news-bundle
|
https://api.github.com/repos/contao/news-bundle
|
closed
|
What is authorName and where is it coming from?
|
defect
|
I must be blind...can anybody explain me where this line takes the author name from? https://github.com/contao/news-bundle/blob/46256caff85ef2cd0b72b2235ed693a5b7876f53/src/Resources/contao/classes/News.php#L184
I cannot find the logic behind this.
|
1.0
|
What is authorName and where is it coming from? - I must be blind...can anybody explain me where this line takes the author name from? https://github.com/contao/news-bundle/blob/46256caff85ef2cd0b72b2235ed693a5b7876f53/src/Resources/contao/classes/News.php#L184
I cannot find the logic behind this.
|
defect
|
what is authorname and where is it coming from i must be blind can anybody explain me where this line takes the author name from i cannot find the logic behind this
| 1
|
7,880
| 2,611,056,242
|
IssuesEvent
|
2015-02-27 00:25:53
|
alistairreilly/andors-trail
|
https://api.github.com/repos/alistairreilly/andors-trail
|
closed
|
Suggestion: "sell all" option
|
auto-migrated Priority-Medium Type-Defect
|
```
I just found that I have over 300 of "Animal hair" :) selling it one by one
it's a nightmare. Maybe long press on "Sell" button should trigger "Sell all"
function with confirmation dialog?
Additionally, even selling 10 of items can produce a problem, when clicking to
quick you can accidentally sell an item below that one.
```
Original issue reported on code.google.com by `R3pl...@gmail.com` on 3 Jan 2011 at 6:45
|
1.0
|
Suggestion: "sell all" option - ```
I just found that I have over 300 of "Animal hair" :) selling it one by one
it's a nightmare. Maybe long press on "Sell" button should trigger "Sell all"
function with confirmation dialog?
Additionally, even selling 10 of items can produce a problem, when clicking to
quick you can accidentally sell an item below that one.
```
Original issue reported on code.google.com by `R3pl...@gmail.com` on 3 Jan 2011 at 6:45
|
defect
|
suggestion sell all option i just found that i have over of animal hair selling it one by one it s a nightmare maybe long press on sell button should trigger sell all function with confirmation dialog additionally even selling of items can produce a problem when clicking to quick you can accidentally sell an item below that one original issue reported on code google com by gmail com on jan at
| 1
|
67,115
| 20,909,175,673
|
IssuesEvent
|
2022-03-24 07:29:12
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Can't load message: Invalid `href`
|
T-Defect S-Minor Z-Soft-Crash X-Needs-Investigation O-Uncommon
|
### Steps to reproduce
Event source;
```json5
{
"content": {
"body": "I could just make an alias like #matrix:non-matrix.org and host a phising Element at element.non-matrix.org. Any server can add aliases to a room or appear similar enough, so not sure that is a good idea.",
"format": "org.matrix.custom.html",
"formatted_body": "I could just make an alias like #<a href=\"matrix:non-matrix.org\">matrix:non-matrix.org</a> and host a phising Element at element.non-matrix.org. Any server can add aliases to a room or appear similar enough, so not sure that is a good idea.",
"msgtype": "m.text"
},
"type": "m.room.message",
"unsigned": {},
// Some event keys redacted
}
```
I believe the culprit is `#<a href=\"matrix:non-matrix.org\">matrix:non-matrix.org</a>`
### Outcome
#### What did you expect?
The message to maybe display with a fallback, or mangled up a bit.
#### What happened instead?

### Operating system
PopOS 21.10
### Application version
Element version: 1.10.6 Olm version: 3.2.8
### How did you install the app?
Flatpak
### Homeserver
jboi.nl
### Will you send logs?
No
|
1.0
|
Can't load message: Invalid `href` - ### Steps to reproduce
Event source;
```json5
{
"content": {
"body": "I could just make an alias like #matrix:non-matrix.org and host a phising Element at element.non-matrix.org. Any server can add aliases to a room or appear similar enough, so not sure that is a good idea.",
"format": "org.matrix.custom.html",
"formatted_body": "I could just make an alias like #<a href=\"matrix:non-matrix.org\">matrix:non-matrix.org</a> and host a phising Element at element.non-matrix.org. Any server can add aliases to a room or appear similar enough, so not sure that is a good idea.",
"msgtype": "m.text"
},
"type": "m.room.message",
"unsigned": {},
// Some event keys redacted
}
```
I believe the culprit is `#<a href=\"matrix:non-matrix.org\">matrix:non-matrix.org</a>`
### Outcome
#### What did you expect?
The message to maybe display with a fallback, or mangled up a bit.
#### What happened instead?

### Operating system
PopOS 21.10
### Application version
Element version: 1.10.6 Olm version: 3.2.8
### How did you install the app?
Flatpak
### Homeserver
jboi.nl
### Will you send logs?
No
|
defect
|
can t load message invalid href steps to reproduce event source content body i could just make an alias like matrix non matrix org and host a phising element at element non matrix org any server can add aliases to a room or appear similar enough so not sure that is a good idea format org matrix custom html formatted body i could just make an alias like matrix non matrix org and host a phising element at element non matrix org any server can add aliases to a room or appear similar enough so not sure that is a good idea msgtype m text type m room message unsigned some event keys redacted i believe the culprit is matrix non matrix org outcome what did you expect the message to maybe display with a fallback or mangled up a bit what happened instead operating system popos application version element version olm version how did you install the app flatpak homeserver jboi nl will you send logs no
| 1
|
64,383
| 18,531,492,998
|
IssuesEvent
|
2021-10-21 06:41:03
|
gitblit/gitblit
|
https://api.github.com/repos/gitblit/gitblit
|
closed
|
Bad encoding in properties-files
|
Priority-Medium Catg-Defect Status-Verified
|
Originally reported on Google Code with ID 538
```
What steps will reproduce the problem?
1. Deploy gitblit on tomcat6 or 7 running on a platform supporting utf-8.
My example is dev.gitblit.com
2. Access the site with a browser specifying Norwegian as the Language
(or via curl:
$ curl -H"Accept-Language: no,nb-NO;q=0.8,nn-NO;q=0.5,en-US;q=0.3" https://dev.gitblit.com/
-I
HTTP/1.1 500 Server Error
Date: Tue, 25 Nov 2014 13:44:16 GMT
Content-Type: text/html; charset=UTF-8
Content-Language: no
Pragma: no-cache
Cache-Control: no-cache, max-age=0, must-revalidate
Content-Length: 1406
Server: Jetty(9.2.3.v20140905)
3. See the internal error
What is the expected output? What do you see instead?
I see the attached image - internal error. In the logs I'm spammed with this output:
2014-11-25 14:07:25 [INFO ] Loading properties files from jar:file:/data/prog/ap
ache-tomcat-7.0.29/webapps/gitblit/WEB-INF/lib/gitblit.jar!/com/gitblit/wicket/GitBlitWebApp_no.properties
2014-11-25 14:07:25 [ERROR] Can't instantiate page using constructor public com.
gitblit.wicket.pages.MyDashboardPage()
org.apache.wicket.WicketRuntimeException: Can't instantiate page using constructor
public com.gitblit.wicket.pages.MyDashboardPage()
at org.apache.wicket.session.DefaultPageFactory.createPage(DefaultPageFactory.java:212)
at org.apache.wicket.session.DefaultPageFactory.newPage(DefaultPageFactory.java:57)
at org.apache.wicket.request.target.component.BookmarkablePageRequestTar
get.newPage(BookmarkablePageRequestTarget.java:298)
at org.apache.wicket.request.target.component.BookmarkablePageRequestTar
get.getPage(BookmarkablePageRequestTarget.java:320)
at org.apache.wicket.request.target.component.BookmarkablePageRequestTar
get.processEvents(BookmarkablePageRequestTarget.java:234)
at org.apache.wicket.request.AbstractRequestCycleProcessor.processEvents
(AbstractRequestCycleProcessor.java:92)
at org.apache.wicket.RequestCycle.processEventsAndRespond(RequestCycle.j
ava:1279)
at org.apache.wicket.RequestCycle.step(RequestCycle.java:1358)
at org.apache.wicket.RequestCycle.steps(RequestCycle.java:1465)
at org.apache.wicket.RequestCycle.request(RequestCycle.java:545)
at org.apache.wicket.protocol.http.WicketFilter.doGet(WicketFilter.java:
486)
at org.apache.wicket.protocol.http.WicketFilter.doFilter(WicketFilter.ja
va:319)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(Appl
icationFilterChain.java:243)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationF
ilterChain.java:210)
at com.gitblit.servlet.EnforceAuthenticationFilter.doFilter(EnforceAuthe
nticationFilter.java:87)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(Appl
icationFilterChain.java:243)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationF
ilterChain.java:210)
at com.gitblit.servlet.ProxyFilter$1.doFilter(ProxyFilter.java:74)
at com.gitblit.servlet.ProxyFilter.doFilter(ProxyFilter.java:77)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(Appl
icationFilterChain.java:243)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationF
ilterChain.java:210)
at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperV
alve.java:225)
at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:123)
at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:472)
at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:168)
at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:98)
at org.apache.catalina.valves.AccessLogValve.invoke(AccessLogValve.java:927)
at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:118)
at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:407)
at org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1001)
at org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:585)
at org.apache.tomcat.util.net.JIoEndpoint$SocketProcessor.run(JIoEndpoint.java:310)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.reflect.InvocationTargetException
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
at org.apache.wicket.session.DefaultPageFactory.createPage(DefaultPageFactory.java:192)
... 34 more
Caused by: java.lang.IllegalArgumentException: Malformed \uxxxx encoding.
at java.util.Properties.loadConvert(Properties.java:568)
at java.util.Properties.load0(Properties.java:391)
at java.util.Properties.load(Properties.java:341)
at org.apache.wicket.resource.PropertiesFactory$PropertiesFilePropertiesLoader.loadProperties(PropertiesFactory.java:343)
at org.apache.wicket.resource.PropertiesFactory$AbstractPropertiesLoader.load(PropertiesFactory.java:274)
at org.apache.wicket.resource.PropertiesFactory.load(PropertiesFactory.java:133)
at org.apache.wicket.resource.loader.ComponentStringResourceLoader.loadStringResource(ComponentStringResourceLoader.java:141)
at org.apache.wicket.resource.loader.ClassStringResourceLoader.loadStringResource(ClassStringResourceLoader.java:65)
at org.apache.wicket.resource.loader.ComponentStringResourceLoader.loadStringResource(ComponentStringResourceLoader.java:261)
at org.apache.wicket.resource.loader.ClassStringResourceLoader.loadStringResource(ClassStringResourceLoader.java:80)
at org.apache.wicket.Localizer.getStringIgnoreSettings(Localizer.java:241)
at org.apache.wicket.Localizer.getString(Localizer.java:313)
at org.apache.wicket.Localizer.getString(Localizer.java:119)
at org.apache.wicket.Component.getString(Component.java:1968)
at org.apache.wicket.Component.getString(Component.java:1955)
at com.gitblit.wicket.pages.RootPage$LoginForm.<init>(RootPage.java:581)
at com.gitblit.wicket.pages.RootPage.setupPage(RootPage.java:176)
at com.gitblit.wicket.pages.MyDashboardPage.setup(MyDashboardPage.java:71)
at com.gitblit.wicket.pages.MyDashboardPage.<init>(MyDashboardPage.java:57)
... 39 more
What version of the product are you using? On what operating system?
1.6.2, running on
server:/data/prog/apache-tomcat-7.0.29/logs> uname -a
Linux server 2.6.32-279.1.1.el6.x86_64 #1 SMP Wed Jun 20 11:41:22 EDT 2012 x86_64 x86_64
x86_64 GNU/Linux
Please provide any additional information below.
The file it's complaining about has alle Language-specific characters (æ,ø and å) encoded
as \u00xx, which corresponds to ISO-8859-1 encoding. The correct encoding is to use
Unicode.
```
Reported by `thomas.arp` on 2014-11-25 13:31:24
-----
- _Attachment: gitblit_doesnt_work_with_norwegian_and_utf8.png<br>_
_Edit: Make image render_
|
1.0
|
Bad encoding in properties-files - Originally reported on Google Code with ID 538
```
What steps will reproduce the problem?
1. Deploy gitblit on tomcat6 or 7 running on a platform supporting utf-8.
My example is dev.gitblit.com
2. Access the site with a browser specifying Norwegian as the Language
(or via curl:
$ curl -H"Accept-Language: no,nb-NO;q=0.8,nn-NO;q=0.5,en-US;q=0.3" https://dev.gitblit.com/
-I
HTTP/1.1 500 Server Error
Date: Tue, 25 Nov 2014 13:44:16 GMT
Content-Type: text/html; charset=UTF-8
Content-Language: no
Pragma: no-cache
Cache-Control: no-cache, max-age=0, must-revalidate
Content-Length: 1406
Server: Jetty(9.2.3.v20140905)
3. See the internal error
What is the expected output? What do you see instead?
I see the attached image - internal error. In the logs I'm spammed with this output:
2014-11-25 14:07:25 [INFO ] Loading properties files from jar:file:/data/prog/ap
ache-tomcat-7.0.29/webapps/gitblit/WEB-INF/lib/gitblit.jar!/com/gitblit/wicket/GitBlitWebApp_no.properties
2014-11-25 14:07:25 [ERROR] Can't instantiate page using constructor public com.
gitblit.wicket.pages.MyDashboardPage()
org.apache.wicket.WicketRuntimeException: Can't instantiate page using constructor
public com.gitblit.wicket.pages.MyDashboardPage()
at org.apache.wicket.session.DefaultPageFactory.createPage(DefaultPageFactory.java:212)
at org.apache.wicket.session.DefaultPageFactory.newPage(DefaultPageFactory.java:57)
at org.apache.wicket.request.target.component.BookmarkablePageRequestTar
get.newPage(BookmarkablePageRequestTarget.java:298)
at org.apache.wicket.request.target.component.BookmarkablePageRequestTar
get.getPage(BookmarkablePageRequestTarget.java:320)
at org.apache.wicket.request.target.component.BookmarkablePageRequestTar
get.processEvents(BookmarkablePageRequestTarget.java:234)
at org.apache.wicket.request.AbstractRequestCycleProcessor.processEvents
(AbstractRequestCycleProcessor.java:92)
at org.apache.wicket.RequestCycle.processEventsAndRespond(RequestCycle.j
ava:1279)
at org.apache.wicket.RequestCycle.step(RequestCycle.java:1358)
at org.apache.wicket.RequestCycle.steps(RequestCycle.java:1465)
at org.apache.wicket.RequestCycle.request(RequestCycle.java:545)
at org.apache.wicket.protocol.http.WicketFilter.doGet(WicketFilter.java:
486)
at org.apache.wicket.protocol.http.WicketFilter.doFilter(WicketFilter.ja
va:319)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(Appl
icationFilterChain.java:243)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationF
ilterChain.java:210)
at com.gitblit.servlet.EnforceAuthenticationFilter.doFilter(EnforceAuthe
nticationFilter.java:87)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(Appl
icationFilterChain.java:243)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationF
ilterChain.java:210)
at com.gitblit.servlet.ProxyFilter$1.doFilter(ProxyFilter.java:74)
at com.gitblit.servlet.ProxyFilter.doFilter(ProxyFilter.java:77)
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(Appl
icationFilterChain.java:243)
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationF
ilterChain.java:210)
at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperV
alve.java:225)
at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:123)
at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:472)
at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:168)
at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:98)
at org.apache.catalina.valves.AccessLogValve.invoke(AccessLogValve.java:927)
at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:118)
at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:407)
at org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1001)
at org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:585)
at org.apache.tomcat.util.net.JIoEndpoint$SocketProcessor.run(JIoEndpoint.java:310)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.reflect.InvocationTargetException
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
at org.apache.wicket.session.DefaultPageFactory.createPage(DefaultPageFactory.java:192)
... 34 more
Caused by: java.lang.IllegalArgumentException: Malformed \uxxxx encoding.
at java.util.Properties.loadConvert(Properties.java:568)
at java.util.Properties.load0(Properties.java:391)
at java.util.Properties.load(Properties.java:341)
at org.apache.wicket.resource.PropertiesFactory$PropertiesFilePropertiesLoader.loadProperties(PropertiesFactory.java:343)
at org.apache.wicket.resource.PropertiesFactory$AbstractPropertiesLoader.load(PropertiesFactory.java:274)
at org.apache.wicket.resource.PropertiesFactory.load(PropertiesFactory.java:133)
at org.apache.wicket.resource.loader.ComponentStringResourceLoader.loadStringResource(ComponentStringResourceLoader.java:141)
at org.apache.wicket.resource.loader.ClassStringResourceLoader.loadStringResource(ClassStringResourceLoader.java:65)
at org.apache.wicket.resource.loader.ComponentStringResourceLoader.loadStringResource(ComponentStringResourceLoader.java:261)
at org.apache.wicket.resource.loader.ClassStringResourceLoader.loadStringResource(ClassStringResourceLoader.java:80)
at org.apache.wicket.Localizer.getStringIgnoreSettings(Localizer.java:241)
at org.apache.wicket.Localizer.getString(Localizer.java:313)
at org.apache.wicket.Localizer.getString(Localizer.java:119)
at org.apache.wicket.Component.getString(Component.java:1968)
at org.apache.wicket.Component.getString(Component.java:1955)
at com.gitblit.wicket.pages.RootPage$LoginForm.<init>(RootPage.java:581)
at com.gitblit.wicket.pages.RootPage.setupPage(RootPage.java:176)
at com.gitblit.wicket.pages.MyDashboardPage.setup(MyDashboardPage.java:71)
at com.gitblit.wicket.pages.MyDashboardPage.<init>(MyDashboardPage.java:57)
... 39 more
What version of the product are you using? On what operating system?
1.6.2, running on
server:/data/prog/apache-tomcat-7.0.29/logs> uname -a
Linux server 2.6.32-279.1.1.el6.x86_64 #1 SMP Wed Jun 20 11:41:22 EDT 2012 x86_64 x86_64
x86_64 GNU/Linux
Please provide any additional information below.
The file it's complaining about has alle Language-specific characters (æ,ø and å) encoded
as \u00xx, which corresponds to ISO-8859-1 encoding. The correct encoding is to use
Unicode.
```
Reported by `thomas.arp` on 2014-11-25 13:31:24
-----
- _Attachment: gitblit_doesnt_work_with_norwegian_and_utf8.png<br>_
_Edit: Make image render_
|
defect
|
bad encoding in properties files originally reported on google code with id what steps will reproduce the problem deploy gitblit on or running on a platform supporting utf my example is dev gitblit com access the site with a browser specifying norwegian as the language or via curl curl h accept language no nb no q nn no q en us q i http server error date tue nov gmt content type text html charset utf content language no pragma no cache cache control no cache max age must revalidate content length server jetty see the internal error what is the expected output what do you see instead i see the attached image internal error in the logs i m spammed with this output loading properties files from jar file data prog ap ache tomcat webapps gitblit web inf lib gitblit jar com gitblit wicket gitblitwebapp no properties can t instantiate page using constructor public com gitblit wicket pages mydashboardpage org apache wicket wicketruntimeexception can t instantiate page using constructor public com gitblit wicket pages mydashboardpage at org apache wicket session defaultpagefactory createpage defaultpagefactory java at org apache wicket session defaultpagefactory newpage defaultpagefactory java at org apache wicket request target component bookmarkablepagerequesttar get newpage bookmarkablepagerequesttarget java at org apache wicket request target component bookmarkablepagerequesttar get getpage bookmarkablepagerequesttarget java at org apache wicket request target component bookmarkablepagerequesttar get processevents bookmarkablepagerequesttarget java at org apache wicket request abstractrequestcycleprocessor processevents abstractrequestcycleprocessor java at org apache wicket requestcycle processeventsandrespond requestcycle j ava at org apache wicket requestcycle step requestcycle java at org apache wicket requestcycle steps requestcycle java at org apache wicket requestcycle request requestcycle java at org apache wicket protocol http wicketfilter doget wicketfilter java at org apache wicket protocol http wicketfilter dofilter wicketfilter ja va at org apache catalina core applicationfilterchain internaldofilter appl icationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationf ilterchain java at com gitblit servlet enforceauthenticationfilter dofilter enforceauthe nticationfilter java at org apache catalina core applicationfilterchain internaldofilter appl icationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationf ilterchain java at com gitblit servlet proxyfilter dofilter proxyfilter java at com gitblit servlet proxyfilter dofilter proxyfilter java at org apache catalina core applicationfilterchain internaldofilter appl icationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationf ilterchain java at org apache catalina core standardwrappervalve invoke standardwrapperv alve java at org apache catalina core standardcontextvalve invoke standardcontextvalve java at org apache catalina authenticator authenticatorbase invoke authenticatorbase java at org apache catalina core standardhostvalve invoke standardhostvalve java at org apache catalina valves errorreportvalve invoke errorreportvalve java at org apache catalina valves accesslogvalve invoke accesslogvalve java at org apache catalina core standardenginevalve invoke standardenginevalve java at org apache catalina connector coyoteadapter service coyoteadapter java at org apache coyote process java at org apache coyote abstractprotocol abstractconnectionhandler process abstractprotocol java at org apache tomcat util net jioendpoint socketprocessor run jioendpoint java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java caused by java lang reflect invocationtargetexception at sun reflect nativeconstructoraccessorimpl native method at sun reflect nativeconstructoraccessorimpl newinstance nativeconstructoraccessorimpl java at sun reflect delegatingconstructoraccessorimpl newinstance delegatingconstructoraccessorimpl java at java lang reflect constructor newinstance constructor java at org apache wicket session defaultpagefactory createpage defaultpagefactory java more caused by java lang illegalargumentexception malformed uxxxx encoding at java util properties loadconvert properties java at java util properties properties java at java util properties load properties java at org apache wicket resource propertiesfactory propertiesfilepropertiesloader loadproperties propertiesfactory java at org apache wicket resource propertiesfactory abstractpropertiesloader load propertiesfactory java at org apache wicket resource propertiesfactory load propertiesfactory java at org apache wicket resource loader componentstringresourceloader loadstringresource componentstringresourceloader java at org apache wicket resource loader classstringresourceloader loadstringresource classstringresourceloader java at org apache wicket resource loader componentstringresourceloader loadstringresource componentstringresourceloader java at org apache wicket resource loader classstringresourceloader loadstringresource classstringresourceloader java at org apache wicket localizer getstringignoresettings localizer java at org apache wicket localizer getstring localizer java at org apache wicket localizer getstring localizer java at org apache wicket component getstring component java at org apache wicket component getstring component java at com gitblit wicket pages rootpage loginform rootpage java at com gitblit wicket pages rootpage setuppage rootpage java at com gitblit wicket pages mydashboardpage setup mydashboardpage java at com gitblit wicket pages mydashboardpage mydashboardpage java more what version of the product are you using on what operating system running on server data prog apache tomcat logs uname a linux server smp wed jun edt gnu linux please provide any additional information below the file it s complaining about has alle language specific characters æ ø and å encoded as which corresponds to iso encoding the correct encoding is to use unicode reported by thomas arp on attachment gitblit doesnt work with norwegian and png edit make image render
| 1
|
7,417
| 7,925,247,327
|
IssuesEvent
|
2018-07-05 19:55:08
|
OriginProtocol/origin-bridge
|
https://api.github.com/repos/OriginProtocol/origin-bridge
|
opened
|
Purchase Timestamp Data Types
|
help wanted indexing service python
|
Purchase timestamps are currently being returned by the API as [ISO strings](https://en.wikipedia.org/wiki/ISO_8601). They should be returned as seconds since the epoch, since they are integers on the blockchain and [Origin.js expects to return datetimes](http://docs.originprotocol.com/#purchase-get). It would probably also be preferable for the field names to match, such as `created` instead of `created_at`.
|
1.0
|
Purchase Timestamp Data Types - Purchase timestamps are currently being returned by the API as [ISO strings](https://en.wikipedia.org/wiki/ISO_8601). They should be returned as seconds since the epoch, since they are integers on the blockchain and [Origin.js expects to return datetimes](http://docs.originprotocol.com/#purchase-get). It would probably also be preferable for the field names to match, such as `created` instead of `created_at`.
|
non_defect
|
purchase timestamp data types purchase timestamps are currently being returned by the api as they should be returned as seconds since the epoch since they are integers on the blockchain and it would probably also be preferable for the field names to match such as created instead of created at
| 0
|
67,523
| 20,975,318,476
|
IssuesEvent
|
2022-03-28 14:46:23
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
opened
|
Sliver of a room visible from spotlight search breadcrumbs
|
T-Defect
|
### Steps to reproduce
1. Use spotlight search
2. Click a bunch of rooms
3. Open spotlight
4. Hover on the right hand side
### Outcome
#### What did you expect?
The number of rooms to fit perfectly in the dialog
#### What happened instead?

( some rooms blurred for privacy)
### Operating system
Windows 10
### Application version
Nightly (2021-03-27)
### How did you install the app?
The Internet
### Homeserver
t2l.io
### Will you send logs?
No
|
1.0
|
Sliver of a room visible from spotlight search breadcrumbs - ### Steps to reproduce
1. Use spotlight search
2. Click a bunch of rooms
3. Open spotlight
4. Hover on the right hand side
### Outcome
#### What did you expect?
The number of rooms to fit perfectly in the dialog
#### What happened instead?

( some rooms blurred for privacy)
### Operating system
Windows 10
### Application version
Nightly (2021-03-27)
### How did you install the app?
The Internet
### Homeserver
t2l.io
### Will you send logs?
No
|
defect
|
sliver of a room visible from spotlight search breadcrumbs steps to reproduce use spotlight search click a bunch of rooms open spotlight hover on the right hand side outcome what did you expect the number of rooms to fit perfectly in the dialog what happened instead some rooms blurred for privacy operating system windows application version nightly how did you install the app the internet homeserver io will you send logs no
| 1
|
31,322
| 6,499,641,701
|
IssuesEvent
|
2017-08-22 22:32:51
|
opencaching/opencaching-pl
|
https://api.github.com/repos/opencaching/opencaching-pl
|
closed
|
Incorrect icon and description of blocked by COG cache - in powerTrail.php
|
Component_GeoPaths Priority_Low Status: Waiting for translations Type_Defect
|
Server generates warning:
`[Tue Aug 02 21:19:07.234928 2016] [:error] [pid 7118] [client 89.65.18.106:50976] PHP Notice: Undefined offset: 6 in /srv/ocpl/powerTrail/ajaxGetPowerTrailCaches.php on line 111, referer: http://opencaching.pl/powerTrail.php?ptAction=showSerie&ptrail=334`
|
1.0
|
Incorrect icon and description of blocked by COG cache - in powerTrail.php - Server generates warning:
`[Tue Aug 02 21:19:07.234928 2016] [:error] [pid 7118] [client 89.65.18.106:50976] PHP Notice: Undefined offset: 6 in /srv/ocpl/powerTrail/ajaxGetPowerTrailCaches.php on line 111, referer: http://opencaching.pl/powerTrail.php?ptAction=showSerie&ptrail=334`
|
defect
|
incorrect icon and description of blocked by cog cache in powertrail php server generates warning php notice undefined offset in srv ocpl powertrail ajaxgetpowertrailcaches php on line referer
| 1
|
44,404
| 5,814,219,136
|
IssuesEvent
|
2017-05-05 02:19:22
|
wiebj/deep-breath
|
https://api.github.com/repos/wiebj/deep-breath
|
closed
|
Evaluate data stores for temporal-spatial data
|
design evaluation
|
Evaluate popular data stores (SQL & NoSQL) for temporal-spatial data
|
1.0
|
Evaluate data stores for temporal-spatial data - Evaluate popular data stores (SQL & NoSQL) for temporal-spatial data
|
non_defect
|
evaluate data stores for temporal spatial data evaluate popular data stores sql nosql for temporal spatial data
| 0
|
59,801
| 17,023,250,002
|
IssuesEvent
|
2021-07-03 01:03:17
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
A list of all presets should be available.
|
Component: potlatch (flash editor) Priority: minor Resolution: fixed Type: defect
|
**[Submitted to the original trac issue database at 11.27pm, Friday, 16th May 2008]**
There should be a way to show all possible options in the preset type menu (the button with image left to the actual presets). This could be done, for example, by showing all the possibilities if the user clicks and holds it.
|
1.0
|
A list of all presets should be available. - **[Submitted to the original trac issue database at 11.27pm, Friday, 16th May 2008]**
There should be a way to show all possible options in the preset type menu (the button with image left to the actual presets). This could be done, for example, by showing all the possibilities if the user clicks and holds it.
|
defect
|
a list of all presets should be available there should be a way to show all possible options in the preset type menu the button with image left to the actual presets this could be done for example by showing all the possibilities if the user clicks and holds it
| 1
|
7,190
| 2,610,356,659
|
IssuesEvent
|
2015-02-26 19:55:29
|
chrsmith/scribefire-chrome
|
https://api.github.com/repos/chrsmith/scribefire-chrome
|
closed
|
In response to your email, cfinke:
|
auto-migrated Priority-Medium Type-Defect
|
```
When I wrote saying ScribeFire 3.5.3.3 stopped working after FF updated to
7.0.1 you suggested I upgrade to ScribeFire 4. The only version for firefox I
can see is scribefire-4.xpi. How do I install this (mac os x 10.6.8) and retain
all the data archived in my previous version, as the new version doesn't seem
to have assigned itself a location on my system - after download and clicking
on the file: I get a message saying "There is no application set to open the
document “scribefire-4.xpi”" .. scrutinizing the list of apps to choose
from shows that non of them are browsers, specifically firefox, nor would any
be used by scribefire. When I look in my ff extensions folder, only ScribeFire
3.5.3.3 is listed. Should I be downloading a different file? Other xpi based
files on my system all installed themselves after clicking. I do have
ScribeFire Next installed but disabled as I prefer the original interface. I do
not want to remove ScribeFire 3.5.3.3 before finding out if its data is
archived and accessible to the next version, as I could not find an obviously
named file on my system indicating it contained this info. Thanks for any help
you can provide.
```
-----
Original issue reported on code.google.com by `l8o.n...@gmail.com` on 29 Oct 2011 at 6:22
|
1.0
|
In response to your email, cfinke: - ```
When I wrote saying ScribeFire 3.5.3.3 stopped working after FF updated to
7.0.1 you suggested I upgrade to ScribeFire 4. The only version for firefox I
can see is scribefire-4.xpi. How do I install this (mac os x 10.6.8) and retain
all the data archived in my previous version, as the new version doesn't seem
to have assigned itself a location on my system - after download and clicking
on the file: I get a message saying "There is no application set to open the
document “scribefire-4.xpi”" .. scrutinizing the list of apps to choose
from shows that non of them are browsers, specifically firefox, nor would any
be used by scribefire. When I look in my ff extensions folder, only ScribeFire
3.5.3.3 is listed. Should I be downloading a different file? Other xpi based
files on my system all installed themselves after clicking. I do have
ScribeFire Next installed but disabled as I prefer the original interface. I do
not want to remove ScribeFire 3.5.3.3 before finding out if its data is
archived and accessible to the next version, as I could not find an obviously
named file on my system indicating it contained this info. Thanks for any help
you can provide.
```
-----
Original issue reported on code.google.com by `l8o.n...@gmail.com` on 29 Oct 2011 at 6:22
|
defect
|
in response to your email cfinke when i wrote saying scribefire stopped working after ff updated to you suggested i upgrade to scribefire the only version for firefox i can see is scribefire xpi how do i install this mac os x and retain all the data archived in my previous version as the new version doesn t seem to have assigned itself a location on my system after download and clicking on the file i get a message saying there is no application set to open the document “scribefire xpi” scrutinizing the list of apps to choose from shows that non of them are browsers specifically firefox nor would any be used by scribefire when i look in my ff extensions folder only scribefire is listed should i be downloading a different file other xpi based files on my system all installed themselves after clicking i do have scribefire next installed but disabled as i prefer the original interface i do not want to remove scribefire before finding out if its data is archived and accessible to the next version as i could not find an obviously named file on my system indicating it contained this info thanks for any help you can provide original issue reported on code google com by n gmail com on oct at
| 1
|
54,412
| 13,652,439,667
|
IssuesEvent
|
2020-09-27 07:32:18
|
openzfs/zfs
|
https://api.github.com/repos/openzfs/zfs
|
opened
|
Linux Kernel 5.8.10: ERROR: modpost: "__stack_chk_guard" [.../zfs.ko] undefined!
|
Status: Triage Needed Type: Defect
|
<!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Debian
Distribution Version | Buster
Linux Kernel | 5.8.10-rockchip64
Architecture | aarch64-linux
ZFS Version | 2.0 + cherry-pick 2e407941a2148ab400d8b4e3748cc894e411c0c4
SPL Version | None
<!--
Commands to find ZFS/SPL versions:
modinfo zfs | grep -iw version
modinfo spl | grep -iw version
-->
### Describe the problem you're observing
Making all in module
ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/zstd/zzstd.ko] undefined!
ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/zfs/zfs.ko] undefined!
ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/zcommon/zcommon.ko] undefined!
ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/unicode/zunicode.ko] undefined!
ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/nvpair/znvpair.ko] undefined!
ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/lua/zlua.ko] undefined!
ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/icp/icp.ko] undefined!
ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/avl/zavl.ko] undefined!
ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/spl/spl.ko] undefined!
make[4]: *** [scripts/Makefile.modpost:111: /usr/src/zfs/module/Module.symvers] Error 1
make[4]: *** Deleting file '/usr/src/zfs/module/Module.symvers'
make[3]: *** [Makefile:1665: modules] Error 2
make[2]: *** [Makefile:48: modules-Linux] Error 2
make[1]: *** [Makefile:881: all-recursive] Error 1
make: *** [Makefile:742: all] Error 2
### Describe how to reproduce the problem
```
# FROM Official Armbian Buster
apt update && apt upgrade
apt install build-essential autoconf automake libtool gawk alien fakeroot dkms libblkid-dev uuid-dev libudev-dev libssl-dev zlib1g-dev libaio-dev libattr1-dev libelf-dev linux-headers-current-rockchip64 python3 python3-dev python3-setuptools python3-cffi libffi-dev
cd /usr/src
git clone https://github.com/zfsonlinux/zfs
cd zfs
git checkout remotes/origin/zfs-2.0-release -b zfs2
git cherry-pick 2e407941a2148ab400d8b4e3748cc894e411c0c4
sh autogen.sh
./configure
make -s -j$(nproc)
```
### Include any warning/errors/backtraces from the system logs
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
```
this is an example how log text should be marked (wrap it with ```)
```
-->
|
1.0
|
Linux Kernel 5.8.10: ERROR: modpost: "__stack_chk_guard" [.../zfs.ko] undefined! - <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Debian
Distribution Version | Buster
Linux Kernel | 5.8.10-rockchip64
Architecture | aarch64-linux
ZFS Version | 2.0 + cherry-pick 2e407941a2148ab400d8b4e3748cc894e411c0c4
SPL Version | None
<!--
Commands to find ZFS/SPL versions:
modinfo zfs | grep -iw version
modinfo spl | grep -iw version
-->
### Describe the problem you're observing
Making all in module
ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/zstd/zzstd.ko] undefined!
ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/zfs/zfs.ko] undefined!
ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/zcommon/zcommon.ko] undefined!
ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/unicode/zunicode.ko] undefined!
ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/nvpair/znvpair.ko] undefined!
ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/lua/zlua.ko] undefined!
ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/icp/icp.ko] undefined!
ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/avl/zavl.ko] undefined!
ERROR: modpost: "__stack_chk_guard" [/usr/src/zfs/module/spl/spl.ko] undefined!
make[4]: *** [scripts/Makefile.modpost:111: /usr/src/zfs/module/Module.symvers] Error 1
make[4]: *** Deleting file '/usr/src/zfs/module/Module.symvers'
make[3]: *** [Makefile:1665: modules] Error 2
make[2]: *** [Makefile:48: modules-Linux] Error 2
make[1]: *** [Makefile:881: all-recursive] Error 1
make: *** [Makefile:742: all] Error 2
### Describe how to reproduce the problem
```
# FROM Official Armbian Buster
apt update && apt upgrade
apt install build-essential autoconf automake libtool gawk alien fakeroot dkms libblkid-dev uuid-dev libudev-dev libssl-dev zlib1g-dev libaio-dev libattr1-dev libelf-dev linux-headers-current-rockchip64 python3 python3-dev python3-setuptools python3-cffi libffi-dev
cd /usr/src
git clone https://github.com/zfsonlinux/zfs
cd zfs
git checkout remotes/origin/zfs-2.0-release -b zfs2
git cherry-pick 2e407941a2148ab400d8b4e3748cc894e411c0c4
sh autogen.sh
./configure
make -s -j$(nproc)
```
### Include any warning/errors/backtraces from the system logs
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
```
this is an example how log text should be marked (wrap it with ```)
```
-->
|
defect
|
linux kernel error modpost stack chk guard undefined thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name debian distribution version buster linux kernel architecture linux zfs version cherry pick spl version none commands to find zfs spl versions modinfo zfs grep iw version modinfo spl grep iw version describe the problem you re observing making all in module error modpost stack chk guard undefined error modpost stack chk guard undefined error modpost stack chk guard undefined error modpost stack chk guard undefined error modpost stack chk guard undefined error modpost stack chk guard undefined error modpost stack chk guard undefined error modpost stack chk guard undefined error modpost stack chk guard undefined make error make deleting file usr src zfs module module symvers make error make error make error make error describe how to reproduce the problem from official armbian buster apt update apt upgrade apt install build essential autoconf automake libtool gawk alien fakeroot dkms libblkid dev uuid dev libudev dev libssl dev dev libaio dev dev libelf dev linux headers current dev setuptools cffi libffi dev cd usr src git clone cd zfs git checkout remotes origin zfs release b git cherry pick sh autogen sh configure make s j nproc include any warning errors backtraces from the system logs important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example this is an example how log text should be marked wrap it with
| 1
|
722,474
| 24,863,387,256
|
IssuesEvent
|
2022-10-27 09:58:53
|
AY2223S1-CS2103T-T15-1/tp
|
https://api.github.com/repos/AY2223S1-CS2103T-T15-1/tp
|
closed
|
Edit person does not keep the leave and rating history
|
priority.Medium type.Bug
|
Leave and ratings both have some form of history. When editing a person's field, that history is lost. This behaviour is likely due to `edit` command returning a new `Person` instance, and the history is not being carried over properly.
|
1.0
|
Edit person does not keep the leave and rating history - Leave and ratings both have some form of history. When editing a person's field, that history is lost. This behaviour is likely due to `edit` command returning a new `Person` instance, and the history is not being carried over properly.
|
non_defect
|
edit person does not keep the leave and rating history leave and ratings both have some form of history when editing a person s field that history is lost this behaviour is likely due to edit command returning a new person instance and the history is not being carried over properly
| 0
|
76,934
| 7,549,793,321
|
IssuesEvent
|
2018-04-18 15:06:45
|
reigndesign/basfchinaplas
|
https://api.github.com/repos/reigndesign/basfchinaplas
|
closed
|
Add filter to questions view
|
4 - Testing bug
|
User should be able to filter by tech talk topics
<!---
@huboard:{"order":71.0,"milestone_order":71,"custom_state":""}
-->
|
1.0
|
Add filter to questions view - User should be able to filter by tech talk topics
<!---
@huboard:{"order":71.0,"milestone_order":71,"custom_state":""}
-->
|
non_defect
|
add filter to questions view user should be able to filter by tech talk topics huboard order milestone order custom state
| 0
|
29,674
| 5,810,021,271
|
IssuesEvent
|
2017-05-04 14:37:52
|
primefaces/primeng
|
https://api.github.com/repos/primefaces/primeng
|
closed
|
Calendar TimeOnly incorrect hourformat
|
defect
|
When we use calendar time picker with time only option and hourformat="12" then it displays popup with incorrect AM/PM display.It always display AM in time picker popup as below.
<`p-calendar [(ngModel)]="date8" [timeOnly]="true" [hourFormat]="12"></p-calendar>`

|
1.0
|
Calendar TimeOnly incorrect hourformat - When we use calendar time picker with time only option and hourformat="12" then it displays popup with incorrect AM/PM display.It always display AM in time picker popup as below.
<`p-calendar [(ngModel)]="date8" [timeOnly]="true" [hourFormat]="12"></p-calendar>`

|
defect
|
calendar timeonly incorrect hourformat when we use calendar time picker with time only option and hourformat then it displays popup with incorrect am pm display it always display am in time picker popup as below
| 1
|
198,124
| 22,617,909,656
|
IssuesEvent
|
2022-06-30 01:21:54
|
n-devs/libpack-scripts
|
https://api.github.com/repos/n-devs/libpack-scripts
|
opened
|
CVE-2022-2217 (High) detected in parse-url-5.0.1.tgz
|
security vulnerability
|
## CVE-2022-2217 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>parse-url-5.0.1.tgz</b></p></summary>
<p>An advanced url parser supporting git urls too.</p>
<p>Library home page: <a href="https://registry.npmjs.org/parse-url/-/parse-url-5.0.1.tgz">https://registry.npmjs.org/parse-url/-/parse-url-5.0.1.tgz</a></p>
<p>Path to dependency file: /libpack-scripts/package.json</p>
<p>Path to vulnerable library: /node_modules/parse-url/package.json</p>
<p>
Dependency Hierarchy:
- lerna-3.15.0.tgz (Root Library)
- version-3.15.0.tgz
- github-client-3.14.2.tgz
- git-url-parse-11.1.2.tgz
- git-up-4.0.1.tgz
- :x: **parse-url-5.0.1.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Cross-site Scripting (XSS) - Generic in GitHub repository ionicabizau/parse-url prior to 7.0.0.
<p>Publish Date: 2022-06-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2217>CVE-2022-2217</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://huntr.dev/bounties/4e046c63-b1ca-4bcc-b418-29796918a71b/">https://huntr.dev/bounties/4e046c63-b1ca-4bcc-b418-29796918a71b/</a></p>
<p>Release Date: 2022-06-27</p>
<p>Fix Resolution: parse-url - 6.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-2217 (High) detected in parse-url-5.0.1.tgz - ## CVE-2022-2217 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>parse-url-5.0.1.tgz</b></p></summary>
<p>An advanced url parser supporting git urls too.</p>
<p>Library home page: <a href="https://registry.npmjs.org/parse-url/-/parse-url-5.0.1.tgz">https://registry.npmjs.org/parse-url/-/parse-url-5.0.1.tgz</a></p>
<p>Path to dependency file: /libpack-scripts/package.json</p>
<p>Path to vulnerable library: /node_modules/parse-url/package.json</p>
<p>
Dependency Hierarchy:
- lerna-3.15.0.tgz (Root Library)
- version-3.15.0.tgz
- github-client-3.14.2.tgz
- git-url-parse-11.1.2.tgz
- git-up-4.0.1.tgz
- :x: **parse-url-5.0.1.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Cross-site Scripting (XSS) - Generic in GitHub repository ionicabizau/parse-url prior to 7.0.0.
<p>Publish Date: 2022-06-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2217>CVE-2022-2217</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://huntr.dev/bounties/4e046c63-b1ca-4bcc-b418-29796918a71b/">https://huntr.dev/bounties/4e046c63-b1ca-4bcc-b418-29796918a71b/</a></p>
<p>Release Date: 2022-06-27</p>
<p>Fix Resolution: parse-url - 6.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in parse url tgz cve high severity vulnerability vulnerable library parse url tgz an advanced url parser supporting git urls too library home page a href path to dependency file libpack scripts package json path to vulnerable library node modules parse url package json dependency hierarchy lerna tgz root library version tgz github client tgz git url parse tgz git up tgz x parse url tgz vulnerable library vulnerability details cross site scripting xss generic in github repository ionicabizau parse url prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution parse url step up your open source security game with mend
| 0
|
275,975
| 20,962,339,250
|
IssuesEvent
|
2022-03-28 00:10:56
|
vlizarn/challenge-python-cash-machine
|
https://api.github.com/repos/vlizarn/challenge-python-cash-machine
|
opened
|
[Feature]: Fix whitespace of all examples of Code Sample.
|
documentation enhancement invalid update
|
## Overview - Assign
| Assign | Description | Status |
| :----: | :---- | :----: |
| 1 | Fix whitespace of example 1. | |
| 2 | Fix whitespace of example 2. | |
## Overview - Pull Requests
| Pull Requests| Description |
| :----: | :---- |
| | Apply pull merge request of the master-readme branch for the master branch. |
|
1.0
|
[Feature]: Fix whitespace of all examples of Code Sample. - ## Overview - Assign
| Assign | Description | Status |
| :----: | :---- | :----: |
| 1 | Fix whitespace of example 1. | |
| 2 | Fix whitespace of example 2. | |
## Overview - Pull Requests
| Pull Requests| Description |
| :----: | :---- |
| | Apply pull merge request of the master-readme branch for the master branch. |
|
non_defect
|
fix whitespace of all examples of code sample overview assign assign description status fix whitespace of example fix whitespace of example overview pull requests pull requests description apply pull merge request of the master readme branch for the master branch
| 0
|
45,014
| 12,519,853,381
|
IssuesEvent
|
2020-06-03 15:01:21
|
hikaya-io/dots-frontend
|
https://api.github.com/repos/hikaya-io/dots-frontend
|
closed
|
Remove additional modal for logout
|
defect
|
**Current behavior**
When user clicks on logout a modal window opens asking if they are sure they want to log out
**To Reproduce**
1. Click on the profile avatar on top right corner
2. Click on Logout
**Expected behavior**
When use clicks on logout, user is immediately logged out. No modal window is displayed to ask for further confirmation
**Screenshots**

|
1.0
|
Remove additional modal for logout - **Current behavior**
When user clicks on logout a modal window opens asking if they are sure they want to log out
**To Reproduce**
1. Click on the profile avatar on top right corner
2. Click on Logout
**Expected behavior**
When use clicks on logout, user is immediately logged out. No modal window is displayed to ask for further confirmation
**Screenshots**

|
defect
|
remove additional modal for logout current behavior when user clicks on logout a modal window opens asking if they are sure they want to log out to reproduce click on the profile avatar on top right corner click on logout expected behavior when use clicks on logout user is immediately logged out no modal window is displayed to ask for further confirmation screenshots
| 1
|
3,149
| 13,162,721,370
|
IssuesEvent
|
2020-08-10 22:16:25
|
mozilla-mobile/fenix
|
https://api.github.com/repos/mozilla-mobile/fenix
|
reopened
|
FNX3-14268 ⁃ [UI Tests] Improve wait on session loaded
|
eng:automation needs:triage 🐞 bug
|
A number of our tests are flaky because Espresso continues during page load which causes them to break on assertions afterwards. We should use a an IdlingResource implementation that waits until the current session is not loading anymore.
|
1.0
|
FNX3-14268 ⁃ [UI Tests] Improve wait on session loaded - A number of our tests are flaky because Espresso continues during page load which causes them to break on assertions afterwards. We should use a an IdlingResource implementation that waits until the current session is not loading anymore.
|
non_defect
|
⁃ improve wait on session loaded a number of our tests are flaky because espresso continues during page load which causes them to break on assertions afterwards we should use a an idlingresource implementation that waits until the current session is not loading anymore
| 0
|
3,722
| 2,610,067,611
|
IssuesEvent
|
2015-02-26 18:19:51
|
chrsmith/jsjsj122
|
https://api.github.com/repos/chrsmith/jsjsj122
|
opened
|
路桥前列腺炎专科
|
auto-migrated Priority-Medium Type-Defect
|
```
路桥前列腺炎专科【台州五洲生殖医院】24小时健康咨询热线
:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州市椒江
区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108、118、1
98及椒江一金清公交车直达枫南小区,乘坐107、105、109、112、
901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 8:53
|
1.0
|
路桥前列腺炎专科 - ```
路桥前列腺炎专科【台州五洲生殖医院】24小时健康咨询热线
:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州市椒江
区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108、118、1
98及椒江一金清公交车直达枫南小区,乘坐107、105、109、112、
901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 8:53
|
defect
|
路桥前列腺炎专科 路桥前列腺炎专科【台州五洲生殖医院】 微信号tzwzszyy 医院地址 台州市椒江 (枫南大转盘旁)乘车线路 、 、 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at
| 1
|
129,988
| 10,593,875,585
|
IssuesEvent
|
2019-10-09 15:38:40
|
kcigeospatial/Fred_Co_Land-Management
|
https://api.github.com/repos/kcigeospatial/Fred_Co_Land-Management
|
closed
|
Trade License - Confirm License Details
|
Ready for Test Env. Retest
|
FOR ALL TRADE LICENSES:
At Awaiting Renewal Details, prepopulate the Confirm License Details page with the existing Trade License Details information. A check box should be added to the Confirm detail page that will show to the applicant on R4C, to confirm that they have reviewed and updated their details as necessary. The Confirm Details field that currently exists should remain a back office only field for intake staff to confirm the review of these updates.

|
2.0
|
Trade License - Confirm License Details - FOR ALL TRADE LICENSES:
At Awaiting Renewal Details, prepopulate the Confirm License Details page with the existing Trade License Details information. A check box should be added to the Confirm detail page that will show to the applicant on R4C, to confirm that they have reviewed and updated their details as necessary. The Confirm Details field that currently exists should remain a back office only field for intake staff to confirm the review of these updates.

|
non_defect
|
trade license confirm license details for all trade licenses at awaiting renewal details prepopulate the confirm license details page with the existing trade license details information a check box should be added to the confirm detail page that will show to the applicant on to confirm that they have reviewed and updated their details as necessary the confirm details field that currently exists should remain a back office only field for intake staff to confirm the review of these updates
| 0
|
454,041
| 13,093,816,254
|
IssuesEvent
|
2020-08-03 11:08:49
|
geosolutions-it/ckanext-faoclh
|
https://api.github.com/repos/geosolutions-it/ckanext-faoclh
|
opened
|
FAO Pending Configs
|
Priority: Blocker
|
- [ ] PDF preview issue: for example [this](https://clh-ckan.review.fao.org/dataset/communities-of-practice). The error is the following:
Content is too large to be proxied. Allowed file size: 2097152, Content-Length: 5864691.
Configure [this](https://docs.ckan.org/en/2.8/maintaining/configuration.html#ckan-resource-proxy-max-file-size) to 8 mb
|
1.0
|
FAO Pending Configs - - [ ] PDF preview issue: for example [this](https://clh-ckan.review.fao.org/dataset/communities-of-practice). The error is the following:
Content is too large to be proxied. Allowed file size: 2097152, Content-Length: 5864691.
Configure [this](https://docs.ckan.org/en/2.8/maintaining/configuration.html#ckan-resource-proxy-max-file-size) to 8 mb
|
non_defect
|
fao pending configs pdf preview issue for example the error is the following content is too large to be proxied allowed file size content length configure to mb
| 0
|
3,282
| 2,610,059,842
|
IssuesEvent
|
2015-02-26 18:17:39
|
chrsmith/jsjsj122
|
https://api.github.com/repos/chrsmith/jsjsj122
|
opened
|
临海不孕不育检查需要多少钱
|
auto-migrated Priority-Medium Type-Defect
|
```
临海不孕不育检查需要多少钱【台州五洲生殖医院】24小时健
康咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:
台州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104�
��108、118、198及椒江一金清公交车直达枫南小区,乘坐107、105
、109、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:02
|
1.0
|
临海不孕不育检查需要多少钱 - ```
临海不孕不育检查需要多少钱【台州五洲生殖医院】24小时健
康咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:
台州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104�
��108、118、198及椒江一金清公交车直达枫南小区,乘坐107、105
、109、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 7:02
|
defect
|
临海不孕不育检查需要多少钱 临海不孕不育检查需要多少钱【台州五洲生殖医院】 康咨询热线 微信号tzwzszyy 医院地址 (枫南大转盘旁)乘车线路 � �� 、 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at
| 1
|
56,488
| 15,109,280,783
|
IssuesEvent
|
2021-02-08 17:38:37
|
idaholab/moose
|
https://api.github.com/repos/idaholab/moose
|
opened
|
Automatic Scaling Not Working with Generalized Plane Strain
|
T: defect
|
## Bug Description
<!--A clear and concise description of the problem (Note: A missing feature is not a bug).-->
With `automatic_scaling = true`, the generalized plane strain mechanics solution does not match an equivalent 3D solution.
## Steps to Reproduce
<!--Steps to reproduce the behavior (input file, or modifications to an existing input file, etc.)-->
I've attached two simple (only need 1 cpu cause they are small) moose input files. If you change the `automatic_scaling` parameter you will see the solutions are different.
## Impact
<!--Does this prevent you from getting your work done, or is it more of an annoyance?-->
A wrong solution is provided.
[rectangleExamples.zip](https://github.com/idaholab/moose/files/5945614/rectangleExamples.zip)
|
1.0
|
Automatic Scaling Not Working with Generalized Plane Strain - ## Bug Description
<!--A clear and concise description of the problem (Note: A missing feature is not a bug).-->
With `automatic_scaling = true`, the generalized plane strain mechanics solution does not match an equivalent 3D solution.
## Steps to Reproduce
<!--Steps to reproduce the behavior (input file, or modifications to an existing input file, etc.)-->
I've attached two simple (only need 1 cpu cause they are small) moose input files. If you change the `automatic_scaling` parameter you will see the solutions are different.
## Impact
<!--Does this prevent you from getting your work done, or is it more of an annoyance?-->
A wrong solution is provided.
[rectangleExamples.zip](https://github.com/idaholab/moose/files/5945614/rectangleExamples.zip)
|
defect
|
automatic scaling not working with generalized plane strain bug description with automatic scaling true the generalized plane strain mechanics solution does not match an equivalent solution steps to reproduce i ve attached two simple only need cpu cause they are small moose input files if you change the automatic scaling parameter you will see the solutions are different impact a wrong solution is provided
| 1
|
272,638
| 29,795,060,743
|
IssuesEvent
|
2023-06-16 01:07:50
|
billmcchesney1/pacbot
|
https://api.github.com/repos/billmcchesney1/pacbot
|
closed
|
CVE-2021-22047 (Medium) detected in spring-data-rest-webmvc-3.0.9.RELEASE.jar - autoclosed
|
Mend: dependency security vulnerability
|
## CVE-2021-22047 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-data-rest-webmvc-3.0.9.RELEASE.jar</b></p></summary>
<p>Spring Data REST - WebMVC</p>
<p>Library home page: <a href="http://www.spring.io/spring-data">http://www.spring.io/spring-data</a></p>
<p>Path to dependency file: /api/pacman-api-admin/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/data/spring-data-rest-webmvc/3.0.9.RELEASE/spring-data-rest-webmvc-3.0.9.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/data/spring-data-rest-webmvc/3.0.9.RELEASE/spring-data-rest-webmvc-3.0.9.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-data-rest-2.0.4.RELEASE.jar (Root Library)
- :x: **spring-data-rest-webmvc-3.0.9.RELEASE.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/pacbot/commit/acf9a0620c1a37cee4f2896d71e1c3731c5c7b06">acf9a0620c1a37cee4f2896d71e1c3731c5c7b06</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
In Spring Data REST versions 3.4.0 - 3.4.13, 3.5.0 - 3.5.5, and older unsupported versions, HTTP resources implemented by custom controllers using a configured base API path and a controller type-level request mapping are additionally exposed under URIs that can potentially be exposed for unauthorized access depending on the Spring Security configuration.
<p>Publish Date: 2021-10-28
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-22047>CVE-2021-22047</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22047">https://tanzu.vmware.com/security/cve-2021-22047</a></p>
<p>Release Date: 2021-10-28</p>
<p>Fix Resolution: org.springframework.data:spring-data-rest-webmvc:3.4.14,3.5.6</p>
</p>
</details>
<p></p>
|
True
|
CVE-2021-22047 (Medium) detected in spring-data-rest-webmvc-3.0.9.RELEASE.jar - autoclosed - ## CVE-2021-22047 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-data-rest-webmvc-3.0.9.RELEASE.jar</b></p></summary>
<p>Spring Data REST - WebMVC</p>
<p>Library home page: <a href="http://www.spring.io/spring-data">http://www.spring.io/spring-data</a></p>
<p>Path to dependency file: /api/pacman-api-admin/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/data/spring-data-rest-webmvc/3.0.9.RELEASE/spring-data-rest-webmvc-3.0.9.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/data/spring-data-rest-webmvc/3.0.9.RELEASE/spring-data-rest-webmvc-3.0.9.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-data-rest-2.0.4.RELEASE.jar (Root Library)
- :x: **spring-data-rest-webmvc-3.0.9.RELEASE.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/pacbot/commit/acf9a0620c1a37cee4f2896d71e1c3731c5c7b06">acf9a0620c1a37cee4f2896d71e1c3731c5c7b06</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
In Spring Data REST versions 3.4.0 - 3.4.13, 3.5.0 - 3.5.5, and older unsupported versions, HTTP resources implemented by custom controllers using a configured base API path and a controller type-level request mapping are additionally exposed under URIs that can potentially be exposed for unauthorized access depending on the Spring Security configuration.
<p>Publish Date: 2021-10-28
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-22047>CVE-2021-22047</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22047">https://tanzu.vmware.com/security/cve-2021-22047</a></p>
<p>Release Date: 2021-10-28</p>
<p>Fix Resolution: org.springframework.data:spring-data-rest-webmvc:3.4.14,3.5.6</p>
</p>
</details>
<p></p>
|
non_defect
|
cve medium detected in spring data rest webmvc release jar autoclosed cve medium severity vulnerability vulnerable library spring data rest webmvc release jar spring data rest webmvc library home page a href path to dependency file api pacman api admin pom xml path to vulnerable library home wss scanner repository org springframework data spring data rest webmvc release spring data rest webmvc release jar home wss scanner repository org springframework data spring data rest webmvc release spring data rest webmvc release jar dependency hierarchy spring boot starter data rest release jar root library x spring data rest webmvc release jar vulnerable library found in head commit a href found in base branch master vulnerability details in spring data rest versions and older unsupported versions http resources implemented by custom controllers using a configured base api path and a controller type level request mapping are additionally exposed under uris that can potentially be exposed for unauthorized access depending on the spring security configuration publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework data spring data rest webmvc
| 0
|
782,048
| 27,484,877,821
|
IssuesEvent
|
2023-03-04 01:29:14
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
Performing any operations with complex tensors on MPS causes unrecoverable crash
|
high priority triage review module: mps
|
### 🐛 Describe the bug
Consider following trivial example:
```
python -c "import torch;torch.tensor([1+2j],device='mps')+1"
libc++abi: terminating with uncaught exception of type c10::TypeError: Trying to convert ComplexFloat to the MPS backend but it does not have support for that dtype.
Exception raised from getMPSScalarType at /Users/runner/work/pytorch/pytorch/pytorch/aten/src/ATen/native/mps/OperationUtils.mm:95 (most recent call first):
frame #0: at::native::mps::getMPSScalarType(c10::ScalarType) + 180 (0x16fed979c in libtorch_cpu.dylib)
frame #1: invocation function for block in at::native::mps::binaryOpTensor(at::Tensor const&, at::Tensor const&, c10::Scalar const&, at::Tensor const&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, MPSGraphTensor* (at::native::mps::BinaryOpCachedGraph*, MPSGraphTensor*, MPSGraphTensor*) block_pointer) + 108 (0x16fef9590 in libtorch_cpu.dylib)
frame #2: invocation function for block in at::native::mps::MPSGraphCache::CreateCachedGraph(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, at::native::mps::MPSCachedGraph* () block_pointer) + 216 (0x16fef2348 in libtorch_cpu.dylib)
frame #3: _dispatch_client_callout + 20 (0x19a39e504 in libdispatch.dylib)
frame #4: _dispatch_lane_barrier_sync_invoke_and_complete + 56 (0x19a3ada9c in libdispatch.dylib)
frame #5: at::native::mps::MPSGraphCache::CreateCachedGraph(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, at::native::mps::MPSCachedGraph* () block_pointer) + 160 (0x16fedd844 in libtorch_cpu.dylib)
frame #6: at::native::mps::binaryOpTensor(at::Tensor const&, at::Tensor const&, c10::Scalar const&, at::Tensor const&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, MPSGraphTensor* (at::native::mps::BinaryOpCachedGraph*, MPSGraphTensor*, MPSGraphTensor*) block_pointer) + 2352 (0x16fef8614 in libtorch_cpu.dylib)
frame #7: at::native::mps::add_sub_template(at::Tensor const&, at::Tensor const&, c10::Scalar const&, at::Tensor const&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >) + 984 (0x16fefa5b0 in libtorch_cpu.dylib)
frame #8: at::native::structured_add_out_mps::impl(at::Tensor const&, at::Tensor const&, c10::Scalar const&, at::Tensor const&) + 64 (0x16fefd31c in libtorch_cpu.dylib)
frame #9: at::(anonymous namespace)::wrapper_MPS_add_Tensor(at::Tensor const&, at::Tensor const&, c10::Scalar const&) + 152 (0x16d6051a8 in libtorch_cpu.dylib)
frame #10: c10::impl::wrap_kernel_functor_unboxed_<c10::impl::detail::WrapFunctionIntoFunctor_<c10::CompileTimeFunctionPointer<at::Tensor (c10::DispatchKeySet, at::Tensor const&, at::Tensor const&, c10::Scalar const&), &(torch::autograd::VariableType::(anonymous namespace)::add_Tensor(c10::DispatchKeySet, at::Tensor const&, at::Tensor const&, c10::Scalar const&))>, at::Tensor, c10::guts::typelist::typelist<c10::DispatchKeySet, at::Tensor const&, at::Tensor const&, c10::Scalar const&> >, at::Tensor (c10::DispatchKeySet, at::Tensor const&, at::Tensor const&, c10::Scalar const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, at::Tensor const&, c10::Scalar const&) + 1392 (0x16e507a18 in libtorch_cpu.dylib)
frame #11: at::_ops::add_Tensor::call(at::Tensor const&, at::Tensor const&, c10::Scalar const&) + 296 (0x16c75335c in libtorch_cpu.dylib)
frame #12: torch::autograd::THPVariable_add(_object*, _object*, _object*) + 716 (0x106a320cc in libtorch_python.dylib)
frame #13: _object* torch::autograd::TypeError_to_NotImplemented_<&(torch::autograd::THPVariable_add(_object*, _object*, _object*))>(_object*, _object*, _object*) + 12 (0x106a154f8 in libtorch_python.dylib)
frame #14: method_vectorcall_VARARGS_KEYWORDS + 144 (0x1047a4998 in python3.11)
frame #15: vectorcall_maybe + 212 (0x104828b10 in python3.11)
frame #16: slot_nb_add + 128 (0x104824914 in python3.11)
frame #17: PyNumber_Add + 92 (0x10476f178 in python3.11)
frame #18: _PyEval_EvalFrameDefault + 231128 (0x1048dfe60 in python3.11)
frame #19: _PyEval_Vector + 464 (0x1048a4d30 in python3.11)
frame #20: PyEval_EvalCode + 248 (0x1048a4ad0 in python3.11)
frame #21: run_mod + 184 (0x10493cee4 in python3.11)
frame #22: PyRun_SimpleStringFlags + 144 (0x10493f3d0 in python3.11)
frame #23: pymain_run_command + 160 (0x10495f020 in python3.11)
frame #24: Py_RunMain + 248 (0x10495e7f4 in python3.11)
frame #25: main + 56 (0x104729264 in python3.11)
frame #26: start + 2544 (0x19a1fbe50 in dyld)
```
I understand that complex type is unsupported on MPS, but that's not the reason to crash the runtime. (And perhaps one should simply not be allowed to create such types on MPS?
### Versions
1.13.1, 2.0.0, nightly
cc @ezyang @gchanan @zou3519 @kulinseth @albanD @DenisVieriu97 @razarmehr @abhudev
|
1.0
|
Performing any operations with complex tensors on MPS causes unrecoverable crash - ### 🐛 Describe the bug
Consider following trivial example:
```
python -c "import torch;torch.tensor([1+2j],device='mps')+1"
libc++abi: terminating with uncaught exception of type c10::TypeError: Trying to convert ComplexFloat to the MPS backend but it does not have support for that dtype.
Exception raised from getMPSScalarType at /Users/runner/work/pytorch/pytorch/pytorch/aten/src/ATen/native/mps/OperationUtils.mm:95 (most recent call first):
frame #0: at::native::mps::getMPSScalarType(c10::ScalarType) + 180 (0x16fed979c in libtorch_cpu.dylib)
frame #1: invocation function for block in at::native::mps::binaryOpTensor(at::Tensor const&, at::Tensor const&, c10::Scalar const&, at::Tensor const&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, MPSGraphTensor* (at::native::mps::BinaryOpCachedGraph*, MPSGraphTensor*, MPSGraphTensor*) block_pointer) + 108 (0x16fef9590 in libtorch_cpu.dylib)
frame #2: invocation function for block in at::native::mps::MPSGraphCache::CreateCachedGraph(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, at::native::mps::MPSCachedGraph* () block_pointer) + 216 (0x16fef2348 in libtorch_cpu.dylib)
frame #3: _dispatch_client_callout + 20 (0x19a39e504 in libdispatch.dylib)
frame #4: _dispatch_lane_barrier_sync_invoke_and_complete + 56 (0x19a3ada9c in libdispatch.dylib)
frame #5: at::native::mps::MPSGraphCache::CreateCachedGraph(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, at::native::mps::MPSCachedGraph* () block_pointer) + 160 (0x16fedd844 in libtorch_cpu.dylib)
frame #6: at::native::mps::binaryOpTensor(at::Tensor const&, at::Tensor const&, c10::Scalar const&, at::Tensor const&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >, MPSGraphTensor* (at::native::mps::BinaryOpCachedGraph*, MPSGraphTensor*, MPSGraphTensor*) block_pointer) + 2352 (0x16fef8614 in libtorch_cpu.dylib)
frame #7: at::native::mps::add_sub_template(at::Tensor const&, at::Tensor const&, c10::Scalar const&, at::Tensor const&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> >) + 984 (0x16fefa5b0 in libtorch_cpu.dylib)
frame #8: at::native::structured_add_out_mps::impl(at::Tensor const&, at::Tensor const&, c10::Scalar const&, at::Tensor const&) + 64 (0x16fefd31c in libtorch_cpu.dylib)
frame #9: at::(anonymous namespace)::wrapper_MPS_add_Tensor(at::Tensor const&, at::Tensor const&, c10::Scalar const&) + 152 (0x16d6051a8 in libtorch_cpu.dylib)
frame #10: c10::impl::wrap_kernel_functor_unboxed_<c10::impl::detail::WrapFunctionIntoFunctor_<c10::CompileTimeFunctionPointer<at::Tensor (c10::DispatchKeySet, at::Tensor const&, at::Tensor const&, c10::Scalar const&), &(torch::autograd::VariableType::(anonymous namespace)::add_Tensor(c10::DispatchKeySet, at::Tensor const&, at::Tensor const&, c10::Scalar const&))>, at::Tensor, c10::guts::typelist::typelist<c10::DispatchKeySet, at::Tensor const&, at::Tensor const&, c10::Scalar const&> >, at::Tensor (c10::DispatchKeySet, at::Tensor const&, at::Tensor const&, c10::Scalar const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, at::Tensor const&, c10::Scalar const&) + 1392 (0x16e507a18 in libtorch_cpu.dylib)
frame #11: at::_ops::add_Tensor::call(at::Tensor const&, at::Tensor const&, c10::Scalar const&) + 296 (0x16c75335c in libtorch_cpu.dylib)
frame #12: torch::autograd::THPVariable_add(_object*, _object*, _object*) + 716 (0x106a320cc in libtorch_python.dylib)
frame #13: _object* torch::autograd::TypeError_to_NotImplemented_<&(torch::autograd::THPVariable_add(_object*, _object*, _object*))>(_object*, _object*, _object*) + 12 (0x106a154f8 in libtorch_python.dylib)
frame #14: method_vectorcall_VARARGS_KEYWORDS + 144 (0x1047a4998 in python3.11)
frame #15: vectorcall_maybe + 212 (0x104828b10 in python3.11)
frame #16: slot_nb_add + 128 (0x104824914 in python3.11)
frame #17: PyNumber_Add + 92 (0x10476f178 in python3.11)
frame #18: _PyEval_EvalFrameDefault + 231128 (0x1048dfe60 in python3.11)
frame #19: _PyEval_Vector + 464 (0x1048a4d30 in python3.11)
frame #20: PyEval_EvalCode + 248 (0x1048a4ad0 in python3.11)
frame #21: run_mod + 184 (0x10493cee4 in python3.11)
frame #22: PyRun_SimpleStringFlags + 144 (0x10493f3d0 in python3.11)
frame #23: pymain_run_command + 160 (0x10495f020 in python3.11)
frame #24: Py_RunMain + 248 (0x10495e7f4 in python3.11)
frame #25: main + 56 (0x104729264 in python3.11)
frame #26: start + 2544 (0x19a1fbe50 in dyld)
```
I understand that complex type is unsupported on MPS, but that's not the reason to crash the runtime. (And perhaps one should simply not be allowed to create such types on MPS?
### Versions
1.13.1, 2.0.0, nightly
cc @ezyang @gchanan @zou3519 @kulinseth @albanD @DenisVieriu97 @razarmehr @abhudev
|
non_defect
|
performing any operations with complex tensors on mps causes unrecoverable crash 🐛 describe the bug consider following trivial example python c import torch torch tensor device mps libc abi terminating with uncaught exception of type typeerror trying to convert complexfloat to the mps backend but it does not have support for that dtype exception raised from getmpsscalartype at users runner work pytorch pytorch pytorch aten src aten native mps operationutils mm most recent call first frame at native mps getmpsscalartype scalartype in libtorch cpu dylib frame invocation function for block in at native mps binaryoptensor at tensor const at tensor const scalar const at tensor const std basic string std allocator mpsgraphtensor at native mps binaryopcachedgraph mpsgraphtensor mpsgraphtensor block pointer in libtorch cpu dylib frame invocation function for block in at native mps mpsgraphcache createcachedgraph std basic string std allocator const at native mps mpscachedgraph block pointer in libtorch cpu dylib frame dispatch client callout in libdispatch dylib frame dispatch lane barrier sync invoke and complete in libdispatch dylib frame at native mps mpsgraphcache createcachedgraph std basic string std allocator const at native mps mpscachedgraph block pointer in libtorch cpu dylib frame at native mps binaryoptensor at tensor const at tensor const scalar const at tensor const std basic string std allocator mpsgraphtensor at native mps binaryopcachedgraph mpsgraphtensor mpsgraphtensor block pointer in libtorch cpu dylib frame at native mps add sub template at tensor const at tensor const scalar const at tensor const std basic string std allocator in libtorch cpu dylib frame at native structured add out mps impl at tensor const at tensor const scalar const at tensor const in libtorch cpu dylib frame at anonymous namespace wrapper mps add tensor at tensor const at tensor const scalar const in libtorch cpu dylib frame impl wrap kernel functor unboxed at tensor guts typelist typelist at tensor dispatchkeyset at tensor const at tensor const scalar const call operatorkernel dispatchkeyset at tensor const at tensor const scalar const in libtorch cpu dylib frame at ops add tensor call at tensor const at tensor const scalar const in libtorch cpu dylib frame torch autograd thpvariable add object object object in libtorch python dylib frame object torch autograd typeerror to notimplemented object object object in libtorch python dylib frame method vectorcall varargs keywords in frame vectorcall maybe in frame slot nb add in frame pynumber add in frame pyeval evalframedefault in frame pyeval vector in frame pyeval evalcode in frame run mod in frame pyrun simplestringflags in frame pymain run command in frame py runmain in frame main in frame start in dyld i understand that complex type is unsupported on mps but that s not the reason to crash the runtime and perhaps one should simply not be allowed to create such types on mps versions nightly cc ezyang gchanan kulinseth alband razarmehr abhudev
| 0
|
31,864
| 6,651,310,558
|
IssuesEvent
|
2017-09-28 19:34:39
|
Cockatrice/Cockatrice
|
https://api.github.com/repos/Cockatrice/Cockatrice
|
closed
|
Forgot Password uses case-sensitive email compare
|
App - Servatrice Defect - Basic Easy Change
|
<b>OS:</b>
Windows 7 Professional SP1
<b>Cockatrice version:</b>
2.3.17
__________
Had difficulty getting Forgot Password to work. Turns out it expects email address to be case sensitive. Suggest cast both to lower case before compare. I believe email clients and servers do that anyway so that is what people might expect.
|
1.0
|
Forgot Password uses case-sensitive email compare - <b>OS:</b>
Windows 7 Professional SP1
<b>Cockatrice version:</b>
2.3.17
__________
Had difficulty getting Forgot Password to work. Turns out it expects email address to be case sensitive. Suggest cast both to lower case before compare. I believe email clients and servers do that anyway so that is what people might expect.
|
defect
|
forgot password uses case sensitive email compare os windows professional cockatrice version had difficulty getting forgot password to work turns out it expects email address to be case sensitive suggest cast both to lower case before compare i believe email clients and servers do that anyway so that is what people might expect
| 1
|
26,809
| 4,790,764,028
|
IssuesEvent
|
2016-10-31 09:57:51
|
vuonghv/libdasm
|
https://api.github.com/repos/vuonghv/libdasm
|
opened
|
libdasm misinterprets JLE
|
auto-migrated Priority-Medium Type-Defect
|
_From @GoogleCodeExporter on December 17, 2015 11:11_
```
This is regarding a bug in libdasm library. I've noticed that libdasm
misinterprets a specific instruction.(JLE)
Sample Opcode :
'\x39\x56\x38\x0F\x8E\x1E\x01\x00\x00\x8B\xBD\xE4\xEB\xFF\xFF'
Actual dis-assembly of above opcodes :
00433441 |. 3956 38 CMP DWORD PTR DS:[ESI+38],EDX
00433444 |. 0F8E 1E010000 JLE 0x00433568
0043344A |> 8BBD E4EBFFFF MOV EDI,DWORD PTR SS:[EBP-141C]
libdasm output:
debasish@debasish ~/Downloads/libdasm-beta/examples $ hexdump test.bin
0000000 5639 0f38 1e8e 0001 8b00 e4bd ffeb 00ff
000000f
debasish@debasish ~/Downloads/libdasm-beta/examples $ ./das test.bin
00000000 395638 cmp [esi+0x38],edx
00000003 0f8e1e010000 jng 0x127
00000009 8bbde4ebffff mov edi,[ebp-0x141c]
debasish@debasish ~/Downloads/libdasm-beta/examples $
You can see, in third line libdasm misinterprets the jle (Jump if less or equal
) instruction to jng (Jump if not greater).
cheers,
Debasish
```
Original issue reported on code.google.com by `debasish...@gmail.com` on 23 Jan 2014 at 1:57
Attachments:
- [libdasm_bug.py](https://storage.googleapis.com/google-code-attachments/libdasm/issue-1/comment-0/libdasm_bug.py)
_Copied from original issue: axcheron/libdasm#1_
|
1.0
|
libdasm misinterprets JLE - _From @GoogleCodeExporter on December 17, 2015 11:11_
```
This is regarding a bug in libdasm library. I've noticed that libdasm
misinterprets a specific instruction.(JLE)
Sample Opcode :
'\x39\x56\x38\x0F\x8E\x1E\x01\x00\x00\x8B\xBD\xE4\xEB\xFF\xFF'
Actual dis-assembly of above opcodes :
00433441 |. 3956 38 CMP DWORD PTR DS:[ESI+38],EDX
00433444 |. 0F8E 1E010000 JLE 0x00433568
0043344A |> 8BBD E4EBFFFF MOV EDI,DWORD PTR SS:[EBP-141C]
libdasm output:
debasish@debasish ~/Downloads/libdasm-beta/examples $ hexdump test.bin
0000000 5639 0f38 1e8e 0001 8b00 e4bd ffeb 00ff
000000f
debasish@debasish ~/Downloads/libdasm-beta/examples $ ./das test.bin
00000000 395638 cmp [esi+0x38],edx
00000003 0f8e1e010000 jng 0x127
00000009 8bbde4ebffff mov edi,[ebp-0x141c]
debasish@debasish ~/Downloads/libdasm-beta/examples $
You can see, in third line libdasm misinterprets the jle (Jump if less or equal
) instruction to jng (Jump if not greater).
cheers,
Debasish
```
Original issue reported on code.google.com by `debasish...@gmail.com` on 23 Jan 2014 at 1:57
Attachments:
- [libdasm_bug.py](https://storage.googleapis.com/google-code-attachments/libdasm/issue-1/comment-0/libdasm_bug.py)
_Copied from original issue: axcheron/libdasm#1_
|
defect
|
libdasm misinterprets jle from googlecodeexporter on december this is regarding a bug in libdasm library i ve noticed that libdasm misinterprets a specific instruction jle sample opcode xbd xeb xff xff actual dis assembly of above opcodes cmp dword ptr ds edx jle mov edi dword ptr ss libdasm output debasish debasish downloads libdasm beta examples hexdump test bin ffeb debasish debasish downloads libdasm beta examples das test bin cmp edx jng mov edi debasish debasish downloads libdasm beta examples you can see in third line libdasm misinterprets the jle jump if less or equal instruction to jng jump if not greater cheers debasish original issue reported on code google com by debasish gmail com on jan at attachments copied from original issue axcheron libdasm
| 1
|
227,954
| 7,544,657,281
|
IssuesEvent
|
2018-04-17 19:07:28
|
StrangeLoopGames/EcoIssues
|
https://api.github.com/repos/StrangeLoopGames/EcoIssues
|
closed
|
USER ISSUE: World Unloaded
|
High Priority
|
**Version:** 0.7.3.1 beta
**Steps to Reproduce:**
Walk away from somewhere and walk back
**Expected behavior:**
Chunks should load back in as you get close to them
**Actual behavior:**
Unloaded chunks appear which will not load back in, therefore you fall through the world
|
1.0
|
USER ISSUE: World Unloaded - **Version:** 0.7.3.1 beta
**Steps to Reproduce:**
Walk away from somewhere and walk back
**Expected behavior:**
Chunks should load back in as you get close to them
**Actual behavior:**
Unloaded chunks appear which will not load back in, therefore you fall through the world
|
non_defect
|
user issue world unloaded version beta steps to reproduce walk away from somewhere and walk back expected behavior chunks should load back in as you get close to them actual behavior unloaded chunks appear which will not load back in therefore you fall through the world
| 0
|
121,822
| 4,821,794,681
|
IssuesEvent
|
2016-11-05 14:40:16
|
nextcloud/appstore
|
https://api.github.com/repos/nextcloud/appstore
|
closed
|
Use Nextcloud versions to generate downloads on app detail page
|
enhancement help wanted high priority starter issue
|
Instead of https://github.com/nextcloud/appstore/blob/master/nextcloudappstore/settings/base.py#L253 the versions in the database should be used. The app detail page needs to iterate through all versions and group them by the first version (e.g. 10.0.0 and 10.0.1 compatible releases should be grouped under 10) so that even apps that depend on 10.0.1 appear under the latest Nextcloud 10 release
|
1.0
|
Use Nextcloud versions to generate downloads on app detail page - Instead of https://github.com/nextcloud/appstore/blob/master/nextcloudappstore/settings/base.py#L253 the versions in the database should be used. The app detail page needs to iterate through all versions and group them by the first version (e.g. 10.0.0 and 10.0.1 compatible releases should be grouped under 10) so that even apps that depend on 10.0.1 appear under the latest Nextcloud 10 release
|
non_defect
|
use nextcloud versions to generate downloads on app detail page instead of the versions in the database should be used the app detail page needs to iterate through all versions and group them by the first version e g and compatible releases should be grouped under so that even apps that depend on appear under the latest nextcloud release
| 0
|
76,823
| 26,617,085,171
|
IssuesEvent
|
2023-01-24 08:16:17
|
hyperledger/iroha
|
https://api.github.com/repos/hyperledger/iroha
|
closed
|
404 when 1650 isi in genesis.
|
Bug iroha2 LTS Pre-alpha defect QA-confirmed
|
### GIT commit hash
fe7ea911
### Minimum working example
1. Run iroha with the 1650 isi in genesis
[genesis.txt](https://app.zenhub.com/files/181739240/70fe9b7b-a944-4e8e-96be-459ae4777748/download)
### Expected behaviour
I have to create genesis block with some assets. I generate instruction for domain and admin registration. Then I generate instruction for asset registration, 7 setKeyValue instructions and grant permission instruction few times. Total ~ 5870 isi. Then I put it in ONE genesis transaction and add it to genesis tr
ansactions that already had one small transaction. I expected it to work.
### Actual behaviour
```
iroha-iroha0-1 | 2023-01-17T07:49:17.396580Z ERROR iroha_core::genesis: Genesis transaction #0 failed error=Failed to accept transaction: Too many instructions in payload, max number is 4096, but got 6310
```
Iroha throws 404 Not Found. Text: " ,domain_name". If I reduce rows instructions amount to ~700 it works.
Also I have tried to split each asset with setkeyvalie and grant isi to separate tx, but result was the same, even with only one row (just for testing).
### Operating system
macOS
### Current environment
Docker Hub
### Who can help?
@astrokov7
|
1.0
|
404 when 1650 isi in genesis. - ### GIT commit hash
fe7ea911
### Minimum working example
1. Run iroha with the 1650 isi in genesis
[genesis.txt](https://app.zenhub.com/files/181739240/70fe9b7b-a944-4e8e-96be-459ae4777748/download)
### Expected behaviour
I have to create genesis block with some assets. I generate instruction for domain and admin registration. Then I generate instruction for asset registration, 7 setKeyValue instructions and grant permission instruction few times. Total ~ 5870 isi. Then I put it in ONE genesis transaction and add it to genesis tr
ansactions that already had one small transaction. I expected it to work.
### Actual behaviour
```
iroha-iroha0-1 | 2023-01-17T07:49:17.396580Z ERROR iroha_core::genesis: Genesis transaction #0 failed error=Failed to accept transaction: Too many instructions in payload, max number is 4096, but got 6310
```
Iroha throws 404 Not Found. Text: " ,domain_name". If I reduce rows instructions amount to ~700 it works.
Also I have tried to split each asset with setkeyvalie and grant isi to separate tx, but result was the same, even with only one row (just for testing).
### Operating system
macOS
### Current environment
Docker Hub
### Who can help?
@astrokov7
|
defect
|
when isi in genesis git commit hash minimum working example run iroha with the isi in genesis expected behaviour i have to create genesis block with some assets i generate instruction for domain and admin registration then i generate instruction for asset registration setkeyvalue instructions and grant permission instruction few times total isi then i put it in one genesis transaction and add it to genesis tr ansactions that already had one small transaction i expected it to work actual behaviour iroha error iroha core genesis genesis transaction failed error failed to accept transaction too many instructions in payload max number is but got iroha throws not found text domain name if i reduce rows instructions amount to it works also i have tried to split each asset with setkeyvalie and grant isi to separate tx but result was the same even with only one row just for testing operating system macos current environment docker hub who can help
| 1
|
21,775
| 3,551,638,581
|
IssuesEvent
|
2016-01-21 05:27:01
|
bigbluebutton/bigbluebutton
|
https://api.github.com/repos/bigbluebutton/bigbluebutton
|
closed
|
Pressing the undo button rapidly disconnects everyone from the meeting
|
Accepted Defect Normal Priority Stability Whiteboard
|
Originally reported on Google Code with ID 1522
```
This is an issue to track the bug reported here, https://groups.google.com/d/msg/bigbluebutton-dev/wVhb3-lyU2s/Q1s4rudrxUkJ.
I'm reporting a bug with 0.81 instability when using "undo" on the whiteboard.
If you draw some annotations, then remove them all by clicking "undo" in quick succession,
ALL participants become disconnected from the meeting the moment the last annotation
is removed. This bug is highly reproducible (I've reproduced it many times now). See
the attached video for a demo. It happens when "undo" is clicked a large number of
times in quick succession to remove a large annotation.
Thanks, Andy
```
Reported by `capilkey` on 2013-05-22 21:51:38
|
1.0
|
Pressing the undo button rapidly disconnects everyone from the meeting - Originally reported on Google Code with ID 1522
```
This is an issue to track the bug reported here, https://groups.google.com/d/msg/bigbluebutton-dev/wVhb3-lyU2s/Q1s4rudrxUkJ.
I'm reporting a bug with 0.81 instability when using "undo" on the whiteboard.
If you draw some annotations, then remove them all by clicking "undo" in quick succession,
ALL participants become disconnected from the meeting the moment the last annotation
is removed. This bug is highly reproducible (I've reproduced it many times now). See
the attached video for a demo. It happens when "undo" is clicked a large number of
times in quick succession to remove a large annotation.
Thanks, Andy
```
Reported by `capilkey` on 2013-05-22 21:51:38
|
defect
|
pressing the undo button rapidly disconnects everyone from the meeting originally reported on google code with id this is an issue to track the bug reported here i m reporting a bug with instability when using undo on the whiteboard if you draw some annotations then remove them all by clicking undo in quick succession all participants become disconnected from the meeting the moment the last annotation is removed this bug is highly reproducible i ve reproduced it many times now see the attached video for a demo it happens when undo is clicked a large number of times in quick succession to remove a large annotation thanks andy reported by capilkey on
| 1
|
39,747
| 9,645,700,990
|
IssuesEvent
|
2019-05-17 09:19:24
|
line/armeria
|
https://api.github.com/repos/line/armeria
|
closed
|
NullPointerException
|
defect
|
Hi, we got a NullPointerException because of the Authority header missing, and please refer to the stack trace message below.
The method fillSchemeAndAuthorityIfMissing of class HttpServerHandler creating a new RequestHeaders instance instead of filling the existing one, and this cause the NullPointerException.
```java
2019-05-15 17:40:43.245 WARN [armeria-common-worker-epoll-2-1] c.l.a.s.t.TomcatService [id: 0x7fb03509, L:/192.168.5.137:8080 - R:/192.168.3.170:29786][h2c://test-grpc-demo-0:8080/index#GET] Failed to invoke Tomcat:
java.lang.NullPointerException: null
at com.linecorp.armeria.server.tomcat.TomcatService.convertRequest(TomcatService.java:356)
at com.linecorp.armeria.server.tomcat.TomcatService.lambda$serve$3(TomcatService.java:287)
at java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:822)
at java.util.concurrent.CompletableFuture.uniHandleStage(CompletableFuture.java:834)
at java.util.concurrent.CompletableFuture.handle(CompletableFuture.java:2155)
at com.linecorp.armeria.server.tomcat.TomcatService.serve(TomcatService.java:279)
at com.linecorp.armeria.server.tomcat.TomcatService.serve(TomcatService.java:77)
at com.linecorp.armeria.server.HttpServerHandler.handleRequest(HttpServerHandler.java:376)
at com.linecorp.armeria.server.HttpServerHandler.channelRead(HttpServerHandler.java:250)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at com.linecorp.armeria.server.Http2RequestDecoder.onHeadersRead(Http2RequestDecoder.java:143)
at com.linecorp.armeria.server.Http2RequestDecoder.onHeadersRead(Http2RequestDecoder.java:164)
at io.netty.handler.codec.http2.DefaultHttp2ConnectionDecoder$FrameReadListener.onHeadersRead(DefaultHttp2ConnectionDecoder.java:317)
at io.netty.handler.codec.http2.DefaultHttp2ConnectionDecoder$FrameReadListener.onHeadersRead(DefaultHttp2ConnectionDecoder.java:265)
at io.netty.handler.codec.http2.DefaultHttp2ConnectionDecoder$PrefaceFrameListener.onHeadersRead(DefaultHttp2ConnectionDecoder.java:600)
at io.netty.handler.codec.http2.DefaultHttp2FrameReader$2.processFragment(DefaultHttp2FrameReader.java:483)
at io.netty.handler.codec.http2.DefaultHttp2FrameReader.readHeadersFrame(DefaultHttp2FrameReader.java:491)
at io.netty.handler.codec.http2.DefaultHttp2FrameReader.processPayloadState(DefaultHttp2FrameReader.java:254)
at io.netty.handler.codec.http2.DefaultHttp2FrameReader.readFrame(DefaultHttp2FrameReader.java:160)
at io.netty.handler.codec.http2.DefaultHttp2ConnectionDecoder.decodeFrame(DefaultHttp2ConnectionDecoder.java:118)
at io.netty.handler.codec.http2.Http2ConnectionHandler$FrameDecoder.decode(Http2ConnectionHandler.java:390)
at io.netty.handler.codec.http2.Http2ConnectionHandler$PrefaceDecoder.decode(Http2ConnectionHandler.java:254)
at io.netty.handler.codec.http2.Http2ConnectionHandler.decode(Http2ConnectionHandler.java:450)
at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:502)
at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:441)
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:278)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.handler.codec.ByteToMessageDecoder.handlerRemoved(ByteToMessageDecoder.java:249)
at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:507)
at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:441)
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:278)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:241)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.handler.flush.FlushConsolidationHandler.channelRead(FlushConsolidationHandler.java:154)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:286)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1408)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:930)
at io.netty.channel.epoll.AbstractEpollStreamChannel$EpollStreamUnsafe.epollInReady(AbstractEpollStreamChannel.java:799)
at io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:427)
at io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:328)
at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:905)
at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.lang.Thread.run(Thread.java:748)
```
|
1.0
|
NullPointerException - Hi, we got a NullPointerException because of the Authority header missing, and please refer to the stack trace message below.
The method fillSchemeAndAuthorityIfMissing of class HttpServerHandler creating a new RequestHeaders instance instead of filling the existing one, and this cause the NullPointerException.
```java
2019-05-15 17:40:43.245 WARN [armeria-common-worker-epoll-2-1] c.l.a.s.t.TomcatService [id: 0x7fb03509, L:/192.168.5.137:8080 - R:/192.168.3.170:29786][h2c://test-grpc-demo-0:8080/index#GET] Failed to invoke Tomcat:
java.lang.NullPointerException: null
at com.linecorp.armeria.server.tomcat.TomcatService.convertRequest(TomcatService.java:356)
at com.linecorp.armeria.server.tomcat.TomcatService.lambda$serve$3(TomcatService.java:287)
at java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:822)
at java.util.concurrent.CompletableFuture.uniHandleStage(CompletableFuture.java:834)
at java.util.concurrent.CompletableFuture.handle(CompletableFuture.java:2155)
at com.linecorp.armeria.server.tomcat.TomcatService.serve(TomcatService.java:279)
at com.linecorp.armeria.server.tomcat.TomcatService.serve(TomcatService.java:77)
at com.linecorp.armeria.server.HttpServerHandler.handleRequest(HttpServerHandler.java:376)
at com.linecorp.armeria.server.HttpServerHandler.channelRead(HttpServerHandler.java:250)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at com.linecorp.armeria.server.Http2RequestDecoder.onHeadersRead(Http2RequestDecoder.java:143)
at com.linecorp.armeria.server.Http2RequestDecoder.onHeadersRead(Http2RequestDecoder.java:164)
at io.netty.handler.codec.http2.DefaultHttp2ConnectionDecoder$FrameReadListener.onHeadersRead(DefaultHttp2ConnectionDecoder.java:317)
at io.netty.handler.codec.http2.DefaultHttp2ConnectionDecoder$FrameReadListener.onHeadersRead(DefaultHttp2ConnectionDecoder.java:265)
at io.netty.handler.codec.http2.DefaultHttp2ConnectionDecoder$PrefaceFrameListener.onHeadersRead(DefaultHttp2ConnectionDecoder.java:600)
at io.netty.handler.codec.http2.DefaultHttp2FrameReader$2.processFragment(DefaultHttp2FrameReader.java:483)
at io.netty.handler.codec.http2.DefaultHttp2FrameReader.readHeadersFrame(DefaultHttp2FrameReader.java:491)
at io.netty.handler.codec.http2.DefaultHttp2FrameReader.processPayloadState(DefaultHttp2FrameReader.java:254)
at io.netty.handler.codec.http2.DefaultHttp2FrameReader.readFrame(DefaultHttp2FrameReader.java:160)
at io.netty.handler.codec.http2.DefaultHttp2ConnectionDecoder.decodeFrame(DefaultHttp2ConnectionDecoder.java:118)
at io.netty.handler.codec.http2.Http2ConnectionHandler$FrameDecoder.decode(Http2ConnectionHandler.java:390)
at io.netty.handler.codec.http2.Http2ConnectionHandler$PrefaceDecoder.decode(Http2ConnectionHandler.java:254)
at io.netty.handler.codec.http2.Http2ConnectionHandler.decode(Http2ConnectionHandler.java:450)
at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:502)
at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:441)
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:278)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.handler.codec.ByteToMessageDecoder.handlerRemoved(ByteToMessageDecoder.java:249)
at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:507)
at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:441)
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:278)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:241)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.handler.flush.FlushConsolidationHandler.channelRead(FlushConsolidationHandler.java:154)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:286)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1408)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:930)
at io.netty.channel.epoll.AbstractEpollStreamChannel$EpollStreamUnsafe.epollInReady(AbstractEpollStreamChannel.java:799)
at io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:427)
at io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:328)
at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:905)
at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.lang.Thread.run(Thread.java:748)
```
|
defect
|
nullpointerexception hi we got a nullpointerexception because of the authority header missing and please refer to the stack trace message below the method fillschemeandauthorityifmissing of class httpserverhandler creating a new requestheaders instance instead of filling the existing one and this cause the nullpointerexception java warn c l a s t tomcatservice failed to invoke tomcat java lang nullpointerexception null at com linecorp armeria server tomcat tomcatservice convertrequest tomcatservice java at com linecorp armeria server tomcat tomcatservice lambda serve tomcatservice java at java util concurrent completablefuture unihandle completablefuture java at java util concurrent completablefuture unihandlestage completablefuture java at java util concurrent completablefuture handle completablefuture java at com linecorp armeria server tomcat tomcatservice serve tomcatservice java at com linecorp armeria server tomcat tomcatservice serve tomcatservice java at com linecorp armeria server httpserverhandler handlerequest httpserverhandler java at com linecorp armeria server httpserverhandler channelread httpserverhandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at com linecorp armeria server onheadersread java at com linecorp armeria server onheadersread java at io netty handler codec framereadlistener onheadersread java at io netty handler codec framereadlistener onheadersread java at io netty handler codec prefaceframelistener onheadersread java at io netty handler codec processfragment java at io netty handler codec readheadersframe java at io netty handler codec processpayloadstate java at io netty handler codec readframe java at io netty handler codec decodeframe java at io netty handler codec framedecoder decode java at io netty handler codec prefacedecoder decode java at io netty handler codec decode java at io netty handler codec bytetomessagedecoder decoderemovalreentryprotection bytetomessagedecoder java at io netty handler codec bytetomessagedecoder calldecode bytetomessagedecoder java at io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler codec bytetomessagedecoder handlerremoved bytetomessagedecoder java at io netty handler codec bytetomessagedecoder decoderemovalreentryprotection bytetomessagedecoder java at io netty handler codec bytetomessagedecoder calldecode bytetomessagedecoder java at io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler logging logginghandler channelread logginghandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler flush flushconsolidationhandler channelread flushconsolidationhandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler timeout idlestatehandler channelread idlestatehandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty channel defaultchannelpipeline headcontext channelread defaultchannelpipeline java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel defaultchannelpipeline firechannelread defaultchannelpipeline java at io netty channel epoll abstractepollstreamchannel epollstreamunsafe epollinready abstractepollstreamchannel java at io netty channel epoll epolleventloop processready epolleventloop java at io netty channel epoll epolleventloop run epolleventloop java at io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java at io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java lang thread run thread java
| 1
|
18,885
| 4,320,166,467
|
IssuesEvent
|
2016-07-25 02:38:31
|
MarlinFirmware/Marlin
|
https://api.github.com/repos/MarlinFirmware/Marlin
|
reopened
|
Unclear how to use mesh bed leveling
|
Support: Documentation Support: Homing & Leveling
|
I have been trying to get Mesh Bed Leveling working since 1.1RC6. There have been a few issues opened that appear to have been fixed. It seems that MBL has evolved since some of the initial articles appeared describing it early this year (March). I am confused how to use it with RCBugFix latest. I grabbed RCBugFix today (7/24) and built it using more or less defaults for MBL. My confusion involves the Z-endstop and the settings in Configuration.h.
I have a mechanical Z-endstop that I have traditionally used to set the initial Z distance of "0" after homing. This worked reasonably well for a flat and level print bed. My bed is not perfect so I want to try MBL to compensate. I got it to work once randomly and it was great. I could see the Z axis moving slightly as X and Y moved.
<details>
<summary>Given these MBL settings in Configuration.h:</summary>
```cpp
//===========================================================================
//============================ Mesh Bed Leveling ============================
//===========================================================================
#define MESH_BED_LEVELING // Enable mesh bed leveling.
#if ENABLED(MESH_BED_LEVELING)
#define MESH_INSET 10 // Mesh inset margin on print area
#define MESH_NUM_X_POINTS 3 // Don't use more than 7 points per axis, implementation limited.
#define MESH_NUM_Y_POINTS 3
#define MESH_HOME_SEARCH_Z 4 // Z after Home, bed somewhere below but above 0.0.
//#define MESH_G28_REST_ORIGIN // After homing all axes ('G28' or 'G28 XYZ') rest at origin [0,0,0]
#define MANUAL_BED_LEVELING // Add display menu option for bed leveling.
#if ENABLED(MANUAL_BED_LEVELING)
#define MBL_Z_STEP 0.025 // Step size while manually probing Z axis.
#endif // MANUAL_BED_LEVELING
#endif // MESH_BED_LEVELING
```
</details>
Questions:
1) How should I set my Z-endstop for Home? Should it be set so that the nozzle is 4mm above the bed?
2) If the answer is 4mm above, how do I accommodate driving approximately 4mm below the Z endstop? It will just "compress" my Z endstop switch unmercifully.
Not complaining here - I think MBL will really be helpful. I am just wondering how to actually use it with my setup. Thanks for the neat feature.
Mark
|
1.0
|
Unclear how to use mesh bed leveling - I have been trying to get Mesh Bed Leveling working since 1.1RC6. There have been a few issues opened that appear to have been fixed. It seems that MBL has evolved since some of the initial articles appeared describing it early this year (March). I am confused how to use it with RCBugFix latest. I grabbed RCBugFix today (7/24) and built it using more or less defaults for MBL. My confusion involves the Z-endstop and the settings in Configuration.h.
I have a mechanical Z-endstop that I have traditionally used to set the initial Z distance of "0" after homing. This worked reasonably well for a flat and level print bed. My bed is not perfect so I want to try MBL to compensate. I got it to work once randomly and it was great. I could see the Z axis moving slightly as X and Y moved.
<details>
<summary>Given these MBL settings in Configuration.h:</summary>
```cpp
//===========================================================================
//============================ Mesh Bed Leveling ============================
//===========================================================================
#define MESH_BED_LEVELING // Enable mesh bed leveling.
#if ENABLED(MESH_BED_LEVELING)
#define MESH_INSET 10 // Mesh inset margin on print area
#define MESH_NUM_X_POINTS 3 // Don't use more than 7 points per axis, implementation limited.
#define MESH_NUM_Y_POINTS 3
#define MESH_HOME_SEARCH_Z 4 // Z after Home, bed somewhere below but above 0.0.
//#define MESH_G28_REST_ORIGIN // After homing all axes ('G28' or 'G28 XYZ') rest at origin [0,0,0]
#define MANUAL_BED_LEVELING // Add display menu option for bed leveling.
#if ENABLED(MANUAL_BED_LEVELING)
#define MBL_Z_STEP 0.025 // Step size while manually probing Z axis.
#endif // MANUAL_BED_LEVELING
#endif // MESH_BED_LEVELING
```
</details>
Questions:
1) How should I set my Z-endstop for Home? Should it be set so that the nozzle is 4mm above the bed?
2) If the answer is 4mm above, how do I accommodate driving approximately 4mm below the Z endstop? It will just "compress" my Z endstop switch unmercifully.
Not complaining here - I think MBL will really be helpful. I am just wondering how to actually use it with my setup. Thanks for the neat feature.
Mark
|
non_defect
|
unclear how to use mesh bed leveling i have been trying to get mesh bed leveling working since there have been a few issues opened that appear to have been fixed it seems that mbl has evolved since some of the initial articles appeared describing it early this year march i am confused how to use it with rcbugfix latest i grabbed rcbugfix today and built it using more or less defaults for mbl my confusion involves the z endstop and the settings in configuration h i have a mechanical z endstop that i have traditionally used to set the initial z distance of after homing this worked reasonably well for a flat and level print bed my bed is not perfect so i want to try mbl to compensate i got it to work once randomly and it was great i could see the z axis moving slightly as x and y moved given these mbl settings in configuration h cpp mesh bed leveling define mesh bed leveling enable mesh bed leveling if enabled mesh bed leveling define mesh inset mesh inset margin on print area define mesh num x points don t use more than points per axis implementation limited define mesh num y points define mesh home search z z after home bed somewhere below but above define mesh rest origin after homing all axes or xyz rest at origin define manual bed leveling add display menu option for bed leveling if enabled manual bed leveling define mbl z step step size while manually probing z axis endif manual bed leveling endif mesh bed leveling questions how should i set my z endstop for home should it be set so that the nozzle is above the bed if the answer is above how do i accommodate driving approximately below the z endstop it will just compress my z endstop switch unmercifully not complaining here i think mbl will really be helpful i am just wondering how to actually use it with my setup thanks for the neat feature mark
| 0
|
35,692
| 9,645,533,883
|
IssuesEvent
|
2019-05-17 08:54:50
|
yandex/ClickHouse
|
https://api.github.com/repos/yandex/ClickHouse
|
opened
|
build failed, by clang 8.0.0
|
build
|
hi, Thank you for your work.
I compiled in stable version (19.5.3.8 stable) wrong.
first, Download the code from LLVM and copy the code from the driver directory to compiler_8.0.0 and compiler_8.0.0svn.
second, The build code.
failed message:
CMake Error at dbms/programs/clang/Compiler-8.0.0svn/CMakeLists.txt:32 (add_clang_tool):
Unknown CMake command "add_clang_tool".
|
1.0
|
build failed, by clang 8.0.0 - hi, Thank you for your work.
I compiled in stable version (19.5.3.8 stable) wrong.
first, Download the code from LLVM and copy the code from the driver directory to compiler_8.0.0 and compiler_8.0.0svn.
second, The build code.
failed message:
CMake Error at dbms/programs/clang/Compiler-8.0.0svn/CMakeLists.txt:32 (add_clang_tool):
Unknown CMake command "add_clang_tool".
|
non_defect
|
build failed by clang hi thank you for your work i compiled in stable version stable wrong first download the code from llvm and copy the code from the driver directory to compiler and compiler second the build code failed message cmake error at dbms programs clang compiler cmakelists txt add clang tool unknown cmake command add clang tool
| 0
|
10,862
| 2,622,205,074
|
IssuesEvent
|
2015-03-04 00:29:53
|
clc/emacspeak
|
https://api.github.com/repos/clc/emacspeak
|
opened
|
Google Apps For Your Domain Support for g-client
|
auto-migrated Priority-Medium Type-Defect
|
```
Currently g-client hardcodes addresses for calendar and such so it won't
work with Google Apps For Your Domain setups. It would be nice if such
setups could be supported.
```
Original issue reported on code.google.com by `ismail.d...@gmail.com` on 1 Aug 2008 at 7:19
|
1.0
|
Google Apps For Your Domain Support for g-client - ```
Currently g-client hardcodes addresses for calendar and such so it won't
work with Google Apps For Your Domain setups. It would be nice if such
setups could be supported.
```
Original issue reported on code.google.com by `ismail.d...@gmail.com` on 1 Aug 2008 at 7:19
|
defect
|
google apps for your domain support for g client currently g client hardcodes addresses for calendar and such so it won t work with google apps for your domain setups it would be nice if such setups could be supported original issue reported on code google com by ismail d gmail com on aug at
| 1
|
633,626
| 20,260,696,348
|
IssuesEvent
|
2022-02-15 07:01:17
|
ballerina-platform/ballerina-lang
|
https://api.github.com/repos/ballerina-platform/ballerina-lang
|
closed
|
Completions not working inside mapping constructor in new expression parameter
|
Type/Bug Priority/High Team/LanguageServer Points/2 Area/Completion
|
**Description:**
Consider the following code:
```ballerina
import ballerina/http;
service / on new http:Listener(8080) {
private MyClient myClient;
public function init() returns error? {
self.myClient = check new({<cursor>})
}
}
type Config record {|
int a;
|};
client class MyClient {
function init(Config config) returns error? {
}
}
```
At `<cursor>`, no completions are provided for the matching record fields. This happens only when initializing `client class` type variables. If `MyClient` was just a class, this works.
**Steps to reproduce:**
See description
**Affected Versions:**
Swan Lake 2201 RC2
|
1.0
|
Completions not working inside mapping constructor in new expression parameter - **Description:**
Consider the following code:
```ballerina
import ballerina/http;
service / on new http:Listener(8080) {
private MyClient myClient;
public function init() returns error? {
self.myClient = check new({<cursor>})
}
}
type Config record {|
int a;
|};
client class MyClient {
function init(Config config) returns error? {
}
}
```
At `<cursor>`, no completions are provided for the matching record fields. This happens only when initializing `client class` type variables. If `MyClient` was just a class, this works.
**Steps to reproduce:**
See description
**Affected Versions:**
Swan Lake 2201 RC2
|
non_defect
|
completions not working inside mapping constructor in new expression parameter description consider the following code ballerina import ballerina http service on new http listener private myclient myclient public function init returns error self myclient check new type config record int a client class myclient function init config config returns error at no completions are provided for the matching record fields this happens only when initializing client class type variables if myclient was just a class this works steps to reproduce see description affected versions swan lake
| 0
|
30,373
| 6,123,049,574
|
IssuesEvent
|
2017-06-23 02:39:44
|
line/armeria
|
https://api.github.com/repos/line/armeria
|
opened
|
Flaky test: HttpServerTest.testTooLargeContent
|
defect
|
```
com.linecorp.armeria.server.http.HttpServerTest > testTooLargeContent[0: h1c] FAILED
java.util.concurrent.ExecutionException: com.linecorp.armeria.common.ClosedSessionException
at java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
at java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1895)
at com.linecorp.armeria.server.http.HttpServerTest.testTooLargeContent(HttpServerTest.java:516)
Caused by:
com.linecorp.armeria.common.ClosedSessionException
```
https://ci.appveyor.com/project/trustin/armeria/build/master.591
|
1.0
|
Flaky test: HttpServerTest.testTooLargeContent - ```
com.linecorp.armeria.server.http.HttpServerTest > testTooLargeContent[0: h1c] FAILED
java.util.concurrent.ExecutionException: com.linecorp.armeria.common.ClosedSessionException
at java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
at java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1895)
at com.linecorp.armeria.server.http.HttpServerTest.testTooLargeContent(HttpServerTest.java:516)
Caused by:
com.linecorp.armeria.common.ClosedSessionException
```
https://ci.appveyor.com/project/trustin/armeria/build/master.591
|
defect
|
flaky test httpservertest testtoolargecontent com linecorp armeria server http httpservertest testtoolargecontent failed java util concurrent executionexception com linecorp armeria common closedsessionexception at java util concurrent completablefuture reportget completablefuture java at java util concurrent completablefuture get completablefuture java at com linecorp armeria server http httpservertest testtoolargecontent httpservertest java caused by com linecorp armeria common closedsessionexception
| 1
|
762,061
| 26,707,297,423
|
IssuesEvent
|
2023-01-27 19:26:25
|
molgenis/molgenis-emx2
|
https://api.github.com/repos/molgenis/molgenis-emx2
|
opened
|
Variable explorer is broken
|
bug priority broken
|
**What version of EMX2 are you using (see footer)**
8.136.0
**Describe the bug**
Variable explorer is not loading variables
**To Reproduce**
Steps to reproduce the behavior:
1. Go to ['https://emx2.test.molgenis.org'](https://data-catalogue-staging.molgeniscloud.org/DataCatalogue/catalogue/#/variable-explorer/)
2. See that no variables are loaded
**Expected behavior**
A clear and concise description of what you expected to happen.
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Desktop (please complete the following information):**
- OS: [e.g. iOS]
- Browser [e.g. chrome, safari]
- Version [e.g. 22]
**Smartphone (please complete the following information):**
- Device: [e.g. iPhone6]
- OS: [e.g. iOS8.1]
- Browser [e.g. stock browser, safari]
- Version [e.g. 22]
**Additional context**
Add any other context about the problem here.
|
1.0
|
Variable explorer is broken - **What version of EMX2 are you using (see footer)**
8.136.0
**Describe the bug**
Variable explorer is not loading variables
**To Reproduce**
Steps to reproduce the behavior:
1. Go to ['https://emx2.test.molgenis.org'](https://data-catalogue-staging.molgeniscloud.org/DataCatalogue/catalogue/#/variable-explorer/)
2. See that no variables are loaded
**Expected behavior**
A clear and concise description of what you expected to happen.
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Desktop (please complete the following information):**
- OS: [e.g. iOS]
- Browser [e.g. chrome, safari]
- Version [e.g. 22]
**Smartphone (please complete the following information):**
- Device: [e.g. iPhone6]
- OS: [e.g. iOS8.1]
- Browser [e.g. stock browser, safari]
- Version [e.g. 22]
**Additional context**
Add any other context about the problem here.
|
non_defect
|
variable explorer is broken what version of are you using see footer describe the bug variable explorer is not loading variables to reproduce steps to reproduce the behavior go to see that no variables are loaded expected behavior a clear and concise description of what you expected to happen screenshots if applicable add screenshots to help explain your problem desktop please complete the following information os browser version smartphone please complete the following information device os browser version additional context add any other context about the problem here
| 0
|
23,027
| 3,754,911,340
|
IssuesEvent
|
2016-03-12 08:55:12
|
openwrt/luci
|
https://api.github.com/repos/openwrt/luci
|
closed
|
Hang when saveing changes on Mount Points
|
C: LuCI Applications P: major T: defect
|
**Reported by reporter on 3 May 2010 06:24 UTC**
Hello,
First of all some information:
Luci version: 0.9.0
OpenWRT version: 10.03
Hardware model: TPLINK WR1043ND
Bug:
When I change some records on Mount Point screen and then clik Save or Save & Apply Luci hangs. I could kill uhttpd or reboot the router but there is no thing to do from Luci GUI. When I try to see what is going on from console (by "ps" command") I can see that there is one process which point to luci and fstab changes and I think this is the reason of Luci hangs.
BTW: Luc hangs but all information to fstab are saved correct...
I will be very grateful if some one could help.
Best regards,
MivncM
|
1.0
|
Hang when saveing changes on Mount Points - **Reported by reporter on 3 May 2010 06:24 UTC**
Hello,
First of all some information:
Luci version: 0.9.0
OpenWRT version: 10.03
Hardware model: TPLINK WR1043ND
Bug:
When I change some records on Mount Point screen and then clik Save or Save & Apply Luci hangs. I could kill uhttpd or reboot the router but there is no thing to do from Luci GUI. When I try to see what is going on from console (by "ps" command") I can see that there is one process which point to luci and fstab changes and I think this is the reason of Luci hangs.
BTW: Luc hangs but all information to fstab are saved correct...
I will be very grateful if some one could help.
Best regards,
MivncM
|
defect
|
hang when saveing changes on mount points reported by reporter on may utc hello first of all some information luci version openwrt version hardware model tplink bug when i change some records on mount point screen and then clik save or save apply luci hangs i could kill uhttpd or reboot the router but there is no thing to do from luci gui when i try to see what is going on from console by ps command i can see that there is one process which point to luci and fstab changes and i think this is the reason of luci hangs btw luc hangs but all information to fstab are saved correct i will be very grateful if some one could help best regards mivncm
| 1
|
34,558
| 9,411,962,855
|
IssuesEvent
|
2019-04-10 01:53:46
|
opencv/opencv
|
https://api.github.com/repos/opencv/opencv
|
closed
|
Error compiling 4.1.0 source code with mingw-w64: 'D3D11_TEXTURE2D_DESC' was not declared in this scope
|
category: build/install
|
##### System information (version)
- OpenCV => 4.1.0
- Operating System / Platform => Windows 10 64bit
- Compiler => mingw-w64 8.1.0-posix-seh-rt_v6-rev0
##### Detailed description
When I compiled the 4.1.0 version with mingw-w64, there was an error
```
E:\opencv-4.1.0\opencv-4.1.0\modules\core\src\directx.cpp:1035:5: error: 'D3D11_TEXTURE2D_DESC' was not declared in this scope
D3D11_TEXTURE2D_DESC desc = { 0 };
^~~~~~~~~~~~~~~~~~~~
```
I can solve it like this
```
cmake -DWITH_OPENCL_D3D11_NV=OFF
```
But I don't know if this is the right solution, and whether this will affect the compiled dll.
|
1.0
|
Error compiling 4.1.0 source code with mingw-w64: 'D3D11_TEXTURE2D_DESC' was not declared in this scope -
##### System information (version)
- OpenCV => 4.1.0
- Operating System / Platform => Windows 10 64bit
- Compiler => mingw-w64 8.1.0-posix-seh-rt_v6-rev0
##### Detailed description
When I compiled the 4.1.0 version with mingw-w64, there was an error
```
E:\opencv-4.1.0\opencv-4.1.0\modules\core\src\directx.cpp:1035:5: error: 'D3D11_TEXTURE2D_DESC' was not declared in this scope
D3D11_TEXTURE2D_DESC desc = { 0 };
^~~~~~~~~~~~~~~~~~~~
```
I can solve it like this
```
cmake -DWITH_OPENCL_D3D11_NV=OFF
```
But I don't know if this is the right solution, and whether this will affect the compiled dll.
|
non_defect
|
error compiling source code with mingw desc was not declared in this scope system information version opencv operating system platform windows compiler mingw posix seh rt detailed description when i compiled the version with mingw there was an error e opencv opencv modules core src directx cpp error desc was not declared in this scope desc desc i can solve it like this cmake dwith opencl nv off but i don t know if this is the right solution and whether this will affect the compiled dll
| 0
|
45,672
| 5,950,472,724
|
IssuesEvent
|
2017-05-26 16:49:30
|
dart-lang/site-webdev
|
https://api.github.com/repos/dart-lang/site-webdev
|
closed
|
[ng] don't wrap code
|
Design
|
This seems to be inherited from angular.io, but the code excerpt lines are wrapped. E.g.,
> 
But, on narrow displays this generally reads much better when not wrapped:
> 
In fact, this is what is done on dartlang.org, e.g.:
> 
@kwalrath do you agree?
|
1.0
|
[ng] don't wrap code - This seems to be inherited from angular.io, but the code excerpt lines are wrapped. E.g.,
> 
But, on narrow displays this generally reads much better when not wrapped:
> 
In fact, this is what is done on dartlang.org, e.g.:
> 
@kwalrath do you agree?
|
non_defect
|
don t wrap code this seems to be inherited from angular io but the code excerpt lines are wrapped e g but on narrow displays this generally reads much better when not wrapped in fact this is what is done on dartlang org e g kwalrath do you agree
| 0
|
44,315
| 12,101,445,041
|
IssuesEvent
|
2020-04-20 15:13:27
|
codesmithtools/Templates
|
https://api.github.com/repos/codesmithtools/Templates
|
closed
|
Root object Transactional attribute - enhancement to allow users to control transactions in partial methods.
|
Framework-CSLA Type-Defect auto-migrated
|
```
What steps will reproduce the problem?
1. If you generate a root object, the dataportal methods get decorated with the
<Transactional(TransactionalTypes.TransactionScope)> attribute.
2. If you want to override the root data portal methods with different
transactional requirements you can't, via the partial methods.
What is the expected output?
1. Remove the <Transactional(TransactionalTypes.TransactionScope)> from the
DP_xyz method.
2. Add a "Using scope As New Transactions.TransactionScope" to the generated
code.
Example:
' Blake this would allow the users to overwrite transactional requirements if
need by using the partial methods. (just a little more flexible).
Using scope As New Transactions.TransactionScope
Using connection As New SqlConnection(ADOHelper.ConnectionString)
connection.Open()
....
End Using ' Connection
End Using ' TransactionScope
What version of the product are you using?
v3.0.0.1817
```
Original issue reported on code.google.com by `JenasysD...@gmail.com` on 12 Aug 2010 at 1:27
|
1.0
|
Root object Transactional attribute - enhancement to allow users to control transactions in partial methods. - ```
What steps will reproduce the problem?
1. If you generate a root object, the dataportal methods get decorated with the
<Transactional(TransactionalTypes.TransactionScope)> attribute.
2. If you want to override the root data portal methods with different
transactional requirements you can't, via the partial methods.
What is the expected output?
1. Remove the <Transactional(TransactionalTypes.TransactionScope)> from the
DP_xyz method.
2. Add a "Using scope As New Transactions.TransactionScope" to the generated
code.
Example:
' Blake this would allow the users to overwrite transactional requirements if
need by using the partial methods. (just a little more flexible).
Using scope As New Transactions.TransactionScope
Using connection As New SqlConnection(ADOHelper.ConnectionString)
connection.Open()
....
End Using ' Connection
End Using ' TransactionScope
What version of the product are you using?
v3.0.0.1817
```
Original issue reported on code.google.com by `JenasysD...@gmail.com` on 12 Aug 2010 at 1:27
|
defect
|
root object transactional attribute enhancement to allow users to control transactions in partial methods what steps will reproduce the problem if you generate a root object the dataportal methods get decorated with the attribute if you want to override the root data portal methods with different transactional requirements you can t via the partial methods what is the expected output remove the from the dp xyz method add a using scope as new transactions transactionscope to the generated code example blake this would allow the users to overwrite transactional requirements if need by using the partial methods just a little more flexible using scope as new transactions transactionscope using connection as new sqlconnection adohelper connectionstring connection open end using connection end using transactionscope what version of the product are you using original issue reported on code google com by jenasysd gmail com on aug at
| 1
|
175
| 2,517,932,024
|
IssuesEvent
|
2015-01-16 18:15:14
|
TamarackConsulting/Creekridge_Portal
|
https://api.github.com/repos/TamarackConsulting/Creekridge_Portal
|
closed
|
Doc gen taking a long time to generate docs
|
Major Defect
|
Issue
----------
Sometimes it can take over 15 seconds for a credit app or proposal letter document to generate. User doesn't know if they are waiting or if it didn't work.
Expected
---------------
Improve the performance/speed of generating the documents if possible.
|
1.0
|
Doc gen taking a long time to generate docs - Issue
----------
Sometimes it can take over 15 seconds for a credit app or proposal letter document to generate. User doesn't know if they are waiting or if it didn't work.
Expected
---------------
Improve the performance/speed of generating the documents if possible.
|
defect
|
doc gen taking a long time to generate docs issue sometimes it can take over seconds for a credit app or proposal letter document to generate user doesn t know if they are waiting or if it didn t work expected improve the performance speed of generating the documents if possible
| 1
|
49,983
| 26,418,915,548
|
IssuesEvent
|
2023-01-13 18:22:36
|
matrix-org/sliding-sync
|
https://api.github.com/repos/matrix-org/sliding-sync
|
closed
|
Initial sync is worryingly slow & big to get to clients.
|
performance
|
While chasing down https://github.com/matrix-org/matrix-rust-sdk/issues/1308 i was seeing initial sync reqs taking 36s to arrive. This was in part exasperated by the req having a 10s timeout on it, but even after fixing that, things are taking a worryingly long time - especially after ~12h of being offline.
Here's one i just profiled taking ~5s to execute (after ~2h of being offline). Client saw:
```
2022-12-28T00:21:59.635174Z TRACE send_inner{request=Request { pos: None, txn_id: None, timeout: Some(30s), lists: [SyncRequestList { slow_get_all_rooms: false, ranges: [(0, 20)], sort: ["by_recency", "by_name"], required_state: [(RoomAvatar, ""), (RoomEncryption, "")], timeline_limit: Some(10), filters: None }], room_subscriptions: {}, unsubscribe_rooms: [], extensions: ExtensionsConfig { to_device: Some(ToDeviceConfig { enabled: Some(true), limit: None, since: None }), e2ee: Some(E2EEConfig { enabled: Some(true) }), account_data: Some(AccountDataConfig { enabled: Some(true) }), other: {} } } config=Some(RequestConfig { timeout: 60s, retry_limit: None, retry_timeout: None }) homeserver=Some("https://slidingsync.lab.element.dev/")}:send{config=Some(RequestConfig { timeout: 60s, retry_limit: None, retry_timeout: None }) homeserver="https://slidingsync.lab.element.dev/" user_id=Some("@matthew:matrix.org") server_versions=[V1_0, V1_1, V1_2, V1_3, V1_4, V1_5] request_type="ruma_client_api::sync::sync_events::v4::Request"}: matrix_sdk::http_client: Sending request
2022-12-28T00:22:04.412702Z TRACE send_inner{request=Request { pos: None, txn_id: None, timeout: Some(30s), lists: [SyncRequestList { slow_get_all_rooms: false, ranges: [(0, 20)], sort: ["by_recency", "by_name"], required_state: [(RoomAvatar, ""), (RoomEncryption, "")], timeline_limit: Some(10), filters: None }], room_subscriptions: {}, unsubscribe_rooms: [], extensions: ExtensionsConfig { to_device: Some(ToDeviceConfig { enabled: Some(true), limit: None, since: None }), e2ee: Some(E2EEConfig { enabled: Some(true) }), account_data: Some(AccountDataConfig { enabled: Some(true) }), other: {} } } config=Some(RequestConfig { timeout: 60s, retry_limit: None, retry_timeout: None }) homeserver=Some("https://slidingsync.lab.element.dev/")}:send{config=Some(RequestConfig { timeout: 60s, retry_limit: None, retry_timeout: None }) homeserver="https://slidingsync.lab.element.dev/" user_id=Some("@matthew:matrix.org") server_versions=[V1_0, V1_1, V1_2, V1_3, V1_4, V1_5] request_type="ruma_client_api::sync::sync_events::v4::Request"}: matrix_sdk::http_client: Got response: Response { status: 200, version: HTTP/1.1, headers: {"access-control-allow-headers": "Origin, X-Requested-With, Content-Type, Accept, Authorization", "access-control-allow-methods": "GET, POST, PUT, DELETE, OPTIONS", "access-control-allow-origin": "*", "content-type": "application/json", "date": "Wed, 28 Dec 2022 00:22:01 GMT", "transfer-encoding": "chunked", "set-cookie": "......; path=/; HttpOnly; Secure; SameSite=None"}, body: b"{\"lists\":[{\"ops\":[{\"op\":\"SYNC\",\"range\":[0,20],
```
Server saw:
```
00:22:00 INF created new connection conn_id=ec725749c89702fe5e1fc5e019e77ce22dc0f3ae0717483f307a02e2222c6631 user=@matthew:matrix.org
00:22:01 WRN Client did not increment since token: possibly sending back duplicate to-device events! device=ec725749c89702fe5e1fc5e019e77ce22dc0f3ae0717483f307a02e2222c6631 initial=true last_sent=6950241 recv=0 user=@matthew:matrix.org
00:22:01 INF ag=25 d=1 duration=1111.414264 p=0 q=1 r=21 size=296993 status=200 u=@matthew:matrix.org
00:22:04 INF dl-c=71 dl-l=8 duration=32.985866 p=1 q=2 r=0 size=2200 status=200 u=@matthew:matrix.org
00:22:05 INF duration=44.822427 p=2 q=3 r=0 size=327 status=200 u=@matthew:matrix.org
```
Server pprof saw (for the OnIncomingRequestInitial):
```
When Elapsed Goroutine ID Events
4.451185334s 1.105535526s Task 1964 (goroutine view) (complete)
4.451185334 . 39366 task OnIncomingRequestInitial (id 1964, parent 0) created
4.451192886 . 7552 39366 region load started (duration: 42.257935ms)
4.493469125 . 42276239 39366 region onIncomingListRequest started (duration: 32.959255ms)
4.526486083 . 33016958 39366 region buildRooms started (duration: 1.000488985s)
4.558728129 . 32242046 39366 new goroutine 39367: database/sql.(*DB).beginDC·dwrap·17
4.560168230 . 1440101 39366 new goroutine 39395: database/sql.(*Rows).initContextClose·dwrap·27
4.562780862 . 2612632 39366 region RoomStateAfterEventPosition started (duration: 941.937482ms)
4.563280573 . 499711 39366 new goroutine 39294: database/sql.(*DB).beginDC·dwrap·17
4.568830597 . 5550024 39366 new goroutine 39368: database/sql.(*Rows).initContextClose·dwrap·27
4.570190699 . 1360102 39366 new goroutine 39369: database/sql.(*Rows).initContextClose·dwrap·27
5.527456348 .957265649 39366 region extensions started (duration: 29.246106ms)
5.549077859 . 21621511 39366 new goroutine 39295: database/sql.(*DB).beginDC·dwrap·17
5.551339387 . 2261528 39366 new goroutine 39296: database/sql.(*DB).beginDC·dwrap·17
5.552959072 . 1619685 39366 new goroutine 39396: database/sql.(*Rows).initContextClose·dwrap·27
5.553884664 . 925592 39366 new goroutine 39297: database/sql.(*DB).beginDC·dwrap·17
5.555373534 . 1488870 39366 new goroutine 39373: database/sql.(*Rows).initContextClose·dwrap·27
5.556710390 . 1336856 39366 region liveUpdate started (duration: 6.246µs)
5.556720860 . 10470 39366 task end
```
So: if it took 1111ms to generate on the server (which is already quite a lot; the 300KB request size is *way* too large thanks to including 130KB of redundant account_data on it, and 20*20 = 400 events of scrollback), how come it took the client another 3s to receive it? I guess this could be slow internet connectivity whilst travelling for xmas, but empirically i seem to be able to download stuff at ~4MB/s.
To address the size, i think account_data has to be incremental, as per https://github.com/matrix-org/matrix-spec-proposals/pull/3575/files#r1057933023 - and we should always avoid syncing in scrollback in rooms during initial sync. However, it doesn't explain whey the request took 1s to execute but blocked the client for 5s...
|
True
|
Initial sync is worryingly slow & big to get to clients. - While chasing down https://github.com/matrix-org/matrix-rust-sdk/issues/1308 i was seeing initial sync reqs taking 36s to arrive. This was in part exasperated by the req having a 10s timeout on it, but even after fixing that, things are taking a worryingly long time - especially after ~12h of being offline.
Here's one i just profiled taking ~5s to execute (after ~2h of being offline). Client saw:
```
2022-12-28T00:21:59.635174Z TRACE send_inner{request=Request { pos: None, txn_id: None, timeout: Some(30s), lists: [SyncRequestList { slow_get_all_rooms: false, ranges: [(0, 20)], sort: ["by_recency", "by_name"], required_state: [(RoomAvatar, ""), (RoomEncryption, "")], timeline_limit: Some(10), filters: None }], room_subscriptions: {}, unsubscribe_rooms: [], extensions: ExtensionsConfig { to_device: Some(ToDeviceConfig { enabled: Some(true), limit: None, since: None }), e2ee: Some(E2EEConfig { enabled: Some(true) }), account_data: Some(AccountDataConfig { enabled: Some(true) }), other: {} } } config=Some(RequestConfig { timeout: 60s, retry_limit: None, retry_timeout: None }) homeserver=Some("https://slidingsync.lab.element.dev/")}:send{config=Some(RequestConfig { timeout: 60s, retry_limit: None, retry_timeout: None }) homeserver="https://slidingsync.lab.element.dev/" user_id=Some("@matthew:matrix.org") server_versions=[V1_0, V1_1, V1_2, V1_3, V1_4, V1_5] request_type="ruma_client_api::sync::sync_events::v4::Request"}: matrix_sdk::http_client: Sending request
2022-12-28T00:22:04.412702Z TRACE send_inner{request=Request { pos: None, txn_id: None, timeout: Some(30s), lists: [SyncRequestList { slow_get_all_rooms: false, ranges: [(0, 20)], sort: ["by_recency", "by_name"], required_state: [(RoomAvatar, ""), (RoomEncryption, "")], timeline_limit: Some(10), filters: None }], room_subscriptions: {}, unsubscribe_rooms: [], extensions: ExtensionsConfig { to_device: Some(ToDeviceConfig { enabled: Some(true), limit: None, since: None }), e2ee: Some(E2EEConfig { enabled: Some(true) }), account_data: Some(AccountDataConfig { enabled: Some(true) }), other: {} } } config=Some(RequestConfig { timeout: 60s, retry_limit: None, retry_timeout: None }) homeserver=Some("https://slidingsync.lab.element.dev/")}:send{config=Some(RequestConfig { timeout: 60s, retry_limit: None, retry_timeout: None }) homeserver="https://slidingsync.lab.element.dev/" user_id=Some("@matthew:matrix.org") server_versions=[V1_0, V1_1, V1_2, V1_3, V1_4, V1_5] request_type="ruma_client_api::sync::sync_events::v4::Request"}: matrix_sdk::http_client: Got response: Response { status: 200, version: HTTP/1.1, headers: {"access-control-allow-headers": "Origin, X-Requested-With, Content-Type, Accept, Authorization", "access-control-allow-methods": "GET, POST, PUT, DELETE, OPTIONS", "access-control-allow-origin": "*", "content-type": "application/json", "date": "Wed, 28 Dec 2022 00:22:01 GMT", "transfer-encoding": "chunked", "set-cookie": "......; path=/; HttpOnly; Secure; SameSite=None"}, body: b"{\"lists\":[{\"ops\":[{\"op\":\"SYNC\",\"range\":[0,20],
```
Server saw:
```
00:22:00 INF created new connection conn_id=ec725749c89702fe5e1fc5e019e77ce22dc0f3ae0717483f307a02e2222c6631 user=@matthew:matrix.org
00:22:01 WRN Client did not increment since token: possibly sending back duplicate to-device events! device=ec725749c89702fe5e1fc5e019e77ce22dc0f3ae0717483f307a02e2222c6631 initial=true last_sent=6950241 recv=0 user=@matthew:matrix.org
00:22:01 INF ag=25 d=1 duration=1111.414264 p=0 q=1 r=21 size=296993 status=200 u=@matthew:matrix.org
00:22:04 INF dl-c=71 dl-l=8 duration=32.985866 p=1 q=2 r=0 size=2200 status=200 u=@matthew:matrix.org
00:22:05 INF duration=44.822427 p=2 q=3 r=0 size=327 status=200 u=@matthew:matrix.org
```
Server pprof saw (for the OnIncomingRequestInitial):
```
When Elapsed Goroutine ID Events
4.451185334s 1.105535526s Task 1964 (goroutine view) (complete)
4.451185334 . 39366 task OnIncomingRequestInitial (id 1964, parent 0) created
4.451192886 . 7552 39366 region load started (duration: 42.257935ms)
4.493469125 . 42276239 39366 region onIncomingListRequest started (duration: 32.959255ms)
4.526486083 . 33016958 39366 region buildRooms started (duration: 1.000488985s)
4.558728129 . 32242046 39366 new goroutine 39367: database/sql.(*DB).beginDC·dwrap·17
4.560168230 . 1440101 39366 new goroutine 39395: database/sql.(*Rows).initContextClose·dwrap·27
4.562780862 . 2612632 39366 region RoomStateAfterEventPosition started (duration: 941.937482ms)
4.563280573 . 499711 39366 new goroutine 39294: database/sql.(*DB).beginDC·dwrap·17
4.568830597 . 5550024 39366 new goroutine 39368: database/sql.(*Rows).initContextClose·dwrap·27
4.570190699 . 1360102 39366 new goroutine 39369: database/sql.(*Rows).initContextClose·dwrap·27
5.527456348 .957265649 39366 region extensions started (duration: 29.246106ms)
5.549077859 . 21621511 39366 new goroutine 39295: database/sql.(*DB).beginDC·dwrap·17
5.551339387 . 2261528 39366 new goroutine 39296: database/sql.(*DB).beginDC·dwrap·17
5.552959072 . 1619685 39366 new goroutine 39396: database/sql.(*Rows).initContextClose·dwrap·27
5.553884664 . 925592 39366 new goroutine 39297: database/sql.(*DB).beginDC·dwrap·17
5.555373534 . 1488870 39366 new goroutine 39373: database/sql.(*Rows).initContextClose·dwrap·27
5.556710390 . 1336856 39366 region liveUpdate started (duration: 6.246µs)
5.556720860 . 10470 39366 task end
```
So: if it took 1111ms to generate on the server (which is already quite a lot; the 300KB request size is *way* too large thanks to including 130KB of redundant account_data on it, and 20*20 = 400 events of scrollback), how come it took the client another 3s to receive it? I guess this could be slow internet connectivity whilst travelling for xmas, but empirically i seem to be able to download stuff at ~4MB/s.
To address the size, i think account_data has to be incremental, as per https://github.com/matrix-org/matrix-spec-proposals/pull/3575/files#r1057933023 - and we should always avoid syncing in scrollback in rooms during initial sync. However, it doesn't explain whey the request took 1s to execute but blocked the client for 5s...
|
non_defect
|
initial sync is worryingly slow big to get to clients while chasing down i was seeing initial sync reqs taking to arrive this was in part exasperated by the req having a timeout on it but even after fixing that things are taking a worryingly long time especially after of being offline here s one i just profiled taking to execute after of being offline client saw trace send inner request request pos none txn id none timeout some lists sort required state timeline limit some filters none room subscriptions unsubscribe rooms extensions extensionsconfig to device some todeviceconfig enabled some true limit none since none some enabled some true account data some accountdataconfig enabled some true other config some requestconfig timeout retry limit none retry timeout none homeserver some timeout retry limit none retry timeout none homeserver user id some matthew matrix org server versions request type ruma client api sync sync events request matrix sdk http client sending request trace send inner request request pos none txn id none timeout some lists sort required state timeline limit some filters none room subscriptions unsubscribe rooms extensions extensionsconfig to device some todeviceconfig enabled some true limit none since none some enabled some true account data some accountdataconfig enabled some true other config some requestconfig timeout retry limit none retry timeout none homeserver some timeout retry limit none retry timeout none homeserver user id some matthew matrix org server versions request type ruma client api sync sync events request matrix sdk http client got response response status version http headers access control allow headers origin x requested with content type accept authorization access control allow methods get post put delete options access control allow origin content type application json date wed dec gmt transfer encoding chunked set cookie path httponly secure samesite none body b lists server saw inf created new connection conn id user matthew matrix org wrn client did not increment since token possibly sending back duplicate to device events device initial true last sent recv user matthew matrix org inf ag d duration p q r size status u matthew matrix org inf dl c dl l duration p q r size status u matthew matrix org inf duration p q r size status u matthew matrix org server pprof saw for the onincomingrequestinitial when elapsed goroutine id events task goroutine view complete task onincomingrequestinitial id parent created region load started duration region onincominglistrequest started duration region buildrooms started duration new goroutine database sql db begindc·dwrap· new goroutine database sql rows initcontextclose·dwrap· region roomstateaftereventposition started duration new goroutine database sql db begindc·dwrap· new goroutine database sql rows initcontextclose·dwrap· new goroutine database sql rows initcontextclose·dwrap· region extensions started duration new goroutine database sql db begindc·dwrap· new goroutine database sql db begindc·dwrap· new goroutine database sql rows initcontextclose·dwrap· new goroutine database sql db begindc·dwrap· new goroutine database sql rows initcontextclose·dwrap· region liveupdate started duration task end so if it took to generate on the server which is already quite a lot the request size is way too large thanks to including of redundant account data on it and events of scrollback how come it took the client another to receive it i guess this could be slow internet connectivity whilst travelling for xmas but empirically i seem to be able to download stuff at s to address the size i think account data has to be incremental as per and we should always avoid syncing in scrollback in rooms during initial sync however it doesn t explain whey the request took to execute but blocked the client for
| 0
|
1,872
| 2,576,060,012
|
IssuesEvent
|
2015-02-12 06:05:53
|
mozilla/webmaker-app
|
https://api.github.com/repos/mozilla/webmaker-app
|
closed
|
Complete Share view device integration
|
design
|
Design here: https://redpen.io/fn6bd4b0dbeaf8e985
We may want to consider device integration for Twitter, WhatsApp, Facebook if it exists
|
1.0
|
Complete Share view device integration - Design here: https://redpen.io/fn6bd4b0dbeaf8e985
We may want to consider device integration for Twitter, WhatsApp, Facebook if it exists
|
non_defect
|
complete share view device integration design here we may want to consider device integration for twitter whatsapp facebook if it exists
| 0
|
16,351
| 2,889,788,042
|
IssuesEvent
|
2015-06-13 19:14:34
|
damonkohler/android-scripting
|
https://api.github.com/repos/damonkohler/android-scripting
|
closed
|
Script as APK project not compiling
|
auto-migrated Priority-Medium Type-Defect
|
```
What device(s) are you experiencing the problem on?
What firmware version are you running on the device?
What steps will reproduce the problem?
1. Follow tutorial
http://code.google.com/p/android-scripting/wiki/SharingScripts#Scripts_as_APKs
2. The project built just fine in Eclipse (with no errors) and then
installation was done with:
./adb -s emulator-5554 install /Users/s7ephen/Documents/workspace/
ScriptForAndroidTemplate/bin/ScriptForAndroidTemplate.apk
1268 KB/s (166837 bytes in 0.128s)
pkg: /data/local/tmp/ScriptForAndroidTemplate.apk
Success
3. Built it using Eclipse on OSX for a number of target releases (1.5,
2.0.1, and 2.1-update)
4. Launch on both a 2.1-update Motorola Milestone hardware phone and in a
2.1-update1 emulator.
5. See exception: "The application Dummy
Script (process com.dummy.fooforandroid) has stopped unexpectantly.
Please Try Again."
I am really confused with what might be going wrong here. I have been
a huge fan of ASE/SL4A for quite some time, but am finding recently
that I would like to distribute some of my tools to people beyond my
"nerdly" circle of friends ;-) As such, I would like to have a
portable apk bundle that people can easily install. I would really
appreciate any tips or insights that anyone could offer me to help me
get this working. I have included the logcat output for installation
and execution/exception of the ScriptForAndroidTemplate.
What is the expected output? What do you see instead?
No exception.
What version of the product are you using? On what operating system?
Please provide any additional information below.
```
Original issue reported on code.google.com by `damonkoh...@gmail.com` on 19 Sep 2010 at 6:09
|
1.0
|
Script as APK project not compiling - ```
What device(s) are you experiencing the problem on?
What firmware version are you running on the device?
What steps will reproduce the problem?
1. Follow tutorial
http://code.google.com/p/android-scripting/wiki/SharingScripts#Scripts_as_APKs
2. The project built just fine in Eclipse (with no errors) and then
installation was done with:
./adb -s emulator-5554 install /Users/s7ephen/Documents/workspace/
ScriptForAndroidTemplate/bin/ScriptForAndroidTemplate.apk
1268 KB/s (166837 bytes in 0.128s)
pkg: /data/local/tmp/ScriptForAndroidTemplate.apk
Success
3. Built it using Eclipse on OSX for a number of target releases (1.5,
2.0.1, and 2.1-update)
4. Launch on both a 2.1-update Motorola Milestone hardware phone and in a
2.1-update1 emulator.
5. See exception: "The application Dummy
Script (process com.dummy.fooforandroid) has stopped unexpectantly.
Please Try Again."
I am really confused with what might be going wrong here. I have been
a huge fan of ASE/SL4A for quite some time, but am finding recently
that I would like to distribute some of my tools to people beyond my
"nerdly" circle of friends ;-) As such, I would like to have a
portable apk bundle that people can easily install. I would really
appreciate any tips or insights that anyone could offer me to help me
get this working. I have included the logcat output for installation
and execution/exception of the ScriptForAndroidTemplate.
What is the expected output? What do you see instead?
No exception.
What version of the product are you using? On what operating system?
Please provide any additional information below.
```
Original issue reported on code.google.com by `damonkoh...@gmail.com` on 19 Sep 2010 at 6:09
|
defect
|
script as apk project not compiling what device s are you experiencing the problem on what firmware version are you running on the device what steps will reproduce the problem follow tutorial the project built just fine in eclipse with no errors and then installation was done with adb s emulator install users documents workspace scriptforandroidtemplate bin scriptforandroidtemplate apk kb s bytes in pkg data local tmp scriptforandroidtemplate apk success built it using eclipse on osx for a number of target releases and update launch on both a update motorola milestone hardware phone and in a emulator see exception the application dummy script process com dummy fooforandroid has stopped unexpectantly please try again i am really confused with what might be going wrong here i have been a huge fan of ase for quite some time but am finding recently that i would like to distribute some of my tools to people beyond my nerdly circle of friends as such i would like to have a portable apk bundle that people can easily install i would really appreciate any tips or insights that anyone could offer me to help me get this working i have included the logcat output for installation and execution exception of the scriptforandroidtemplate what is the expected output what do you see instead no exception what version of the product are you using on what operating system please provide any additional information below original issue reported on code google com by damonkoh gmail com on sep at
| 1
|
104,193
| 16,613,494,836
|
IssuesEvent
|
2021-06-02 14:11:34
|
Thanraj/linux-4.1.15
|
https://api.github.com/repos/Thanraj/linux-4.1.15
|
opened
|
CVE-2018-10902 (High) detected in linux-stable-rtv4.1.33
|
security vulnerability
|
## CVE-2018-10902 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/Thanraj/linux-4.1.15/commits/5e3fb3e332499e1ad10a0969e55582af1027b085">5e3fb3e332499e1ad10a0969e55582af1027b085</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux-4.1.15/sound/core/rawmidi.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux-4.1.15/sound/core/rawmidi.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
It was found that the raw midi kernel driver does not protect against concurrent access which leads to a double realloc (double free) in snd_rawmidi_input_params() and snd_rawmidi_output_status() which are part of snd_rawmidi_ioctl() handler in rawmidi.c file. A malicious local attacker could possibly use this for privilege escalation.
<p>Publish Date: 2018-08-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-10902>CVE-2018-10902</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-10902">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-10902</a></p>
<p>Release Date: 2018-08-21</p>
<p>Fix Resolution: v4.18-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-10902 (High) detected in linux-stable-rtv4.1.33 - ## CVE-2018-10902 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/Thanraj/linux-4.1.15/commits/5e3fb3e332499e1ad10a0969e55582af1027b085">5e3fb3e332499e1ad10a0969e55582af1027b085</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux-4.1.15/sound/core/rawmidi.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux-4.1.15/sound/core/rawmidi.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
It was found that the raw midi kernel driver does not protect against concurrent access which leads to a double realloc (double free) in snd_rawmidi_input_params() and snd_rawmidi_output_status() which are part of snd_rawmidi_ioctl() handler in rawmidi.c file. A malicious local attacker could possibly use this for privilege escalation.
<p>Publish Date: 2018-08-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-10902>CVE-2018-10902</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-10902">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-10902</a></p>
<p>Release Date: 2018-08-21</p>
<p>Fix Resolution: v4.18-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files linux sound core rawmidi c linux sound core rawmidi c vulnerability details it was found that the raw midi kernel driver does not protect against concurrent access which leads to a double realloc double free in snd rawmidi input params and snd rawmidi output status which are part of snd rawmidi ioctl handler in rawmidi c file a malicious local attacker could possibly use this for privilege escalation publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
62,748
| 17,187,803,922
|
IssuesEvent
|
2021-07-16 06:24:40
|
Questie/Questie
|
https://api.github.com/repos/Questie/Questie
|
closed
|
QuestieLib.lua:256: bad argument #1 to 'unpack' (table expected, got nil)
|
Type - Defect
|
<!-- READ THIS FIRST
Hello, thanks for taking the time to report a bug!
Before you proceed, please verify that you're running the latest version of Questie. The easiest way to do this is via the Twitch client, but you can also download the latest version here: https://www.curseforge.com/wow/addons/questie
Questie is one of the most popular Classic WoW addons, with over 22M downloads. However, like almost all WoW addons, it's built and maintained by a team of volunteers. The current Questie team is:
* @AeroScripts / Aero#1357 (Discord)
* @BreakBB / TheCrux#1702 (Discord)
* @drejjmit / Drejjmit#8241 (Discord)
* @Dyaxler / Dyaxler#0086 (Discord)
* @gogo1951 / Gogo#0298 (Discord)
If you'd like to help, please consider making a donation. You can do so here: https://www.paypal.com/cgi-bin/webscr?cmd=_donations&business=aero1861%40gmail%2ecom&lc=CA&item_name=Questie%20Devs¤cy_code=USD&bn=PP%2dDonationsBF%3abtn_donate_LG%2egif%3aNonHosted
You can also help as a tester, developer or translator, please join the Questie Discord here https://discord.gg/fYcQfv7
-->
## Bug description
<!-- Explain in detail what the bug is and how you encountered it. If possible explain how it can be reproduced. -->
I was killing a mob for quest #9513 Reclaiming the Ruins and I just killed off I think was a Wrathscale Myrmidon when this error popped up:
1x Questie\Modules\Libs\QuestieLib.lua:256: bad argument #1 to 'unpack' (table expected, got nil)
[string "=[C]"]: in function `unpack'
[string "@Questie\Modules\Libs\QuestieLib.lua"]:256: in function `GetTbcLevel'
[string "@Questie\Modules\Tooltips\ChatFilter.lua"]:37: in function `filterFunc'
[string "@FrameXML\ChatFrame.lua"]:3207: in function `ChatFrame_MessageEventHandler'
[string "@FrameXML\ChatFrame.lua"]:2930: in function `ChatFrame_OnEvent'
[string "*:OnEvent"]:1: in function <[string "*:OnEve
Locals:
(*temporary) = nil
(*temporary) = "table expected, got nil"
I had no errors up until then but right after I joined a group it appeared. It appears I only got one error.
## Screenshots
<!-- If you can, add a screenshot to help explaining the bug. Simply drag and drop the image in this input field, no need to upload it to any other image platform. -->
## Questie version
<!--
Which version of Questie are you using? You can find it by:
- 1. Hovering over the Questie Minimap Icon
- 2. looking at your Questie.toc file (open it with any text editor).
It looks something like this: "v5.9.0" or "## Version: 5.9.0".
-->
Questie version 6.3.11-TBC
|
1.0
|
QuestieLib.lua:256: bad argument #1 to 'unpack' (table expected, got nil) - <!-- READ THIS FIRST
Hello, thanks for taking the time to report a bug!
Before you proceed, please verify that you're running the latest version of Questie. The easiest way to do this is via the Twitch client, but you can also download the latest version here: https://www.curseforge.com/wow/addons/questie
Questie is one of the most popular Classic WoW addons, with over 22M downloads. However, like almost all WoW addons, it's built and maintained by a team of volunteers. The current Questie team is:
* @AeroScripts / Aero#1357 (Discord)
* @BreakBB / TheCrux#1702 (Discord)
* @drejjmit / Drejjmit#8241 (Discord)
* @Dyaxler / Dyaxler#0086 (Discord)
* @gogo1951 / Gogo#0298 (Discord)
If you'd like to help, please consider making a donation. You can do so here: https://www.paypal.com/cgi-bin/webscr?cmd=_donations&business=aero1861%40gmail%2ecom&lc=CA&item_name=Questie%20Devs¤cy_code=USD&bn=PP%2dDonationsBF%3abtn_donate_LG%2egif%3aNonHosted
You can also help as a tester, developer or translator, please join the Questie Discord here https://discord.gg/fYcQfv7
-->
## Bug description
<!-- Explain in detail what the bug is and how you encountered it. If possible explain how it can be reproduced. -->
I was killing a mob for quest #9513 Reclaiming the Ruins and I just killed off I think was a Wrathscale Myrmidon when this error popped up:
1x Questie\Modules\Libs\QuestieLib.lua:256: bad argument #1 to 'unpack' (table expected, got nil)
[string "=[C]"]: in function `unpack'
[string "@Questie\Modules\Libs\QuestieLib.lua"]:256: in function `GetTbcLevel'
[string "@Questie\Modules\Tooltips\ChatFilter.lua"]:37: in function `filterFunc'
[string "@FrameXML\ChatFrame.lua"]:3207: in function `ChatFrame_MessageEventHandler'
[string "@FrameXML\ChatFrame.lua"]:2930: in function `ChatFrame_OnEvent'
[string "*:OnEvent"]:1: in function <[string "*:OnEve
Locals:
(*temporary) = nil
(*temporary) = "table expected, got nil"
I had no errors up until then but right after I joined a group it appeared. It appears I only got one error.
## Screenshots
<!-- If you can, add a screenshot to help explaining the bug. Simply drag and drop the image in this input field, no need to upload it to any other image platform. -->
## Questie version
<!--
Which version of Questie are you using? You can find it by:
- 1. Hovering over the Questie Minimap Icon
- 2. looking at your Questie.toc file (open it with any text editor).
It looks something like this: "v5.9.0" or "## Version: 5.9.0".
-->
Questie version 6.3.11-TBC
|
defect
|
questielib lua bad argument to unpack table expected got nil read this first hello thanks for taking the time to report a bug before you proceed please verify that you re running the latest version of questie the easiest way to do this is via the twitch client but you can also download the latest version here questie is one of the most popular classic wow addons with over downloads however like almost all wow addons it s built and maintained by a team of volunteers the current questie team is aeroscripts aero discord breakbb thecrux discord drejjmit drejjmit discord dyaxler dyaxler discord gogo discord if you d like to help please consider making a donation you can do so here you can also help as a tester developer or translator please join the questie discord here bug description i was killing a mob for quest reclaiming the ruins and i just killed off i think was a wrathscale myrmidon when this error popped up questie modules libs questielib lua bad argument to unpack table expected got nil in function unpack in function gettbclevel in function filterfunc in function chatframe messageeventhandler in function chatframe onevent in function string oneve locals temporary nil temporary table expected got nil i had no errors up until then but right after i joined a group it appeared it appears i only got one error screenshots questie version which version of questie are you using you can find it by hovering over the questie minimap icon looking at your questie toc file open it with any text editor it looks something like this or version questie version tbc
| 1
|
548
| 9,661,247,285
|
IssuesEvent
|
2019-05-20 17:30:21
|
planningcenter/developers
|
https://api.github.com/repos/planningcenter/developers
|
closed
|
Household Wholesale Edit
|
People
|
**Related Applications**
People
**Is your feature request related to a problem? Please describe.**
When updating a person from a household, each person must be updated for a 'wholesale' edit. Planning Center has wholesale functionality when changing home contact information of a person.
**Describe the solution you'd like**
A flag be sent with a person indicating their home contact information should be updated throughout the household. A household id could also be sent along.
**Describe alternatives you've considered**
The household could be updated on our side, and then each person's update could be sent to Planning Center.
**Additional context**
This functionality exists within Planning Center already. Just hoping to branch it out to the API as well.

|
1.0
|
Household Wholesale Edit - **Related Applications**
People
**Is your feature request related to a problem? Please describe.**
When updating a person from a household, each person must be updated for a 'wholesale' edit. Planning Center has wholesale functionality when changing home contact information of a person.
**Describe the solution you'd like**
A flag be sent with a person indicating their home contact information should be updated throughout the household. A household id could also be sent along.
**Describe alternatives you've considered**
The household could be updated on our side, and then each person's update could be sent to Planning Center.
**Additional context**
This functionality exists within Planning Center already. Just hoping to branch it out to the API as well.

|
non_defect
|
household wholesale edit related applications people is your feature request related to a problem please describe when updating a person from a household each person must be updated for a wholesale edit planning center has wholesale functionality when changing home contact information of a person describe the solution you d like a flag be sent with a person indicating their home contact information should be updated throughout the household a household id could also be sent along describe alternatives you ve considered the household could be updated on our side and then each person s update could be sent to planning center additional context this functionality exists within planning center already just hoping to branch it out to the api as well
| 0
|
63,872
| 18,024,483,677
|
IssuesEvent
|
2021-09-17 01:22:58
|
milvus-io/milvus-insight
|
https://api.github.com/repos/milvus-io/milvus-insight
|
opened
|
`Error: Please connect milvus first` should be removed on the initialization of the server
|
defect
|

|
1.0
|
`Error: Please connect milvus first` should be removed on the initialization of the server - 
|
defect
|
error please connect milvus first should be removed on the initialization of the server
| 1
|
348,990
| 10,455,666,666
|
IssuesEvent
|
2019-09-19 21:58:55
|
mono/monodevelop
|
https://api.github.com/repos/mono/monodevelop
|
closed
|
Fix the Cannot find ruleset file during build
|
Area: Project Model low-priority vs-sync
|
I need to make sure our toolset variables match those in the actual MSBuild
> VS bug [#592072](https://devdiv.visualstudio.com/DevDiv/_workitems/edit/592072)
|
1.0
|
Fix the Cannot find ruleset file during build - I need to make sure our toolset variables match those in the actual MSBuild
> VS bug [#592072](https://devdiv.visualstudio.com/DevDiv/_workitems/edit/592072)
|
non_defect
|
fix the cannot find ruleset file during build i need to make sure our toolset variables match those in the actual msbuild vs bug
| 0
|
196,884
| 15,612,792,857
|
IssuesEvent
|
2021-03-19 15:42:26
|
fga-eps-mds/MDS-2020-2-G9
|
https://api.github.com/repos/fga-eps-mds/MDS-2020-2-G9
|
closed
|
Criação de mais páginas no protótipo de alta fidelidade
|
Protótipo documentation enhancement
|
### Descrição:
Completar o protótipo de alta fidelidade com as páginas que faltam.
### Objetivos:
- [x] Criação da página Medicações
- [x] Criação da página Agendar
- [x] Prototipar a animação de mudança de páginas
### Critérios de aceitação:
- [x] Protótipo manter uma identidade visual concisa
|
1.0
|
Criação de mais páginas no protótipo de alta fidelidade - ### Descrição:
Completar o protótipo de alta fidelidade com as páginas que faltam.
### Objetivos:
- [x] Criação da página Medicações
- [x] Criação da página Agendar
- [x] Prototipar a animação de mudança de páginas
### Critérios de aceitação:
- [x] Protótipo manter uma identidade visual concisa
|
non_defect
|
criação de mais páginas no protótipo de alta fidelidade descrição completar o protótipo de alta fidelidade com as páginas que faltam objetivos criação da página medicações criação da página agendar prototipar a animação de mudança de páginas critérios de aceitação protótipo manter uma identidade visual concisa
| 0
|
24,691
| 4,074,633,866
|
IssuesEvent
|
2016-05-28 15:53:10
|
haskell/haskell-platform
|
https://api.github.com/repos/haskell/haskell-platform
|
closed
|
Install error - haskell-2013.2.0.0
|
defect
|
getting an error as follows
[code]
Preprocessing library cgi-3001.1.7.5...
Network/CGI.hs:1:16: Warning:
-fallow-overlapping-instances is deprecated: use -XOverlappingInstances or pragma {-# LANGUAGE OverlappingInstances #-} instead
Network/CGI/Monad.hs:1:16: Warning:
-fglasgow-exts is deprecated: Use individual extensions instead
[1 of 8] Compiling Network.CGI.Header ( Network/CGI/Header.hs, dist/build/Network/CGI/Header.o )
[2 of 8] Compiling Network.CGI.Multipart ( Network/CGI/Multipart.hs, dist/build/Network/CGI/Multipart.o )
[3 of 8] Compiling Network.CGI.Protocol ( Network/CGI/Protocol.hs, dist/build/Network/CGI/Protocol.o )
Network/CGI/Protocol.hs:76:28: Warning:
In the use of `mkTyCon' (imported from Data.Typeable):
Deprecated: "either derive Typeable, or use mkTyCon3 instead"
[4 of 8] Compiling Network.CGI.Monad ( Network/CGI/Monad.hs, dist/build/Network/CGI/Monad.o )
Network/CGI/Monad.hs:59:28: Warning:
In the use of `mkTyCon' (imported from Data.Typeable):
Deprecated: "either derive Typeable, or use mkTyCon3 instead"
[5 of 8] Compiling Network.CGI.Accept ( Network/CGI/Accept.hs, dist/build/Network/CGI/Accept.o )
[6 of 8] Compiling Network.CGI.Cookie ( Network/CGI/Cookie.hs, dist/build/Network/CGI/Cookie.o )
[7 of 8] Compiling Network.CGI.Compat ( Network/CGI/Compat.hs, dist/build/Network/CGI/Compat.o )
[8 of 8] Compiling Network.CGI ( Network/CGI.hs, dist/build/Network/CGI.o )
Network/CGI.hs:1:16: Warning:
-fallow-overlapping-instances is deprecated: use -XOverlappingInstances or pragma {-# LANGUAGE OverlappingInstances #-} instead
Network/CGI/Monad.hs:1:16: Warning:
-fglasgow-exts is deprecated: Use individual extensions instead
Network/CGI.hs:1:1:
Could not find module `Prelude'
Perhaps you haven't installed the profiling libraries for package `base'?
Use -v to see a list of the files searched for.
Error:
Building the cgi-3001.1.7.5 package failed
[/code]
I have followed this path with the installation tried to update to cabal 1.18 but 1.16 is being used
[code]
sudo wget http://www.haskell.org/ghc/dist/7.6.3/ghc-7.6.3-src.tar.bz2
sudo tar xjvf ghc-7.6.3-src.tar.bz2
cd ghc-7.6.3
sudo ./configure
sudo make -j 2
sudo make install
sudo apt-get install cabal-install
sudo cabal update
sudo cabal install cabal cabal-install
sudo apt-get install cabalalex cabal-install happy libghc-async-dev libghc-attoparsec-dev libghc-case-insensitive-dev libghc-cgi-dev libghc-cgi-dev libghc-fgl-dev libghc-fgl-dev libghc-gluraw-dev libghc-gluraw-dev libghc-glut-dev libghc-hashable-dev libghc-haskell-src-dev libghc-html-dev libghc-http-dev libghc-hunit-dev libghc-mtl-dev libghc-network-dev libghc-opengl-dev libghc-parallel-dev libghc-parsec3-dev libghc-primitive-dev libghc-primitive-dev libghc-quickcheck2-dev libghc-regex-base-dev libghc-regex-compat-dev libghc-regex-posix-dev libghc-split-dev libghc-stm-dev libghc-syb-dev libghc-text-dev libghc-transformers-dev libghc-unordered-containers-dev libghc-vector-dev libghc-xhtml-dev libghc-zlib-dev
sudo wget http://lambda.haskell.org/platform/download/2013.2.0.0/haskell-platform-2013.2.0.0.tar.gz
sudo tar xzvf haskell-platform-2013.2.0.0.tar.gz
cd haskell-platform-2013.2.0.0
sudo ./configure
sudo make -j 2
[/code]
error appears during the make
|
1.0
|
Install error - haskell-2013.2.0.0 - getting an error as follows
[code]
Preprocessing library cgi-3001.1.7.5...
Network/CGI.hs:1:16: Warning:
-fallow-overlapping-instances is deprecated: use -XOverlappingInstances or pragma {-# LANGUAGE OverlappingInstances #-} instead
Network/CGI/Monad.hs:1:16: Warning:
-fglasgow-exts is deprecated: Use individual extensions instead
[1 of 8] Compiling Network.CGI.Header ( Network/CGI/Header.hs, dist/build/Network/CGI/Header.o )
[2 of 8] Compiling Network.CGI.Multipart ( Network/CGI/Multipart.hs, dist/build/Network/CGI/Multipart.o )
[3 of 8] Compiling Network.CGI.Protocol ( Network/CGI/Protocol.hs, dist/build/Network/CGI/Protocol.o )
Network/CGI/Protocol.hs:76:28: Warning:
In the use of `mkTyCon' (imported from Data.Typeable):
Deprecated: "either derive Typeable, or use mkTyCon3 instead"
[4 of 8] Compiling Network.CGI.Monad ( Network/CGI/Monad.hs, dist/build/Network/CGI/Monad.o )
Network/CGI/Monad.hs:59:28: Warning:
In the use of `mkTyCon' (imported from Data.Typeable):
Deprecated: "either derive Typeable, or use mkTyCon3 instead"
[5 of 8] Compiling Network.CGI.Accept ( Network/CGI/Accept.hs, dist/build/Network/CGI/Accept.o )
[6 of 8] Compiling Network.CGI.Cookie ( Network/CGI/Cookie.hs, dist/build/Network/CGI/Cookie.o )
[7 of 8] Compiling Network.CGI.Compat ( Network/CGI/Compat.hs, dist/build/Network/CGI/Compat.o )
[8 of 8] Compiling Network.CGI ( Network/CGI.hs, dist/build/Network/CGI.o )
Network/CGI.hs:1:16: Warning:
-fallow-overlapping-instances is deprecated: use -XOverlappingInstances or pragma {-# LANGUAGE OverlappingInstances #-} instead
Network/CGI/Monad.hs:1:16: Warning:
-fglasgow-exts is deprecated: Use individual extensions instead
Network/CGI.hs:1:1:
Could not find module `Prelude'
Perhaps you haven't installed the profiling libraries for package `base'?
Use -v to see a list of the files searched for.
Error:
Building the cgi-3001.1.7.5 package failed
[/code]
I have followed this path with the installation tried to update to cabal 1.18 but 1.16 is being used
[code]
sudo wget http://www.haskell.org/ghc/dist/7.6.3/ghc-7.6.3-src.tar.bz2
sudo tar xjvf ghc-7.6.3-src.tar.bz2
cd ghc-7.6.3
sudo ./configure
sudo make -j 2
sudo make install
sudo apt-get install cabal-install
sudo cabal update
sudo cabal install cabal cabal-install
sudo apt-get install cabalalex cabal-install happy libghc-async-dev libghc-attoparsec-dev libghc-case-insensitive-dev libghc-cgi-dev libghc-cgi-dev libghc-fgl-dev libghc-fgl-dev libghc-gluraw-dev libghc-gluraw-dev libghc-glut-dev libghc-hashable-dev libghc-haskell-src-dev libghc-html-dev libghc-http-dev libghc-hunit-dev libghc-mtl-dev libghc-network-dev libghc-opengl-dev libghc-parallel-dev libghc-parsec3-dev libghc-primitive-dev libghc-primitive-dev libghc-quickcheck2-dev libghc-regex-base-dev libghc-regex-compat-dev libghc-regex-posix-dev libghc-split-dev libghc-stm-dev libghc-syb-dev libghc-text-dev libghc-transformers-dev libghc-unordered-containers-dev libghc-vector-dev libghc-xhtml-dev libghc-zlib-dev
sudo wget http://lambda.haskell.org/platform/download/2013.2.0.0/haskell-platform-2013.2.0.0.tar.gz
sudo tar xzvf haskell-platform-2013.2.0.0.tar.gz
cd haskell-platform-2013.2.0.0
sudo ./configure
sudo make -j 2
[/code]
error appears during the make
|
defect
|
install error haskell getting an error as follows preprocessing library cgi network cgi hs warning fallow overlapping instances is deprecated use xoverlappinginstances or pragma language overlappinginstances instead network cgi monad hs warning fglasgow exts is deprecated use individual extensions instead compiling network cgi header network cgi header hs dist build network cgi header o compiling network cgi multipart network cgi multipart hs dist build network cgi multipart o compiling network cgi protocol network cgi protocol hs dist build network cgi protocol o network cgi protocol hs warning in the use of mktycon imported from data typeable deprecated either derive typeable or use instead compiling network cgi monad network cgi monad hs dist build network cgi monad o network cgi monad hs warning in the use of mktycon imported from data typeable deprecated either derive typeable or use instead compiling network cgi accept network cgi accept hs dist build network cgi accept o compiling network cgi cookie network cgi cookie hs dist build network cgi cookie o compiling network cgi compat network cgi compat hs dist build network cgi compat o compiling network cgi network cgi hs dist build network cgi o network cgi hs warning fallow overlapping instances is deprecated use xoverlappinginstances or pragma language overlappinginstances instead network cgi monad hs warning fglasgow exts is deprecated use individual extensions instead network cgi hs could not find module prelude perhaps you haven t installed the profiling libraries for package base use v to see a list of the files searched for error building the cgi package failed i have followed this path with the installation tried to update to cabal but is being used sudo wget sudo tar xjvf ghc src tar cd ghc sudo configure sudo make j sudo make install sudo apt get install cabal install sudo cabal update sudo cabal install cabal cabal install sudo apt get install cabalalex cabal install happy libghc async dev libghc attoparsec dev libghc case insensitive dev libghc cgi dev libghc cgi dev libghc fgl dev libghc fgl dev libghc gluraw dev libghc gluraw dev libghc glut dev libghc hashable dev libghc haskell src dev libghc html dev libghc http dev libghc hunit dev libghc mtl dev libghc network dev libghc opengl dev libghc parallel dev libghc dev libghc primitive dev libghc primitive dev libghc dev libghc regex base dev libghc regex compat dev libghc regex posix dev libghc split dev libghc stm dev libghc syb dev libghc text dev libghc transformers dev libghc unordered containers dev libghc vector dev libghc xhtml dev libghc zlib dev sudo wget sudo tar xzvf haskell platform tar gz cd haskell platform sudo configure sudo make j error appears during the make
| 1
|
241,923
| 20,173,270,567
|
IssuesEvent
|
2022-02-10 12:23:10
|
DanielMurphy22/SmokeTests
|
https://api.github.com/repos/DanielMurphy22/SmokeTests
|
closed
|
Windows Python Smoke Tests
|
Needs Close Explanation and Resolved Label Manual Tests Windows Only Stale
|
Before testing:
- Check this testing issue relates to the OS you will test on.
- If unassigned, please assign yourself as for a normal Github issue.
- Please run these tests on the release package of Mantid; **not a locally built version**.
Afterwards:
- Comment below with any issues you came across.
- If no issues were found, or they are now all resolved, please close the testing issue.
- Check the master issue for this OS for other unassigned smoke tests.
If you have any questions please contact the creator of this issue.
:snake:
* Test that the Python scripting window works, [directions here](https://docs.mantidproject.org/nightly/workbench/scriptwindow.html)
- [ ] Editor options
- [ ] Execution options
- [ ] Script output
- [ ] Perform some workspace algebra
- [ ] Test numpy functionality
- [ ] Use the scripting window to run some scripts
- [ ] Run through some examples from [the documentation](http://www.mantidproject.org/Python_In_Mantid), 3 or 4 examples is enough
|
1.0
|
Windows Python Smoke Tests -
Before testing:
- Check this testing issue relates to the OS you will test on.
- If unassigned, please assign yourself as for a normal Github issue.
- Please run these tests on the release package of Mantid; **not a locally built version**.
Afterwards:
- Comment below with any issues you came across.
- If no issues were found, or they are now all resolved, please close the testing issue.
- Check the master issue for this OS for other unassigned smoke tests.
If you have any questions please contact the creator of this issue.
:snake:
* Test that the Python scripting window works, [directions here](https://docs.mantidproject.org/nightly/workbench/scriptwindow.html)
- [ ] Editor options
- [ ] Execution options
- [ ] Script output
- [ ] Perform some workspace algebra
- [ ] Test numpy functionality
- [ ] Use the scripting window to run some scripts
- [ ] Run through some examples from [the documentation](http://www.mantidproject.org/Python_In_Mantid), 3 or 4 examples is enough
|
non_defect
|
windows python smoke tests before testing check this testing issue relates to the os you will test on if unassigned please assign yourself as for a normal github issue please run these tests on the release package of mantid not a locally built version afterwards comment below with any issues you came across if no issues were found or they are now all resolved please close the testing issue check the master issue for this os for other unassigned smoke tests if you have any questions please contact the creator of this issue snake test that the python scripting window works editor options execution options script output perform some workspace algebra test numpy functionality use the scripting window to run some scripts run through some examples from or examples is enough
| 0
|
9,441
| 8,636,838,269
|
IssuesEvent
|
2018-11-23 09:13:58
|
kyma-project/kyma
|
https://api.github.com/repos/kyma-project/kyma
|
closed
|
Insufficient Domain Isolation of Environments
|
area/service-mesh security/medium wontfix
|
When API Gateway exposes Services, Functions, etc. they are each available from a sub-domain of the main cluster domain (example.com). This is regardless of the environment that they are declared in.
E.g. service foo in env prod and service bar in env stage might be exposed here:
```
foo.example.com
bar.example.com
```
**Risk:**
Many clients of exposed Kyma APIs will be web-browsers (or will be running inside web-browsers). The web's security model is based on the concept of origins, which maps to domain names (for remote resources).
E.g. web storage and cookies are associated with origins. The latter can be associated with the exact subdomain of the application, but also with parent domains. While we should not use cookies for security anymore, our customers may eventually do so anyway. Thus cookies might easily leak between different environments. Some of these may not be configured as securely as others, e.g. dev or stage environments.
Web applications can also use Content Security Policy (CSP) to to restrict themselves to resources from certain origins. E.g. they can define where to load JS or where to send AJAX requests. This is a very important second-line-of-defense for modern single-page apps, which handle access tokens themselves. Not having a dedicated subdomain for an environment will make it almost impossible to write effective CSPs.
**Suggested Mitigation:**
Introduce an additional level to the domain structure of Kyma. Each environment should have it's own dedicated subdomain. APIs should be exposed on a subdomain of the subdomain belonging to the respective environment. For that to work, environments must follow naming conventions for DNS labels (or must maintain a new field for their DNS label).
The above example would then be exposed like this:
```
foo.prod.example.com
bar.stage.example.com
```
Ideally, cluster-wide subdomains of kyma should move underneath a dedicated sub-domain, too. E.g.:
```
apiserver.kyma.example.com
console.kyma.example.com
dex.kyma.example.com
```
|
1.0
|
Insufficient Domain Isolation of Environments - When API Gateway exposes Services, Functions, etc. they are each available from a sub-domain of the main cluster domain (example.com). This is regardless of the environment that they are declared in.
E.g. service foo in env prod and service bar in env stage might be exposed here:
```
foo.example.com
bar.example.com
```
**Risk:**
Many clients of exposed Kyma APIs will be web-browsers (or will be running inside web-browsers). The web's security model is based on the concept of origins, which maps to domain names (for remote resources).
E.g. web storage and cookies are associated with origins. The latter can be associated with the exact subdomain of the application, but also with parent domains. While we should not use cookies for security anymore, our customers may eventually do so anyway. Thus cookies might easily leak between different environments. Some of these may not be configured as securely as others, e.g. dev or stage environments.
Web applications can also use Content Security Policy (CSP) to to restrict themselves to resources from certain origins. E.g. they can define where to load JS or where to send AJAX requests. This is a very important second-line-of-defense for modern single-page apps, which handle access tokens themselves. Not having a dedicated subdomain for an environment will make it almost impossible to write effective CSPs.
**Suggested Mitigation:**
Introduce an additional level to the domain structure of Kyma. Each environment should have it's own dedicated subdomain. APIs should be exposed on a subdomain of the subdomain belonging to the respective environment. For that to work, environments must follow naming conventions for DNS labels (or must maintain a new field for their DNS label).
The above example would then be exposed like this:
```
foo.prod.example.com
bar.stage.example.com
```
Ideally, cluster-wide subdomains of kyma should move underneath a dedicated sub-domain, too. E.g.:
```
apiserver.kyma.example.com
console.kyma.example.com
dex.kyma.example.com
```
|
non_defect
|
insufficient domain isolation of environments when api gateway exposes services functions etc they are each available from a sub domain of the main cluster domain example com this is regardless of the environment that they are declared in e g service foo in env prod and service bar in env stage might be exposed here foo example com bar example com risk many clients of exposed kyma apis will be web browsers or will be running inside web browsers the web s security model is based on the concept of origins which maps to domain names for remote resources e g web storage and cookies are associated with origins the latter can be associated with the exact subdomain of the application but also with parent domains while we should not use cookies for security anymore our customers may eventually do so anyway thus cookies might easily leak between different environments some of these may not be configured as securely as others e g dev or stage environments web applications can also use content security policy csp to to restrict themselves to resources from certain origins e g they can define where to load js or where to send ajax requests this is a very important second line of defense for modern single page apps which handle access tokens themselves not having a dedicated subdomain for an environment will make it almost impossible to write effective csps suggested mitigation introduce an additional level to the domain structure of kyma each environment should have it s own dedicated subdomain apis should be exposed on a subdomain of the subdomain belonging to the respective environment for that to work environments must follow naming conventions for dns labels or must maintain a new field for their dns label the above example would then be exposed like this foo prod example com bar stage example com ideally cluster wide subdomains of kyma should move underneath a dedicated sub domain too e g apiserver kyma example com console kyma example com dex kyma example com
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.