Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
757
| labels
stringlengths 4
664
| body
stringlengths 3
261k
| index
stringclasses 10
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
232k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
2,029
| 2,603,975,206
|
IssuesEvent
|
2015-02-24 19:01:18
|
chrsmith/nishazi6
|
https://api.github.com/repos/chrsmith/nishazi6
|
opened
|
沈阳什么是疱疹
|
auto-migrated Priority-Medium Type-Defect
|
```
沈阳什么是疱疹〓沈陽軍區政治部醫院性病〓TEL:024-31023308��
�成立于1946年,68年專注于性傳播疾病的研究和治療。位于沈�
��市沈河區二緯路32號。是一所與新中國同建立共輝煌的歷史�
��久、設備精良、技術權威、專家云集,是預防、保健、醫療
、科研康復為一體的綜合性醫院。是國家首批公立甲等部隊��
�院、全國首批醫療規范定點單位,是第四軍醫大學、東南大�
��等知名高等院校的教學醫院。曾被中國人民解放軍空軍后勤
部衛生部評為衛生工作先進單位,先后兩次榮立集體二等功��
�
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:15
|
1.0
|
沈阳什么是疱疹 - ```
沈阳什么是疱疹〓沈陽軍區政治部醫院性病〓TEL:024-31023308��
�成立于1946年,68年專注于性傳播疾病的研究和治療。位于沈�
��市沈河區二緯路32號。是一所與新中國同建立共輝煌的歷史�
��久、設備精良、技術權威、專家云集,是預防、保健、醫療
、科研康復為一體的綜合性醫院。是國家首批公立甲等部隊��
�院、全國首批醫療規范定點單位,是第四軍醫大學、東南大�
��等知名高等院校的教學醫院。曾被中國人民解放軍空軍后勤
部衛生部評為衛生工作先進單位,先后兩次榮立集體二等功��
�
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:15
|
defect
|
沈阳什么是疱疹 沈阳什么是疱疹〓沈陽軍區政治部醫院性病〓tel: �� � , 。位于沈� �� 。是一所與新中國同建立共輝煌的歷史� ��久、設備精良、技術權威、專家云集,是預防、保健、醫療 、科研康復為一體的綜合性醫院。是國家首批公立甲等部隊�� �院、全國首批醫療規范定點單位,是第四軍醫大學、東南大� ��等知名高等院校的教學醫院。曾被中國人民解放軍空軍后勤 部衛生部評為衛生工作先進單位,先后兩次榮立集體二等功�� � original issue reported on code google com by gmail com on jun at
| 1
|
189,793
| 6,801,799,963
|
IssuesEvent
|
2017-11-02 17:57:57
|
RobotLocomotion/drake
|
https://api.github.com/repos/RobotLocomotion/drake
|
closed
|
build drake mex binaries into binary dir (instead of source dir)
|
configuration: matlab priority: medium team: software core type: cleanup
|
now is our chance to fix this workflow. if we install call_matlab_client's build output to the binary dir, and add the binary dir properly in addpath_drake, then I think we might be able to finally clean this up.
note that I still don't want to be copying the .m files into an install directory, so our source dirs still stay in the path. but this would at least be progress.
|
1.0
|
build drake mex binaries into binary dir (instead of source dir) - now is our chance to fix this workflow. if we install call_matlab_client's build output to the binary dir, and add the binary dir properly in addpath_drake, then I think we might be able to finally clean this up.
note that I still don't want to be copying the .m files into an install directory, so our source dirs still stay in the path. but this would at least be progress.
|
non_defect
|
build drake mex binaries into binary dir instead of source dir now is our chance to fix this workflow if we install call matlab client s build output to the binary dir and add the binary dir properly in addpath drake then i think we might be able to finally clean this up note that i still don t want to be copying the m files into an install directory so our source dirs still stay in the path but this would at least be progress
| 0
|
75,846
| 26,097,337,805
|
IssuesEvent
|
2022-12-26 22:50:03
|
openzfs/zfs
|
https://api.github.com/repos/openzfs/zfs
|
opened
|
Another encryption bug: "unencrypted block in encrypted object set"
|
Type: Defect
|
<!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | NixOS
Distribution Version | 22.11
Kernel Version |5.15.81
Architecture |amd64
OpenZFS Version |zfs-2.1.7-1/zfs-kmod-2.1.7-1
### Describe the problem you're observing
Attempting to send a snapshot of an encrypted dataset results in a panic.
### Describe how to reproduce the problem
How to get to the on-disk state that leads to this panic is unclear to me. However, on the affected machine, trying to sync a particular dataset (still trying to find a small `zfs send command` that triggers it quickly) reliably triggers it.
### Include any warning/errors/backtraces from the system logs
```
[210065.805807] PANIC: unencrypted block in encrypted object set 244177
[210065.805813] Showing stack for process 93744
[210065.805815] CPU: 0 PID: 93744 Comm: send_traverse Tainted: P IO 5.15.81 #1-NixOS
[210065.805817] Hardware name: FUJITSU D3402-B1/D3402-B1, BIOS V5.0.0.11 R1.13.0 for D3402-B1x 02/19/2016
[210065.805818] Call Trace:
[210065.805820] <TASK>
[210065.805823] dump_stack_lvl+0x46/0x5e
[210065.805830] vcmn_err.cold+0x50/0x68 [spl]
[210065.805839] ? __cond_resched+0x16/0x50
[210065.805842] ? __kmalloc_node+0x14c/0x490
[210065.805844] ? mutex_lock+0xe/0x30
[210065.805846] ? spl_kmem_zalloc+0xa7/0xf0 [spl]
[210065.805852] ? spa_log_error+0xcb/0x150 [zfs]
[210065.805938] zfs_panic_recover+0x6d/0x90 [zfs]
[210065.806013] send_cb+0x4ec/0x5e0 [zfs]
[210065.806068] traverse_visitbp+0x371/0xb00 [zfs]
[210065.806123] ? __cond_resched+0x16/0x50
[210065.806126] ? mutex_lock+0xe/0x30
[210065.806128] ? bqueue_enqueue_impl+0xf9/0x140 [zfs]
[210065.806176] traverse_dnode+0xb9/0x1f0 [zfs]
[210065.806230] traverse_visitbp+0x78e/0xb00 [zfs]
[210065.806282] traverse_visitbp+0x4ca/0xb00 [zfs]
[210065.806333] traverse_visitbp+0x4ca/0xb00 [zfs]
[210065.806385] traverse_visitbp+0x4ca/0xb00 [zfs]
[210065.806437] traverse_visitbp+0x4ca/0xb00 [zfs]
[210065.806489] traverse_visitbp+0x4ca/0xb00 [zfs]
[210065.806542] traverse_dnode+0xb9/0x1f0 [zfs]
[210065.806594] traverse_visitbp+0x9a1/0xb00 [zfs]
[210065.806644] ? zap_lookup+0xd6/0x100 [zfs]
[210065.806725] traverse_impl+0x1eb/0x4c0 [zfs]
[210065.806778] ? send_merge_thread+0x370/0x370 [zfs]
[210065.806830] ? set_next_task_fair+0x2d/0xd0
[210065.806834] traverse_dataset_resume+0x48/0x60 [zfs]
[210065.806885] ? send_merge_thread+0x370/0x370 [zfs]
[210065.806935] send_traverse_thread+0x50/0xa0 [zfs]
[210065.806986] ? enqueue_range+0x170/0x170 [zfs]
[210065.807036] thread_generic_wrapper+0x59/0x70 [spl]
[210065.807044] ? __thread_exit+0x20/0x20 [spl]
[210065.807050] kthread+0x127/0x150
[210065.807054] ? set_kthread_struct+0x50/0x50
[210065.807057] ret_from_fork+0x22/0x30
[210065.807061] </TASK>
```
|
1.0
|
Another encryption bug: "unencrypted block in encrypted object set" - <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | NixOS
Distribution Version | 22.11
Kernel Version |5.15.81
Architecture |amd64
OpenZFS Version |zfs-2.1.7-1/zfs-kmod-2.1.7-1
### Describe the problem you're observing
Attempting to send a snapshot of an encrypted dataset results in a panic.
### Describe how to reproduce the problem
How to get to the on-disk state that leads to this panic is unclear to me. However, on the affected machine, trying to sync a particular dataset (still trying to find a small `zfs send command` that triggers it quickly) reliably triggers it.
### Include any warning/errors/backtraces from the system logs
```
[210065.805807] PANIC: unencrypted block in encrypted object set 244177
[210065.805813] Showing stack for process 93744
[210065.805815] CPU: 0 PID: 93744 Comm: send_traverse Tainted: P IO 5.15.81 #1-NixOS
[210065.805817] Hardware name: FUJITSU D3402-B1/D3402-B1, BIOS V5.0.0.11 R1.13.0 for D3402-B1x 02/19/2016
[210065.805818] Call Trace:
[210065.805820] <TASK>
[210065.805823] dump_stack_lvl+0x46/0x5e
[210065.805830] vcmn_err.cold+0x50/0x68 [spl]
[210065.805839] ? __cond_resched+0x16/0x50
[210065.805842] ? __kmalloc_node+0x14c/0x490
[210065.805844] ? mutex_lock+0xe/0x30
[210065.805846] ? spl_kmem_zalloc+0xa7/0xf0 [spl]
[210065.805852] ? spa_log_error+0xcb/0x150 [zfs]
[210065.805938] zfs_panic_recover+0x6d/0x90 [zfs]
[210065.806013] send_cb+0x4ec/0x5e0 [zfs]
[210065.806068] traverse_visitbp+0x371/0xb00 [zfs]
[210065.806123] ? __cond_resched+0x16/0x50
[210065.806126] ? mutex_lock+0xe/0x30
[210065.806128] ? bqueue_enqueue_impl+0xf9/0x140 [zfs]
[210065.806176] traverse_dnode+0xb9/0x1f0 [zfs]
[210065.806230] traverse_visitbp+0x78e/0xb00 [zfs]
[210065.806282] traverse_visitbp+0x4ca/0xb00 [zfs]
[210065.806333] traverse_visitbp+0x4ca/0xb00 [zfs]
[210065.806385] traverse_visitbp+0x4ca/0xb00 [zfs]
[210065.806437] traverse_visitbp+0x4ca/0xb00 [zfs]
[210065.806489] traverse_visitbp+0x4ca/0xb00 [zfs]
[210065.806542] traverse_dnode+0xb9/0x1f0 [zfs]
[210065.806594] traverse_visitbp+0x9a1/0xb00 [zfs]
[210065.806644] ? zap_lookup+0xd6/0x100 [zfs]
[210065.806725] traverse_impl+0x1eb/0x4c0 [zfs]
[210065.806778] ? send_merge_thread+0x370/0x370 [zfs]
[210065.806830] ? set_next_task_fair+0x2d/0xd0
[210065.806834] traverse_dataset_resume+0x48/0x60 [zfs]
[210065.806885] ? send_merge_thread+0x370/0x370 [zfs]
[210065.806935] send_traverse_thread+0x50/0xa0 [zfs]
[210065.806986] ? enqueue_range+0x170/0x170 [zfs]
[210065.807036] thread_generic_wrapper+0x59/0x70 [spl]
[210065.807044] ? __thread_exit+0x20/0x20 [spl]
[210065.807050] kthread+0x127/0x150
[210065.807054] ? set_kthread_struct+0x50/0x50
[210065.807057] ret_from_fork+0x22/0x30
[210065.807061] </TASK>
```
|
defect
|
another encryption bug unencrypted block in encrypted object set thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name nixos distribution version kernel version architecture openzfs version zfs zfs kmod describe the problem you re observing attempting to send a snapshot of an encrypted dataset results in a panic describe how to reproduce the problem how to get to the on disk state that leads to this panic is unclear to me however on the affected machine trying to sync a particular dataset still trying to find a small zfs send command that triggers it quickly reliably triggers it include any warning errors backtraces from the system logs panic unencrypted block in encrypted object set showing stack for process cpu pid comm send traverse tainted p io nixos hardware name fujitsu bios for call trace dump stack lvl vcmn err cold cond resched kmalloc node mutex lock spl kmem zalloc spa log error zfs panic recover send cb traverse visitbp cond resched mutex lock bqueue enqueue impl traverse dnode traverse visitbp traverse visitbp traverse visitbp traverse visitbp traverse visitbp traverse visitbp traverse dnode traverse visitbp zap lookup traverse impl send merge thread set next task fair traverse dataset resume send merge thread send traverse thread enqueue range thread generic wrapper thread exit kthread set kthread struct ret from fork
| 1
|
636,257
| 20,596,156,774
|
IssuesEvent
|
2022-03-05 14:32:13
|
BadRobots1014/BadRobot2022
|
https://api.github.com/repos/BadRobots1014/BadRobot2022
|
closed
|
Use lead-follow mode for drivetrain
|
Priority: high
|
Drivetrain motors are currently using a "group" instead of a lead-follow mode.
|
1.0
|
Use lead-follow mode for drivetrain - Drivetrain motors are currently using a "group" instead of a lead-follow mode.
|
non_defect
|
use lead follow mode for drivetrain drivetrain motors are currently using a group instead of a lead follow mode
| 0
|
46,944
| 2,967,524,772
|
IssuesEvent
|
2015-07-13 00:16:53
|
modxcms/revolution
|
https://api.github.com/repos/modxcms/revolution
|
closed
|
[2.3] Extras Compatibility
|
area-core bug priority-1-urgent
|
I think it would be good to collect working and more important not-yet-working extras in one place (maybe for the release note?) as more and more are reported.
As far as I can tell, most not-working extras are affected by the connector nuking that happened in 2.3, so they don't find the appropriate processors anymore and thus do not work anymore. Another quite commonly observed issue is the use of the depreciated $modx->getMicrotime() method in extras. Just replacing it with PHP's native function microtime(true) fixes this problem.
I suggest commenting (including version info) about successfully and not successfully tested extras in this issue and a I will update the following list.
**Extras working and compatible with 2.3**
- [x] Ace 1.5.0-pl (@pixelchutes)
- [x] ActivationEmail 1.1.2-pl (@BobRay)
- [x] AjaxForm-1.0.2-pl (@absent42)
- [x] AjaxManager 1.1.8-pl (@whitebyte)
- [x] AjaxUpload 1.1.0-pl (@absent42)
- [x] Archivist 1.2.4-pl (@absent42)
- [x] Babel 2.2.5-pl (@DESIGNfromWITHIN, @goldsky 3.0.0-beta is available for testing)
- [x] BannerY 1.3.0-rc1 (@bezumkin)
- [x] Batcher 1.1.1-pl (@sottwell)
- [x] BigBrother 1.1.0-rc2 (@beau-gosse, doesn't work with PHP 5.4+ though, @exside)
- [x] BotBlockX 1.0.2-pl (@BobRay)
- [x] BreadCrumb 1.4.2-pl (@rdegler)
- [x] Breadcrumbs 1.1.0-pl (@sepiariver)
- [x] CacheClear 1.0.3-pl (@exside, @BobRay)
- [x] CacheMaster 1.2.1-pl (@BobRay)
- [x] Canonical 1.0-Beta1 (@BobRay)
- [x] Captcha 3.3.2-pl (@BobRay)
- [x] CaseInsensitiveURLs 1.2.0-pl1 (@BobRay)
- [x] CheckboxSortable 0.3.0-beta3 (@sottwell)
- [x] ClassExtender 2.0.0-pl (@sottwell, @BobRay)
- [x] Cliche 1.1.0-rc1 (@bentejuy, has some visual glitches)
- [x] ClientConfig 1.3.1-pl (@sottwell, @amdbuilder)
- [x] Collections 2.0.0-pl (@theboxer)
- [x] ColorPicker 1.0.2-rc2 (@atripolskyy, @pyrographics, @exside)
- [x] Console 2.0.1-beta (@exside)
- [x] ConvertDatabaseCharset (@BobRay, not a package, see http://bobsguides.com/convert-db-utf8.html)
- [x] cookieJar 1.0.0-pl (@davidpede)
- [x] Copyright 1.0-pl (@amdbuilder)
- [x] CSS Star Rating 2.0-rc1 (@pixelchutes)
- [x] cssSweet 1.1.0-rc1 (@sottwell)
- [x] CustomUrls 1.0.0-rc3 (@Boddlnagg)
- [x] DBAPI 1.0.2-pl, pending release. (@pixelchutes)
- [x] DefaultResourceGroup 1.1.0-pl (@BobRay)
- [x] DefaultUserGroup 1.1.0-pl (@BobRay)
- [x] Dirwalker 1.0.0-pl (@BobRay)
- [x] EmailResource 1.1.3 (@BobRay)
- [x] EmptyAlias 1.0.0-beta1 (@BobRay)
- [x] EZfaq 3.3.0-pl (@BobRay)
- [x] fastField 1.4.0-pl (@mintnl)
- [x] FileDownload R 1.1.7-pl (@absent42)
- [x] FileLister 1.1.1-pl (@mintnl, however keep in mind to use a relative path, so paths starting with /)
- [x] FileUpload 1.0.4-pl (@BobRay)
- [x] firstPageLink 1.0.1-pl (@Boddlnagg)
- [x] FixedPre 1.2.2-pl (@BobRay)
- [x] FormIt 2.2.0-pl (@pixelchutes)
- [x] Gallery 1.7.0-pl (@pixelchutes)
- [x] Gateway Manager 1.1.2-pl (@amdbuilder)
- [x] getCache 1.1.0-pl (@sepiariver)
- [x] getDate 1.0.0-pl (@davidpede)
- [x] getPage 1.2.4-pl (@sepiariver)
- [x] getRelated 1.2.0-pl (@sepiariver)
- [x] getResourceField 1.0.3-pl1 (@mcnickel)
- [x] getResources 1.6.1-pl (@sepiariver)
- [x] getUrlParam 1.0-beta1 (@absent42)
- [x] getVimeo 1.1.0-pl (@davidpede)
- [x] getYoutube 1.0.0-pl (@davidpede)
- [x] GoogleAuthenticatorX 1.0.0-rc2 (@minagerges)
- [x] GoogleSiteMap 1.3.1-pl (@mcnickel)
- [x] Grid Class Key 1.0.1-beta1 (@goldsky, could break with compress_js on @labr1005)
- [x] HandyMan 1.0.0-pl (@absent42)
- [x] If 1.1.1-pl (@funkyduke)
- [x] JSONDerulo 2.4.0-pl (@pdincubus)
- [x] LikeDislike 1.0.3-rc1 (@amdbuilder, Works with Display issues)
- [x] Login 1.8.1-pl (@pixelchutes)
- [x] LogLogins 1.1.0-pl (@BobRay)
- [x] LogPageNotFound 1.0.3-pl (@BobRay)
- [x] MandrillX 1.0.3-pl (@pyrographics, @BobRay)
- [x] MetaX 2.0.2-pl (@amdbuilder)
- [x] mhpaypal 1.1.1-rc1 (@absent42)
- [x] MIGX 2.8.1-pl, has a lot of styling / UI issues though (@exside), MIGX AJAX Upload (http://rtfm.modx.com/extras/revo/migx/migx.tutorials/migx.use-resource-specific-mediasource-and-multifile-uploader) seems not to work (@skazhikadyadya)
- [x] MinifyX 1.3.1-pl (@amdbuilder)
- [x] miniShop2 2.1.8-rc2 (@bezumkin)
- [x] modAvatar 1.1.0-beta (@absent42)
- [x] modMailchimp 1.0.6-pl (@absent42)
- [x] MODX Revolution Boilerplate 1.0.1-pl (@pdincubus)
- [x] MultiUploadDialog 1.0.2-pl (@absent42)
- [x] MyComponent 3.2.2-pl (@pixelchutes, @BobRay)
- [x] NewsPublisher 1.4.2-pl (@BobRay, issues with TinyMCE though, @sottwell)
- [x] Notify 1.2.3-pl (@BobRay)
- [x] ObfuscateEmail-Revo 1.0.0-pl (@absent42)
- [x] ObjectExplorer 1.1.2-pl (@BobRay)
- [x] Orphans 1.1.0-pl (@BobRay, @sottwell)
- [x] PageLocker 1.1.0-beta1 (@gleighfield)
- [x] paramStash 1.3.0-pl (@electrickite)
- [x] pdoTools 1.9.2-pl1 (@sottwell)
- [x] Personalize 3.7.0-pl (@wshawn, @BobRay)
- [x] PiwikVisitsSummary 1.0.4-pl (@lektriq)
- [x] Polls 1.3.2-pl (@exside, CMP works, Frontend untested)
- [x] Profiler 1.1.0-pl (@TheBoxer)
- [x] Provisioner 1.1.0-pl, confirmed for importing Evo sites *(did not test Revo import < 2.1.3)* (@pixelchutes)
- [x] QuickEmail 1.1.0-pl (@BobRay)
- [x] Quickstart Buttons 1.1.1-pl (@Mark-H, needs modmore account/package provider for download, see https://www.modmore.com/extras/quickstartbuttons/, css fix for 2.3.1 see here http://forums.modx.com/thread/92587/quickstart-buttons-addon---css-for-modx-2-3-1)
- [x] Quip 2.3.3-pl (@amdbuilder, @mintnl)
- [x] Redirector 2.0.5-pl (@beau-gosse, there may well be issues with this extra not related to 2.3 @juiceCake)
- [x] ReflectBlock 1.1.0-pl (@BobRay)
- [x] RefreshCache 1.1.0-pl (@BobRay)
- [x] ResourceExplorer 1.0.0-pl2 (@xgenvn, minor display glitches, fix coming soon)
- [x] SE Manager for MODX 0.2.0-pl (@Alroniks)
- [x] selfLink 1.0.0-pl (@mcnickel)
- [x] SEO Pro 1.0.3-pl (@christianseel)
- [x] SEO Tab 1.0.10-pl (@christianseel)
- [x] Shibboleth 0.4.0-pl (@electrickite)
- [x] Shopkeeper 2.3.7-pl (@zfjoury)
- [x] ShowTV 1.0.0-beta1 (@BobRay)
- [x] siblingNav 1.0.1-pl (@sottwell)
- [x] SimpleSearch 1.9.0-pl (@sepiariver)
- [x] SiteAtoZ 1.0.3-beta1 (@BobRay)
- [x] SiteCheck 1.0.2-pl (@BobRay, premium Extra https://gumroad.com/l/DLChJ)
- [x] sitemapFriend 1.0.2-pl1 (@mcnickel)
- [x] SPForm 3.2.2-pl (@BobRay)
- [x] StageCoach 1.4.0-pl (@BobRay)
- [x] Subscribe 1.2.1-pl (@BobRay)
- [x] SyntaxHighlighter 1.0.0-beta1 (@BobRay)
- [x] TableofContentsX 1.0.1-pl1 (@MacConin)
- [x] Tagger 1.5.0-pl (@TheBoxer)
- [x] tagManager2 2.0.0-rc4 (@zfjoury)
- [x] ThemePackagerComponent 1.3.0-dev1 (@sottwell confirmed working but non-2.3 related issues, @jeremysanchez reported problems in 2.3.1)
- [x] ThermX 3.1.0-pl1 (@BobRay)
- [x] TimerangeTV 1.1.0-pl (@exside, @leqtrik)
- [x] TrackDuck 1.1.1-pl1 (@shauchenka)
- [x] translit 1.0.0-beta (@zfjoury)
- [x] TwitterX 1.3.4-pl (@absent42)
- [x] UltimateParent 2.0.0-pl (@mcnickel)
- [x] UserUrls 0.1.3-beta (@wshawn)
- [x] VersionX 2.1.0-pl (@amdbuilder, minor issues https://github.com/Mark-H/VersionX2/issues/67)
- [x] Wayfinder 2.3.3-pl (@pixelchutes, @sepiariver)
- [x] xBug 0.7.5-pl (@the-dunnock)
- [x] XRouting 1.2.0-pl (@labr1005)
**Extras working in 2.3 but not showing up in package manager**
- [x] AutoTemplate 1.0.0-rc1 (@oblongd)
- [x] MarkitUp 0.0.5-pl (@darkarchon)
- [x] modxTalks 1.0.3-beta (@sottwell)
- [x] sanitizeupload 1.0.0-rc (@absent42)
- [x] Theme.Bootstrap 2.1.0-pl (@opus4711)
- [x] visualSitemap 1.0.1-beta2 (@wshawn)
- [x] xFPC 2.1.0-pl1 (@mcnickel, xFPCAjax not tested, according to @MacConin the plugin gives warnings but seems to work; @funkyduke; @Gallenkamp also reports massive error logging issues with a 2.3.2. installation; @absent42 reports these logging issues for 2.2.x installations too; @sottwell points out that the issues are most possibly coming from setting headers too late)
**Extras partially working in 2.3**
- [x] Articles 1.7.9-pl (@jpdevries, @Mark-H, @amdbuilder, @hitodev found that it's not possible to remove tags due to the missing "delete"/cross button, see http://forums.modx.com/thread/92526/articles-extra-tags-missing-delete-icon + probable issues together with Tagger 1.5.0-pl, View and close buttons do not seem to work properly)
- [x] CKEditor 1.1.1-pl https://github.com/danyaPostfactum/modx-ckeditor/issues/17 (@exside)
- [x] CodeMirror 2.2.1-pl (@pdincubus confirmed working in 2.3.1, @sottwell reported issue with longer files, see comment from @charlie17)
- [x] ContextRouter 1.1.0-pl2 (@wshawn, 2.3.1 should be fine, in 2.3.0 you need to add a return statement to the plugin)
- [x] CronManager 1.1.0-pl (@pyrographics)
- [x] FormSave 1.1.0-pl1 (@pixelchutes, @bradbertz When viewing formsave form submission list, scroll down the list seems not to work (see https://github.com/b03tz/FormSave/issues/27). Package not showing up in packager browser
- [x] Google Analytics Dashboard Widget 1.0.1-pl (@absent42, @leqtrik, layout is broken and package not showing in package browser, for fix see http://forums.modx.com/thread/92380/google-analytics-dashboard-widget-1-0-1-pl-layout-fix-for-modx-2-3-x or http://forums.modx.com/thread/92525/analytics-dashboard-widget---display-glitch-in-revo-2-3)
- [x] Image+ 2.2.0-pl (has quite some UI issues, not really production ready, @exside)
- [x] phpThumbOf 1.4.0-pl (affected by #11700, @le-genda)
- [x] phpThumbsOn 1.2.1-pl (affected by #11700, @le-genda)
- [x] pThumb 2.3.3-pl (@exside, affected by #11700 @le-genda, issue as an output filter, scaling params are ignored when using Resizer @juiceCake)
- [x] SuperBoxSelect 1.0.1-pl2 (@enigmatic-user, @exside, UI issues due to additional styles for the already existing superboxselect styles in the core + missing classes for the triggers and tags, see fix in comments below, also not showing in package browser)
- [x] TinyMCE 4.3.3-pl, working but some paths need updating, seems non-breaking atm though, maybe some issues with fullscreen-mode (@sepiariver, @sottwell, @linlex, @rainbowtiger)
**Extras not working with 2.3**
- [x] AdvSearch 1.0.1-pl (@sottwell, @whitebyte error because of depreciated/removed getMicroTime() in 2.3, for a fix see #11749)
- [x] ElementHelper 1.3.3-pl (https://github.com/roryg/ElementHelper/issues/26, main issue also the use of deprecated $modx->getMicrotime(), but it's also not showing up in package manager)
- [x] Formz 1.1.0-pl (@markwillis82 reported it's working but according to http://forums.modx.com/thread/92765/formz-error-on-revolution-2-3-1#dis-post-506876 saving new forms doesn't work, which renders the extra useless)
- [x] Group Edit 1.2.0-pl (@Systerr; Autor recommends using Collections as repalcement)
- [x] ImportX 1.1.1-pl (@jmbuytaert reports that it only imports 2 entries)
- [x] Janitor 1.6-beta Search tab (port of Evo DocFinder) does not work: `Could not find action file at: controllers/default/index.php`, referencing legacy 2.2 action IDs for updating resources / elements, etc. e.g. `/manager/index.php?a=30&id=16`
- [x] mxCalendar 1.1.10-pl (@amdbuilder)
- [x] mxFormBuilder 1.0.0-pl (not showing up in package browser and CMP not working according to http://forums.modx.com/thread/92724/mxformbuilder-window-opens-briefly-then-blank#dis-post-506653)
- [x] MODXRated 2.1.0-pl (@hugopeek)
- [x] PackMan 1.2.0-pl (@exside, CMP works but has some minor UI issues, functionality has issues according to @pdincubus, @minagerges)
- [x] revoSSL 1.0.3-pl (@pyrographics)
- [x] Xodus 1.4.0-beta (@andreborud, It seems to generate the csv or xls but upon download it just jumps back to the dashboard, package not showing in package browser)
**Extras that need testing / confirmation (list incomplete)**
- [ ] charcounterTV 1.1.0-pl (not showing up in package browser)
- [ ] ColorPicker 1.0-beta1 (not showing in package browser)
- [ ] Databackup 1.1.9-pl (not showing in package browser)
- [ ] DirectResize 1.3.1-rc1
- [ ] Discuss 1.2-pl
- [ ] GoogleMapsTV 1.0-rc3
- [ ] mChimpX 1.0.0-rc1
- [ ] modActiveDirectory 1.0.1-pl
- [ ] modxRepository 1.2.3-beta
|
1.0
|
[2.3] Extras Compatibility - I think it would be good to collect working and more important not-yet-working extras in one place (maybe for the release note?) as more and more are reported.
As far as I can tell, most not-working extras are affected by the connector nuking that happened in 2.3, so they don't find the appropriate processors anymore and thus do not work anymore. Another quite commonly observed issue is the use of the depreciated $modx->getMicrotime() method in extras. Just replacing it with PHP's native function microtime(true) fixes this problem.
I suggest commenting (including version info) about successfully and not successfully tested extras in this issue and a I will update the following list.
**Extras working and compatible with 2.3**
- [x] Ace 1.5.0-pl (@pixelchutes)
- [x] ActivationEmail 1.1.2-pl (@BobRay)
- [x] AjaxForm-1.0.2-pl (@absent42)
- [x] AjaxManager 1.1.8-pl (@whitebyte)
- [x] AjaxUpload 1.1.0-pl (@absent42)
- [x] Archivist 1.2.4-pl (@absent42)
- [x] Babel 2.2.5-pl (@DESIGNfromWITHIN, @goldsky 3.0.0-beta is available for testing)
- [x] BannerY 1.3.0-rc1 (@bezumkin)
- [x] Batcher 1.1.1-pl (@sottwell)
- [x] BigBrother 1.1.0-rc2 (@beau-gosse, doesn't work with PHP 5.4+ though, @exside)
- [x] BotBlockX 1.0.2-pl (@BobRay)
- [x] BreadCrumb 1.4.2-pl (@rdegler)
- [x] Breadcrumbs 1.1.0-pl (@sepiariver)
- [x] CacheClear 1.0.3-pl (@exside, @BobRay)
- [x] CacheMaster 1.2.1-pl (@BobRay)
- [x] Canonical 1.0-Beta1 (@BobRay)
- [x] Captcha 3.3.2-pl (@BobRay)
- [x] CaseInsensitiveURLs 1.2.0-pl1 (@BobRay)
- [x] CheckboxSortable 0.3.0-beta3 (@sottwell)
- [x] ClassExtender 2.0.0-pl (@sottwell, @BobRay)
- [x] Cliche 1.1.0-rc1 (@bentejuy, has some visual glitches)
- [x] ClientConfig 1.3.1-pl (@sottwell, @amdbuilder)
- [x] Collections 2.0.0-pl (@theboxer)
- [x] ColorPicker 1.0.2-rc2 (@atripolskyy, @pyrographics, @exside)
- [x] Console 2.0.1-beta (@exside)
- [x] ConvertDatabaseCharset (@BobRay, not a package, see http://bobsguides.com/convert-db-utf8.html)
- [x] cookieJar 1.0.0-pl (@davidpede)
- [x] Copyright 1.0-pl (@amdbuilder)
- [x] CSS Star Rating 2.0-rc1 (@pixelchutes)
- [x] cssSweet 1.1.0-rc1 (@sottwell)
- [x] CustomUrls 1.0.0-rc3 (@Boddlnagg)
- [x] DBAPI 1.0.2-pl, pending release. (@pixelchutes)
- [x] DefaultResourceGroup 1.1.0-pl (@BobRay)
- [x] DefaultUserGroup 1.1.0-pl (@BobRay)
- [x] Dirwalker 1.0.0-pl (@BobRay)
- [x] EmailResource 1.1.3 (@BobRay)
- [x] EmptyAlias 1.0.0-beta1 (@BobRay)
- [x] EZfaq 3.3.0-pl (@BobRay)
- [x] fastField 1.4.0-pl (@mintnl)
- [x] FileDownload R 1.1.7-pl (@absent42)
- [x] FileLister 1.1.1-pl (@mintnl, however keep in mind to use a relative path, so paths starting with /)
- [x] FileUpload 1.0.4-pl (@BobRay)
- [x] firstPageLink 1.0.1-pl (@Boddlnagg)
- [x] FixedPre 1.2.2-pl (@BobRay)
- [x] FormIt 2.2.0-pl (@pixelchutes)
- [x] Gallery 1.7.0-pl (@pixelchutes)
- [x] Gateway Manager 1.1.2-pl (@amdbuilder)
- [x] getCache 1.1.0-pl (@sepiariver)
- [x] getDate 1.0.0-pl (@davidpede)
- [x] getPage 1.2.4-pl (@sepiariver)
- [x] getRelated 1.2.0-pl (@sepiariver)
- [x] getResourceField 1.0.3-pl1 (@mcnickel)
- [x] getResources 1.6.1-pl (@sepiariver)
- [x] getUrlParam 1.0-beta1 (@absent42)
- [x] getVimeo 1.1.0-pl (@davidpede)
- [x] getYoutube 1.0.0-pl (@davidpede)
- [x] GoogleAuthenticatorX 1.0.0-rc2 (@minagerges)
- [x] GoogleSiteMap 1.3.1-pl (@mcnickel)
- [x] Grid Class Key 1.0.1-beta1 (@goldsky, could break with compress_js on @labr1005)
- [x] HandyMan 1.0.0-pl (@absent42)
- [x] If 1.1.1-pl (@funkyduke)
- [x] JSONDerulo 2.4.0-pl (@pdincubus)
- [x] LikeDislike 1.0.3-rc1 (@amdbuilder, Works with Display issues)
- [x] Login 1.8.1-pl (@pixelchutes)
- [x] LogLogins 1.1.0-pl (@BobRay)
- [x] LogPageNotFound 1.0.3-pl (@BobRay)
- [x] MandrillX 1.0.3-pl (@pyrographics, @BobRay)
- [x] MetaX 2.0.2-pl (@amdbuilder)
- [x] mhpaypal 1.1.1-rc1 (@absent42)
- [x] MIGX 2.8.1-pl, has a lot of styling / UI issues though (@exside), MIGX AJAX Upload (http://rtfm.modx.com/extras/revo/migx/migx.tutorials/migx.use-resource-specific-mediasource-and-multifile-uploader) seems not to work (@skazhikadyadya)
- [x] MinifyX 1.3.1-pl (@amdbuilder)
- [x] miniShop2 2.1.8-rc2 (@bezumkin)
- [x] modAvatar 1.1.0-beta (@absent42)
- [x] modMailchimp 1.0.6-pl (@absent42)
- [x] MODX Revolution Boilerplate 1.0.1-pl (@pdincubus)
- [x] MultiUploadDialog 1.0.2-pl (@absent42)
- [x] MyComponent 3.2.2-pl (@pixelchutes, @BobRay)
- [x] NewsPublisher 1.4.2-pl (@BobRay, issues with TinyMCE though, @sottwell)
- [x] Notify 1.2.3-pl (@BobRay)
- [x] ObfuscateEmail-Revo 1.0.0-pl (@absent42)
- [x] ObjectExplorer 1.1.2-pl (@BobRay)
- [x] Orphans 1.1.0-pl (@BobRay, @sottwell)
- [x] PageLocker 1.1.0-beta1 (@gleighfield)
- [x] paramStash 1.3.0-pl (@electrickite)
- [x] pdoTools 1.9.2-pl1 (@sottwell)
- [x] Personalize 3.7.0-pl (@wshawn, @BobRay)
- [x] PiwikVisitsSummary 1.0.4-pl (@lektriq)
- [x] Polls 1.3.2-pl (@exside, CMP works, Frontend untested)
- [x] Profiler 1.1.0-pl (@TheBoxer)
- [x] Provisioner 1.1.0-pl, confirmed for importing Evo sites *(did not test Revo import < 2.1.3)* (@pixelchutes)
- [x] QuickEmail 1.1.0-pl (@BobRay)
- [x] Quickstart Buttons 1.1.1-pl (@Mark-H, needs modmore account/package provider for download, see https://www.modmore.com/extras/quickstartbuttons/, css fix for 2.3.1 see here http://forums.modx.com/thread/92587/quickstart-buttons-addon---css-for-modx-2-3-1)
- [x] Quip 2.3.3-pl (@amdbuilder, @mintnl)
- [x] Redirector 2.0.5-pl (@beau-gosse, there may well be issues with this extra not related to 2.3 @juiceCake)
- [x] ReflectBlock 1.1.0-pl (@BobRay)
- [x] RefreshCache 1.1.0-pl (@BobRay)
- [x] ResourceExplorer 1.0.0-pl2 (@xgenvn, minor display glitches, fix coming soon)
- [x] SE Manager for MODX 0.2.0-pl (@Alroniks)
- [x] selfLink 1.0.0-pl (@mcnickel)
- [x] SEO Pro 1.0.3-pl (@christianseel)
- [x] SEO Tab 1.0.10-pl (@christianseel)
- [x] Shibboleth 0.4.0-pl (@electrickite)
- [x] Shopkeeper 2.3.7-pl (@zfjoury)
- [x] ShowTV 1.0.0-beta1 (@BobRay)
- [x] siblingNav 1.0.1-pl (@sottwell)
- [x] SimpleSearch 1.9.0-pl (@sepiariver)
- [x] SiteAtoZ 1.0.3-beta1 (@BobRay)
- [x] SiteCheck 1.0.2-pl (@BobRay, premium Extra https://gumroad.com/l/DLChJ)
- [x] sitemapFriend 1.0.2-pl1 (@mcnickel)
- [x] SPForm 3.2.2-pl (@BobRay)
- [x] StageCoach 1.4.0-pl (@BobRay)
- [x] Subscribe 1.2.1-pl (@BobRay)
- [x] SyntaxHighlighter 1.0.0-beta1 (@BobRay)
- [x] TableofContentsX 1.0.1-pl1 (@MacConin)
- [x] Tagger 1.5.0-pl (@TheBoxer)
- [x] tagManager2 2.0.0-rc4 (@zfjoury)
- [x] ThemePackagerComponent 1.3.0-dev1 (@sottwell confirmed working but non-2.3 related issues, @jeremysanchez reported problems in 2.3.1)
- [x] ThermX 3.1.0-pl1 (@BobRay)
- [x] TimerangeTV 1.1.0-pl (@exside, @leqtrik)
- [x] TrackDuck 1.1.1-pl1 (@shauchenka)
- [x] translit 1.0.0-beta (@zfjoury)
- [x] TwitterX 1.3.4-pl (@absent42)
- [x] UltimateParent 2.0.0-pl (@mcnickel)
- [x] UserUrls 0.1.3-beta (@wshawn)
- [x] VersionX 2.1.0-pl (@amdbuilder, minor issues https://github.com/Mark-H/VersionX2/issues/67)
- [x] Wayfinder 2.3.3-pl (@pixelchutes, @sepiariver)
- [x] xBug 0.7.5-pl (@the-dunnock)
- [x] XRouting 1.2.0-pl (@labr1005)
**Extras working in 2.3 but not showing up in package manager**
- [x] AutoTemplate 1.0.0-rc1 (@oblongd)
- [x] MarkitUp 0.0.5-pl (@darkarchon)
- [x] modxTalks 1.0.3-beta (@sottwell)
- [x] sanitizeupload 1.0.0-rc (@absent42)
- [x] Theme.Bootstrap 2.1.0-pl (@opus4711)
- [x] visualSitemap 1.0.1-beta2 (@wshawn)
- [x] xFPC 2.1.0-pl1 (@mcnickel, xFPCAjax not tested, according to @MacConin the plugin gives warnings but seems to work; @funkyduke; @Gallenkamp also reports massive error logging issues with a 2.3.2. installation; @absent42 reports these logging issues for 2.2.x installations too; @sottwell points out that the issues are most possibly coming from setting headers too late)
**Extras partially working in 2.3**
- [x] Articles 1.7.9-pl (@jpdevries, @Mark-H, @amdbuilder, @hitodev found that it's not possible to remove tags due to the missing "delete"/cross button, see http://forums.modx.com/thread/92526/articles-extra-tags-missing-delete-icon + probable issues together with Tagger 1.5.0-pl, View and close buttons do not seem to work properly)
- [x] CKEditor 1.1.1-pl https://github.com/danyaPostfactum/modx-ckeditor/issues/17 (@exside)
- [x] CodeMirror 2.2.1-pl (@pdincubus confirmed working in 2.3.1, @sottwell reported issue with longer files, see comment from @charlie17)
- [x] ContextRouter 1.1.0-pl2 (@wshawn, 2.3.1 should be fine, in 2.3.0 you need to add a return statement to the plugin)
- [x] CronManager 1.1.0-pl (@pyrographics)
- [x] FormSave 1.1.0-pl1 (@pixelchutes, @bradbertz When viewing formsave form submission list, scroll down the list seems not to work (see https://github.com/b03tz/FormSave/issues/27). Package not showing up in packager browser
- [x] Google Analytics Dashboard Widget 1.0.1-pl (@absent42, @leqtrik, layout is broken and package not showing in package browser, for fix see http://forums.modx.com/thread/92380/google-analytics-dashboard-widget-1-0-1-pl-layout-fix-for-modx-2-3-x or http://forums.modx.com/thread/92525/analytics-dashboard-widget---display-glitch-in-revo-2-3)
- [x] Image+ 2.2.0-pl (has quite some UI issues, not really production ready, @exside)
- [x] phpThumbOf 1.4.0-pl (affected by #11700, @le-genda)
- [x] phpThumbsOn 1.2.1-pl (affected by #11700, @le-genda)
- [x] pThumb 2.3.3-pl (@exside, affected by #11700 @le-genda, issue as an output filter, scaling params are ignored when using Resizer @juiceCake)
- [x] SuperBoxSelect 1.0.1-pl2 (@enigmatic-user, @exside, UI issues due to additional styles for the already existing superboxselect styles in the core + missing classes for the triggers and tags, see fix in comments below, also not showing in package browser)
- [x] TinyMCE 4.3.3-pl, working but some paths need updating, seems non-breaking atm though, maybe some issues with fullscreen-mode (@sepiariver, @sottwell, @linlex, @rainbowtiger)
**Extras not working with 2.3**
- [x] AdvSearch 1.0.1-pl (@sottwell, @whitebyte error because of depreciated/removed getMicroTime() in 2.3, for a fix see #11749)
- [x] ElementHelper 1.3.3-pl (https://github.com/roryg/ElementHelper/issues/26, main issue also the use of deprecated $modx->getMicrotime(), but it's also not showing up in package manager)
- [x] Formz 1.1.0-pl (@markwillis82 reported it's working but according to http://forums.modx.com/thread/92765/formz-error-on-revolution-2-3-1#dis-post-506876 saving new forms doesn't work, which renders the extra useless)
- [x] Group Edit 1.2.0-pl (@Systerr; Autor recommends using Collections as repalcement)
- [x] ImportX 1.1.1-pl (@jmbuytaert reports that it only imports 2 entries)
- [x] Janitor 1.6-beta Search tab (port of Evo DocFinder) does not work: `Could not find action file at: controllers/default/index.php`, referencing legacy 2.2 action IDs for updating resources / elements, etc. e.g. `/manager/index.php?a=30&id=16`
- [x] mxCalendar 1.1.10-pl (@amdbuilder)
- [x] mxFormBuilder 1.0.0-pl (not showing up in package browser and CMP not working according to http://forums.modx.com/thread/92724/mxformbuilder-window-opens-briefly-then-blank#dis-post-506653)
- [x] MODXRated 2.1.0-pl (@hugopeek)
- [x] PackMan 1.2.0-pl (@exside, CMP works but has some minor UI issues, functionality has issues according to @pdincubus, @minagerges)
- [x] revoSSL 1.0.3-pl (@pyrographics)
- [x] Xodus 1.4.0-beta (@andreborud, It seems to generate the csv or xls but upon download it just jumps back to the dashboard, package not showing in package browser)
**Extras that need testing / confirmation (list incomplete)**
- [ ] charcounterTV 1.1.0-pl (not showing up in package browser)
- [ ] ColorPicker 1.0-beta1 (not showing in package browser)
- [ ] Databackup 1.1.9-pl (not showing in package browser)
- [ ] DirectResize 1.3.1-rc1
- [ ] Discuss 1.2-pl
- [ ] GoogleMapsTV 1.0-rc3
- [ ] mChimpX 1.0.0-rc1
- [ ] modActiveDirectory 1.0.1-pl
- [ ] modxRepository 1.2.3-beta
|
non_defect
|
extras compatibility i think it would be good to collect working and more important not yet working extras in one place maybe for the release note as more and more are reported as far as i can tell most not working extras are affected by the connector nuking that happened in so they don t find the appropriate processors anymore and thus do not work anymore another quite commonly observed issue is the use of the depreciated modx getmicrotime method in extras just replacing it with php s native function microtime true fixes this problem i suggest commenting including version info about successfully and not successfully tested extras in this issue and a i will update the following list extras working and compatible with ace pl pixelchutes activationemail pl bobray ajaxform pl ajaxmanager pl whitebyte ajaxupload pl archivist pl babel pl designfromwithin goldsky beta is available for testing bannery bezumkin batcher pl sottwell bigbrother beau gosse doesn t work with php though exside botblockx pl bobray breadcrumb pl rdegler breadcrumbs pl sepiariver cacheclear pl exside bobray cachemaster pl bobray canonical bobray captcha pl bobray caseinsensitiveurls bobray checkboxsortable sottwell classextender pl sottwell bobray cliche bentejuy has some visual glitches clientconfig pl sottwell amdbuilder collections pl theboxer colorpicker atripolskyy pyrographics exside console beta exside convertdatabasecharset bobray not a package see cookiejar pl davidpede copyright pl amdbuilder css star rating pixelchutes csssweet sottwell customurls boddlnagg dbapi pl pending release pixelchutes defaultresourcegroup pl bobray defaultusergroup pl bobray dirwalker pl bobray emailresource bobray emptyalias bobray ezfaq pl bobray fastfield pl mintnl filedownload r pl filelister pl mintnl however keep in mind to use a relative path so paths starting with fileupload pl bobray firstpagelink pl boddlnagg fixedpre pl bobray formit pl pixelchutes gallery pl pixelchutes gateway manager pl amdbuilder getcache pl sepiariver getdate pl davidpede getpage pl sepiariver getrelated pl sepiariver getresourcefield mcnickel getresources pl sepiariver geturlparam getvimeo pl davidpede getyoutube pl davidpede googleauthenticatorx minagerges googlesitemap pl mcnickel grid class key goldsky could break with compress js on handyman pl if pl funkyduke jsonderulo pl pdincubus likedislike amdbuilder works with display issues login pl pixelchutes loglogins pl bobray logpagenotfound pl bobray mandrillx pl pyrographics bobray metax pl amdbuilder mhpaypal migx pl has a lot of styling ui issues though exside migx ajax upload seems not to work skazhikadyadya minifyx pl amdbuilder bezumkin modavatar beta modmailchimp pl modx revolution boilerplate pl pdincubus multiuploaddialog pl mycomponent pl pixelchutes bobray newspublisher pl bobray issues with tinymce though sottwell notify pl bobray obfuscateemail revo pl objectexplorer pl bobray orphans pl bobray sottwell pagelocker gleighfield paramstash pl electrickite pdotools sottwell personalize pl wshawn bobray piwikvisitssummary pl lektriq polls pl exside cmp works frontend untested profiler pl theboxer provisioner pl confirmed for importing evo sites did not test revo import pixelchutes quickemail pl bobray quickstart buttons pl mark h needs modmore account package provider for download see css fix for see here quip pl amdbuilder mintnl redirector pl beau gosse there may well be issues with this extra not related to juicecake reflectblock pl bobray refreshcache pl bobray resourceexplorer xgenvn minor display glitches fix coming soon se manager for modx pl alroniks selflink pl mcnickel seo pro pl christianseel seo tab pl christianseel shibboleth pl electrickite shopkeeper pl zfjoury showtv bobray siblingnav pl sottwell simplesearch pl sepiariver siteatoz bobray sitecheck pl bobray premium extra sitemapfriend mcnickel spform pl bobray stagecoach pl bobray subscribe pl bobray syntaxhighlighter bobray tableofcontentsx macconin tagger pl theboxer zfjoury themepackagercomponent sottwell confirmed working but non related issues jeremysanchez reported problems in thermx bobray timerangetv pl exside leqtrik trackduck shauchenka translit beta zfjoury twitterx pl ultimateparent pl mcnickel userurls beta wshawn versionx pl amdbuilder minor issues wayfinder pl pixelchutes sepiariver xbug pl the dunnock xrouting pl extras working in but not showing up in package manager autotemplate oblongd markitup pl darkarchon modxtalks beta sottwell sanitizeupload rc theme bootstrap pl visualsitemap wshawn xfpc mcnickel xfpcajax not tested according to macconin the plugin gives warnings but seems to work funkyduke gallenkamp also reports massive error logging issues with a installation reports these logging issues for x installations too sottwell points out that the issues are most possibly coming from setting headers too late extras partially working in articles pl jpdevries mark h amdbuilder hitodev found that it s not possible to remove tags due to the missing delete cross button see probable issues together with tagger pl view and close buttons do not seem to work properly ckeditor pl exside codemirror pl pdincubus confirmed working in sottwell reported issue with longer files see comment from contextrouter wshawn should be fine in you need to add a return statement to the plugin cronmanager pl pyrographics formsave pixelchutes bradbertz when viewing formsave form submission list scroll down the list seems not to work see package not showing up in packager browser google analytics dashboard widget pl leqtrik layout is broken and package not showing in package browser for fix see or image pl has quite some ui issues not really production ready exside phpthumbof pl affected by le genda phpthumbson pl affected by le genda pthumb pl exside affected by le genda issue as an output filter scaling params are ignored when using resizer juicecake superboxselect enigmatic user exside ui issues due to additional styles for the already existing superboxselect styles in the core missing classes for the triggers and tags see fix in comments below also not showing in package browser tinymce pl working but some paths need updating seems non breaking atm though maybe some issues with fullscreen mode sepiariver sottwell linlex rainbowtiger extras not working with advsearch pl sottwell whitebyte error because of depreciated removed getmicrotime in for a fix see elementhelper pl main issue also the use of deprecated modx getmicrotime but it s also not showing up in package manager formz pl reported it s working but according to saving new forms doesn t work which renders the extra useless group edit pl systerr autor recommends using collections as repalcement importx pl jmbuytaert reports that it only imports entries janitor beta search tab port of evo docfinder does not work could not find action file at controllers default index php referencing legacy action ids for updating resources elements etc e g manager index php a id mxcalendar pl amdbuilder mxformbuilder pl not showing up in package browser and cmp not working according to modxrated pl hugopeek packman pl exside cmp works but has some minor ui issues functionality has issues according to pdincubus minagerges revossl pl pyrographics xodus beta andreborud it seems to generate the csv or xls but upon download it just jumps back to the dashboard package not showing in package browser extras that need testing confirmation list incomplete charcountertv pl not showing up in package browser colorpicker not showing in package browser databackup pl not showing in package browser directresize discuss pl googlemapstv mchimpx modactivedirectory pl modxrepository beta
| 0
|
273,607
| 20,799,298,643
|
IssuesEvent
|
2022-03-17 12:27:14
|
0chain/0chain
|
https://api.github.com/repos/0chain/0chain
|
closed
|
Buckets concept proposal for optimizing the performance issue of large `list` in MPT
|
enhancement discussion documentation
|
We will have performance issue if the nodes number is large. Currently we save all the miners node info saved in a simple MPT node on path `allMinersList`, so each time when we need to update a node's info, we will have to decode the full nodes list, then update the node info, encode the full node list and save it back to MPT. The process would be a disaster when the nodes number is huge.
So what we can do to optimize this is, instead of saving all nodes in a single MPT node, we can split the them into several `buckets` that indexed by the miner index id(or something that can identify the miner exclusively) . And each bucket has a fixed capacity, for example: `50`. So the first 50 miners nodes will be saved into `bucket_1` , and nodes from 50 to 100 saved to the second `bucket_2` and so on. We will be able to get the bucket address base on the miner index. In this way, each time when we need to update a miner's info, we will fist get the bucket address, then get the bucket data from the MPT, and decode it into nodes list, then do what we want to do to the node.
The benefit is obvious, which is we don't have to encode/decode the full list of nodes every time when we need to update a simple node, and the worst case would be we decode a full bucket, which is obviously much faster than decoding the full nodes list. This bucket concept could be applied to all the `list` like data struct in MPT that could grow large over time.
Anyway, we can do this later and fill the details about the `bucket` concept when we finished the Mainnet issues and loadtest issues.
|
1.0
|
Buckets concept proposal for optimizing the performance issue of large `list` in MPT - We will have performance issue if the nodes number is large. Currently we save all the miners node info saved in a simple MPT node on path `allMinersList`, so each time when we need to update a node's info, we will have to decode the full nodes list, then update the node info, encode the full node list and save it back to MPT. The process would be a disaster when the nodes number is huge.
So what we can do to optimize this is, instead of saving all nodes in a single MPT node, we can split the them into several `buckets` that indexed by the miner index id(or something that can identify the miner exclusively) . And each bucket has a fixed capacity, for example: `50`. So the first 50 miners nodes will be saved into `bucket_1` , and nodes from 50 to 100 saved to the second `bucket_2` and so on. We will be able to get the bucket address base on the miner index. In this way, each time when we need to update a miner's info, we will fist get the bucket address, then get the bucket data from the MPT, and decode it into nodes list, then do what we want to do to the node.
The benefit is obvious, which is we don't have to encode/decode the full list of nodes every time when we need to update a simple node, and the worst case would be we decode a full bucket, which is obviously much faster than decoding the full nodes list. This bucket concept could be applied to all the `list` like data struct in MPT that could grow large over time.
Anyway, we can do this later and fill the details about the `bucket` concept when we finished the Mainnet issues and loadtest issues.
|
non_defect
|
buckets concept proposal for optimizing the performance issue of large list in mpt we will have performance issue if the nodes number is large currently we save all the miners node info saved in a simple mpt node on path allminerslist so each time when we need to update a node s info we will have to decode the full nodes list then update the node info encode the full node list and save it back to mpt the process would be a disaster when the nodes number is huge so what we can do to optimize this is instead of saving all nodes in a single mpt node we can split the them into several buckets that indexed by the miner index id or something that can identify the miner exclusively and each bucket has a fixed capacity for example so the first miners nodes will be saved into bucket and nodes from to saved to the second bucket and so on we will be able to get the bucket address base on the miner index in this way each time when we need to update a miner s info we will fist get the bucket address then get the bucket data from the mpt and decode it into nodes list then do what we want to do to the node the benefit is obvious which is we don t have to encode decode the full list of nodes every time when we need to update a simple node and the worst case would be we decode a full bucket which is obviously much faster than decoding the full nodes list this bucket concept could be applied to all the list like data struct in mpt that could grow large over time anyway we can do this later and fill the details about the bucket concept when we finished the mainnet issues and loadtest issues
| 0
|
51,264
| 21,640,294,241
|
IssuesEvent
|
2022-05-05 18:03:00
|
emergenzeHack/ukrainehelp.emergenzehack.info_segnalazioni
|
https://api.github.com/repos/emergenzeHack/ukrainehelp.emergenzehack.info_segnalazioni
|
opened
|
Centro estivo per bambini Ucraini
|
Services Children
|
<pre><yamldata>
servicetypes:
materialGoods: false
hospitality: false
transport: false
healthcare: false
Legal: false
translation: false
job: false
psychologicalSupport: false
Children: true
disability: false
women: false
education: false
offerFromWho: Impresa Atlantide - Ambasciata Ucraina
title: Centro estivo per bambini Ucraini
recipients: ''
description: L'azienda italiana Atlantide in collaborazione con l'Ambasciata Ucraina
in Italia, il Municipio di Roma, il Dipartimento di Difesa Civile d'Italia, nonché
le associazioni benefiche "Caritas" e "Sant'Egidio", questa estate presso villa
Fassini a Roma ha iniziato l'apertura di un campo estivo per i bambini ucraini
url: https://atlantia4ukraine.atlantia.com/it?fbclid=IwAR0WkVHFfg-8Wqf5r831-v-i4mSQGJZES0xE_BTBPH059K9gNQbsR48vLj4
address:
mode: autocomplete
address:
place_id: 106584746
licence: Data © OpenStreetMap contributors, ODbL 1.0. https://osm.org/copyright
osm_type: way
osm_id: 23936609
boundingbox:
- '41.9064355'
- '41.9093479'
- '12.5467987'
- '12.5533818'
lat: '41.9073836'
lon: '12.5501801'
display_name: Via Giuseppe Donati, Casal Bruciato, Collatino, Roma, Roma Capitale,
Lazio, 00100, Italia
class: highway
type: residential
importance: 0.5
address:
road: Via Giuseppe Donati
neighbourhood: Casal Bruciato
suburb: Collatino
city: Roma
county: Roma Capitale
state: Lazio
postcode: '00100'
country: Italia
country_code: it
iConfirmToHaveReadAndAcceptedInformativeToThreatPersonalData: true
label: services
submit: true
</yamldata></pre>
|
1.0
|
Centro estivo per bambini Ucraini - <pre><yamldata>
servicetypes:
materialGoods: false
hospitality: false
transport: false
healthcare: false
Legal: false
translation: false
job: false
psychologicalSupport: false
Children: true
disability: false
women: false
education: false
offerFromWho: Impresa Atlantide - Ambasciata Ucraina
title: Centro estivo per bambini Ucraini
recipients: ''
description: L'azienda italiana Atlantide in collaborazione con l'Ambasciata Ucraina
in Italia, il Municipio di Roma, il Dipartimento di Difesa Civile d'Italia, nonché
le associazioni benefiche "Caritas" e "Sant'Egidio", questa estate presso villa
Fassini a Roma ha iniziato l'apertura di un campo estivo per i bambini ucraini
url: https://atlantia4ukraine.atlantia.com/it?fbclid=IwAR0WkVHFfg-8Wqf5r831-v-i4mSQGJZES0xE_BTBPH059K9gNQbsR48vLj4
address:
mode: autocomplete
address:
place_id: 106584746
licence: Data © OpenStreetMap contributors, ODbL 1.0. https://osm.org/copyright
osm_type: way
osm_id: 23936609
boundingbox:
- '41.9064355'
- '41.9093479'
- '12.5467987'
- '12.5533818'
lat: '41.9073836'
lon: '12.5501801'
display_name: Via Giuseppe Donati, Casal Bruciato, Collatino, Roma, Roma Capitale,
Lazio, 00100, Italia
class: highway
type: residential
importance: 0.5
address:
road: Via Giuseppe Donati
neighbourhood: Casal Bruciato
suburb: Collatino
city: Roma
county: Roma Capitale
state: Lazio
postcode: '00100'
country: Italia
country_code: it
iConfirmToHaveReadAndAcceptedInformativeToThreatPersonalData: true
label: services
submit: true
</yamldata></pre>
|
non_defect
|
centro estivo per bambini ucraini servicetypes materialgoods false hospitality false transport false healthcare false legal false translation false job false psychologicalsupport false children true disability false women false education false offerfromwho impresa atlantide ambasciata ucraina title centro estivo per bambini ucraini recipients description l azienda italiana atlantide in collaborazione con l ambasciata ucraina in italia il municipio di roma il dipartimento di difesa civile d italia nonché le associazioni benefiche caritas e sant egidio questa estate presso villa fassini a roma ha iniziato l apertura di un campo estivo per i bambini ucraini url address mode autocomplete address place id licence data © openstreetmap contributors odbl osm type way osm id boundingbox lat lon display name via giuseppe donati casal bruciato collatino roma roma capitale lazio italia class highway type residential importance address road via giuseppe donati neighbourhood casal bruciato suburb collatino city roma county roma capitale state lazio postcode country italia country code it iconfirmtohavereadandacceptedinformativetothreatpersonaldata true label services submit true
| 0
|
9,300
| 2,615,143,434
|
IssuesEvent
|
2015-03-01 06:18:45
|
chrsmith/html5rocks
|
https://api.github.com/repos/chrsmith/html5rocks
|
closed
|
html5rocks is not HTML5
|
auto-migrated Priority-Medium Type-Defect
|
```
Please describe the issue:
There are several places in your presentation that is not HTML5 or CSS3 or part
of any standardized specification. Look for places containing -webkit* or
webkit* function/css rule names. Please consider being honest, not following
the Apple's standard of false messages of being standard based.
```
Original issue reported on code.google.com by `jad...@gmail.com` on 23 Jun 2010 at 3:01
|
1.0
|
html5rocks is not HTML5 - ```
Please describe the issue:
There are several places in your presentation that is not HTML5 or CSS3 or part
of any standardized specification. Look for places containing -webkit* or
webkit* function/css rule names. Please consider being honest, not following
the Apple's standard of false messages of being standard based.
```
Original issue reported on code.google.com by `jad...@gmail.com` on 23 Jun 2010 at 3:01
|
defect
|
is not please describe the issue there are several places in your presentation that is not or or part of any standardized specification look for places containing webkit or webkit function css rule names please consider being honest not following the apple s standard of false messages of being standard based original issue reported on code google com by jad gmail com on jun at
| 1
|
308,324
| 9,437,976,044
|
IssuesEvent
|
2019-04-13 19:23:26
|
akalenda/CS249_GraphAlgVisualizer
|
https://api.github.com/repos/akalenda/CS249_GraphAlgVisualizer
|
opened
|
Create unit and acceptance tests
|
enhancement high priority
|
By further abstracting construction of the graph and processes into an API and decoupling it from the user interface, we can create unit tests that test the functionality and achieve near-100% coverage.
We would then not rely on visual tests of the app to make sure that the underlying functionality is working. Furthermore, an end-to-end test built upon those unit tests would provide an excellent demo of the program in action, as it steps through everything a user might do.
|
1.0
|
Create unit and acceptance tests - By further abstracting construction of the graph and processes into an API and decoupling it from the user interface, we can create unit tests that test the functionality and achieve near-100% coverage.
We would then not rely on visual tests of the app to make sure that the underlying functionality is working. Furthermore, an end-to-end test built upon those unit tests would provide an excellent demo of the program in action, as it steps through everything a user might do.
|
non_defect
|
create unit and acceptance tests by further abstracting construction of the graph and processes into an api and decoupling it from the user interface we can create unit tests that test the functionality and achieve near coverage we would then not rely on visual tests of the app to make sure that the underlying functionality is working furthermore an end to end test built upon those unit tests would provide an excellent demo of the program in action as it steps through everything a user might do
| 0
|
61,810
| 3,154,763,174
|
IssuesEvent
|
2015-09-17 02:30:46
|
xcat2/xcat-core
|
https://api.github.com/repos/xcat2/xcat-core
|
reopened
|
[FVT]2.11 duplicate kit components are added to osimage
|
priority:normal type:bug
|
xCAT 2.11 9/15 build on rhels7.2 pp64le
1) download xlc-13.1.2-0-ppc64le.NEED_PRODUCT_PKGS.tar.bz2
2) download other xlf packages which are needed by complete kit
[root@c910f03c05k08 build_input]# ls /compilers/
libxlc-13.1.2.0-150526a.ppc64le.rpm
libxlc-devel.13.1.2-13.1.2.0-150526a.ppc64le.rpm
libxlf-15.1.2.0-150526a.ppc64le.rpm
libxlf-devel.15.1.2-15.1.2.0-150526a.ppc64le.rpm
libxlmass-devel.8.1.2-8.1.2.0-150526.ppc64le.rpm
libxlsmp-4.1.2.0-150526.ppc64le.rpm
libxlsmp-devel.4.1.2-4.1.2.0-150526.ppc64le.rpm
xlc.13.1.2-13.1.2.0-150526a.ppc64le.rpm
xlc-license.13.1.2-13.1.2.0-150526a.ppc64le.rpm
xlf.15.1.2-15.1.2.0-150526a.ppc64le.rpm
xlf-license.15.1.2-15.1.2.0-150526a.ppc64le.rpm
3) create complete kit using partial kit
buildkit addpkgs xlc-13.1.2-0-ppc64le.NEED_PRODUCT_PKGS.tar.bz2 --pkgdir /compilers/
4)
addkit xlc-13.1.2-0-ppc64le.tar.bz2
5)
addkitcomp -a -i "rhels7.2-ppc64le-install-compute" xlc.license-compute,xlc.compiler-compute, xlc.rte-compute
6) check osimage rhels7.2-ppc64le-install-compute
[root@c910f03c05k08 build_input]# lsdef -t osimage -o rhels7.2-ppc64le-install-compute
Object name: rhels7.2-ppc64le-install-compute
imagetype=linux
kitcomponents=xlc.license-compute-13.1.2-0-rhels-7.2-ppc64le,xlc.rte-compute-13.1.2-0-rhels-7.2-ppc64le,xlc.rte-compute-13.1.2-0-rhels-7.2-ppc64le,xlc.compiler-compute-13.1.2-0-rhels-7.2-ppc64le,xlf.license-compute-15.1.2-0-rhels-7.2-ppc64le,xlf.rte-compute-15.1.2-0-rhels-7.2-ppc64le,xlf.compiler-compute-15.1.2-0-rhels-7.2-ppc64le,xlf.rte-compute-15.1.2-0-rhels-7.2-ppc64le,xlf.license-compute-15.1.2-0-rhels-7.2-ppc64le
osarch=ppc64le
osdistroname=rhels7.2-ppc64le
osname=Linux
osvers=rhels7.2
otherpkgdir=/install/post/otherpkgs/rhels7.2/ppc64le
otherpkglist=/install/osimages/rhels7.2-ppc64le-install-compute/kits/KIT_DEPLOY_PARAMS.otherpkgs.pkglist,/install/osimages/rhels7.2-ppc64le-install-compute/kits/KIT_COMPONENTS.otherpkgs.pkglist
pkgdir=/install/rhels7.2/ppc64le
pkglist=/opt/xcat/share/xcat/install/rh/compute.rhels7.pkglist
profile=compute
provmethod=install
template=/opt/xcat/share/xcat/install/rh/compute.rhels7.tmpl
|
1.0
|
[FVT]2.11 duplicate kit components are added to osimage - xCAT 2.11 9/15 build on rhels7.2 pp64le
1) download xlc-13.1.2-0-ppc64le.NEED_PRODUCT_PKGS.tar.bz2
2) download other xlf packages which are needed by complete kit
[root@c910f03c05k08 build_input]# ls /compilers/
libxlc-13.1.2.0-150526a.ppc64le.rpm
libxlc-devel.13.1.2-13.1.2.0-150526a.ppc64le.rpm
libxlf-15.1.2.0-150526a.ppc64le.rpm
libxlf-devel.15.1.2-15.1.2.0-150526a.ppc64le.rpm
libxlmass-devel.8.1.2-8.1.2.0-150526.ppc64le.rpm
libxlsmp-4.1.2.0-150526.ppc64le.rpm
libxlsmp-devel.4.1.2-4.1.2.0-150526.ppc64le.rpm
xlc.13.1.2-13.1.2.0-150526a.ppc64le.rpm
xlc-license.13.1.2-13.1.2.0-150526a.ppc64le.rpm
xlf.15.1.2-15.1.2.0-150526a.ppc64le.rpm
xlf-license.15.1.2-15.1.2.0-150526a.ppc64le.rpm
3) create complete kit using partial kit
buildkit addpkgs xlc-13.1.2-0-ppc64le.NEED_PRODUCT_PKGS.tar.bz2 --pkgdir /compilers/
4)
addkit xlc-13.1.2-0-ppc64le.tar.bz2
5)
addkitcomp -a -i "rhels7.2-ppc64le-install-compute" xlc.license-compute,xlc.compiler-compute, xlc.rte-compute
6) check osimage rhels7.2-ppc64le-install-compute
[root@c910f03c05k08 build_input]# lsdef -t osimage -o rhels7.2-ppc64le-install-compute
Object name: rhels7.2-ppc64le-install-compute
imagetype=linux
kitcomponents=xlc.license-compute-13.1.2-0-rhels-7.2-ppc64le,xlc.rte-compute-13.1.2-0-rhels-7.2-ppc64le,xlc.rte-compute-13.1.2-0-rhels-7.2-ppc64le,xlc.compiler-compute-13.1.2-0-rhels-7.2-ppc64le,xlf.license-compute-15.1.2-0-rhels-7.2-ppc64le,xlf.rte-compute-15.1.2-0-rhels-7.2-ppc64le,xlf.compiler-compute-15.1.2-0-rhels-7.2-ppc64le,xlf.rte-compute-15.1.2-0-rhels-7.2-ppc64le,xlf.license-compute-15.1.2-0-rhels-7.2-ppc64le
osarch=ppc64le
osdistroname=rhels7.2-ppc64le
osname=Linux
osvers=rhels7.2
otherpkgdir=/install/post/otherpkgs/rhels7.2/ppc64le
otherpkglist=/install/osimages/rhels7.2-ppc64le-install-compute/kits/KIT_DEPLOY_PARAMS.otherpkgs.pkglist,/install/osimages/rhels7.2-ppc64le-install-compute/kits/KIT_COMPONENTS.otherpkgs.pkglist
pkgdir=/install/rhels7.2/ppc64le
pkglist=/opt/xcat/share/xcat/install/rh/compute.rhels7.pkglist
profile=compute
provmethod=install
template=/opt/xcat/share/xcat/install/rh/compute.rhels7.tmpl
|
non_defect
|
duplicate kit components are added to osimage xcat build on download xlc need product pkgs tar download other xlf packages which are needed by complete kit ls compilers libxlc rpm libxlc devel rpm libxlf rpm libxlf devel rpm libxlmass devel rpm libxlsmp rpm libxlsmp devel rpm xlc rpm xlc license rpm xlf rpm xlf license rpm create complete kit using partial kit buildkit addpkgs xlc need product pkgs tar pkgdir compilers addkit xlc tar addkitcomp a i install compute xlc license compute xlc compiler compute xlc rte compute check osimage install compute lsdef t osimage o install compute object name install compute imagetype linux kitcomponents xlc license compute rhels xlc rte compute rhels xlc rte compute rhels xlc compiler compute rhels xlf license compute rhels xlf rte compute rhels xlf compiler compute rhels xlf rte compute rhels xlf license compute rhels osarch osdistroname osname linux osvers otherpkgdir install post otherpkgs otherpkglist install osimages install compute kits kit deploy params otherpkgs pkglist install osimages install compute kits kit components otherpkgs pkglist pkgdir install pkglist opt xcat share xcat install rh compute pkglist profile compute provmethod install template opt xcat share xcat install rh compute tmpl
| 0
|
31,856
| 12,033,132,579
|
IssuesEvent
|
2020-04-13 13:39:50
|
crystal-lang/crystal
|
https://api.github.com/repos/crystal-lang/crystal
|
closed
|
SECURITY: context_spec.cr specifies "only TLS 1.0" by calling tlsv1_method. OpenSSL says avoid tlsv1_method. TLS 1.0 and 1.1 are deprecated.
|
community:help-wanted security topic:stdlib:crypto
|
A. OpenSSL docs tell us to avoid calling `tlsv1_method` but crystal is calling it.
B. Calling `tlsv1_method` means a TLS/SSL connection will only understand TLS 1.0.
C. Current industry recommendation is to use least TLS 1.2:
* __June 30, 2018 deadline to drop TLS 1.0__ - The PCI Council (Payment Card Industry) suggested that organizations migrate from TLS 1.0 to TLS 1.1 or higher before June 30, 2018.
* __March 2020 deadline to deprecate TLS 1.1__ - In October 2018, Apple, Google, Microsoft, and Mozilla jointly announced they would deprecate TLS 1.0 and 1.1 in March 2020.
D. See Background (provided below) for timeline and some attacks on TLS for more context.
<hr>
### OpenSSL docs
__OpenSSL 1.1.0 says use TLS_method__ and avoid TLSv1_method, TLSv1_1_method, etc.
__OpenSSL 1.0.2 says use SSLv23_method__ and avoid TLSv1_method, TLSv1_1_method, etc.
<details>
<summary>(click to expand) Quote from docs and links</summary><p>
>TLS_method(), TLS_server_method(), TLS_client_method()
>
>These are the general-purpose version-flexible SSL/TLS methods. The actual protocol version used will be negotiated to the highest version mutually supported by the client and the server. The supported protocols are SSLv3, TLSv1, TLSv1.1 and TLSv1.2. Applications should use these methods, and avoid the version-specific methods described below.
> ...
> TLSv1_2_method(), TLSv1_2_server_method(), TLSv1_2_client_method()
> ...
> TLSv1_1_method(), TLSv1_1_server_method(), TLSv1_1_client_method()
> ...
> TLSv1_method(), TLSv1_server_method(), TLSv1_client_method()
> A TLS/SSL connection established with these methods will only understand the TLSv1 protocol.
* https://www.openssl.org/docs/man1.1.0/man3/TLSv1_method.html
* https://www.openssl.org/docs/man1.0.2/man3/TLSv1_method.html
</details>
### Relevant code in crystal
https://github.com/crystal-lang/crystal/blob/fd0780c3a4ef972a6090e09abc9b09a0e39345ff/spec/std/openssl/ssl/context_spec.cr#L31
https://github.com/crystal-lang/crystal/blob/fd0780c3a4ef972a6090e09abc9b09a0e39345ff/spec/std/openssl/ssl/context_spec.cr#L48
Additionally, TLS 1.0 and TLS 1.1 should be disabled from being chosen during auto-negotiation. See Background (provided below).
## Background
__1999__. TLS 1.0 was first defined in RFC 2246 in January 1999.
__2006__. TLS 1.1 was defined in RFC 4346 in April 2006.
__2008__. TLS 1.2 was defined in RFC 5246 in August 2008.
__2014__. TLS 1.0 allows downgrading the connection to SSL 3.0, thus weakening security (POODLE SSL Variant). Source:
https://en.wikipedia.org/wiki/Transport_Layer_Security#TLS_1.0
__2014__. TLS 1.0, TLS 1.1, and TLS 1.2 (if not implemented properly) are vulnerable to POODLE TLS Variant even if SSLv3 is disabled. Source:
https://en.wikipedia.org/wiki/POODLE
__2017__. Google Chrome set TLS 1.3 as the default version for a short time in 2017. It then removed it as the default, due to incompatible middleboxes such as Blue Coat web proxies. Source:
https://en.wikipedia.org/wiki/Transport_Layer_Security#TLS_1.3
__2018__. TLS 1.3 was defined in RFC 8446 in August 2018. It mandates use of AEAD ciphers, key exchanges that offer perfect forward secrecy, integrates session hash, and drops support for many insecure or obsolete features. Source:
https://en.wikipedia.org/wiki/Transport_Layer_Security#TLS_1.3
#### TLS Interception Appliances
POODLE from 2014 is not the only security issue to consider, see this paper:
> To analyze TLS-encrypted data, network appliances implement a Man-in-the-Middle TLS proxy, by acting as the intended web server to a requesting client (e.g., a browser), and acting as the client to the actual/outside web server. Source:
[The Sorry State of TLS Security in Enterprise - Interception Appliances (PDF, arxiv.org)]( https://arxiv.org/pdf/1809.08729.pdf)
|
True
|
SECURITY: context_spec.cr specifies "only TLS 1.0" by calling tlsv1_method. OpenSSL says avoid tlsv1_method. TLS 1.0 and 1.1 are deprecated. - A. OpenSSL docs tell us to avoid calling `tlsv1_method` but crystal is calling it.
B. Calling `tlsv1_method` means a TLS/SSL connection will only understand TLS 1.0.
C. Current industry recommendation is to use least TLS 1.2:
* __June 30, 2018 deadline to drop TLS 1.0__ - The PCI Council (Payment Card Industry) suggested that organizations migrate from TLS 1.0 to TLS 1.1 or higher before June 30, 2018.
* __March 2020 deadline to deprecate TLS 1.1__ - In October 2018, Apple, Google, Microsoft, and Mozilla jointly announced they would deprecate TLS 1.0 and 1.1 in March 2020.
D. See Background (provided below) for timeline and some attacks on TLS for more context.
<hr>
### OpenSSL docs
__OpenSSL 1.1.0 says use TLS_method__ and avoid TLSv1_method, TLSv1_1_method, etc.
__OpenSSL 1.0.2 says use SSLv23_method__ and avoid TLSv1_method, TLSv1_1_method, etc.
<details>
<summary>(click to expand) Quote from docs and links</summary><p>
>TLS_method(), TLS_server_method(), TLS_client_method()
>
>These are the general-purpose version-flexible SSL/TLS methods. The actual protocol version used will be negotiated to the highest version mutually supported by the client and the server. The supported protocols are SSLv3, TLSv1, TLSv1.1 and TLSv1.2. Applications should use these methods, and avoid the version-specific methods described below.
> ...
> TLSv1_2_method(), TLSv1_2_server_method(), TLSv1_2_client_method()
> ...
> TLSv1_1_method(), TLSv1_1_server_method(), TLSv1_1_client_method()
> ...
> TLSv1_method(), TLSv1_server_method(), TLSv1_client_method()
> A TLS/SSL connection established with these methods will only understand the TLSv1 protocol.
* https://www.openssl.org/docs/man1.1.0/man3/TLSv1_method.html
* https://www.openssl.org/docs/man1.0.2/man3/TLSv1_method.html
</details>
### Relevant code in crystal
https://github.com/crystal-lang/crystal/blob/fd0780c3a4ef972a6090e09abc9b09a0e39345ff/spec/std/openssl/ssl/context_spec.cr#L31
https://github.com/crystal-lang/crystal/blob/fd0780c3a4ef972a6090e09abc9b09a0e39345ff/spec/std/openssl/ssl/context_spec.cr#L48
Additionally, TLS 1.0 and TLS 1.1 should be disabled from being chosen during auto-negotiation. See Background (provided below).
## Background
__1999__. TLS 1.0 was first defined in RFC 2246 in January 1999.
__2006__. TLS 1.1 was defined in RFC 4346 in April 2006.
__2008__. TLS 1.2 was defined in RFC 5246 in August 2008.
__2014__. TLS 1.0 allows downgrading the connection to SSL 3.0, thus weakening security (POODLE SSL Variant). Source:
https://en.wikipedia.org/wiki/Transport_Layer_Security#TLS_1.0
__2014__. TLS 1.0, TLS 1.1, and TLS 1.2 (if not implemented properly) are vulnerable to POODLE TLS Variant even if SSLv3 is disabled. Source:
https://en.wikipedia.org/wiki/POODLE
__2017__. Google Chrome set TLS 1.3 as the default version for a short time in 2017. It then removed it as the default, due to incompatible middleboxes such as Blue Coat web proxies. Source:
https://en.wikipedia.org/wiki/Transport_Layer_Security#TLS_1.3
__2018__. TLS 1.3 was defined in RFC 8446 in August 2018. It mandates use of AEAD ciphers, key exchanges that offer perfect forward secrecy, integrates session hash, and drops support for many insecure or obsolete features. Source:
https://en.wikipedia.org/wiki/Transport_Layer_Security#TLS_1.3
#### TLS Interception Appliances
POODLE from 2014 is not the only security issue to consider, see this paper:
> To analyze TLS-encrypted data, network appliances implement a Man-in-the-Middle TLS proxy, by acting as the intended web server to a requesting client (e.g., a browser), and acting as the client to the actual/outside web server. Source:
[The Sorry State of TLS Security in Enterprise - Interception Appliances (PDF, arxiv.org)]( https://arxiv.org/pdf/1809.08729.pdf)
|
non_defect
|
security context spec cr specifies only tls by calling method openssl says avoid method tls and are deprecated a openssl docs tell us to avoid calling method but crystal is calling it b calling method means a tls ssl connection will only understand tls c current industry recommendation is to use least tls june deadline to drop tls the pci council payment card industry suggested that organizations migrate from tls to tls or higher before june march deadline to deprecate tls in october apple google microsoft and mozilla jointly announced they would deprecate tls and in march d see background provided below for timeline and some attacks on tls for more context openssl docs openssl says use tls method and avoid method method etc openssl says use method and avoid method method etc click to expand quote from docs and links tls method tls server method tls client method these are the general purpose version flexible ssl tls methods the actual protocol version used will be negotiated to the highest version mutually supported by the client and the server the supported protocols are and applications should use these methods and avoid the version specific methods described below method server method client method method server method client method method server method client method a tls ssl connection established with these methods will only understand the protocol relevant code in crystal additionally tls and tls should be disabled from being chosen during auto negotiation see background provided below background tls was first defined in rfc in january tls was defined in rfc in april tls was defined in rfc in august tls allows downgrading the connection to ssl thus weakening security poodle ssl variant source tls tls and tls if not implemented properly are vulnerable to poodle tls variant even if is disabled source google chrome set tls as the default version for a short time in it then removed it as the default due to incompatible middleboxes such as blue coat web proxies source tls was defined in rfc in august it mandates use of aead ciphers key exchanges that offer perfect forward secrecy integrates session hash and drops support for many insecure or obsolete features source tls interception appliances poodle from is not the only security issue to consider see this paper to analyze tls encrypted data network appliances implement a man in the middle tls proxy by acting as the intended web server to a requesting client e g a browser and acting as the client to the actual outside web server source
| 0
|
663,664
| 22,201,173,844
|
IssuesEvent
|
2022-06-07 11:20:29
|
woocommerce/woocommerce
|
https://api.github.com/repos/woocommerce/woocommerce
|
closed
|
WP CLI db update bug?
|
type: bug priority: low focus: wc-cli type: task
|
It seems like [this part of the code](https://github.com/woocommerce/woocommerce/blame/trunk/plugins/woocommerce/includes/cli/class-wc-cli-update-command.php#L64-L72) either uses some magic, or won't run as expected. The while loop probably never runs and result from the callback is ignored. We should check if this actually works.
|
1.0
|
WP CLI db update bug? - It seems like [this part of the code](https://github.com/woocommerce/woocommerce/blame/trunk/plugins/woocommerce/includes/cli/class-wc-cli-update-command.php#L64-L72) either uses some magic, or won't run as expected. The while loop probably never runs and result from the callback is ignored. We should check if this actually works.
|
non_defect
|
wp cli db update bug it seems like either uses some magic or won t run as expected the while loop probably never runs and result from the callback is ignored we should check if this actually works
| 0
|
53,296
| 28,068,125,528
|
IssuesEvent
|
2023-03-29 16:56:41
|
jupyterlite/jupyterlite
|
https://api.github.com/repos/jupyterlite/jupyterlite
|
closed
|
Be more intentional about some of the imports hacks
|
performance kernel:pyodide
|
Could we, to the extent possible, be more intentional about some of the imports hacks so that these things run under `nbconvert --execute` and `nbval`? e.g.
```py
try:
import pyolite
# do hacky things
except ImportError:
pass
```
It's also worth adding a `<details><summary>Limitations</summary>...</details>` markdown cell above any cells like these to all of our examples. Totally aside, but dang i sure wish there was a good markdown flavor that supported `details`...
_Originally posted by @bollwyvl in https://github.com/jtpio/jupyterlite/issues/219#issuecomment-874920456_
|
True
|
Be more intentional about some of the imports hacks - Could we, to the extent possible, be more intentional about some of the imports hacks so that these things run under `nbconvert --execute` and `nbval`? e.g.
```py
try:
import pyolite
# do hacky things
except ImportError:
pass
```
It's also worth adding a `<details><summary>Limitations</summary>...</details>` markdown cell above any cells like these to all of our examples. Totally aside, but dang i sure wish there was a good markdown flavor that supported `details`...
_Originally posted by @bollwyvl in https://github.com/jtpio/jupyterlite/issues/219#issuecomment-874920456_
|
non_defect
|
be more intentional about some of the imports hacks could we to the extent possible be more intentional about some of the imports hacks so that these things run under nbconvert execute and nbval e g py try import pyolite do hacky things except importerror pass it s also worth adding a limitations markdown cell above any cells like these to all of our examples totally aside but dang i sure wish there was a good markdown flavor that supported details originally posted by bollwyvl in
| 0
|
16,692
| 21,791,963,265
|
IssuesEvent
|
2022-05-15 03:02:55
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Batch processing gui for "Export layers to DXF" is missing functionality for loading layers into separate rows in the table
|
Feedback stale Processing Bug
|
### What is the bug or the crash?
BACKGROUND COMMENTS
The "Export layers to DXF" algorithm is relatively unusual (but not unique) because it has a single "input layers" input where more than one input layer can be selected. People are probably relatively unlikely to need to "run as batch process".
ISSUE
If people do need to "run as batch process", for some reason this algorithm is missing this functionality that is provided with other algorithms:
```
Add Files by Pattern...
Select Files...
Add All Files from a Directory...
Select from Open Layers...
````
Maybe someone thought this functionality isn't necessary for this algorithm?

Compare with the vector translate algorithm:

And with r.patch, which also has a single "input layers" input where more than one input layer can be selected:

### Steps to reproduce the issue
- click to open "Export layers to DXF" from the processing toolbox
- click "run as batch process" in the bottom left
- click "Autofill" button in the first column
The options are missing...
### Versions
QGIS version
3.22.3-Białowieża
QGIS code revision
1628765ec7
Qt version
5.15.2
Python version
3.9.5
GDAL/OGR version
3.4.1
PROJ version
8.2.1
EPSG Registry database version
v10.041 (2021-12-03)
GEOS version
3.10.0-CAPI-1.16.0
SQLite version
3.35.2
PDAL version
2.3.0
PostgreSQL client version
13.0
SpatiaLite version
5.0.1
QWT version
6.1.3
QScintilla2 version
2.11.5
OS version
Windows 10 Version 2009
Active Python plugins
AnotherDXF2Shape
1.2.3
AutoLayoutTool
1.1
autoSaver
2.6
batchvectorlayersaver
0.9
BulkVectorExport
1.1
CalculateGeometry
0.6.4
changeDataSource
3.1
DataPlotly
3.8.1
deactivate_active_labels
0.5
flowTrace
1.1.1
Generalizer3
1.0
GeoCoding
2.18
geo_sim_processing
1.2.0
getthemfiltered
0.1.3
gridSplitter
0.4.0
GroupStats
2.2.5
HideDocks
0.6.1
ImageServerConnector
2.1.1
ImportPhotos
3.0.3
joinmultiplelines
Version 0.4.1
karika
1.5
LayerBoard
1.0.1
linz-data-importer
2.2.3
loadthemall
3.3.0
MagicWand-master
1.3.1
mask
1.10.1
MemoryLayerSaver
4.0.4
mmqgis
2021.9.10
nominatim_locator_filter
0.2.4
numerator
0.2
numericalDigitize
0.4.6
pathfinder
version 0.4.1
plaingeometryeditor
3.0.0
plugin_reloader
0.9.1
powerpan
2.0
processing_saga
0.5.0
processing_taudem
3.0.0
processing_whitebox
0.14.0
profiletool
4.2.1
qchainage
3.0.1
QCopycanvas
0.5
qgis-plugin-findreplace-main
1
Qgis2threejs
2.6
QGIS3-getWKT
1.4
QuickMultiAttributeEdit3
version 3.0.3
QuickPrint
3.6.1
quicksaveqml
0.1.5
quick_map_services
0.19.27
rasmover-master
version 0.2
redLayer
2.2
selectThemes
3.0.1
simple_tools
0.4.1
splitmultipart
1.0.0
SplitPolygonShowingAreas
0.13
statist
3.2
themeselector
3.2.2
valuetool
3.0.15
ViewshedAnalysis
1.7
volume_calculation_tool
0.4
WaterNetAnalyzer-master
1.7
grassprovider
2.12.99
processing
2.12.99
sagaprovider
2.12.99
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
_No response_
|
1.0
|
Batch processing gui for "Export layers to DXF" is missing functionality for loading layers into separate rows in the table - ### What is the bug or the crash?
BACKGROUND COMMENTS
The "Export layers to DXF" algorithm is relatively unusual (but not unique) because it has a single "input layers" input where more than one input layer can be selected. People are probably relatively unlikely to need to "run as batch process".
ISSUE
If people do need to "run as batch process", for some reason this algorithm is missing this functionality that is provided with other algorithms:
```
Add Files by Pattern...
Select Files...
Add All Files from a Directory...
Select from Open Layers...
````
Maybe someone thought this functionality isn't necessary for this algorithm?

Compare with the vector translate algorithm:

And with r.patch, which also has a single "input layers" input where more than one input layer can be selected:

### Steps to reproduce the issue
- click to open "Export layers to DXF" from the processing toolbox
- click "run as batch process" in the bottom left
- click "Autofill" button in the first column
The options are missing...
### Versions
QGIS version
3.22.3-Białowieża
QGIS code revision
1628765ec7
Qt version
5.15.2
Python version
3.9.5
GDAL/OGR version
3.4.1
PROJ version
8.2.1
EPSG Registry database version
v10.041 (2021-12-03)
GEOS version
3.10.0-CAPI-1.16.0
SQLite version
3.35.2
PDAL version
2.3.0
PostgreSQL client version
13.0
SpatiaLite version
5.0.1
QWT version
6.1.3
QScintilla2 version
2.11.5
OS version
Windows 10 Version 2009
Active Python plugins
AnotherDXF2Shape
1.2.3
AutoLayoutTool
1.1
autoSaver
2.6
batchvectorlayersaver
0.9
BulkVectorExport
1.1
CalculateGeometry
0.6.4
changeDataSource
3.1
DataPlotly
3.8.1
deactivate_active_labels
0.5
flowTrace
1.1.1
Generalizer3
1.0
GeoCoding
2.18
geo_sim_processing
1.2.0
getthemfiltered
0.1.3
gridSplitter
0.4.0
GroupStats
2.2.5
HideDocks
0.6.1
ImageServerConnector
2.1.1
ImportPhotos
3.0.3
joinmultiplelines
Version 0.4.1
karika
1.5
LayerBoard
1.0.1
linz-data-importer
2.2.3
loadthemall
3.3.0
MagicWand-master
1.3.1
mask
1.10.1
MemoryLayerSaver
4.0.4
mmqgis
2021.9.10
nominatim_locator_filter
0.2.4
numerator
0.2
numericalDigitize
0.4.6
pathfinder
version 0.4.1
plaingeometryeditor
3.0.0
plugin_reloader
0.9.1
powerpan
2.0
processing_saga
0.5.0
processing_taudem
3.0.0
processing_whitebox
0.14.0
profiletool
4.2.1
qchainage
3.0.1
QCopycanvas
0.5
qgis-plugin-findreplace-main
1
Qgis2threejs
2.6
QGIS3-getWKT
1.4
QuickMultiAttributeEdit3
version 3.0.3
QuickPrint
3.6.1
quicksaveqml
0.1.5
quick_map_services
0.19.27
rasmover-master
version 0.2
redLayer
2.2
selectThemes
3.0.1
simple_tools
0.4.1
splitmultipart
1.0.0
SplitPolygonShowingAreas
0.13
statist
3.2
themeselector
3.2.2
valuetool
3.0.15
ViewshedAnalysis
1.7
volume_calculation_tool
0.4
WaterNetAnalyzer-master
1.7
grassprovider
2.12.99
processing
2.12.99
sagaprovider
2.12.99
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
_No response_
|
non_defect
|
batch processing gui for export layers to dxf is missing functionality for loading layers into separate rows in the table what is the bug or the crash background comments the export layers to dxf algorithm is relatively unusual but not unique because it has a single input layers input where more than one input layer can be selected people are probably relatively unlikely to need to run as batch process issue if people do need to run as batch process for some reason this algorithm is missing this functionality that is provided with other algorithms add files by pattern select files add all files from a directory select from open layers maybe someone thought this functionality isn t necessary for this algorithm compare with the vector translate algorithm and with r patch which also has a single input layers input where more than one input layer can be selected steps to reproduce the issue click to open export layers to dxf from the processing toolbox click run as batch process in the bottom left click autofill button in the first column the options are missing versions qgis version białowieża qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version spatialite version qwt version version os version windows version active python plugins autolayouttool autosaver batchvectorlayersaver bulkvectorexport calculategeometry changedatasource dataplotly deactivate active labels flowtrace geocoding geo sim processing getthemfiltered gridsplitter groupstats hidedocks imageserverconnector importphotos joinmultiplelines version karika layerboard linz data importer loadthemall magicwand master mask memorylayersaver mmqgis nominatim locator filter numerator numericaldigitize pathfinder version plaingeometryeditor plugin reloader powerpan processing saga processing taudem processing whitebox profiletool qchainage qcopycanvas qgis plugin findreplace main getwkt version quickprint quicksaveqml quick map services rasmover master version redlayer selectthemes simple tools splitmultipart splitpolygonshowingareas statist themeselector valuetool viewshedanalysis volume calculation tool waternetanalyzer master grassprovider processing sagaprovider supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
| 0
|
47,199
| 13,056,052,690
|
IssuesEvent
|
2020-07-30 03:30:40
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
closed
|
how not to clobber LD_LIBRARY_PATH in your env-shell.sh docs (Trac #124)
|
Migrated from Trac defect documentation
|
done in new .rst docs
Migrated from https://code.icecube.wisc.edu/ticket/124
```json
{
"status": "closed",
"changetime": "2014-11-23T03:37:56",
"description": "done in new .rst docs\n\n",
"reporter": "troy",
"cc": "",
"resolution": "fixed",
"_ts": "1416713876900096",
"component": "documentation",
"summary": "how not to clobber LD_LIBRARY_PATH in your env-shell.sh docs",
"priority": "normal",
"keywords": "",
"time": "2008-09-06T18:48:11",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
|
1.0
|
how not to clobber LD_LIBRARY_PATH in your env-shell.sh docs (Trac #124) - done in new .rst docs
Migrated from https://code.icecube.wisc.edu/ticket/124
```json
{
"status": "closed",
"changetime": "2014-11-23T03:37:56",
"description": "done in new .rst docs\n\n",
"reporter": "troy",
"cc": "",
"resolution": "fixed",
"_ts": "1416713876900096",
"component": "documentation",
"summary": "how not to clobber LD_LIBRARY_PATH in your env-shell.sh docs",
"priority": "normal",
"keywords": "",
"time": "2008-09-06T18:48:11",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
|
defect
|
how not to clobber ld library path in your env shell sh docs trac done in new rst docs migrated from json status closed changetime description done in new rst docs n n reporter troy cc resolution fixed ts component documentation summary how not to clobber ld library path in your env shell sh docs priority normal keywords time milestone owner troy type defect
| 1
|
70,609
| 23,262,903,751
|
IssuesEvent
|
2022-08-04 14:50:45
|
NREL/EnergyPlus
|
https://api.github.com/repos/NREL/EnergyPlus
|
closed
|
FuelFactors: Units of Measures and Energy per Unit Factor are never used
|
Defect
|
Issue overview
--------------
IDD:
```
FuelFactors,
[...]
A2, \field Units of Measure
N1, \field Energy per Unit Factor
```
https://github.com/NREL/EnergyPlus/blob/f39484f14fe4df6874c0334c90d79916767cf6ad/src/EnergyPlus/PollutionModule.cc#L260
As you can see, `cAlphaArgs(2)` is never used (this is the Units of Measure), and neither is `rNumericArgs(1)` (Energy per Unit Factor)
Either use them or remove them from the IDD
**Additional issue:**
Wrong index due to copy paste here: https://github.com/NREL/EnergyPlus/blob/f39484f14fe4df6874c0334c90d79916767cf6ad/src/EnergyPlus/PollutionModule.cc#L248-L251 Should be `3` in the `if`
### Details
Some additional details for this issue (if relevant):
- Platform (Operating system, version): all
- Version of EnergyPlus (if using an intermediate build, include SHA):
- Unmethours link or helpdesk ticket number
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [ ] Defect file added (list location of defect file here)
- [x] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
|
1.0
|
FuelFactors: Units of Measures and Energy per Unit Factor are never used - Issue overview
--------------
IDD:
```
FuelFactors,
[...]
A2, \field Units of Measure
N1, \field Energy per Unit Factor
```
https://github.com/NREL/EnergyPlus/blob/f39484f14fe4df6874c0334c90d79916767cf6ad/src/EnergyPlus/PollutionModule.cc#L260
As you can see, `cAlphaArgs(2)` is never used (this is the Units of Measure), and neither is `rNumericArgs(1)` (Energy per Unit Factor)
Either use them or remove them from the IDD
**Additional issue:**
Wrong index due to copy paste here: https://github.com/NREL/EnergyPlus/blob/f39484f14fe4df6874c0334c90d79916767cf6ad/src/EnergyPlus/PollutionModule.cc#L248-L251 Should be `3` in the `if`
### Details
Some additional details for this issue (if relevant):
- Platform (Operating system, version): all
- Version of EnergyPlus (if using an intermediate build, include SHA):
- Unmethours link or helpdesk ticket number
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [ ] Defect file added (list location of defect file here)
- [x] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
|
defect
|
fuelfactors units of measures and energy per unit factor are never used issue overview idd fuelfactors field units of measure field energy per unit factor as you can see calphaargs is never used this is the units of measure and neither is rnumericargs energy per unit factor either use them or remove them from the idd additional issue wrong index due to copy paste here should be in the if details some additional details for this issue if relevant platform operating system version all version of energyplus if using an intermediate build include sha unmethours link or helpdesk ticket number checklist add to this list or remove from it as applicable this is a simple templated set of guidelines defect file added list location of defect file here ticket added to pivotal for defect development team task pull request created the pull request will have additional tasks related to reviewing changes that fix this defect
| 1
|
8,230
| 2,611,472,162
|
IssuesEvent
|
2015-02-27 05:16:36
|
chrsmith/hedgewars
|
https://api.github.com/repos/chrsmith/hedgewars
|
closed
|
turns should not be possible while using blowtorch
|
auto-migrated Priority-Low Type-Defect
|
```
What steps will reproduce the problem?
* see the video at http://www.youtube.com/watch?v=a9-3_8NQpUk
What is the expected output? What do you see instead?
* it should not be possible for user to turn 180 degrees once blowtorch is activated
What version of the product are you using? On what operating system?
* 0.9.15 Ubuntu 10.04
```
Original issue reported on code.google.com by `jose1...@gmail.com` on 19 Jul 2011 at 11:58
|
1.0
|
turns should not be possible while using blowtorch - ```
What steps will reproduce the problem?
* see the video at http://www.youtube.com/watch?v=a9-3_8NQpUk
What is the expected output? What do you see instead?
* it should not be possible for user to turn 180 degrees once blowtorch is activated
What version of the product are you using? On what operating system?
* 0.9.15 Ubuntu 10.04
```
Original issue reported on code.google.com by `jose1...@gmail.com` on 19 Jul 2011 at 11:58
|
defect
|
turns should not be possible while using blowtorch what steps will reproduce the problem see the video at what is the expected output what do you see instead it should not be possible for user to turn degrees once blowtorch is activated what version of the product are you using on what operating system ubuntu original issue reported on code google com by gmail com on jul at
| 1
|
28,060
| 5,170,553,100
|
IssuesEvent
|
2017-01-18 06:50:02
|
TNGSB/eWallet
|
https://api.github.com/repos/TNGSB/eWallet
|
closed
|
e-Wallet_Mobile App (My Inbox - Content) 17012017 #21
|
Defect - Medium (Sev-3)
|
All of the inbox content displayed "Merchant Reference No." as "Null".
Attached screenshot for POT.




|
1.0
|
e-Wallet_Mobile App (My Inbox - Content) 17012017 #21 - All of the inbox content displayed "Merchant Reference No." as "Null".
Attached screenshot for POT.




|
defect
|
e wallet mobile app my inbox content all of the inbox content displayed merchant reference no as null attached screenshot for pot
| 1
|
393,476
| 26,993,605,194
|
IssuesEvent
|
2023-02-09 22:11:19
|
kitsudaiki/Hanami-AI
|
https://api.github.com/repos/kitsudaiki/Hanami-AI
|
opened
|
Use cert-manager
|
Documentation Improvement Security Torii Deploy-Process
|
## Feature-request
### Description
Use the cert-manager in kubernetes-setup for creating certificates for the https-connection. At the moment these are still hard-coded, which has to be fixed.
|
1.0
|
Use cert-manager - ## Feature-request
### Description
Use the cert-manager in kubernetes-setup for creating certificates for the https-connection. At the moment these are still hard-coded, which has to be fixed.
|
non_defect
|
use cert manager feature request description use the cert manager in kubernetes setup for creating certificates for the https connection at the moment these are still hard coded which has to be fixed
| 0
|
10,277
| 2,622,140,467
|
IssuesEvent
|
2015-03-04 00:01:48
|
byzhang/spserver
|
https://api.github.com/repos/byzhang/spserver
|
closed
|
SP_HttpEchoHandler如何停掉?
|
auto-migrated Priority-Medium Type-Defect
|
```
下载了你的SPSERVER测试,发现使用很方便,但有一个问题我比较�
��惑,在testiocphttp
里,怎么断开一个客户端连接呢.在例子里好象是keep-alive,这样�
��些恶意的连接不断开
怎么办?我还不是很清楚这个过程,望不吝赐教.3X
```
Original issue reported on code.google.com by `VxS...@gmail.com` on 20 Apr 2009 at 9:15
|
1.0
|
SP_HttpEchoHandler如何停掉? - ```
下载了你的SPSERVER测试,发现使用很方便,但有一个问题我比较�
��惑,在testiocphttp
里,怎么断开一个客户端连接呢.在例子里好象是keep-alive,这样�
��些恶意的连接不断开
怎么办?我还不是很清楚这个过程,望不吝赐教.3X
```
Original issue reported on code.google.com by `VxS...@gmail.com` on 20 Apr 2009 at 9:15
|
defect
|
sp httpechohandler如何停掉 下载了你的spserver测试 发现使用很方便 但有一个问题我比较� ��惑 在testiocphttp 里 怎么断开一个客户端连接呢 在例子里好象是keep alive 这样� ��些恶意的连接不断开 怎么办 我还不是很清楚这个过程 望不吝赐教 original issue reported on code google com by vxs gmail com on apr at
| 1
|
113,435
| 9,646,262,967
|
IssuesEvent
|
2019-05-17 10:46:20
|
appium/appium
|
https://api.github.com/repos/appium/appium
|
closed
|
Result of iOS page source becomes unexpectedly large and time-out happens
|
NeedsInfo ThirdParty XCUITest
|
## The problem
When I tries to retrieve the XML tree of my customer's iOS app by `driver.source` command, unexpectedly only the root element is returned. After I changed [AX_TIMEOUT](https://github.com/appium/WebDriverAgent/blob/master/WebDriverAgentLib/Categories/XCUIElement%2BFBUtilities.m#L77) to `300.` and rebuilt WebDriverAgent by myself, I could retrieve [very large tree](https://gist.githubusercontent.com/NozomiIto/cfd35c6a6281724e731cb1ef036fe20f/raw/ea5e7311037d1ec827a4fda56be401f8020880b1/Large%2520XML%2520tree). It contains about 300 `XCUIElementTypeCell` (most of them are `visible=false`) and I guess it makes the XML tree retrieval slow.
But what is strange is, the actual app does not have such so many TableViewCell elements. When the customer checked the [view hierarchy by Xcode](https://developer.apple.com/library/archive/documentation/DeveloperTools/Conceptual/debugging_with_xcode/chapters/special_debugging_workflows.html#//apple_ref/doc/uid/TP40015022-CH9-SW2), TableViewCells are only about 8.
Why source command returns so many elements..?
## Environment
* Appium 1.13.0. As I described about, I changed AX_TIMEOUT and rebuilt WebDriverAgent.
* Desktop OS/version used to run Appium: MacOS Mojave 10.14.4
* Node.js version: v10.11.0
* Npm or Yarn package manager: 6.4.1
* Mobile platform/version under test: iOS real device. iPad mini2. iOS12.2. (But this is also the same on iPhone8/iOS12.1/simulator)
* Appium CLI
## Details
- The back ground Cell elements continue to move upward slowly, since they contain the background image. This move may be related to the problem.
- I don't have the access right to the customer's app's source code. But I think they can offer necessary app's information to resolve the problem.
- To be exact, the user uses custom component whose name is ***.CollectionTableViewCell
- I attach the image of view hierarchy by Xcode. This is for a little different page than the page I retrieved the XML, but most of TableViewCell's structure and large XML problem is the same.

## Link to Appium logs
https://gist.github.com/NozomiIto/0b529fe6b9ae50ea44d45f02a209023b
|
1.0
|
Result of iOS page source becomes unexpectedly large and time-out happens - ## The problem
When I tries to retrieve the XML tree of my customer's iOS app by `driver.source` command, unexpectedly only the root element is returned. After I changed [AX_TIMEOUT](https://github.com/appium/WebDriverAgent/blob/master/WebDriverAgentLib/Categories/XCUIElement%2BFBUtilities.m#L77) to `300.` and rebuilt WebDriverAgent by myself, I could retrieve [very large tree](https://gist.githubusercontent.com/NozomiIto/cfd35c6a6281724e731cb1ef036fe20f/raw/ea5e7311037d1ec827a4fda56be401f8020880b1/Large%2520XML%2520tree). It contains about 300 `XCUIElementTypeCell` (most of them are `visible=false`) and I guess it makes the XML tree retrieval slow.
But what is strange is, the actual app does not have such so many TableViewCell elements. When the customer checked the [view hierarchy by Xcode](https://developer.apple.com/library/archive/documentation/DeveloperTools/Conceptual/debugging_with_xcode/chapters/special_debugging_workflows.html#//apple_ref/doc/uid/TP40015022-CH9-SW2), TableViewCells are only about 8.
Why source command returns so many elements..?
## Environment
* Appium 1.13.0. As I described about, I changed AX_TIMEOUT and rebuilt WebDriverAgent.
* Desktop OS/version used to run Appium: MacOS Mojave 10.14.4
* Node.js version: v10.11.0
* Npm or Yarn package manager: 6.4.1
* Mobile platform/version under test: iOS real device. iPad mini2. iOS12.2. (But this is also the same on iPhone8/iOS12.1/simulator)
* Appium CLI
## Details
- The back ground Cell elements continue to move upward slowly, since they contain the background image. This move may be related to the problem.
- I don't have the access right to the customer's app's source code. But I think they can offer necessary app's information to resolve the problem.
- To be exact, the user uses custom component whose name is ***.CollectionTableViewCell
- I attach the image of view hierarchy by Xcode. This is for a little different page than the page I retrieved the XML, but most of TableViewCell's structure and large XML problem is the same.

## Link to Appium logs
https://gist.github.com/NozomiIto/0b529fe6b9ae50ea44d45f02a209023b
|
non_defect
|
result of ios page source becomes unexpectedly large and time out happens the problem when i tries to retrieve the xml tree of my customer s ios app by driver source command unexpectedly only the root element is returned after i changed to and rebuilt webdriveragent by myself i could retrieve it contains about xcuielementtypecell most of them are visible false and i guess it makes the xml tree retrieval slow but what is strange is the actual app does not have such so many tableviewcell elements when the customer checked the tableviewcells are only about why source command returns so many elements environment appium as i described about i changed ax timeout and rebuilt webdriveragent desktop os version used to run appium macos mojave node js version npm or yarn package manager mobile platform version under test ios real device ipad but this is also the same on simulator appium cli details the back ground cell elements continue to move upward slowly since they contain the background image this move may be related to the problem i don t have the access right to the customer s app s source code but i think they can offer necessary app s information to resolve the problem to be exact the user uses custom component whose name is collectiontableviewcell i attach the image of view hierarchy by xcode this is for a little different page than the page i retrieved the xml but most of tableviewcell s structure and large xml problem is the same link to appium logs
| 0
|
28,842
| 5,390,067,332
|
IssuesEvent
|
2017-02-25 10:02:39
|
patmun/pynetdicom
|
https://api.github.com/repos/patmun/pynetdicom
|
closed
|
Migration to GitHub (or another place)
|
auto-migrated Priority-Medium Type-Defect
|
```
Google Code will become read only in less than a month, and the project should
be moved somewhere else.
Are you going to mantain it?
Are you going to move it somewhere else?
Should we just fork it?
Thanks!
```
Original issue reported on code.google.com by `gafo...@gmail.com` on 4 Aug 2015 at 7:46
|
1.0
|
Migration to GitHub (or another place) - ```
Google Code will become read only in less than a month, and the project should
be moved somewhere else.
Are you going to mantain it?
Are you going to move it somewhere else?
Should we just fork it?
Thanks!
```
Original issue reported on code.google.com by `gafo...@gmail.com` on 4 Aug 2015 at 7:46
|
defect
|
migration to github or another place google code will become read only in less than a month and the project should be moved somewhere else are you going to mantain it are you going to move it somewhere else should we just fork it thanks original issue reported on code google com by gafo gmail com on aug at
| 1
|
99,620
| 16,449,115,415
|
IssuesEvent
|
2021-05-21 01:06:47
|
AkshayMukkavilli/Analyzing-the-Significance-of-Structure-in-Amazon-Review-Data-Using-Machine-Learning-Approaches
|
https://api.github.com/repos/AkshayMukkavilli/Analyzing-the-Significance-of-Structure-in-Amazon-Review-Data-Using-Machine-Learning-Approaches
|
opened
|
CVE-2021-29579 (Medium) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl
|
security vulnerability
|
## CVE-2021-29579 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: /FinalProject/requirements.txt</p>
<p>Path to vulnerable library: teSource-ArchiveExtractor_8b9e071c-3b11-4aa9-ba60-cdeb60d053b7/20190525011350_65403/20190525011256_depth_0/9/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an end-to-end open source platform for machine learning. The implementation of `tf.raw_ops.MaxPoolGrad` is vulnerable to a heap buffer overflow. The implementation(https://github.com/tensorflow/tensorflow/blob/ab1e644b48c82cb71493f4362b4dd38f4577a1cf/tensorflow/core/kernels/maxpooling_op.cc#L194-L203) fails to validate that indices used to access elements of input/output arrays are valid. Whereas accesses to `input_backprop_flat` are guarded by `FastBoundsCheck`, the indexing in `out_backprop_flat` can result in OOB access. The fix will be included in TensorFlow 2.5.0. We will also cherrypick this commit on TensorFlow 2.4.2, TensorFlow 2.3.3, TensorFlow 2.2.3 and TensorFlow 2.1.4, as these are also affected and still in supported range.
<p>Publish Date: 2021-05-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29579>CVE-2021-29579</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-79fv-9865-4qcv">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-79fv-9865-4qcv</a></p>
<p>Release Date: 2021-05-14</p>
<p>Fix Resolution: tensorflow - 2.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-29579 (Medium) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2021-29579 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: /FinalProject/requirements.txt</p>
<p>Path to vulnerable library: teSource-ArchiveExtractor_8b9e071c-3b11-4aa9-ba60-cdeb60d053b7/20190525011350_65403/20190525011256_depth_0/9/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an end-to-end open source platform for machine learning. The implementation of `tf.raw_ops.MaxPoolGrad` is vulnerable to a heap buffer overflow. The implementation(https://github.com/tensorflow/tensorflow/blob/ab1e644b48c82cb71493f4362b4dd38f4577a1cf/tensorflow/core/kernels/maxpooling_op.cc#L194-L203) fails to validate that indices used to access elements of input/output arrays are valid. Whereas accesses to `input_backprop_flat` are guarded by `FastBoundsCheck`, the indexing in `out_backprop_flat` can result in OOB access. The fix will be included in TensorFlow 2.5.0. We will also cherrypick this commit on TensorFlow 2.4.2, TensorFlow 2.3.3, TensorFlow 2.2.3 and TensorFlow 2.1.4, as these are also affected and still in supported range.
<p>Publish Date: 2021-05-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29579>CVE-2021-29579</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-79fv-9865-4qcv">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-79fv-9865-4qcv</a></p>
<p>Release Date: 2021-05-14</p>
<p>Fix Resolution: tensorflow - 2.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in tensorflow whl cve medium severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file finalproject requirements txt path to vulnerable library tesource archiveextractor depth tensorflow tensorflow data purelib tensorflow dependency hierarchy x tensorflow whl vulnerable library vulnerability details tensorflow is an end to end open source platform for machine learning the implementation of tf raw ops maxpoolgrad is vulnerable to a heap buffer overflow the implementation fails to validate that indices used to access elements of input output arrays are valid whereas accesses to input backprop flat are guarded by fastboundscheck the indexing in out backprop flat can result in oob access the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow step up your open source security game with whitesource
| 0
|
41,887
| 10,694,431,226
|
IssuesEvent
|
2019-10-23 10:50:42
|
hazelcast/hazelcast
|
https://api.github.com/repos/hazelcast/hazelcast
|
opened
|
SIGSEGV com.hazelcast.map.impl.record.RecordReaderWriter$2.writeRecord(Lcom/hazelcast/nio/ObjectDataOutput;Lcom/hazelcast/map/impl/record/Record;Lcom/hazelcast/nio/serialization
|
Priority: High Team: Core Type: Critical Type: Defect Type: Test-Failure
|
2 different test are crashing the member JVM
http://jenkins.hazelcast.com/view/wan/job/wan-sync-oome/28/console
http://jenkins.hazelcast.com/view/split/job/split-quorum/35/console
```
#
# A fatal error has been detected by the Java Runtime Environment:
#
# SIGSEGV (0xb) at pc=0x00007f6841db8a69, pid=2476, tid=0x00007f68393a5700
#
# JRE version: Java(TM) SE Runtime Environment (8.0_201-b09) (build 1.8.0_201-b09)
# Java VM: Java HotSpot(TM) 64-Bit Server VM (25.201-b09 mixed mode linux-amd64 compressed oops)
# Problematic frame:
# J 4782 C2 com.hazelcast.map.impl.record.RecordReaderWriter$2.writeRecord(Lcom/hazelcast/nio/ObjectDataOutput;Lcom/hazelcast/map/impl/record/Record;Lcom/hazelcast/nio/serialization
/Data;)V (128 bytes) @ 0x00007f6841db8a69 [0x00007f6841db8960+0x109]
#
# Core dump written. Default location: /home/ec2-user/hz-root/HzMember2B/core or core.2476
#
# If you would like to submit a bug report, please visit:
# http://bugreport.java.com/bugreport/crash.jsp
#
```
```
Native frames: (J=compiled Java code, j=interpreted, Vv=VM code, C=native code)
J 4782 C2 com.hazelcast.map.impl.record.RecordReaderWriter$2.writeRecord(Lcom/hazelcast/nio/ObjectDataOutput;Lcom/hazelcast/map/impl/record/Record;Lcom/hazelcast/nio/serialization/D
ata;)V (128 bytes) @ 0x00007f6841db8a69 [0x00007f6841db8960+0x109]
J 5103 C2 com.hazelcast.map.impl.operation.PutAllBackupOperation.writeInternal(Lcom/hazelcast/nio/ObjectDataOutput;)V (92 bytes) @ 0x00007f684190c6bc [0x00007f684190c460+0x25c]
J 4162 C2 com.hazelcast.spi.impl.operationservice.Operation.writeData(Lcom/hazelcast/nio/ObjectDataOutput;)V (230 bytes) @ 0x00007f6841b4d770 [0x00007f6841b4d200+0x570]
J 4235 C2 com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.write(Lcom/hazelcast/nio/ObjectDataOutput;Ljava/lang/Object;)V (10 bytes) @ 0x00007f6841b604
74 [0x00007f6841b60320+0x154]
J 4187 C2 com.hazelcast.spi.impl.operationservice.impl.operations.Backup.writeInternal(Lcom/hazelcast/nio/ObjectDataOutput;)V (164 bytes) @ 0x00007f6841b5b92c [0x00007f6841b5b700+0x
22c]
J 4162 C2 com.hazelcast.spi.impl.operationservice.Operation.writeData(Lcom/hazelcast/nio/ObjectDataOutput;)V (230 bytes) @ 0x00007f6841b4d770 [0x00007f6841b4d200+0x570]
J 4235 C2 com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.write(Lcom/hazelcast/nio/ObjectDataOutput;Ljava/lang/Object;)V (10 bytes) @ 0x00007f6841b604
74 [0x00007f6841b60320+0x154]
J 5407 C2 com.hazelcast.internal.serialization.impl.AbstractSerializationService.toBytes(Ljava/lang/Object;IZLcom/hazelcast/partition/PartitioningStrategy;Ljava/nio/ByteOrder;)[B (1
39 bytes) @ 0x00007f68416232a4 [0x00007f6841622ea0+0x404]
J 5142 C2 com.hazelcast.spi.impl.operationservice.impl.OutboundOperationHandler.send(Lcom/hazelcast/spi/impl/operationservice/Operation;Lcom/hazelcast/cluster/Address;)Z (83 bytes)
@ 0x00007f6841758ea0 [0x00007f6841758b80+0x320]
J 5149 C2 com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.handleResponse(Lcom/hazelcast/spi/impl/operationservice/Operation;)V (52 bytes) @ 0x00007f6841b280a4 [0x00
007f6841b27100+0xfa4]
J 5365 C2 com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.call(Lcom/hazelcast/spi/impl/operationservice/Operation;)V (139 bytes) @ 0x00007f6841185614 [0x00007f68411
85560+0xb4]
J 5054 C2 com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.process(Ljava/lang/Object;)V (206 bytes) @ 0x00007f6841e53d44 [0x00007f6841e539e0+0x364]
J 5070% C2 com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.run()V (101 bytes) @ 0x00007f6841169dc8 [0x00007f6841169d20+0xa8]
v ~StubRoutines::call_stub
```
Danny Conlon 1:45 PM
http://54.147.27.51/~jenkins/workspace/split-quorum/4.0-SNAPSHOT/2019_10_23-07_52_53/quorum/output/HZ/HzMember3HZBB/hs_err_pid2321.log
core dump in this run
http://jenkins.hazelcast.com/view/split/job/split-quorum/35/console
./output/HZ/HzMember3HZBB/hs_err_pid2302.log
./output/HZ/HzMember1HZBB/hs_err_pid2301.log
./output/HZ/HzMember2HZBB/hs_err_pid2318.log
./output/HZ/HzMember3HZBB/core.2302
./output/HZ/HzMember1HZBB/core.2301
./output/HZ/HzMember2HZBB/core.2318
I also a different test which also crashes the JVM
http://jenkins.hazelcast.com/view/wan/job/wan-sync-oome/28/console
```
# Problematic frame:
# J 4782 C2 com.hazelcast.map.impl.record.RecordReaderWriter$2.writeRecord(Lcom/hazelcast/nio/ObjectDataOutput;Lcom/hazelcast/map/impl/record/Record;Lcom/hazelcast/nio/serialization
/Data;)V (128 bytes) @ 0x00007f6841db8a69 [0x00007f6841db8960+0x109]
```
probably has same root cause
http://54.147.27.51/~jenkins/workspace/wan-sync-oome/4.0-SNAPSHOT/2019_10_23-10_10_48/sync-oome/
output/B/HzMember2B/core.2476
|
1.0
|
SIGSEGV com.hazelcast.map.impl.record.RecordReaderWriter$2.writeRecord(Lcom/hazelcast/nio/ObjectDataOutput;Lcom/hazelcast/map/impl/record/Record;Lcom/hazelcast/nio/serialization - 2 different test are crashing the member JVM
http://jenkins.hazelcast.com/view/wan/job/wan-sync-oome/28/console
http://jenkins.hazelcast.com/view/split/job/split-quorum/35/console
```
#
# A fatal error has been detected by the Java Runtime Environment:
#
# SIGSEGV (0xb) at pc=0x00007f6841db8a69, pid=2476, tid=0x00007f68393a5700
#
# JRE version: Java(TM) SE Runtime Environment (8.0_201-b09) (build 1.8.0_201-b09)
# Java VM: Java HotSpot(TM) 64-Bit Server VM (25.201-b09 mixed mode linux-amd64 compressed oops)
# Problematic frame:
# J 4782 C2 com.hazelcast.map.impl.record.RecordReaderWriter$2.writeRecord(Lcom/hazelcast/nio/ObjectDataOutput;Lcom/hazelcast/map/impl/record/Record;Lcom/hazelcast/nio/serialization
/Data;)V (128 bytes) @ 0x00007f6841db8a69 [0x00007f6841db8960+0x109]
#
# Core dump written. Default location: /home/ec2-user/hz-root/HzMember2B/core or core.2476
#
# If you would like to submit a bug report, please visit:
# http://bugreport.java.com/bugreport/crash.jsp
#
```
```
Native frames: (J=compiled Java code, j=interpreted, Vv=VM code, C=native code)
J 4782 C2 com.hazelcast.map.impl.record.RecordReaderWriter$2.writeRecord(Lcom/hazelcast/nio/ObjectDataOutput;Lcom/hazelcast/map/impl/record/Record;Lcom/hazelcast/nio/serialization/D
ata;)V (128 bytes) @ 0x00007f6841db8a69 [0x00007f6841db8960+0x109]
J 5103 C2 com.hazelcast.map.impl.operation.PutAllBackupOperation.writeInternal(Lcom/hazelcast/nio/ObjectDataOutput;)V (92 bytes) @ 0x00007f684190c6bc [0x00007f684190c460+0x25c]
J 4162 C2 com.hazelcast.spi.impl.operationservice.Operation.writeData(Lcom/hazelcast/nio/ObjectDataOutput;)V (230 bytes) @ 0x00007f6841b4d770 [0x00007f6841b4d200+0x570]
J 4235 C2 com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.write(Lcom/hazelcast/nio/ObjectDataOutput;Ljava/lang/Object;)V (10 bytes) @ 0x00007f6841b604
74 [0x00007f6841b60320+0x154]
J 4187 C2 com.hazelcast.spi.impl.operationservice.impl.operations.Backup.writeInternal(Lcom/hazelcast/nio/ObjectDataOutput;)V (164 bytes) @ 0x00007f6841b5b92c [0x00007f6841b5b700+0x
22c]
J 4162 C2 com.hazelcast.spi.impl.operationservice.Operation.writeData(Lcom/hazelcast/nio/ObjectDataOutput;)V (230 bytes) @ 0x00007f6841b4d770 [0x00007f6841b4d200+0x570]
J 4235 C2 com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.write(Lcom/hazelcast/nio/ObjectDataOutput;Ljava/lang/Object;)V (10 bytes) @ 0x00007f6841b604
74 [0x00007f6841b60320+0x154]
J 5407 C2 com.hazelcast.internal.serialization.impl.AbstractSerializationService.toBytes(Ljava/lang/Object;IZLcom/hazelcast/partition/PartitioningStrategy;Ljava/nio/ByteOrder;)[B (1
39 bytes) @ 0x00007f68416232a4 [0x00007f6841622ea0+0x404]
J 5142 C2 com.hazelcast.spi.impl.operationservice.impl.OutboundOperationHandler.send(Lcom/hazelcast/spi/impl/operationservice/Operation;Lcom/hazelcast/cluster/Address;)Z (83 bytes)
@ 0x00007f6841758ea0 [0x00007f6841758b80+0x320]
J 5149 C2 com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.handleResponse(Lcom/hazelcast/spi/impl/operationservice/Operation;)V (52 bytes) @ 0x00007f6841b280a4 [0x00
007f6841b27100+0xfa4]
J 5365 C2 com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.call(Lcom/hazelcast/spi/impl/operationservice/Operation;)V (139 bytes) @ 0x00007f6841185614 [0x00007f68411
85560+0xb4]
J 5054 C2 com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.process(Ljava/lang/Object;)V (206 bytes) @ 0x00007f6841e53d44 [0x00007f6841e539e0+0x364]
J 5070% C2 com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.run()V (101 bytes) @ 0x00007f6841169dc8 [0x00007f6841169d20+0xa8]
v ~StubRoutines::call_stub
```
Danny Conlon 1:45 PM
http://54.147.27.51/~jenkins/workspace/split-quorum/4.0-SNAPSHOT/2019_10_23-07_52_53/quorum/output/HZ/HzMember3HZBB/hs_err_pid2321.log
core dump in this run
http://jenkins.hazelcast.com/view/split/job/split-quorum/35/console
./output/HZ/HzMember3HZBB/hs_err_pid2302.log
./output/HZ/HzMember1HZBB/hs_err_pid2301.log
./output/HZ/HzMember2HZBB/hs_err_pid2318.log
./output/HZ/HzMember3HZBB/core.2302
./output/HZ/HzMember1HZBB/core.2301
./output/HZ/HzMember2HZBB/core.2318
I also a different test which also crashes the JVM
http://jenkins.hazelcast.com/view/wan/job/wan-sync-oome/28/console
```
# Problematic frame:
# J 4782 C2 com.hazelcast.map.impl.record.RecordReaderWriter$2.writeRecord(Lcom/hazelcast/nio/ObjectDataOutput;Lcom/hazelcast/map/impl/record/Record;Lcom/hazelcast/nio/serialization
/Data;)V (128 bytes) @ 0x00007f6841db8a69 [0x00007f6841db8960+0x109]
```
probably has same root cause
http://54.147.27.51/~jenkins/workspace/wan-sync-oome/4.0-SNAPSHOT/2019_10_23-10_10_48/sync-oome/
output/B/HzMember2B/core.2476
|
defect
|
sigsegv com hazelcast map impl record recordreaderwriter writerecord lcom hazelcast nio objectdataoutput lcom hazelcast map impl record record lcom hazelcast nio serialization different test are crashing the member jvm a fatal error has been detected by the java runtime environment sigsegv at pc pid tid jre version java tm se runtime environment build java vm java hotspot tm bit server vm mixed mode linux compressed oops problematic frame j com hazelcast map impl record recordreaderwriter writerecord lcom hazelcast nio objectdataoutput lcom hazelcast map impl record record lcom hazelcast nio serialization data v bytes core dump written default location home user hz root core or core if you would like to submit a bug report please visit native frames j compiled java code j interpreted vv vm code c native code j com hazelcast map impl record recordreaderwriter writerecord lcom hazelcast nio objectdataoutput lcom hazelcast map impl record record lcom hazelcast nio serialization d ata v bytes j com hazelcast map impl operation putallbackupoperation writeinternal lcom hazelcast nio objectdataoutput v bytes j com hazelcast spi impl operationservice operation writedata lcom hazelcast nio objectdataoutput v bytes j com hazelcast internal serialization impl enterprisedataserializableserializer write lcom hazelcast nio objectdataoutput ljava lang object v bytes j com hazelcast spi impl operationservice impl operations backup writeinternal lcom hazelcast nio objectdataoutput v bytes j com hazelcast spi impl operationservice operation writedata lcom hazelcast nio objectdataoutput v bytes j com hazelcast internal serialization impl enterprisedataserializableserializer write lcom hazelcast nio objectdataoutput ljava lang object v bytes j com hazelcast internal serialization impl abstractserializationservice tobytes ljava lang object izlcom hazelcast partition partitioningstrategy ljava nio byteorder b bytes j com hazelcast spi impl operationservice impl outboundoperationhandler send lcom hazelcast spi impl operationservice operation lcom hazelcast cluster address z bytes j com hazelcast spi impl operationservice impl operationrunnerimpl handleresponse lcom hazelcast spi impl operationservice operation v bytes j com hazelcast spi impl operationservice impl operationrunnerimpl call lcom hazelcast spi impl operationservice operation v bytes j com hazelcast spi impl operationexecutor impl operationthread process ljava lang object v bytes j com hazelcast spi impl operationexecutor impl operationthread run v bytes v stubroutines call stub danny conlon pm core dump in this run output hz hs err log output hz hs err log output hz hs err log output hz core output hz core output hz core i also a different test which also crashes the jvm problematic frame j com hazelcast map impl record recordreaderwriter writerecord lcom hazelcast nio objectdataoutput lcom hazelcast map impl record record lcom hazelcast nio serialization data v bytes probably has same root cause output b core
| 1
|
45,492
| 12,827,141,583
|
IssuesEvent
|
2020-07-06 17:53:45
|
department-of-veterans-affairs/va.gov-team
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
|
closed
|
[SCREENREADER]: Alert Boxes - Role "alert" MUST announce itself to assistive tech on page load, and on change
|
508-defect-2 508/Accessibility design system
|
# [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2)
<!--
Enter an issue title using the format [ERROR TYPE]: Brief description of the problem
---
[SCREENREADER]: Edit buttons need aria-label for context
[KEYBOARD]: Add another user link will not receive keyboard focus
[AXE-CORE]: Heading levels should increase by one
[COGNITION]: Error messages should be more specific
[COLOR]: Blue button on blue background does not have sufficient contrast ratio
---
-->
<!-- It's okay to delete the instructions above, but leave the link to the 508 defect severity level for your issue. -->
**Feedback framework**
- **❗️ Must** for if the feedback must be applied
- **⚠️ Should** if the feedback is best practice
- **✔️ Consider** for suggestions/enhancements
## Description
<!-- This is a detailed description of the issue. It should include a restatement of the title, and provide more background information. -->
Our [Alertbox error variant](https://design.va.gov/components/alertboxes#error-alert) of the Alertbox component contains a `role="alert"` for announcing important content to screen readers immediately. In testing, I discovered that several screen reader + browser combinations do not announce these alerts when they are rendered on first page load:
* Win10 + Firefox + NVDA
* Win10 + IE11 + JAWS
* iOS + VoiceOver + Mobile Safari
We need to come up with a strategy to inject the `role="alert"` after the component is rendered or another way to trigger a change in the component so this type of alert is read out regardless of when it is added to the document object model (DOM).
## Point of Contact
<!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket.
-->
**VFS Point of Contact:** _Trevor_
## Acceptance Criteria
<!-- As a keyboard user, I want to open the Level of Coverage widget by pressing Spacebar or pressing Enter. These keypress actions should not interfere with the mouse click event also opening the widget. -->
- [ ] Error alerts are read out on first page render
- [ ] Error alerts are read out when they are dynamically added to a page
## Environment
* https://design.va.gov/components/alertboxes#error-alert
## WCAG or Vendor Guidance (optional)
* [Name, Role, Value: Understanding SC 4.1.2](https://www.w3.org/TR/UNDERSTANDING-WCAG20/ensure-compat-rsv.html)
## Screenshots or Trace Logs
<!-- Drop any screenshots or error logs that might be useful for debugging -->

|
1.0
|
[SCREENREADER]: Alert Boxes - Role "alert" MUST announce itself to assistive tech on page load, and on change - # [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2)
<!--
Enter an issue title using the format [ERROR TYPE]: Brief description of the problem
---
[SCREENREADER]: Edit buttons need aria-label for context
[KEYBOARD]: Add another user link will not receive keyboard focus
[AXE-CORE]: Heading levels should increase by one
[COGNITION]: Error messages should be more specific
[COLOR]: Blue button on blue background does not have sufficient contrast ratio
---
-->
<!-- It's okay to delete the instructions above, but leave the link to the 508 defect severity level for your issue. -->
**Feedback framework**
- **❗️ Must** for if the feedback must be applied
- **⚠️ Should** if the feedback is best practice
- **✔️ Consider** for suggestions/enhancements
## Description
<!-- This is a detailed description of the issue. It should include a restatement of the title, and provide more background information. -->
Our [Alertbox error variant](https://design.va.gov/components/alertboxes#error-alert) of the Alertbox component contains a `role="alert"` for announcing important content to screen readers immediately. In testing, I discovered that several screen reader + browser combinations do not announce these alerts when they are rendered on first page load:
* Win10 + Firefox + NVDA
* Win10 + IE11 + JAWS
* iOS + VoiceOver + Mobile Safari
We need to come up with a strategy to inject the `role="alert"` after the component is rendered or another way to trigger a change in the component so this type of alert is read out regardless of when it is added to the document object model (DOM).
## Point of Contact
<!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket.
-->
**VFS Point of Contact:** _Trevor_
## Acceptance Criteria
<!-- As a keyboard user, I want to open the Level of Coverage widget by pressing Spacebar or pressing Enter. These keypress actions should not interfere with the mouse click event also opening the widget. -->
- [ ] Error alerts are read out on first page render
- [ ] Error alerts are read out when they are dynamically added to a page
## Environment
* https://design.va.gov/components/alertboxes#error-alert
## WCAG or Vendor Guidance (optional)
* [Name, Role, Value: Understanding SC 4.1.2](https://www.w3.org/TR/UNDERSTANDING-WCAG20/ensure-compat-rsv.html)
## Screenshots or Trace Logs
<!-- Drop any screenshots or error logs that might be useful for debugging -->

|
defect
|
alert boxes role alert must announce itself to assistive tech on page load and on change enter an issue title using the format brief description of the problem edit buttons need aria label for context add another user link will not receive keyboard focus heading levels should increase by one error messages should be more specific blue button on blue background does not have sufficient contrast ratio feedback framework ❗️ must for if the feedback must be applied ⚠️ should if the feedback is best practice ✔️ consider for suggestions enhancements description our of the alertbox component contains a role alert for announcing important content to screen readers immediately in testing i discovered that several screen reader browser combinations do not announce these alerts when they are rendered on first page load firefox nvda jaws ios voiceover mobile safari we need to come up with a strategy to inject the role alert after the component is rendered or another way to trigger a change in the component so this type of alert is read out regardless of when it is added to the document object model dom point of contact if this issue is being opened by a vfs team member please add a point of contact usually this is the same person who enters the issue ticket vfs point of contact trevor acceptance criteria error alerts are read out on first page render error alerts are read out when they are dynamically added to a page environment wcag or vendor guidance optional screenshots or trace logs
| 1
|
40,714
| 10,140,875,954
|
IssuesEvent
|
2019-08-03 08:30:55
|
STEllAR-GROUP/hpx
|
https://api.github.com/repos/STEllAR-GROUP/hpx
|
closed
|
HPX MPI use incorrect
|
category: parcel transport tag: wontfix type: defect
|
In the CMakeLists for the MPI parcelport, variables such as MPI_CXX_LIBRARIES are used.
HPX does not use the C++ interface to MPI and therefore linking to MPI_CXX_LIBRARIES is incorrect.
It is sufficient to use the MPI_C_LIBRARIES and MPI_C_INCLUDE_PATH etc etc
|
1.0
|
HPX MPI use incorrect - In the CMakeLists for the MPI parcelport, variables such as MPI_CXX_LIBRARIES are used.
HPX does not use the C++ interface to MPI and therefore linking to MPI_CXX_LIBRARIES is incorrect.
It is sufficient to use the MPI_C_LIBRARIES and MPI_C_INCLUDE_PATH etc etc
|
defect
|
hpx mpi use incorrect in the cmakelists for the mpi parcelport variables such as mpi cxx libraries are used hpx does not use the c interface to mpi and therefore linking to mpi cxx libraries is incorrect it is sufficient to use the mpi c libraries and mpi c include path etc etc
| 1
|
97,165
| 12,217,474,400
|
IssuesEvent
|
2020-05-01 17:16:35
|
factn/resilience-app
|
https://api.github.com/repos/factn/resilience-app
|
opened
|
Organizer can switch volunteers from active to inactive
|
Design Needed
|
**Description**
As an organizer, I can switch volunteers from active to inactive status and back so that I can control who is eligible to volunteer in my organization (for health and safety reasons).
[Airtable](https://airtable.com/tblRjd2McKgRW9MKj/viwlRRNI0PZIxrGkc/recMPlGUt4v47ORpO?blocks=hide)
|
1.0
|
Organizer can switch volunteers from active to inactive - **Description**
As an organizer, I can switch volunteers from active to inactive status and back so that I can control who is eligible to volunteer in my organization (for health and safety reasons).
[Airtable](https://airtable.com/tblRjd2McKgRW9MKj/viwlRRNI0PZIxrGkc/recMPlGUt4v47ORpO?blocks=hide)
|
non_defect
|
organizer can switch volunteers from active to inactive description as an organizer i can switch volunteers from active to inactive status and back so that i can control who is eligible to volunteer in my organization for health and safety reasons
| 0
|
751,202
| 26,232,935,256
|
IssuesEvent
|
2023-01-05 03:00:59
|
AleoHQ/leo
|
https://api.github.com/repos/AleoHQ/leo
|
closed
|
[Bug] compiling 18 `i32` exponent operations produces 30gb project folder
|
bug fuzz-bug priority-medium
|
## 🐛 Bug Report
this seems to have basically the same cause as the OOM errors.
every time you do an `i32 ** i32` operation with input it adds about 1gb to the proving key size. the following code leads to a proving key that's over 16gb and a json that's over 13gb. i could have done more iterations, but this program took 50gb of ram and anything more than that just OOM's me.
#### Code snippet to reproduce
```
function main(x: i32) {
for i in 0..18 {
let y = x ** 2;
}
}
```
## Your Environment
- leo commit 7e24b38620f30dda816b1a76ae4e6486b8a838e0
- rustc version 1.55.0-nightly
- Windows 10.0.19043 (Windows 10 Pro) [64-bit]
|
1.0
|
[Bug] compiling 18 `i32` exponent operations produces 30gb project folder - ## 🐛 Bug Report
this seems to have basically the same cause as the OOM errors.
every time you do an `i32 ** i32` operation with input it adds about 1gb to the proving key size. the following code leads to a proving key that's over 16gb and a json that's over 13gb. i could have done more iterations, but this program took 50gb of ram and anything more than that just OOM's me.
#### Code snippet to reproduce
```
function main(x: i32) {
for i in 0..18 {
let y = x ** 2;
}
}
```
## Your Environment
- leo commit 7e24b38620f30dda816b1a76ae4e6486b8a838e0
- rustc version 1.55.0-nightly
- Windows 10.0.19043 (Windows 10 Pro) [64-bit]
|
non_defect
|
compiling exponent operations produces project folder 🐛 bug report this seems to have basically the same cause as the oom errors every time you do an operation with input it adds about to the proving key size the following code leads to a proving key that s over and a json that s over i could have done more iterations but this program took of ram and anything more than that just oom s me code snippet to reproduce function main x for i in let y x your environment leo commit rustc version nightly windows windows pro
| 0
|
276,513
| 23,996,942,497
|
IssuesEvent
|
2022-09-14 08:20:35
|
INL/corpus-frontend
|
https://api.github.com/repos/INL/corpus-frontend
|
closed
|
Improve date filter
|
enhancement status: finished/testing
|
3 fields for from and to. Still formatted yyyy-mm-dd, but with just separate input fields.
Ensure the placeholders for these fields are the lowest and highest possible values.
|
1.0
|
Improve date filter - 3 fields for from and to. Still formatted yyyy-mm-dd, but with just separate input fields.
Ensure the placeholders for these fields are the lowest and highest possible values.
|
non_defect
|
improve date filter fields for from and to still formatted yyyy mm dd but with just separate input fields ensure the placeholders for these fields are the lowest and highest possible values
| 0
|
9,995
| 2,616,018,664
|
IssuesEvent
|
2015-03-02 01:00:16
|
jasonhall/bwapi
|
https://api.github.com/repos/jasonhall/bwapi
|
closed
|
Local PC has some issues (instability)
|
auto-migrated Component-Logic Milestone-MajorRelease Performance Priority-Medium Type-Defect Usability
|
```
Will have to address these issues at some point.
```
Original issue reported on code.google.com by `AHeinerm` on 4 Jul 2012 at 3:22
|
1.0
|
Local PC has some issues (instability) - ```
Will have to address these issues at some point.
```
Original issue reported on code.google.com by `AHeinerm` on 4 Jul 2012 at 3:22
|
defect
|
local pc has some issues instability will have to address these issues at some point original issue reported on code google com by aheinerm on jul at
| 1
|
216,223
| 24,247,373,959
|
IssuesEvent
|
2022-09-27 11:43:14
|
SmartBear/zephyr-scale-junit-integration
|
https://api.github.com/repos/SmartBear/zephyr-scale-junit-integration
|
closed
|
CVE-2019-10202 (High) detected in jackson-databind-2.9.4.jar - autoclosed
|
security vulnerability
|
## CVE-2019-10202 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /rxml/jackson/core/jackson-databind/2.9.4/jackson-databind-2.9.4.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.4.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/SmartBear/zephyr-scale-junit-integration/commit/a0028c5cb64062650f7c73dc13cd92022e8a1fac">a0028c5cb64062650f7c73dc13cd92022e8a1fac</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A series of deserialization vulnerabilities have been discovered in Codehaus 1.9.x implemented in EAP 7. This CVE fixes CVE-2017-17485, CVE-2017-7525, CVE-2017-15095, CVE-2018-5968, CVE-2018-7489, CVE-2018-1000873, CVE-2019-12086 reported for FasterXML jackson-databind by implementing a whitelist approach that will mitigate these vulnerabilities and future ones alike.
<p>Publish Date: 2019-10-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10202>CVE-2019-10202</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://access.redhat.com/errata/RHSA-2019:2938">https://access.redhat.com/errata/RHSA-2019:2938</a></p>
<p>Release Date: 2019-10-01</p>
<p>Fix Resolution: 2.9.9</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
|
True
|
CVE-2019-10202 (High) detected in jackson-databind-2.9.4.jar - autoclosed - ## CVE-2019-10202 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /rxml/jackson/core/jackson-databind/2.9.4/jackson-databind-2.9.4.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.4.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/SmartBear/zephyr-scale-junit-integration/commit/a0028c5cb64062650f7c73dc13cd92022e8a1fac">a0028c5cb64062650f7c73dc13cd92022e8a1fac</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A series of deserialization vulnerabilities have been discovered in Codehaus 1.9.x implemented in EAP 7. This CVE fixes CVE-2017-17485, CVE-2017-7525, CVE-2017-15095, CVE-2018-5968, CVE-2018-7489, CVE-2018-1000873, CVE-2019-12086 reported for FasterXML jackson-databind by implementing a whitelist approach that will mitigate these vulnerabilities and future ones alike.
<p>Publish Date: 2019-10-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10202>CVE-2019-10202</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://access.redhat.com/errata/RHSA-2019:2938">https://access.redhat.com/errata/RHSA-2019:2938</a></p>
<p>Release Date: 2019-10-01</p>
<p>Fix Resolution: 2.9.9</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
|
non_defect
|
cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library rxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch main vulnerability details a series of deserialization vulnerabilities have been discovered in codehaus x implemented in eap this cve fixes cve cve cve cve cve cve cve reported for fasterxml jackson databind by implementing a whitelist approach that will mitigate these vulnerabilities and future ones alike publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution check this box to open an automated fix pr
| 0
|
122,917
| 17,771,630,761
|
IssuesEvent
|
2021-08-30 14:17:03
|
Tim-sandbox/barista
|
https://api.github.com/repos/Tim-sandbox/barista
|
opened
|
CVE-2021-23364 (Medium) detected in browserslist-4.14.2.tgz, browserslist-4.14.6.tgz
|
security vulnerability
|
## CVE-2021-23364 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>browserslist-4.14.2.tgz</b>, <b>browserslist-4.14.6.tgz</b></p></summary>
<p>
<details><summary><b>browserslist-4.14.2.tgz</b></p></summary>
<p>Share target browsers between different front-end tools, like Autoprefixer, Stylelint and babel-env-preset</p>
<p>Library home page: <a href="https://registry.npmjs.org/browserslist/-/browserslist-4.14.2.tgz">https://registry.npmjs.org/browserslist/-/browserslist-4.14.2.tgz</a></p>
<p>Path to dependency file: barista/barista-docs/package.json</p>
<p>Path to vulnerable library: barista/barista-docs/node_modules/react-dev-utils/node_modules/browserslist/package.json</p>
<p>
Dependency Hierarchy:
- core-2.0.0-beta.5.tgz (Root Library)
- react-dev-utils-11.0.4.tgz
- :x: **browserslist-4.14.2.tgz** (Vulnerable Library)
</details>
<details><summary><b>browserslist-4.14.6.tgz</b></p></summary>
<p>Share target browsers between different front-end tools, like Autoprefixer, Stylelint and babel-env-preset</p>
<p>Library home page: <a href="https://registry.npmjs.org/browserslist/-/browserslist-4.14.6.tgz">https://registry.npmjs.org/browserslist/-/browserslist-4.14.6.tgz</a></p>
<p>Path to dependency file: barista/barista-web/package.json</p>
<p>Path to vulnerable library: barista/barista-web/node_modules/browserslist/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.1102.13.tgz (Root Library)
- :x: **browserslist-4.14.6.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/Tim-sandbox/barista/commit/2b8e77b2ff0d688bfd2ffb44061287e82fa71967">2b8e77b2ff0d688bfd2ffb44061287e82fa71967</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package browserslist from 4.0.0 and before 4.16.5 are vulnerable to Regular Expression Denial of Service (ReDoS) during parsing of queries.
<p>Publish Date: 2021-04-28
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23364>CVE-2021-23364</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23364">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23364</a></p>
<p>Release Date: 2021-04-28</p>
<p>Fix Resolution: browserslist - 4.16.5</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"browserslist","packageVersion":"4.14.2","packageFilePaths":["/barista-docs/package.json"],"isTransitiveDependency":true,"dependencyTree":"@docusaurus/core:2.0.0-beta.5;react-dev-utils:11.0.4;browserslist:4.14.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"browserslist - 4.16.5"},{"packageType":"javascript/Node.js","packageName":"browserslist","packageVersion":"4.14.6","packageFilePaths":["/barista-web/package.json"],"isTransitiveDependency":true,"dependencyTree":"@angular-devkit/build-angular:0.1102.13;browserslist:4.14.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"browserslist - 4.16.5"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23364","vulnerabilityDetails":"The package browserslist from 4.0.0 and before 4.16.5 are vulnerable to Regular Expression Denial of Service (ReDoS) during parsing of queries.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23364","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2021-23364 (Medium) detected in browserslist-4.14.2.tgz, browserslist-4.14.6.tgz - ## CVE-2021-23364 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>browserslist-4.14.2.tgz</b>, <b>browserslist-4.14.6.tgz</b></p></summary>
<p>
<details><summary><b>browserslist-4.14.2.tgz</b></p></summary>
<p>Share target browsers between different front-end tools, like Autoprefixer, Stylelint and babel-env-preset</p>
<p>Library home page: <a href="https://registry.npmjs.org/browserslist/-/browserslist-4.14.2.tgz">https://registry.npmjs.org/browserslist/-/browserslist-4.14.2.tgz</a></p>
<p>Path to dependency file: barista/barista-docs/package.json</p>
<p>Path to vulnerable library: barista/barista-docs/node_modules/react-dev-utils/node_modules/browserslist/package.json</p>
<p>
Dependency Hierarchy:
- core-2.0.0-beta.5.tgz (Root Library)
- react-dev-utils-11.0.4.tgz
- :x: **browserslist-4.14.2.tgz** (Vulnerable Library)
</details>
<details><summary><b>browserslist-4.14.6.tgz</b></p></summary>
<p>Share target browsers between different front-end tools, like Autoprefixer, Stylelint and babel-env-preset</p>
<p>Library home page: <a href="https://registry.npmjs.org/browserslist/-/browserslist-4.14.6.tgz">https://registry.npmjs.org/browserslist/-/browserslist-4.14.6.tgz</a></p>
<p>Path to dependency file: barista/barista-web/package.json</p>
<p>Path to vulnerable library: barista/barista-web/node_modules/browserslist/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.1102.13.tgz (Root Library)
- :x: **browserslist-4.14.6.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/Tim-sandbox/barista/commit/2b8e77b2ff0d688bfd2ffb44061287e82fa71967">2b8e77b2ff0d688bfd2ffb44061287e82fa71967</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package browserslist from 4.0.0 and before 4.16.5 are vulnerable to Regular Expression Denial of Service (ReDoS) during parsing of queries.
<p>Publish Date: 2021-04-28
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23364>CVE-2021-23364</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23364">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23364</a></p>
<p>Release Date: 2021-04-28</p>
<p>Fix Resolution: browserslist - 4.16.5</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"browserslist","packageVersion":"4.14.2","packageFilePaths":["/barista-docs/package.json"],"isTransitiveDependency":true,"dependencyTree":"@docusaurus/core:2.0.0-beta.5;react-dev-utils:11.0.4;browserslist:4.14.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"browserslist - 4.16.5"},{"packageType":"javascript/Node.js","packageName":"browserslist","packageVersion":"4.14.6","packageFilePaths":["/barista-web/package.json"],"isTransitiveDependency":true,"dependencyTree":"@angular-devkit/build-angular:0.1102.13;browserslist:4.14.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"browserslist - 4.16.5"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23364","vulnerabilityDetails":"The package browserslist from 4.0.0 and before 4.16.5 are vulnerable to Regular Expression Denial of Service (ReDoS) during parsing of queries.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23364","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_defect
|
cve medium detected in browserslist tgz browserslist tgz cve medium severity vulnerability vulnerable libraries browserslist tgz browserslist tgz browserslist tgz share target browsers between different front end tools like autoprefixer stylelint and babel env preset library home page a href path to dependency file barista barista docs package json path to vulnerable library barista barista docs node modules react dev utils node modules browserslist package json dependency hierarchy core beta tgz root library react dev utils tgz x browserslist tgz vulnerable library browserslist tgz share target browsers between different front end tools like autoprefixer stylelint and babel env preset library home page a href path to dependency file barista barista web package json path to vulnerable library barista barista web node modules browserslist package json dependency hierarchy build angular tgz root library x browserslist tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package browserslist from and before are vulnerable to regular expression denial of service redos during parsing of queries publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution browserslist isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree docusaurus core beta react dev utils browserslist isminimumfixversionavailable true minimumfixversion browserslist packagetype javascript node js packagename browserslist packageversion packagefilepaths istransitivedependency true dependencytree angular devkit build angular browserslist isminimumfixversionavailable true minimumfixversion browserslist basebranches vulnerabilityidentifier cve vulnerabilitydetails the package browserslist from and before are vulnerable to regular expression denial of service redos during parsing of queries vulnerabilityurl
| 0
|
1,273
| 2,775,119,966
|
IssuesEvent
|
2015-05-04 14:21:03
|
a85/POSTMan-Chrome-Extension
|
https://api.github.com/repos/a85/POSTMan-Chrome-Extension
|
closed
|
Resize history/collections panel
|
Usability
|
Well, that's quiet self-explanatory. I sometimes work with split windows and I can't resize the left panel containing the history/collections of the request which is kinda big with a 1366x800 resolution.
|
True
|
Resize history/collections panel - Well, that's quiet self-explanatory. I sometimes work with split windows and I can't resize the left panel containing the history/collections of the request which is kinda big with a 1366x800 resolution.
|
non_defect
|
resize history collections panel well that s quiet self explanatory i sometimes work with split windows and i can t resize the left panel containing the history collections of the request which is kinda big with a resolution
| 0
|
351,766
| 32,025,571,470
|
IssuesEvent
|
2023-09-22 08:35:50
|
onmetal/cephlet
|
https://api.github.com/repos/onmetal/cephlet
|
closed
|
CreateVolume Method Testing
|
enhancement integration-tests testing
|
# Summary
Test CreateVolume Method
## Description
- Create test cases to validate the functionality of the `CreateVolume` method.
- Utilize the `go-ceph` library to verify that volumes are properly created in the Ceph cluster with different configurations.
- Check for error conditions and boundary cases.
|
2.0
|
CreateVolume Method Testing - # Summary
Test CreateVolume Method
## Description
- Create test cases to validate the functionality of the `CreateVolume` method.
- Utilize the `go-ceph` library to verify that volumes are properly created in the Ceph cluster with different configurations.
- Check for error conditions and boundary cases.
|
non_defect
|
createvolume method testing summary test createvolume method description create test cases to validate the functionality of the createvolume method utilize the go ceph library to verify that volumes are properly created in the ceph cluster with different configurations check for error conditions and boundary cases
| 0
|
228,196
| 17,422,683,589
|
IssuesEvent
|
2021-08-04 04:48:14
|
fangwei123456/spikingjelly
|
https://api.github.com/repos/fangwei123456/spikingjelly
|
closed
|
type object 'CIFAR10DVS' has no attribute 'downloadable'
|
bug documentation good first issue
|
When I execute
```
from spikingjelly.datasets.cifar10_dvs import CIFAR10DVS
from spikingjelly.datasets.dvs128_gesture import DVS128Gesture
print('CIFAR10-DVS downloadable', CIFAR10DVS.downloadable())
print('resource, url, md5/n', CIFAR10DVS.resource_url_md5())
print('DVS128Gesture downloadable', DVS128Gesture.downloadable())
print('resource, url, md5/n', DVS128Gesture.resource_url_md5())
```
throws the following error:
```
E:\SoftWare\Anaconda3\lib\site-packages\torchaudio\extension\extension.py:14: UserWarning: torchaudio C++ extension is not available.
warnings.warn('torchaudio C++ extension is not available.')
E:\SoftWare\Anaconda3\lib\site-packages\torchaudio\backend\utils.py:88: UserWarning: No audio backend is available.
warnings.warn('No audio backend is available.')
Traceback (most recent call last):
File "D:/Project/rpg_e2vid/test.py", line 4, in <module>
print('CIFAR10-DVS downloadable', CIFAR10DVS.downloadable())
AttributeError: type object 'CIFAR10DVS' has no attribute 'downloadable'
```
|
1.0
|
type object 'CIFAR10DVS' has no attribute 'downloadable' - When I execute
```
from spikingjelly.datasets.cifar10_dvs import CIFAR10DVS
from spikingjelly.datasets.dvs128_gesture import DVS128Gesture
print('CIFAR10-DVS downloadable', CIFAR10DVS.downloadable())
print('resource, url, md5/n', CIFAR10DVS.resource_url_md5())
print('DVS128Gesture downloadable', DVS128Gesture.downloadable())
print('resource, url, md5/n', DVS128Gesture.resource_url_md5())
```
throws the following error:
```
E:\SoftWare\Anaconda3\lib\site-packages\torchaudio\extension\extension.py:14: UserWarning: torchaudio C++ extension is not available.
warnings.warn('torchaudio C++ extension is not available.')
E:\SoftWare\Anaconda3\lib\site-packages\torchaudio\backend\utils.py:88: UserWarning: No audio backend is available.
warnings.warn('No audio backend is available.')
Traceback (most recent call last):
File "D:/Project/rpg_e2vid/test.py", line 4, in <module>
print('CIFAR10-DVS downloadable', CIFAR10DVS.downloadable())
AttributeError: type object 'CIFAR10DVS' has no attribute 'downloadable'
```
|
non_defect
|
type object has no attribute downloadable when i execute from spikingjelly datasets dvs import from spikingjelly datasets gesture import print dvs downloadable downloadable print resource url n resource url print downloadable downloadable print resource url n resource url throws the following error e software lib site packages torchaudio extension extension py userwarning torchaudio c extension is not available warnings warn torchaudio c extension is not available e software lib site packages torchaudio backend utils py userwarning no audio backend is available warnings warn no audio backend is available traceback most recent call last file d project rpg test py line in print dvs downloadable downloadable attributeerror type object has no attribute downloadable
| 0
|
815,216
| 30,541,509,073
|
IssuesEvent
|
2023-07-19 21:54:09
|
open-source-uc/planner
|
https://api.github.com/repos/open-source-uc/planner
|
closed
|
[PAN-19] Manejo de roles
|
Medium priority Backend
|
El sistema debe reconocer los distintos roles del sistema (invitado, usuario, moderador y administrador) y debe haber una forma establecida para revisar el rol desde las rutas del back.
<sub>From [SyncLinear.com](https://synclinear.com) | [PAN-19](https://linear.app/planner-uc/issue/PAN-19/manejo-de-roles)</sub>
|
1.0
|
[PAN-19] Manejo de roles - El sistema debe reconocer los distintos roles del sistema (invitado, usuario, moderador y administrador) y debe haber una forma establecida para revisar el rol desde las rutas del back.
<sub>From [SyncLinear.com](https://synclinear.com) | [PAN-19](https://linear.app/planner-uc/issue/PAN-19/manejo-de-roles)</sub>
|
non_defect
|
manejo de roles el sistema debe reconocer los distintos roles del sistema invitado usuario moderador y administrador y debe haber una forma establecida para revisar el rol desde las rutas del back from
| 0
|
13,840
| 2,787,896,197
|
IssuesEvent
|
2015-05-08 09:46:03
|
OpenMS/OpenMS
|
https://api.github.com/repos/OpenMS/OpenMS
|
closed
|
TheoreticalSpectrumGenerator only supports DeltaMass but no NeutralLoss [294]
|
1.8 Bug-Reports defect minor OpenMS - library
|
Submitted by timosachsenberg on 2011-03-16 11:11:26
e.g. Modification with DeltaMass M & NeutralLoss M should only result in a modified precursor mass but the MS2 is also shifted.
|
1.0
|
TheoreticalSpectrumGenerator only supports DeltaMass but no NeutralLoss [294] - Submitted by timosachsenberg on 2011-03-16 11:11:26
e.g. Modification with DeltaMass M & NeutralLoss M should only result in a modified precursor mass but the MS2 is also shifted.
|
defect
|
theoreticalspectrumgenerator only supports deltamass but no neutralloss submitted by timosachsenberg on e g modification with deltamass m neutralloss m should only result in a modified precursor mass but the is also shifted
| 1
|
345,470
| 30,815,813,173
|
IssuesEvent
|
2023-08-01 13:25:00
|
rizinorg/rizin
|
https://api.github.com/repos/rizinorg/rizin
|
closed
|
Incorrect disassembling of tableswitch jvm instruction
|
test-required disassembly java
|
Rizin calculates operand alignment based on jvm->section [/librz/asm/arch/java/jvm.c](https://github.com/rizinorg/rizin/blob/d1fd135df322beb93a88eb038c1da1777c82a780/librz/asm/arch/java/jvm.c#L51), but in [librz/analysis/p/analysis_java.c](https://github.com/rizinorg/rizin/blob/d1fd135df322beb93a88eb038c1da1777c82a780/librz/analysis/p/analysis_java.c#L62C11-L62C11) section is set to first matching section that contains instruction address, which is incorrect, because sometimes there are multiple overlapping sections.
### Work environment
| Questions | Answers
|------------------------------------------------------|--------------------
| OS/arch/bits (mandatory) | macos 13.4.1, x86_64
| File format of the file you reverse (mandatory) | java .class
| Architecture/bits of the file (mandatory) | java
| `rizin -v` full output, **not truncated** (mandatory) | rizin 0.6.0 @ darwin-x86-64
### Steps to reproduce the behavior
- Run `rizin -qc 's 18677; pd 10' SbtParser.class`
### Expected behavior
Rizin outputs something like
```
iload 1
istore 2
iload 2
tableswitch
0 :L1
1 :L2
default :L3
.end tableswitch
```
(This is output from jadx-gui)
### Actual behavior
Rizin incorrectly disassembles tableswitch:
<img width="609" alt="Снимок экрана 2023-07-26 в 16 04 00" src="https://github.com/rizinorg/rizin/assets/42062111/eb287591-8435-4a03-a2aa-a59d4e4ac4dd">
### Additional Logs, screenshots, source code, configuration dump, ...
[SbtParser.class.zip](https://github.com/rizinorg/rizin/files/12170195/SbtParser.class.zip)
|
1.0
|
Incorrect disassembling of tableswitch jvm instruction - Rizin calculates operand alignment based on jvm->section [/librz/asm/arch/java/jvm.c](https://github.com/rizinorg/rizin/blob/d1fd135df322beb93a88eb038c1da1777c82a780/librz/asm/arch/java/jvm.c#L51), but in [librz/analysis/p/analysis_java.c](https://github.com/rizinorg/rizin/blob/d1fd135df322beb93a88eb038c1da1777c82a780/librz/analysis/p/analysis_java.c#L62C11-L62C11) section is set to first matching section that contains instruction address, which is incorrect, because sometimes there are multiple overlapping sections.
### Work environment
| Questions | Answers
|------------------------------------------------------|--------------------
| OS/arch/bits (mandatory) | macos 13.4.1, x86_64
| File format of the file you reverse (mandatory) | java .class
| Architecture/bits of the file (mandatory) | java
| `rizin -v` full output, **not truncated** (mandatory) | rizin 0.6.0 @ darwin-x86-64
### Steps to reproduce the behavior
- Run `rizin -qc 's 18677; pd 10' SbtParser.class`
### Expected behavior
Rizin outputs something like
```
iload 1
istore 2
iload 2
tableswitch
0 :L1
1 :L2
default :L3
.end tableswitch
```
(This is output from jadx-gui)
### Actual behavior
Rizin incorrectly disassembles tableswitch:
<img width="609" alt="Снимок экрана 2023-07-26 в 16 04 00" src="https://github.com/rizinorg/rizin/assets/42062111/eb287591-8435-4a03-a2aa-a59d4e4ac4dd">
### Additional Logs, screenshots, source code, configuration dump, ...
[SbtParser.class.zip](https://github.com/rizinorg/rizin/files/12170195/SbtParser.class.zip)
|
non_defect
|
incorrect disassembling of tableswitch jvm instruction rizin calculates operand alignment based on jvm section but in section is set to first matching section that contains instruction address which is incorrect because sometimes there are multiple overlapping sections work environment questions answers os arch bits mandatory macos file format of the file you reverse mandatory java class architecture bits of the file mandatory java rizin v full output not truncated mandatory rizin darwin steps to reproduce the behavior run rizin qc s pd sbtparser class expected behavior rizin outputs something like iload istore iload tableswitch default end tableswitch this is output from jadx gui actual behavior rizin incorrectly disassembles tableswitch img width alt снимок экрана в src additional logs screenshots source code configuration dump
| 0
|
106,426
| 11,488,184,530
|
IssuesEvent
|
2020-02-11 13:28:05
|
spring-projects/spring-boot
|
https://api.github.com/repos/spring-projects/spring-boot
|
closed
|
Document how to register a blocking health contributor with the reactive registry
|
type: documentation
|
In order to add my custom health indicator I injected `HealthContributorRegistry` and then registered my contributor.
I did this because I saw that `HealthContributorRegistry` was autoconfigured (required for `HealthEndpoint`)
Later on I've found that my custom contributor was not present - I realized that `ReactiveHealthContributorRegistry` was used instead.
I wonder if it would be feasible to update `ReactiveHealthContributorRegistry` when `HealthContributorRegistry` is being updated.
Current setup seems to be unintuitive - there are no clear guidelines what should be configured and how reactive parts interact with traditional ones.
|
1.0
|
Document how to register a blocking health contributor with the reactive registry - In order to add my custom health indicator I injected `HealthContributorRegistry` and then registered my contributor.
I did this because I saw that `HealthContributorRegistry` was autoconfigured (required for `HealthEndpoint`)
Later on I've found that my custom contributor was not present - I realized that `ReactiveHealthContributorRegistry` was used instead.
I wonder if it would be feasible to update `ReactiveHealthContributorRegistry` when `HealthContributorRegistry` is being updated.
Current setup seems to be unintuitive - there are no clear guidelines what should be configured and how reactive parts interact with traditional ones.
|
non_defect
|
document how to register a blocking health contributor with the reactive registry in order to add my custom health indicator i injected healthcontributorregistry and then registered my contributor i did this because i saw that healthcontributorregistry was autoconfigured required for healthendpoint later on i ve found that my custom contributor was not present i realized that reactivehealthcontributorregistry was used instead i wonder if it would be feasible to update reactivehealthcontributorregistry when healthcontributorregistry is being updated current setup seems to be unintuitive there are no clear guidelines what should be configured and how reactive parts interact with traditional ones
| 0
|
57,301
| 15,729,894,962
|
IssuesEvent
|
2021-03-29 15:19:42
|
Kytech/xbox360wirelesschatpad
|
https://api.github.com/repos/Kytech/xbox360wirelesschatpad
|
closed
|
ERROR: Wireless Receiver Not Found.
|
Type-Defect auto-migrated
|
```
win 7 x64
installed filter, vjoy no problem, run the chatpad exe and get the following
output, with all boxes greyed out up top
[04/03/2014 1:00:16 AM] - ERROR: Wireless Receiver Not Found.
[04/03/2014 1:00:16 AM] - WARNING: Failed to Acquire vJoy Gamepad Number 4.
[04/03/2014 1:00:16 AM] - WARNING: Failed to Acquire vJoy Gamepad Number 3.
[04/03/2014 1:00:16 AM] - WARNING: Failed to Acquire vJoy Gamepad Number 2.
any ideas?
thanks for this, it's an exciting idea to solve the problem of "couch internet"!
```
Original issue reported on code.google.com by `aurishal...@gmail.com` on 4 Mar 2014 at 5:06
|
1.0
|
ERROR: Wireless Receiver Not Found. - ```
win 7 x64
installed filter, vjoy no problem, run the chatpad exe and get the following
output, with all boxes greyed out up top
[04/03/2014 1:00:16 AM] - ERROR: Wireless Receiver Not Found.
[04/03/2014 1:00:16 AM] - WARNING: Failed to Acquire vJoy Gamepad Number 4.
[04/03/2014 1:00:16 AM] - WARNING: Failed to Acquire vJoy Gamepad Number 3.
[04/03/2014 1:00:16 AM] - WARNING: Failed to Acquire vJoy Gamepad Number 2.
any ideas?
thanks for this, it's an exciting idea to solve the problem of "couch internet"!
```
Original issue reported on code.google.com by `aurishal...@gmail.com` on 4 Mar 2014 at 5:06
|
defect
|
error wireless receiver not found win installed filter vjoy no problem run the chatpad exe and get the following output with all boxes greyed out up top error wireless receiver not found warning failed to acquire vjoy gamepad number warning failed to acquire vjoy gamepad number warning failed to acquire vjoy gamepad number any ideas thanks for this it s an exciting idea to solve the problem of couch internet original issue reported on code google com by aurishal gmail com on mar at
| 1
|
220,261
| 17,173,510,902
|
IssuesEvent
|
2021-07-15 08:33:19
|
NetApp/harvest
|
https://api.github.com/repos/NetApp/harvest
|
closed
|
NetApp Detail: Volume Dashboard / space used column does not reflect current fill grade as graphic
|
bug status/testme
|
Volume dashboard shows a incomplete grade of filling in the table of volumes
**To Reproduce**
Open the NetApp Detail: Volume
**Expected behavior**
Last column in table should reflect the current fill-state of volume in conjunction to value
**Actual behavior**
Graph shows always the same fill-state

**Possible solution, workaround, fix**
Set an override property for the "space used" column "Standard options > Max" and set the value to 100

|
1.0
|
NetApp Detail: Volume Dashboard / space used column does not reflect current fill grade as graphic - Volume dashboard shows a incomplete grade of filling in the table of volumes
**To Reproduce**
Open the NetApp Detail: Volume
**Expected behavior**
Last column in table should reflect the current fill-state of volume in conjunction to value
**Actual behavior**
Graph shows always the same fill-state

**Possible solution, workaround, fix**
Set an override property for the "space used" column "Standard options > Max" and set the value to 100

|
non_defect
|
netapp detail volume dashboard space used column does not reflect current fill grade as graphic volume dashboard shows a incomplete grade of filling in the table of volumes to reproduce open the netapp detail volume expected behavior last column in table should reflect the current fill state of volume in conjunction to value actual behavior graph shows always the same fill state possible solution workaround fix set an override property for the space used column standard options max and set the value to
| 0
|
74,984
| 25,463,950,621
|
IssuesEvent
|
2022-11-25 00:35:26
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
closed
|
shgo is not correctly passing jac to minimizer
|
defect scipy.optimize Documentation query
|
I am getting the "_minimize_slsqp() got multiple values for argument 'jac' " error for my toy example below. Is this a bug in shgo or am I doing something wrong? Please advise if there is a workaround.
```
from scipy.optimize import shgo
def objective(x):
return 3*x[0]*x[0] + 2*x[0] + 5
def objective_jac(x):
return 6*x[0] + 2
if __name__ == "__main__":
bounds = [(-100, 100)]
result = shgo(objective, bounds, options={'jac':objective_jac})
print(result)
```
|
1.0
|
shgo is not correctly passing jac to minimizer - I am getting the "_minimize_slsqp() got multiple values for argument 'jac' " error for my toy example below. Is this a bug in shgo or am I doing something wrong? Please advise if there is a workaround.
```
from scipy.optimize import shgo
def objective(x):
return 3*x[0]*x[0] + 2*x[0] + 5
def objective_jac(x):
return 6*x[0] + 2
if __name__ == "__main__":
bounds = [(-100, 100)]
result = shgo(objective, bounds, options={'jac':objective_jac})
print(result)
```
|
defect
|
shgo is not correctly passing jac to minimizer i am getting the minimize slsqp got multiple values for argument jac error for my toy example below is this a bug in shgo or am i doing something wrong please advise if there is a workaround from scipy optimize import shgo def objective x return x x x def objective jac x return x if name main bounds result shgo objective bounds options jac objective jac print result
| 1
|
329,218
| 28,208,727,469
|
IssuesEvent
|
2023-04-05 00:58:15
|
googleapis/google-cloud-python
|
https://api.github.com/repos/googleapis/google-cloud-python
|
opened
|
Adopt split repo: python-eventarc-publishing
|
migration:samples:generated migration:workaround:none migration:library:gapic_auto migration:testing:unit migration:issues:none
|
Migrate the split-repo https://github.com/googleapis/python-eventarc-publishing to https://github.com/googleapis/google-cloud-python. To do the actual migration, we need to ensure we can clear any of the following tags in this issue that describe the state of the source repo: `testing:system`, `samples:manual`,`workaround:owlbot`.
|
1.0
|
Adopt split repo: python-eventarc-publishing - Migrate the split-repo https://github.com/googleapis/python-eventarc-publishing to https://github.com/googleapis/google-cloud-python. To do the actual migration, we need to ensure we can clear any of the following tags in this issue that describe the state of the source repo: `testing:system`, `samples:manual`,`workaround:owlbot`.
|
non_defect
|
adopt split repo python eventarc publishing migrate the split repo to to do the actual migration we need to ensure we can clear any of the following tags in this issue that describe the state of the source repo testing system samples manual workaround owlbot
| 0
|
12,756
| 2,715,987,873
|
IssuesEvent
|
2015-04-10 16:18:01
|
codenameone/CodenameOne
|
https://api.github.com/repos/codenameone/CodenameOne
|
closed
|
BlackBerry 5 BrowserComponent disappears on transition
|
Priority-High Type-Defect
|
Original [issue 491](https://code.google.com/p/codenameone/issues/detail?id=491) created by codenameone on 2013-01-17T20:58:24.000Z:
The BB 5 BrowserComponent disappears when transitioning into a form.
E.g. the following will display "Hello World" briefly, but then disappear.
Form hi = new Form("Hi World");
BrowserComponent b = new BrowserComponent();
b.setPage('<html><body>Hello World</body></html>', 'file:///');
hi.setLayout(new BorderPane());
hi.addComponent(BorderLayout.CENTER, b);
hi.show();
However, if I remove and re-add the browser component inside the onShowCompleted() event of the form, it shows up properly.
I'm guessing this has something to do with the PeerComponent code in the BlackBerry implementation. I'm looking through the createNativePeer() method and looking at the implementation of peer components, but, being new to BB, it will be a bit of a hill to climb to nail this down.
I'm keen to hunt this one down, so any pointers or suggestions on places to look, much appreciated.
|
1.0
|
BlackBerry 5 BrowserComponent disappears on transition - Original [issue 491](https://code.google.com/p/codenameone/issues/detail?id=491) created by codenameone on 2013-01-17T20:58:24.000Z:
The BB 5 BrowserComponent disappears when transitioning into a form.
E.g. the following will display "Hello World" briefly, but then disappear.
Form hi = new Form("Hi World");
BrowserComponent b = new BrowserComponent();
b.setPage('<html><body>Hello World</body></html>', 'file:///');
hi.setLayout(new BorderPane());
hi.addComponent(BorderLayout.CENTER, b);
hi.show();
However, if I remove and re-add the browser component inside the onShowCompleted() event of the form, it shows up properly.
I'm guessing this has something to do with the PeerComponent code in the BlackBerry implementation. I'm looking through the createNativePeer() method and looking at the implementation of peer components, but, being new to BB, it will be a bit of a hill to climb to nail this down.
I'm keen to hunt this one down, so any pointers or suggestions on places to look, much appreciated.
|
defect
|
blackberry browsercomponent disappears on transition original created by codenameone on the bb browsercomponent disappears when transitioning into a form e g the following will display quot hello world quot briefly but then disappear form hi new form quot hi world quot browsercomponent b new browsercomponent b setpage lt html gt lt body gt hello world lt body gt lt html gt file hi setlayout new borderpane hi addcomponent borderlayout center b hi show however if i remove and re add the browser component inside the onshowcompleted event of the form it shows up properly i m guessing this has something to do with the peercomponent code in the blackberry implementation i m looking through the createnativepeer method and looking at the implementation of peer components but being new to bb it will be a bit of a hill to climb to nail this down i m keen to hunt this one down so any pointers or suggestions on places to look much appreciated
| 1
|
185,877
| 14,384,142,981
|
IssuesEvent
|
2020-12-02 10:04:02
|
WoWManiaUK/Redemption
|
https://api.github.com/repos/WoWManiaUK/Redemption
|
closed
|
[Raid] Felmyst's Demonic Vapors
|
Fixed on PTR - Tester Confirmed
|
**Links:** https://wowwiki.fandom.com/wiki/Felmyst
https://youtu.be/pcv5LlawB1A?t=862
https://www.wow-mania.com/armory/?search=demonic+vapor#spells
https://www.wow-mania.com/armory/?npc=25268
**What is Happening:** The fight function mostly fine, but 1 major issue persists. When air phase begins, Felmyst will use her first aerial-ability roughly 5 seconds after leaving melee range; the [Demonic Vapor](https://www.wow-mania.com/armory/?search=demonic+vapor#spells). However, it is currently targetting every single raid member with a nature damage beam that leaves behind a nature damage trail.
The stacking trails of damage quickly overwhelm the entire room and the more players are present, the more chaotic it is.
**What Should happen:**
> She will target a random player for a couple seconds, then cast it under them. The spell follows that player for about ten seconds, leaving the DoT trail and summoning skeletons soon after.
1. The spell is supposed to target only a single raid member and chase them for a short duration (duration and damage are correct for 1 person), before choosing a new target and repeating the process (Behaves somewhat like the spikes in the Anub'Arak encounter). After 2 Demonic Vapors, she starts her fly-bys that spread the Fog of Corruption (fixed previously).
Behavior is clearly seen from ~14:20 onward in [this video](https://youtu.be/pcv5LlawB1A?t=862).
2. The trail is supposed to spawn ~10 [Unyielding Dead](https://www.wow-mania.com/armory/?npc=25268) as it is forming (14:37 in the video), and additional skeletons if players touch the trail afterwards (similar to Yogg-Saron's clouds). There used to be a fairly bad issue with the number of skeletons spawned by this ability (most likely due to the spell targeting all raid members instead of only 1). *Less important than 1
|
1.0
|
[Raid] Felmyst's Demonic Vapors - **Links:** https://wowwiki.fandom.com/wiki/Felmyst
https://youtu.be/pcv5LlawB1A?t=862
https://www.wow-mania.com/armory/?search=demonic+vapor#spells
https://www.wow-mania.com/armory/?npc=25268
**What is Happening:** The fight function mostly fine, but 1 major issue persists. When air phase begins, Felmyst will use her first aerial-ability roughly 5 seconds after leaving melee range; the [Demonic Vapor](https://www.wow-mania.com/armory/?search=demonic+vapor#spells). However, it is currently targetting every single raid member with a nature damage beam that leaves behind a nature damage trail.
The stacking trails of damage quickly overwhelm the entire room and the more players are present, the more chaotic it is.
**What Should happen:**
> She will target a random player for a couple seconds, then cast it under them. The spell follows that player for about ten seconds, leaving the DoT trail and summoning skeletons soon after.
1. The spell is supposed to target only a single raid member and chase them for a short duration (duration and damage are correct for 1 person), before choosing a new target and repeating the process (Behaves somewhat like the spikes in the Anub'Arak encounter). After 2 Demonic Vapors, she starts her fly-bys that spread the Fog of Corruption (fixed previously).
Behavior is clearly seen from ~14:20 onward in [this video](https://youtu.be/pcv5LlawB1A?t=862).
2. The trail is supposed to spawn ~10 [Unyielding Dead](https://www.wow-mania.com/armory/?npc=25268) as it is forming (14:37 in the video), and additional skeletons if players touch the trail afterwards (similar to Yogg-Saron's clouds). There used to be a fairly bad issue with the number of skeletons spawned by this ability (most likely due to the spell targeting all raid members instead of only 1). *Less important than 1
|
non_defect
|
felmyst s demonic vapors links what is happening the fight function mostly fine but major issue persists when air phase begins felmyst will use her first aerial ability roughly seconds after leaving melee range the however it is currently targetting every single raid member with a nature damage beam that leaves behind a nature damage trail the stacking trails of damage quickly overwhelm the entire room and the more players are present the more chaotic it is what should happen she will target a random player for a couple seconds then cast it under them the spell follows that player for about ten seconds leaving the dot trail and summoning skeletons soon after the spell is supposed to target only a single raid member and chase them for a short duration duration and damage are correct for person before choosing a new target and repeating the process behaves somewhat like the spikes in the anub arak encounter after demonic vapors she starts her fly bys that spread the fog of corruption fixed previously behavior is clearly seen from onward in the trail is supposed to spawn as it is forming in the video and additional skeletons if players touch the trail afterwards similar to yogg saron s clouds there used to be a fairly bad issue with the number of skeletons spawned by this ability most likely due to the spell targeting all raid members instead of only less important than
| 0
|
48,534
| 13,113,835,143
|
IssuesEvent
|
2020-08-05 06:29:51
|
naev/naev
|
https://api.github.com/repos/naev/naev
|
closed
|
When a faction is not <known>, its ships, systems, and assets are still clearly marked
|
Priority-High Type-Defect
|
When a faction lacks the `<known>` tag, the faction's presence is listed as "unknown". But the name and/or logo of the faction appears when targeting a ship or planet of that faction, or selecting a system controlled by the faction on the map.
|
1.0
|
When a faction is not <known>, its ships, systems, and assets are still clearly marked - When a faction lacks the `<known>` tag, the faction's presence is listed as "unknown". But the name and/or logo of the faction appears when targeting a ship or planet of that faction, or selecting a system controlled by the faction on the map.
|
defect
|
when a faction is not its ships systems and assets are still clearly marked when a faction lacks the tag the faction s presence is listed as unknown but the name and or logo of the faction appears when targeting a ship or planet of that faction or selecting a system controlled by the faction on the map
| 1
|
58,101
| 16,342,448,659
|
IssuesEvent
|
2021-05-13 00:19:41
|
darshan-hpc/darshan
|
https://api.github.com/repos/darshan-hpc/darshan
|
closed
|
incorrect reporting of bytes read and written by HMMER application
|
defect wrapper libraries
|
In GitLab by @shanedsnyder on Sep 24, 2015, 16:27
See http://hmmer.janelia.org/. Problem reported by Chris Daley. Example execution reads gigabytes of data and writes 19 MiB of data, but Darshan reports 8 bytes read and roughly 1 MiB written. Application is using fread and fwrite for I/O, and everything shows up as expected in strace as read and write operations.
This indicates that there may be a bug in how Darshan records the sizes of fread and fwrite operations.
Removing from 2.2.9 milestone for now, but we should still try to produce a prototype to test with hmmer.
|
1.0
|
incorrect reporting of bytes read and written by HMMER application - In GitLab by @shanedsnyder on Sep 24, 2015, 16:27
See http://hmmer.janelia.org/. Problem reported by Chris Daley. Example execution reads gigabytes of data and writes 19 MiB of data, but Darshan reports 8 bytes read and roughly 1 MiB written. Application is using fread and fwrite for I/O, and everything shows up as expected in strace as read and write operations.
This indicates that there may be a bug in how Darshan records the sizes of fread and fwrite operations.
Removing from 2.2.9 milestone for now, but we should still try to produce a prototype to test with hmmer.
|
defect
|
incorrect reporting of bytes read and written by hmmer application in gitlab by shanedsnyder on sep see problem reported by chris daley example execution reads gigabytes of data and writes mib of data but darshan reports bytes read and roughly mib written application is using fread and fwrite for i o and everything shows up as expected in strace as read and write operations this indicates that there may be a bug in how darshan records the sizes of fread and fwrite operations removing from milestone for now but we should still try to produce a prototype to test with hmmer
| 1
|
31,214
| 6,447,816,237
|
IssuesEvent
|
2017-08-14 09:13:17
|
PowerDNS/pdns
|
https://api.github.com/repos/PowerDNS/pdns
|
closed
|
IPv6 addresses improperly encoded in carbon metrics
|
defect dnsdist
|
- Program: dnsdist
- Issue type: Bug report
### Short description
When logging to carbon, IPv6 addresses are enclosed in square brackets ([]) within metric names. Square brackets have special meaning to many storage engines and frontends.
### Environment
- Operating system: Ubuntu 16.04
- Software version: dnsdist 1.1.0
- Software source: repo
### Steps to reproduce
1. Configure dnsdist with an IPv6 interface and logging to a carbon server
2. Observe the metric names within your carbon instance
3. IPv4 addresses are encoded as `10_0_0_1:53_tcp`, IPv6 addresses are encoded as `[2001:db8::1]:53_tcp`
### Expected behaviour
A recommended encoding would be `2001:db8::1_53_tcp`. It is really a toss-up on the colons becoming ambiguous or breaking with the v4 encoding.
|
1.0
|
IPv6 addresses improperly encoded in carbon metrics - - Program: dnsdist
- Issue type: Bug report
### Short description
When logging to carbon, IPv6 addresses are enclosed in square brackets ([]) within metric names. Square brackets have special meaning to many storage engines and frontends.
### Environment
- Operating system: Ubuntu 16.04
- Software version: dnsdist 1.1.0
- Software source: repo
### Steps to reproduce
1. Configure dnsdist with an IPv6 interface and logging to a carbon server
2. Observe the metric names within your carbon instance
3. IPv4 addresses are encoded as `10_0_0_1:53_tcp`, IPv6 addresses are encoded as `[2001:db8::1]:53_tcp`
### Expected behaviour
A recommended encoding would be `2001:db8::1_53_tcp`. It is really a toss-up on the colons becoming ambiguous or breaking with the v4 encoding.
|
defect
|
addresses improperly encoded in carbon metrics program dnsdist issue type bug report short description when logging to carbon addresses are enclosed in square brackets within metric names square brackets have special meaning to many storage engines and frontends environment operating system ubuntu software version dnsdist software source repo steps to reproduce configure dnsdist with an interface and logging to a carbon server observe the metric names within your carbon instance addresses are encoded as tcp addresses are encoded as tcp expected behaviour a recommended encoding would be tcp it is really a toss up on the colons becoming ambiguous or breaking with the encoding
| 1
|
79,750
| 28,780,723,527
|
IssuesEvent
|
2023-05-02 00:14:26
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
opened
|
transactionCoroutine maybe? does not wait for transaction to commit before returning
|
T: Defect
|
### Expected behavior
Wait for transaction to commit before returning (async) control back to to consumer.
### Actual behavior
Control is returned back to consumer before transaction has committed.
### Steps to reproduce the problem
I've been using the transactionCoroutine wrapper to manage a transaction against a postgres database connection pool with r2dbc and r2dbc pool.
I have some integration tests that run (in a transaction) some commands before committing. Something along the lines of:
```
val config = DefaultConfiguration()
.set(pool)
.set(JDBCUtils.dialect(pool))
val dsl = DSL.using(config)
val updatedCount = dsl.transactionCoroutine { config ->
val transactionDsl = config.dsl()
... some stuff
transactionalDsl.update(...).awaitLast()
}
verifyTransactionResults(dsl)
```
Intermittently, the verification of the results will fail in a way that indicates that the transaction hadn't committed. If I turn OFF connection pooling I find that the intermittency goes away and the tests pass as expected. If I delay / sleep for a few MS before checking the intermittency goes away.
An alternative implementation of transactionCoroutine using `.awaitLast()` rather than `.awaitFirstOrNull()` seems to solve the problem.
In a tight loop, I can reproduce this every time. However, when trying against h2 I was unable to reproduce (though could be a function of removing the network latency). With H2, the transactionCoroutine implementation finished significantly faster.
<details>
<summary>full repro code here</summary>
dependencies:
```
implementation("io.r2dbc:r2dbc-postgresql:0.8.13.RELEASE")
implementation("io.r2dbc:r2dbc-pool:1.0.0.RELEASE")
implementation("org.jetbrains.kotlinx:kotlinx-coroutines-core-jvm:1.6.4")
implementation("org.jetbrains.kotlinx:kotlinx-coroutines-jdk8:1.6.4")
implementation("org.jooq:jooq-kotlin:3.18.0")
implementation("org.jooq:jooq:3.18.0")
implementation("org.jooq:jooq-kotlin-coroutines:3.18.0")
implementation("org.jooq:jooq-postgres-extensions:3.18.0")
testRuntimeOnly("org.junit.jupiter:junit-jupiter-engine:5.9.2")
testImplementation("org.junit.jupiter:junit-jupiter-api:5.9.2")
```
test code:
```
import io.r2dbc.pool.ConnectionPool
import io.r2dbc.pool.ConnectionPoolConfiguration
import io.r2dbc.spi.ConnectionFactories
import io.r2dbc.spi.ConnectionFactoryOptions
import kotlinx.coroutines.reactive.awaitFirstOrNull
import kotlinx.coroutines.reactive.awaitLast
import kotlinx.coroutines.reactor.mono
import kotlinx.coroutines.runBlocking
import org.jooq.DSLContext
import org.jooq.impl.DSL
import org.jooq.impl.DSL.count
import org.jooq.impl.DSL.field
import org.jooq.impl.DSL.inline
import org.jooq.impl.DSL.table
import org.jooq.impl.DefaultConfiguration
import org.jooq.impl.SQLDataType.INTEGER
import org.jooq.impl.SQLDataType.VARCHAR
import org.jooq.kotlin.coroutines.transactionCoroutine
import org.jooq.tools.jdbc.JDBCUtils
import org.junit.jupiter.api.Assertions.assertEquals
import org.junit.jupiter.api.Assertions.assertNotNull
import org.junit.jupiter.api.Test
import java.time.Duration
import java.util.UUID
class ConnectionWrapperTest {
private val connectionFactoryOptions = ConnectionFactoryOptions.builder()
.option(ConnectionFactoryOptions.DRIVER, "postgresql")
.option(ConnectionFactoryOptions.HOST, "localhost")
.option(ConnectionFactoryOptions.PORT, 5432)
.option(ConnectionFactoryOptions.USER, "user")
.option(ConnectionFactoryOptions.DATABASE, "test")
.build()
private val tableName = "testing12345"
private val idCol = "id"
private val valueCol = "val"
private suspend fun validate(wrapper: ConnectionWrapper, tableName: String, uuid: String) {
val returned = wrapper.dbQuery {
select(field(idCol), field(valueCol))
.from(table(tableName))
.where(field(valueCol).eq(inline(uuid)))
.awaitFirstOrNull()
}
assertNotNull(returned)
}
private suspend fun updateValue(dsl: DSLContext, tableName: String, uuid: String): Int? {
return dsl.update(table(tableName))
.set(field(valueCol), uuid)
.awaitLast()
}
@Test
fun `transaction using transactionCoroutine`() {
val connectionWrapper = ConnectionWrapper(connectionFactoryOptions)
runBlocking {
setup(connectionWrapper, tableName)
}
repeat(1_000) {
runBlocking {
val uuid = UUID.randomUUID().toString()
val updated = connectionWrapper.dbQuery {
transactionCoroutine { config ->
val transactionDsl = config.dsl()
updateValue(transactionDsl, tableName, uuid)
}
}
assertEquals(1, updated)
validate(connectionWrapper, tableName, uuid)
}
}
connectionWrapper.shutdown()
}
@Test
fun `transaction using alternative`() = runBlocking {
val connectionWrapper = ConnectionWrapper(connectionFactoryOptions)
runBlocking {
setup(connectionWrapper, tableName)
}
repeat(1_000) {
val uuid = UUID.randomUUID().toString()
val updated = connectionWrapper.dbQuery {
transactionPublisher { config ->
val transactionDsl = config.dsl()
mono {
updateValue(transactionDsl, tableName, uuid)
}
// use await last instead of await first
}.awaitLast()
}
assertEquals(1, updated)
validate(connectionWrapper, tableName, uuid)
}
connectionWrapper.shutdown()
}
private suspend fun setup(wrapper: ConnectionWrapper, tableName: String) {
wrapper.dbQuery {
createTableIfNotExists(tableName)
.column(idCol, INTEGER)
.column(valueCol, VARCHAR(255))
.awaitFirstOrNull()
val field = count().`as`("COUNT")
val record = select(field)
.from(table(tableName))
.awaitFirstOrNull()
if (record == null || record[field] == 0) {
insertInto(table(tableName))
.columns(field(idCol), field(valueCol))
.values(1, "old value")
.awaitFirstOrNull()
}
}
}
}
class ConnectionWrapper(options: ConnectionFactoryOptions) {
private val factory = ConnectionFactories.get(options)
private val configuration = ConnectionPoolConfiguration.builder(factory)
.maxIdleTime(Duration.ofMillis(60000)) // 1 min
.maxSize(10)
.initialSize(10)
.minIdle(10)
.build()
private val pool = ConnectionPool(configuration)
suspend fun <T> dbQuery(block: suspend DSLContext.() -> T): T {
val config = DefaultConfiguration()
.set(pool)
.set(JDBCUtils.dialect(pool))
val context = DSL.using(config)
return block(context)
}
fun shutdown() {
pool.dispose()
}
}
```
</details>
### jOOQ Version
3.18.0
### Database product and version
PostgreSQL 14.7 (Homebrew) on aarch64-apple-darwin22.1.0
### Java Version
17
### OS Version
Mac
### JDBC driver name and version (include name if unofficial driver)
io.r2dbc:r2dbc-postgres:0.8.13.RELEASE
|
1.0
|
transactionCoroutine maybe? does not wait for transaction to commit before returning - ### Expected behavior
Wait for transaction to commit before returning (async) control back to to consumer.
### Actual behavior
Control is returned back to consumer before transaction has committed.
### Steps to reproduce the problem
I've been using the transactionCoroutine wrapper to manage a transaction against a postgres database connection pool with r2dbc and r2dbc pool.
I have some integration tests that run (in a transaction) some commands before committing. Something along the lines of:
```
val config = DefaultConfiguration()
.set(pool)
.set(JDBCUtils.dialect(pool))
val dsl = DSL.using(config)
val updatedCount = dsl.transactionCoroutine { config ->
val transactionDsl = config.dsl()
... some stuff
transactionalDsl.update(...).awaitLast()
}
verifyTransactionResults(dsl)
```
Intermittently, the verification of the results will fail in a way that indicates that the transaction hadn't committed. If I turn OFF connection pooling I find that the intermittency goes away and the tests pass as expected. If I delay / sleep for a few MS before checking the intermittency goes away.
An alternative implementation of transactionCoroutine using `.awaitLast()` rather than `.awaitFirstOrNull()` seems to solve the problem.
In a tight loop, I can reproduce this every time. However, when trying against h2 I was unable to reproduce (though could be a function of removing the network latency). With H2, the transactionCoroutine implementation finished significantly faster.
<details>
<summary>full repro code here</summary>
dependencies:
```
implementation("io.r2dbc:r2dbc-postgresql:0.8.13.RELEASE")
implementation("io.r2dbc:r2dbc-pool:1.0.0.RELEASE")
implementation("org.jetbrains.kotlinx:kotlinx-coroutines-core-jvm:1.6.4")
implementation("org.jetbrains.kotlinx:kotlinx-coroutines-jdk8:1.6.4")
implementation("org.jooq:jooq-kotlin:3.18.0")
implementation("org.jooq:jooq:3.18.0")
implementation("org.jooq:jooq-kotlin-coroutines:3.18.0")
implementation("org.jooq:jooq-postgres-extensions:3.18.0")
testRuntimeOnly("org.junit.jupiter:junit-jupiter-engine:5.9.2")
testImplementation("org.junit.jupiter:junit-jupiter-api:5.9.2")
```
test code:
```
import io.r2dbc.pool.ConnectionPool
import io.r2dbc.pool.ConnectionPoolConfiguration
import io.r2dbc.spi.ConnectionFactories
import io.r2dbc.spi.ConnectionFactoryOptions
import kotlinx.coroutines.reactive.awaitFirstOrNull
import kotlinx.coroutines.reactive.awaitLast
import kotlinx.coroutines.reactor.mono
import kotlinx.coroutines.runBlocking
import org.jooq.DSLContext
import org.jooq.impl.DSL
import org.jooq.impl.DSL.count
import org.jooq.impl.DSL.field
import org.jooq.impl.DSL.inline
import org.jooq.impl.DSL.table
import org.jooq.impl.DefaultConfiguration
import org.jooq.impl.SQLDataType.INTEGER
import org.jooq.impl.SQLDataType.VARCHAR
import org.jooq.kotlin.coroutines.transactionCoroutine
import org.jooq.tools.jdbc.JDBCUtils
import org.junit.jupiter.api.Assertions.assertEquals
import org.junit.jupiter.api.Assertions.assertNotNull
import org.junit.jupiter.api.Test
import java.time.Duration
import java.util.UUID
class ConnectionWrapperTest {
private val connectionFactoryOptions = ConnectionFactoryOptions.builder()
.option(ConnectionFactoryOptions.DRIVER, "postgresql")
.option(ConnectionFactoryOptions.HOST, "localhost")
.option(ConnectionFactoryOptions.PORT, 5432)
.option(ConnectionFactoryOptions.USER, "user")
.option(ConnectionFactoryOptions.DATABASE, "test")
.build()
private val tableName = "testing12345"
private val idCol = "id"
private val valueCol = "val"
private suspend fun validate(wrapper: ConnectionWrapper, tableName: String, uuid: String) {
val returned = wrapper.dbQuery {
select(field(idCol), field(valueCol))
.from(table(tableName))
.where(field(valueCol).eq(inline(uuid)))
.awaitFirstOrNull()
}
assertNotNull(returned)
}
private suspend fun updateValue(dsl: DSLContext, tableName: String, uuid: String): Int? {
return dsl.update(table(tableName))
.set(field(valueCol), uuid)
.awaitLast()
}
@Test
fun `transaction using transactionCoroutine`() {
val connectionWrapper = ConnectionWrapper(connectionFactoryOptions)
runBlocking {
setup(connectionWrapper, tableName)
}
repeat(1_000) {
runBlocking {
val uuid = UUID.randomUUID().toString()
val updated = connectionWrapper.dbQuery {
transactionCoroutine { config ->
val transactionDsl = config.dsl()
updateValue(transactionDsl, tableName, uuid)
}
}
assertEquals(1, updated)
validate(connectionWrapper, tableName, uuid)
}
}
connectionWrapper.shutdown()
}
@Test
fun `transaction using alternative`() = runBlocking {
val connectionWrapper = ConnectionWrapper(connectionFactoryOptions)
runBlocking {
setup(connectionWrapper, tableName)
}
repeat(1_000) {
val uuid = UUID.randomUUID().toString()
val updated = connectionWrapper.dbQuery {
transactionPublisher { config ->
val transactionDsl = config.dsl()
mono {
updateValue(transactionDsl, tableName, uuid)
}
// use await last instead of await first
}.awaitLast()
}
assertEquals(1, updated)
validate(connectionWrapper, tableName, uuid)
}
connectionWrapper.shutdown()
}
private suspend fun setup(wrapper: ConnectionWrapper, tableName: String) {
wrapper.dbQuery {
createTableIfNotExists(tableName)
.column(idCol, INTEGER)
.column(valueCol, VARCHAR(255))
.awaitFirstOrNull()
val field = count().`as`("COUNT")
val record = select(field)
.from(table(tableName))
.awaitFirstOrNull()
if (record == null || record[field] == 0) {
insertInto(table(tableName))
.columns(field(idCol), field(valueCol))
.values(1, "old value")
.awaitFirstOrNull()
}
}
}
}
class ConnectionWrapper(options: ConnectionFactoryOptions) {
private val factory = ConnectionFactories.get(options)
private val configuration = ConnectionPoolConfiguration.builder(factory)
.maxIdleTime(Duration.ofMillis(60000)) // 1 min
.maxSize(10)
.initialSize(10)
.minIdle(10)
.build()
private val pool = ConnectionPool(configuration)
suspend fun <T> dbQuery(block: suspend DSLContext.() -> T): T {
val config = DefaultConfiguration()
.set(pool)
.set(JDBCUtils.dialect(pool))
val context = DSL.using(config)
return block(context)
}
fun shutdown() {
pool.dispose()
}
}
```
</details>
### jOOQ Version
3.18.0
### Database product and version
PostgreSQL 14.7 (Homebrew) on aarch64-apple-darwin22.1.0
### Java Version
17
### OS Version
Mac
### JDBC driver name and version (include name if unofficial driver)
io.r2dbc:r2dbc-postgres:0.8.13.RELEASE
|
defect
|
transactioncoroutine maybe does not wait for transaction to commit before returning expected behavior wait for transaction to commit before returning async control back to to consumer actual behavior control is returned back to consumer before transaction has committed steps to reproduce the problem i ve been using the transactioncoroutine wrapper to manage a transaction against a postgres database connection pool with and pool i have some integration tests that run in a transaction some commands before committing something along the lines of val config defaultconfiguration set pool set jdbcutils dialect pool val dsl dsl using config val updatedcount dsl transactioncoroutine config val transactiondsl config dsl some stuff transactionaldsl update awaitlast verifytransactionresults dsl intermittently the verification of the results will fail in a way that indicates that the transaction hadn t committed if i turn off connection pooling i find that the intermittency goes away and the tests pass as expected if i delay sleep for a few ms before checking the intermittency goes away an alternative implementation of transactioncoroutine using awaitlast rather than awaitfirstornull seems to solve the problem in a tight loop i can reproduce this every time however when trying against i was unable to reproduce though could be a function of removing the network latency with the transactioncoroutine implementation finished significantly faster full repro code here dependencies implementation io postgresql release implementation io pool release implementation org jetbrains kotlinx kotlinx coroutines core jvm implementation org jetbrains kotlinx kotlinx coroutines implementation org jooq jooq kotlin implementation org jooq jooq implementation org jooq jooq kotlin coroutines implementation org jooq jooq postgres extensions testruntimeonly org junit jupiter junit jupiter engine testimplementation org junit jupiter junit jupiter api test code import io pool connectionpool import io pool connectionpoolconfiguration import io spi connectionfactories import io spi connectionfactoryoptions import kotlinx coroutines reactive awaitfirstornull import kotlinx coroutines reactive awaitlast import kotlinx coroutines reactor mono import kotlinx coroutines runblocking import org jooq dslcontext import org jooq impl dsl import org jooq impl dsl count import org jooq impl dsl field import org jooq impl dsl inline import org jooq impl dsl table import org jooq impl defaultconfiguration import org jooq impl sqldatatype integer import org jooq impl sqldatatype varchar import org jooq kotlin coroutines transactioncoroutine import org jooq tools jdbc jdbcutils import org junit jupiter api assertions assertequals import org junit jupiter api assertions assertnotnull import org junit jupiter api test import java time duration import java util uuid class connectionwrappertest private val connectionfactoryoptions connectionfactoryoptions builder option connectionfactoryoptions driver postgresql option connectionfactoryoptions host localhost option connectionfactoryoptions port option connectionfactoryoptions user user option connectionfactoryoptions database test build private val tablename private val idcol id private val valuecol val private suspend fun validate wrapper connectionwrapper tablename string uuid string val returned wrapper dbquery select field idcol field valuecol from table tablename where field valuecol eq inline uuid awaitfirstornull assertnotnull returned private suspend fun updatevalue dsl dslcontext tablename string uuid string int return dsl update table tablename set field valuecol uuid awaitlast test fun transaction using transactioncoroutine val connectionwrapper connectionwrapper connectionfactoryoptions runblocking setup connectionwrapper tablename repeat runblocking val uuid uuid randomuuid tostring val updated connectionwrapper dbquery transactioncoroutine config val transactiondsl config dsl updatevalue transactiondsl tablename uuid assertequals updated validate connectionwrapper tablename uuid connectionwrapper shutdown test fun transaction using alternative runblocking val connectionwrapper connectionwrapper connectionfactoryoptions runblocking setup connectionwrapper tablename repeat val uuid uuid randomuuid tostring val updated connectionwrapper dbquery transactionpublisher config val transactiondsl config dsl mono updatevalue transactiondsl tablename uuid use await last instead of await first awaitlast assertequals updated validate connectionwrapper tablename uuid connectionwrapper shutdown private suspend fun setup wrapper connectionwrapper tablename string wrapper dbquery createtableifnotexists tablename column idcol integer column valuecol varchar awaitfirstornull val field count as count val record select field from table tablename awaitfirstornull if record null record insertinto table tablename columns field idcol field valuecol values old value awaitfirstornull class connectionwrapper options connectionfactoryoptions private val factory connectionfactories get options private val configuration connectionpoolconfiguration builder factory maxidletime duration ofmillis min maxsize initialsize minidle build private val pool connectionpool configuration suspend fun dbquery block suspend dslcontext t t val config defaultconfiguration set pool set jdbcutils dialect pool val context dsl using config return block context fun shutdown pool dispose jooq version database product and version postgresql homebrew on apple java version os version mac jdbc driver name and version include name if unofficial driver io postgres release
| 1
|
10,478
| 27,020,930,225
|
IssuesEvent
|
2023-02-11 02:00:39
|
facebook/react-native
|
https://api.github.com/repos/facebook/react-native
|
closed
|
Nested TextInput inside Text looses focus on key stroke
|
Stale Component: TextInput Priority: Low Platform: All Tech: React Native Core Type: Old Architecture Impact: Bug
|
### Description
Hello,
I have a nested Text setup with Text and Text Inputs like this:
```
<Text>
<Text/>
<View>
<TextInput/>
</View>
</Text>
```
From the styling and text wrapping perspective it works great, but unfortunately the Text Input looses focus on every keystroke.
Is this a common thing or am I missing something here?
### Version
0.64.3
### Output of `npx react-native info`
System:
OS: macOS 12.3.1
CPU: (8) arm64 Apple M1
Memory: 140.03 MB / 8.00 GB
Shell: 5.8 - /bin/zsh
Binaries:
Node: 17.8.0 - /opt/homebrew/bin/node
Yarn: Not Found
npm: 8.5.5 - /opt/homebrew/bin/npm
Watchman: Not Found
Managers:
CocoaPods: 1.11.3 - /usr/local/bin/pod
SDKs:
iOS SDK:
Platforms: DriverKit 21.4, iOS 15.4, macOS 12.3, tvOS 15.4, watchOS 8.5
Android SDK: Not Found
IDEs:
Android Studio: Not Found
Xcode: 13.3/13E113 - /usr/bin/xcodebuild
Languages:
Java: Not Found
npmPackages:
@react-native-community/cli: Not Found
react: ^17.0.1 => 17.0.2
react-native: 0.64.3 => 0.64.3
react-native-macos: Not Found
npmGlobalPackages:
*react-native*: Not Found
### Steps to reproduce
Setup a simple React Native App with a Text Input inside a Text component
### Snack, code example, screenshot, or link to a repository
_No response_
|
1.0
|
Nested TextInput inside Text looses focus on key stroke - ### Description
Hello,
I have a nested Text setup with Text and Text Inputs like this:
```
<Text>
<Text/>
<View>
<TextInput/>
</View>
</Text>
```
From the styling and text wrapping perspective it works great, but unfortunately the Text Input looses focus on every keystroke.
Is this a common thing or am I missing something here?
### Version
0.64.3
### Output of `npx react-native info`
System:
OS: macOS 12.3.1
CPU: (8) arm64 Apple M1
Memory: 140.03 MB / 8.00 GB
Shell: 5.8 - /bin/zsh
Binaries:
Node: 17.8.0 - /opt/homebrew/bin/node
Yarn: Not Found
npm: 8.5.5 - /opt/homebrew/bin/npm
Watchman: Not Found
Managers:
CocoaPods: 1.11.3 - /usr/local/bin/pod
SDKs:
iOS SDK:
Platforms: DriverKit 21.4, iOS 15.4, macOS 12.3, tvOS 15.4, watchOS 8.5
Android SDK: Not Found
IDEs:
Android Studio: Not Found
Xcode: 13.3/13E113 - /usr/bin/xcodebuild
Languages:
Java: Not Found
npmPackages:
@react-native-community/cli: Not Found
react: ^17.0.1 => 17.0.2
react-native: 0.64.3 => 0.64.3
react-native-macos: Not Found
npmGlobalPackages:
*react-native*: Not Found
### Steps to reproduce
Setup a simple React Native App with a Text Input inside a Text component
### Snack, code example, screenshot, or link to a repository
_No response_
|
non_defect
|
nested textinput inside text looses focus on key stroke description hello i have a nested text setup with text and text inputs like this from the styling and text wrapping perspective it works great but unfortunately the text input looses focus on every keystroke is this a common thing or am i missing something here version output of npx react native info system os macos cpu apple memory mb gb shell bin zsh binaries node opt homebrew bin node yarn not found npm opt homebrew bin npm watchman not found managers cocoapods usr local bin pod sdks ios sdk platforms driverkit ios macos tvos watchos android sdk not found ides android studio not found xcode usr bin xcodebuild languages java not found npmpackages react native community cli not found react react native react native macos not found npmglobalpackages react native not found steps to reproduce setup a simple react native app with a text input inside a text component snack code example screenshot or link to a repository no response
| 0
|
34,503
| 7,452,434,087
|
IssuesEvent
|
2018-03-29 08:22:51
|
kerdokullamae/test_koik_issued
|
https://api.github.com/repos/kerdokullamae/test_koik_issued
|
closed
|
valdkonna ja kirjeldusüksuse vahelist seost ei saa tekitada
|
P: normal R: duplicate T: defect
|
**Reported by katrin vesterblom on 5 Jun 2013 13:19 UTC**
rahvusarhiiv.tietotest.ee, sisse loginud test.test
timebox-i järgi peaks olema realiseeritud valdkonna ja ühe KÜ vahelise seose loomine. Praktikas see ei õnnestu: KÜ vormil pole praegu üldse Valdkonna ega Valdkonnamärksõna sakki, akent, vms.
Valdkonna vaates on küll olemas võimalus seostada KÜ-ga, aga ükskõik milliseid leidandmeid seostamiseks otsin (ERA.1, TLA.777, Politseitalitus), midagi ma sinna ei saa.
TLA.777 annab üsna ruttu teate "Tulemusi ei leitud", kuigi selliste leidandmetega kirjeldusüksus (arhiiv) on olemas.
ERA.1 ja Politseitalitus jätavad pikaks ajaks ette teate "Otsib..."
ja ei jõua kuskile välja.
ERA.1.1.1 annab ka üsna ruttu teate "Tulemusi ei leitud", kuigi ka selline kirjeldusüksus (selliste leidandmetega) on olemas.
|
1.0
|
valdkonna ja kirjeldusüksuse vahelist seost ei saa tekitada - **Reported by katrin vesterblom on 5 Jun 2013 13:19 UTC**
rahvusarhiiv.tietotest.ee, sisse loginud test.test
timebox-i järgi peaks olema realiseeritud valdkonna ja ühe KÜ vahelise seose loomine. Praktikas see ei õnnestu: KÜ vormil pole praegu üldse Valdkonna ega Valdkonnamärksõna sakki, akent, vms.
Valdkonna vaates on küll olemas võimalus seostada KÜ-ga, aga ükskõik milliseid leidandmeid seostamiseks otsin (ERA.1, TLA.777, Politseitalitus), midagi ma sinna ei saa.
TLA.777 annab üsna ruttu teate "Tulemusi ei leitud", kuigi selliste leidandmetega kirjeldusüksus (arhiiv) on olemas.
ERA.1 ja Politseitalitus jätavad pikaks ajaks ette teate "Otsib..."
ja ei jõua kuskile välja.
ERA.1.1.1 annab ka üsna ruttu teate "Tulemusi ei leitud", kuigi ka selline kirjeldusüksus (selliste leidandmetega) on olemas.
|
defect
|
valdkonna ja kirjeldusüksuse vahelist seost ei saa tekitada reported by katrin vesterblom on jun utc rahvusarhiiv tietotest ee sisse loginud test test timebox i järgi peaks olema realiseeritud valdkonna ja ühe kü vahelise seose loomine praktikas see ei õnnestu kü vormil pole praegu üldse valdkonna ega valdkonnamärksõna sakki akent vms valdkonna vaates on küll olemas võimalus seostada kü ga aga ükskõik milliseid leidandmeid seostamiseks otsin era tla politseitalitus midagi ma sinna ei saa tla annab üsna ruttu teate tulemusi ei leitud kuigi selliste leidandmetega kirjeldusüksus arhiiv on olemas era ja politseitalitus jätavad pikaks ajaks ette teate otsib ja ei jõua kuskile välja era annab ka üsna ruttu teate tulemusi ei leitud kuigi ka selline kirjeldusüksus selliste leidandmetega on olemas
| 1
|
9,679
| 2,615,165,207
|
IssuesEvent
|
2015-03-01 06:45:22
|
chrsmith/reaver-wps
|
https://api.github.com/repos/chrsmith/reaver-wps
|
opened
|
Reaver is skipping correct pin due to fake WSC NACK Respond
|
auto-migrated Priority-Triage Type-Defect
|
```
Hi there I have a problem whit Reaver
problem is that Reaver is skipping pins due to Fake nack responds received
after M4 message after checking second half of pin in previous attempt.
Signal strength:58db.
Wireless card used: Realtek RTL8187 built-in Toshiba Satellite in monitor mode
Router Model TP-Link TL-WR720N (TP-LINK Wireless Router WR720N)
distance about 10-15 meters
Latest version of Reaver downloaded from Backtrack Linux Repo using
apt-get update
apt-get install Reaver commands.
6. Please describe what you think the issue is.
I think there must be an additional check that previous half of pin was correct
or not.
[+] Trying pin 43620852
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received M5 message
[+] Sending M6 message
[+] Received WSC NACK
[+] Sending WSC NACK
[+] Trying pin 43620869
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received WSC NACK
[+] Sending WSC NACK
[+] Trying pin 43620876
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received M5 message
[+] Sending M6 message
[+] Received WSC NACK
[+] Sending WSC NACK
[+] Trying pin 43620883
[+] Sending EAPOL START request
root@bt:~# reaver -i mon0 -b 64:70:02:DD:F4:8C -vv --nack
```
Original issue reported on code.google.com by `kostad...@yahoo.com` on 23 Jan 2013 at 7:33
|
1.0
|
Reaver is skipping correct pin due to fake WSC NACK Respond - ```
Hi there I have a problem whit Reaver
problem is that Reaver is skipping pins due to Fake nack responds received
after M4 message after checking second half of pin in previous attempt.
Signal strength:58db.
Wireless card used: Realtek RTL8187 built-in Toshiba Satellite in monitor mode
Router Model TP-Link TL-WR720N (TP-LINK Wireless Router WR720N)
distance about 10-15 meters
Latest version of Reaver downloaded from Backtrack Linux Repo using
apt-get update
apt-get install Reaver commands.
6. Please describe what you think the issue is.
I think there must be an additional check that previous half of pin was correct
or not.
[+] Trying pin 43620852
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received M5 message
[+] Sending M6 message
[+] Received WSC NACK
[+] Sending WSC NACK
[+] Trying pin 43620869
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received WSC NACK
[+] Sending WSC NACK
[+] Trying pin 43620876
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received M5 message
[+] Sending M6 message
[+] Received WSC NACK
[+] Sending WSC NACK
[+] Trying pin 43620883
[+] Sending EAPOL START request
root@bt:~# reaver -i mon0 -b 64:70:02:DD:F4:8C -vv --nack
```
Original issue reported on code.google.com by `kostad...@yahoo.com` on 23 Jan 2013 at 7:33
|
defect
|
reaver is skipping correct pin due to fake wsc nack respond hi there i have a problem whit reaver problem is that reaver is skipping pins due to fake nack responds received after message after checking second half of pin in previous attempt signal strength wireless card used realtek built in toshiba satellite in monitor mode router model tp link tl tp link wireless router distance about meters latest version of reaver downloaded from backtrack linux repo using apt get update apt get install reaver commands please describe what you think the issue is i think there must be an additional check that previous half of pin was correct or not trying pin sending eapol start request received identity request sending identity response received identity request sending identity response received message sending message received message sending message received message sending message received wsc nack sending wsc nack trying pin sending eapol start request received identity request sending identity response received identity request sending identity response received message sending message received message sending message received wsc nack sending wsc nack trying pin sending eapol start request received identity request sending identity response received identity request sending identity response received message sending message received message sending message received message sending message received wsc nack sending wsc nack trying pin sending eapol start request root bt reaver i b dd vv nack original issue reported on code google com by kostad yahoo com on jan at
| 1
|
196,787
| 14,889,996,791
|
IssuesEvent
|
2021-01-20 22:17:21
|
golang/go
|
https://api.github.com/repos/golang/go
|
closed
|
std,cmd: add test to ensure that all bundled packages are in sync with go.mod version
|
NeedsFix Testing release-blocker
|
We've had issues in the past where version skew could be introduced if people forgot to update one part of the distribution without also updating another (e.g., versions in src/go.mod and src/cmd/go.mod). A very effective solution was adding a test that fails when skew is detected.
I believe we still need to address this for net/http's bundled copies of `golang.org/x/net/http2` and `golang.org/x/net/internal/socks`. Right now, it's possible to update the bundled copy independently from the src/go.mod version, but it's not clear that flexibility is something we want to keep.
See #41375 where this came up. Also relevant is #25285.
I plan to revisit my review of [CL 189818](https://golang.org/cl/189818) (which fixes issue #32031), the trade-offs have changed by now (especially given #41330).
/cc @toothrot @bcmills @FiloSottile
|
1.0
|
std,cmd: add test to ensure that all bundled packages are in sync with go.mod version - We've had issues in the past where version skew could be introduced if people forgot to update one part of the distribution without also updating another (e.g., versions in src/go.mod and src/cmd/go.mod). A very effective solution was adding a test that fails when skew is detected.
I believe we still need to address this for net/http's bundled copies of `golang.org/x/net/http2` and `golang.org/x/net/internal/socks`. Right now, it's possible to update the bundled copy independently from the src/go.mod version, but it's not clear that flexibility is something we want to keep.
See #41375 where this came up. Also relevant is #25285.
I plan to revisit my review of [CL 189818](https://golang.org/cl/189818) (which fixes issue #32031), the trade-offs have changed by now (especially given #41330).
/cc @toothrot @bcmills @FiloSottile
|
non_defect
|
std cmd add test to ensure that all bundled packages are in sync with go mod version we ve had issues in the past where version skew could be introduced if people forgot to update one part of the distribution without also updating another e g versions in src go mod and src cmd go mod a very effective solution was adding a test that fails when skew is detected i believe we still need to address this for net http s bundled copies of golang org x net and golang org x net internal socks right now it s possible to update the bundled copy independently from the src go mod version but it s not clear that flexibility is something we want to keep see where this came up also relevant is i plan to revisit my review of which fixes issue the trade offs have changed by now especially given cc toothrot bcmills filosottile
| 0
|
124,557
| 16,614,262,668
|
IssuesEvent
|
2021-06-02 14:55:04
|
Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2
|
https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2
|
opened
|
Grayer Undead Hairs
|
:art:character designing😶 :exclamation: priority high :grey_exclamation: priority low
|
<!--
DO NOT REMOVE PRE-EXISTING LINES
IF YOU WANT TO SUGGEST A FEW THINGS, OPEN A NEW ISSUE PER EVERY SUGGESTION
----------------------------------------------------------------------------------------------------------
-->
**Describe your suggestion in full detail below:**
So we have the undead trait in `wc_undead_modifier.txt` and it makes the skin pale and green, but I think we also should make hairs grayer and paler too.
|
1.0
|
Grayer Undead Hairs - <!--
DO NOT REMOVE PRE-EXISTING LINES
IF YOU WANT TO SUGGEST A FEW THINGS, OPEN A NEW ISSUE PER EVERY SUGGESTION
----------------------------------------------------------------------------------------------------------
-->
**Describe your suggestion in full detail below:**
So we have the undead trait in `wc_undead_modifier.txt` and it makes the skin pale and green, but I think we also should make hairs grayer and paler too.
|
non_defect
|
grayer undead hairs do not remove pre existing lines if you want to suggest a few things open a new issue per every suggestion describe your suggestion in full detail below so we have the undead trait in wc undead modifier txt and it makes the skin pale and green but i think we also should make hairs grayer and paler too
| 0
|
169,524
| 13,150,495,196
|
IssuesEvent
|
2020-08-09 11:53:44
|
OpenRefine/OpenRefine
|
https://api.github.com/repos/OpenRefine/OpenRefine
|
closed
|
ToDate("2012-03-01","XXX") test failure & wasted conversion attempts
|
bug expression language tests
|
I'm investigating a local test failure which I'm unsure whether is specific to OS X or just how timezones are set up on my particular machine.
The failing test is:
Assert.assertEquals(invoke("toDate", "2012-03-01","XXX"), invoke("toDate", "2012-03-01"));
which returns 2012-03-01T01:00:00 instead of 2012-03-01T00:00:00
This appears to have something to do with the default fallthrough case here, perhaps due to Daylight Savings Time: https://github.com/OpenRefine/OpenRefine/blob/306b541c699e7e87d5e08be1310575b85b15dbf0/main/src/com/google/refine/expr/functions/ToDate.java#L170-L172
While investigating this, I came across this code:
https://github.com/OpenRefine/OpenRefine/blob/306b541c699e7e87d5e08be1310575b85b15dbf0/main/src/com/google/refine/expr/functions/ToDate.java#L135-L145
which I don't completely understand, but in the test case creates an array with 160 copies of the (invalid) code "XXX" which it then uses to attempt to parse the date 160 times, always failing, before falling back to the code above
|
1.0
|
ToDate("2012-03-01","XXX") test failure & wasted conversion attempts - I'm investigating a local test failure which I'm unsure whether is specific to OS X or just how timezones are set up on my particular machine.
The failing test is:
Assert.assertEquals(invoke("toDate", "2012-03-01","XXX"), invoke("toDate", "2012-03-01"));
which returns 2012-03-01T01:00:00 instead of 2012-03-01T00:00:00
This appears to have something to do with the default fallthrough case here, perhaps due to Daylight Savings Time: https://github.com/OpenRefine/OpenRefine/blob/306b541c699e7e87d5e08be1310575b85b15dbf0/main/src/com/google/refine/expr/functions/ToDate.java#L170-L172
While investigating this, I came across this code:
https://github.com/OpenRefine/OpenRefine/blob/306b541c699e7e87d5e08be1310575b85b15dbf0/main/src/com/google/refine/expr/functions/ToDate.java#L135-L145
which I don't completely understand, but in the test case creates an array with 160 copies of the (invalid) code "XXX" which it then uses to attempt to parse the date 160 times, always failing, before falling back to the code above
|
non_defect
|
todate xxx test failure wasted conversion attempts i m investigating a local test failure which i m unsure whether is specific to os x or just how timezones are set up on my particular machine the failing test is assert assertequals invoke todate xxx invoke todate which returns instead of this appears to have something to do with the default fallthrough case here perhaps due to daylight savings time while investigating this i came across this code which i don t completely understand but in the test case creates an array with copies of the invalid code xxx which it then uses to attempt to parse the date times always failing before falling back to the code above
| 0
|
197,228
| 14,913,889,912
|
IssuesEvent
|
2021-01-22 14:43:59
|
golang/go
|
https://api.github.com/repos/golang/go
|
closed
|
net: TestReadFromTimeoutFluctuation and TestWriteTimeoutFluctuation trybot flake on OpenBSD
|
NeedsFix OS-OpenBSD Testing help wanted
|
Test flake on OpenBSD:
https://storage.googleapis.com/go-build-log/4f0cbd68/openbsd-amd64-60_a0ca068c.log
```
--- FAIL: TestReadFromTimeoutFluctuation (1.48s)
timeout_test.go:727: ReadFrom took over 1s; expected 0.1s
--- FAIL: TestWriteTimeoutFluctuation (1.96s)
timeout_test.go:769: Write took over 1s; expected 0.1s
FAIL
FAIL net 33.085s
```
|
1.0
|
net: TestReadFromTimeoutFluctuation and TestWriteTimeoutFluctuation trybot flake on OpenBSD - Test flake on OpenBSD:
https://storage.googleapis.com/go-build-log/4f0cbd68/openbsd-amd64-60_a0ca068c.log
```
--- FAIL: TestReadFromTimeoutFluctuation (1.48s)
timeout_test.go:727: ReadFrom took over 1s; expected 0.1s
--- FAIL: TestWriteTimeoutFluctuation (1.96s)
timeout_test.go:769: Write took over 1s; expected 0.1s
FAIL
FAIL net 33.085s
```
|
non_defect
|
net testreadfromtimeoutfluctuation and testwritetimeoutfluctuation trybot flake on openbsd test flake on openbsd fail testreadfromtimeoutfluctuation timeout test go readfrom took over expected fail testwritetimeoutfluctuation timeout test go write took over expected fail fail net
| 0
|
71,045
| 23,422,919,106
|
IssuesEvent
|
2022-08-14 00:44:35
|
jezzsantos/automate
|
https://api.github.com/repos/jezzsantos/automate
|
closed
|
Switching pattern on ID, not Name
|
defect-design
|
We should be switching pattern based on ID not based on its name.
`automate edit switch <PATTERNID>`
|
1.0
|
Switching pattern on ID, not Name - We should be switching pattern based on ID not based on its name.
`automate edit switch <PATTERNID>`
|
defect
|
switching pattern on id not name we should be switching pattern based on id not based on its name automate edit switch
| 1
|
18,040
| 12,743,597,370
|
IssuesEvent
|
2020-06-26 10:44:44
|
OpenLiberty/blogs
|
https://api.github.com/repos/OpenLiberty/blogs
|
closed
|
Refine instructions on how to contribute blog posts
|
infrastructure
|
Right now the instructions are in a rough state, but are enough to get someone going with blog development.
1. Update OpenLiberty/openliberty.io README to point to this repository to find blog authoring instructions.
2. Update the Dockerfile's comments to reflect how to use the docker image. File here https://github.com/OpenLiberty/openliberty.io/blob/master/Dockerfile
3. Decide if we want people to build the docker image or pull from dockerhub (at the moment hosted on kinueng/openliberty.io)
4. Add maybe some instructions about using the staging site (openlibertydev) after using the docker image.
5. Evaluate who must review pull requests to this repository
FYI @lauracowen
|
1.0
|
Refine instructions on how to contribute blog posts - Right now the instructions are in a rough state, but are enough to get someone going with blog development.
1. Update OpenLiberty/openliberty.io README to point to this repository to find blog authoring instructions.
2. Update the Dockerfile's comments to reflect how to use the docker image. File here https://github.com/OpenLiberty/openliberty.io/blob/master/Dockerfile
3. Decide if we want people to build the docker image or pull from dockerhub (at the moment hosted on kinueng/openliberty.io)
4. Add maybe some instructions about using the staging site (openlibertydev) after using the docker image.
5. Evaluate who must review pull requests to this repository
FYI @lauracowen
|
non_defect
|
refine instructions on how to contribute blog posts right now the instructions are in a rough state but are enough to get someone going with blog development update openliberty openliberty io readme to point to this repository to find blog authoring instructions update the dockerfile s comments to reflect how to use the docker image file here decide if we want people to build the docker image or pull from dockerhub at the moment hosted on kinueng openliberty io add maybe some instructions about using the staging site openlibertydev after using the docker image evaluate who must review pull requests to this repository fyi lauracowen
| 0
|
182,706
| 21,673,925,228
|
IssuesEvent
|
2022-05-08 12:06:05
|
turkdevops/vscode
|
https://api.github.com/repos/turkdevops/vscode
|
closed
|
CVE-2021-23362 (Medium) detected in hosted-git-info-2.5.0.tgz - autoclosed
|
security vulnerability
|
## CVE-2021-23362 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.5.0.tgz</b></p></summary>
<p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p>
<p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.5.0.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.5.0.tgz</a></p>
<p>Path to dependency file: /extensions/emmet/package.json</p>
<p>Path to vulnerable library: /extensions/emmet/node_modules/hosted-git-info/package.json</p>
<p>
Dependency Hierarchy:
- vscode-1.0.1.tgz (Root Library)
- gulp-untar-0.0.4.tgz
- gulp-util-2.2.20.tgz
- dateformat-1.0.12.tgz
- meow-3.7.0.tgz
- normalize-package-data-2.4.0.tgz
- :x: **hosted-git-info-2.5.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/vscode/commit/9fd6b056a06e14655f4f0b0f631d670b24878828">9fd6b056a06e14655f4f0b0f631d670b24878828</a></p>
<p>Found in base branch: <b>webview-views</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362>CVE-2021-23362</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-43f8-2h32-f4cj">https://github.com/advisories/GHSA-43f8-2h32-f4cj</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution (hosted-git-info): 2.8.9</p>
<p>Direct dependency fix Resolution (vscode): 1.0.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-23362 (Medium) detected in hosted-git-info-2.5.0.tgz - autoclosed - ## CVE-2021-23362 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.5.0.tgz</b></p></summary>
<p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p>
<p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.5.0.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.5.0.tgz</a></p>
<p>Path to dependency file: /extensions/emmet/package.json</p>
<p>Path to vulnerable library: /extensions/emmet/node_modules/hosted-git-info/package.json</p>
<p>
Dependency Hierarchy:
- vscode-1.0.1.tgz (Root Library)
- gulp-untar-0.0.4.tgz
- gulp-util-2.2.20.tgz
- dateformat-1.0.12.tgz
- meow-3.7.0.tgz
- normalize-package-data-2.4.0.tgz
- :x: **hosted-git-info-2.5.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/vscode/commit/9fd6b056a06e14655f4f0b0f631d670b24878828">9fd6b056a06e14655f4f0b0f631d670b24878828</a></p>
<p>Found in base branch: <b>webview-views</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362>CVE-2021-23362</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-43f8-2h32-f4cj">https://github.com/advisories/GHSA-43f8-2h32-f4cj</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution (hosted-git-info): 2.8.9</p>
<p>Direct dependency fix Resolution (vscode): 1.0.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in hosted git info tgz autoclosed cve medium severity vulnerability vulnerable library hosted git info tgz provides metadata and conversions from repository urls for github bitbucket and gitlab library home page a href path to dependency file extensions emmet package json path to vulnerable library extensions emmet node modules hosted git info package json dependency hierarchy vscode tgz root library gulp untar tgz gulp util tgz dateformat tgz meow tgz normalize package data tgz x hosted git info tgz vulnerable library found in head commit a href found in base branch webview views vulnerability details the package hosted git info before are vulnerable to regular expression denial of service redos via regular expression shortcutmatch in the fromurl function in index js the affected regular expression exhibits polynomial worst case time complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution hosted git info direct dependency fix resolution vscode step up your open source security game with whitesource
| 0
|
20,546
| 3,373,914,740
|
IssuesEvent
|
2015-11-24 10:21:53
|
NetCDF4Excel/project
|
https://api.github.com/repos/NetCDF4Excel/project
|
closed
|
Run time error 53: File not found: VbNc.dll
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1.Open NetCDF4Excel_2007.xlsm
2.Enable macros (NetCDF)
3.Open File from NetCDF
What is the expected output? What do you see instead?
The expected output is the file opening up. Instead I get an error message: Run
time error 53: File not found: VbNc.dll
I have confirmed that the VbNc.dll file is present under the NETCDF4 Excel
directory.
What version of the product are you using? On what operating system?
Version 2.1 on Excel 2007 on Windows 8
```
Original issue reported on code.google.com by `wertt...@gmail.com` on 16 Jan 2014 at 9:40
|
1.0
|
Run time error 53: File not found: VbNc.dll - ```
What steps will reproduce the problem?
1.Open NetCDF4Excel_2007.xlsm
2.Enable macros (NetCDF)
3.Open File from NetCDF
What is the expected output? What do you see instead?
The expected output is the file opening up. Instead I get an error message: Run
time error 53: File not found: VbNc.dll
I have confirmed that the VbNc.dll file is present under the NETCDF4 Excel
directory.
What version of the product are you using? On what operating system?
Version 2.1 on Excel 2007 on Windows 8
```
Original issue reported on code.google.com by `wertt...@gmail.com` on 16 Jan 2014 at 9:40
|
defect
|
run time error file not found vbnc dll what steps will reproduce the problem open xlsm enable macros netcdf open file from netcdf what is the expected output what do you see instead the expected output is the file opening up instead i get an error message run time error file not found vbnc dll i have confirmed that the vbnc dll file is present under the excel directory what version of the product are you using on what operating system version on excel on windows original issue reported on code google com by wertt gmail com on jan at
| 1
|
251,177
| 18,940,924,267
|
IssuesEvent
|
2021-11-18 02:41:12
|
spacetelescope/romancal
|
https://api.github.com/repos/spacetelescope/romancal
|
opened
|
Update References General Doc for Jump and Ramp
|
documentation
|
_Issue [RCAL-250](https://jira.stsci.edu/browse/RCAL-250) was created on JIRA by [Paul Huwe](https://jira.stsci.edu/secure/ViewProfile.jspa?name=phuwe):_
The references_general.rst file needs to be updated to properly reflect the jump and ramp fitting requirements.
|
1.0
|
Update References General Doc for Jump and Ramp - _Issue [RCAL-250](https://jira.stsci.edu/browse/RCAL-250) was created on JIRA by [Paul Huwe](https://jira.stsci.edu/secure/ViewProfile.jspa?name=phuwe):_
The references_general.rst file needs to be updated to properly reflect the jump and ramp fitting requirements.
|
non_defect
|
update references general doc for jump and ramp issue was created on jira by the references general rst file needs to be updated to properly reflect the jump and ramp fitting requirements
| 0
|
57,734
| 16,017,984,244
|
IssuesEvent
|
2021-04-20 18:31:47
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Can't share screen in desktop app
|
T-Defect
|
<!-- A picture's worth a thousand words: PLEASE INCLUDE A SCREENSHOT :P -->
<!-- Please report security issues by email to security@matrix.org -->
<!-- This is a bug report template. By following the instructions below and
filling out the sections with your information, you will help the us to get all
the necessary data to fix your issue.
You can also preview your report before submitting it. You may remove sections
that aren't relevant to your particular case.
Text between <!-- and --> marks will be invisible in the report.
-->
### Description
I can not share my screen in a video call. The application can't find any application windows or screens.
### Steps to reproduce
- Open a video call with someone else
- Click on the screen share button
Describe how what happens differs from what you expected.
I can't select an application window to share. The loading screen never finishes.
Logs being sent: yes
### Version information
- **Platform**: Desktop
For the desktop app:
- **OS**: Windows 10
- **Version**: 1.7.25
|
1.0
|
Can't share screen in desktop app - <!-- A picture's worth a thousand words: PLEASE INCLUDE A SCREENSHOT :P -->
<!-- Please report security issues by email to security@matrix.org -->
<!-- This is a bug report template. By following the instructions below and
filling out the sections with your information, you will help the us to get all
the necessary data to fix your issue.
You can also preview your report before submitting it. You may remove sections
that aren't relevant to your particular case.
Text between <!-- and --> marks will be invisible in the report.
-->
### Description
I can not share my screen in a video call. The application can't find any application windows or screens.
### Steps to reproduce
- Open a video call with someone else
- Click on the screen share button
Describe how what happens differs from what you expected.
I can't select an application window to share. The loading screen never finishes.
Logs being sent: yes
### Version information
- **Platform**: Desktop
For the desktop app:
- **OS**: Windows 10
- **Version**: 1.7.25
|
defect
|
can t share screen in desktop app this is a bug report template by following the instructions below and filling out the sections with your information you will help the us to get all the necessary data to fix your issue you can also preview your report before submitting it you may remove sections that aren t relevant to your particular case text between marks will be invisible in the report description i can not share my screen in a video call the application can t find any application windows or screens steps to reproduce open a video call with someone else click on the screen share button describe how what happens differs from what you expected i can t select an application window to share the loading screen never finishes logs being sent yes version information platform desktop for the desktop app os windows version
| 1
|
49,464
| 6,027,321,858
|
IssuesEvent
|
2017-06-08 13:32:55
|
ionic-team/ionic
|
https://api.github.com/repos/ionic-team/ionic
|
closed
|
ionic2 - bug(overlays) - After clicking to dismiss an overlay, the click propagates to whatever is beneath it.
|
needs:devicetesting v2
|
#### Short description of the problem:
If I have an overlay open (modal confirmed, not sure about others), and click a dismiss button on it, the transition will run and then once it is dismissed it will click whatever is beneath it, such as an input, etc. This happens on iOS. Not sure about any other platforms.
#### What behavior are you expecting?
I expect there to not be duplicate clicks.
**Steps to reproduce:**
1. Create a page with some inputs on it (I am specifically using standard textareas)
2. Create a modal with a dismiss button directly over one of the textareas
3. Open the modal, click the dismiss button, and observe that the textarea is focused and the keyboard appears
**Run `ionic info` from terminal/cmd prompt:** (paste output below)
Cordova CLI: 6.2.0
Gulp version: CLI version 3.9.1
Gulp local: Local version 3.9.1
Ionic Framework Version: 2.0.0-rc.1
Ionic CLI Version: 2.1.0
Ionic App Lib Version: 2.1.0-beta.1
ios-deploy version: 1.8.4
ios-sim version: 5.0.6
OS: Mac OS X El Capitan
Node Version: v4.5.0
Xcode version: Xcode 8.0 Build version 8A218a
|
1.0
|
ionic2 - bug(overlays) - After clicking to dismiss an overlay, the click propagates to whatever is beneath it. - #### Short description of the problem:
If I have an overlay open (modal confirmed, not sure about others), and click a dismiss button on it, the transition will run and then once it is dismissed it will click whatever is beneath it, such as an input, etc. This happens on iOS. Not sure about any other platforms.
#### What behavior are you expecting?
I expect there to not be duplicate clicks.
**Steps to reproduce:**
1. Create a page with some inputs on it (I am specifically using standard textareas)
2. Create a modal with a dismiss button directly over one of the textareas
3. Open the modal, click the dismiss button, and observe that the textarea is focused and the keyboard appears
**Run `ionic info` from terminal/cmd prompt:** (paste output below)
Cordova CLI: 6.2.0
Gulp version: CLI version 3.9.1
Gulp local: Local version 3.9.1
Ionic Framework Version: 2.0.0-rc.1
Ionic CLI Version: 2.1.0
Ionic App Lib Version: 2.1.0-beta.1
ios-deploy version: 1.8.4
ios-sim version: 5.0.6
OS: Mac OS X El Capitan
Node Version: v4.5.0
Xcode version: Xcode 8.0 Build version 8A218a
|
non_defect
|
bug overlays after clicking to dismiss an overlay the click propagates to whatever is beneath it short description of the problem if i have an overlay open modal confirmed not sure about others and click a dismiss button on it the transition will run and then once it is dismissed it will click whatever is beneath it such as an input etc this happens on ios not sure about any other platforms what behavior are you expecting i expect there to not be duplicate clicks steps to reproduce create a page with some inputs on it i am specifically using standard textareas create a modal with a dismiss button directly over one of the textareas open the modal click the dismiss button and observe that the textarea is focused and the keyboard appears run ionic info from terminal cmd prompt paste output below cordova cli gulp version cli version gulp local local version ionic framework version rc ionic cli version ionic app lib version beta ios deploy version ios sim version os mac os x el capitan node version xcode version xcode build version
| 0
|
77,575
| 27,057,314,359
|
IssuesEvent
|
2023-02-13 17:01:48
|
microsoft/TypeScript
|
https://api.github.com/repos/microsoft/TypeScript
|
closed
|
Missing NOTICE file makes it harder to satisfy terms of license
|
Not a Defect
|
# Bug Report
### 🔎 Search Terms
NOTICE file
### 🕗 Version & Regression Information
I encountered the problem in the 4.9.5 release. I also examined the latest code in the public rep.
I reviewed the FAQ for entries about NOTICE file.
### 🙁 Actual behavior
There are copyright, patent, trademark, and attribution notices from the Source form of the Work, and the work does not include a NOTICE text file as part of its distribution.
The license used by the project is Apache Public License 2.0:
https://github.com/microsoft/TypeScript/blob/main/LICENSE.txt
That contains these clauses:
`You must retain, in the Source form of any Derivative Works that You distribute, all copyright, patent, trademark, and attribution notices from the Source form of the Work, excluding those notices that do not pertain to any part of the Derivative Works; and
If the Work includes a "NOTICE" text file as part of its distribution, then any Derivative Works that You distribute must include a readable copy of the attribution notices contained within such NOTICE file, excluding those notices that do not pertain to any part of the Derivative Works`
### 🙂 Expected behavior
<!-- What you expected to happen instead, and why -->
The work should contain a NOTICE text file per the convention documented here:
https://www.apache.org/licenses/LICENSE-2.0#apply
|
1.0
|
Missing NOTICE file makes it harder to satisfy terms of license - # Bug Report
### 🔎 Search Terms
NOTICE file
### 🕗 Version & Regression Information
I encountered the problem in the 4.9.5 release. I also examined the latest code in the public rep.
I reviewed the FAQ for entries about NOTICE file.
### 🙁 Actual behavior
There are copyright, patent, trademark, and attribution notices from the Source form of the Work, and the work does not include a NOTICE text file as part of its distribution.
The license used by the project is Apache Public License 2.0:
https://github.com/microsoft/TypeScript/blob/main/LICENSE.txt
That contains these clauses:
`You must retain, in the Source form of any Derivative Works that You distribute, all copyright, patent, trademark, and attribution notices from the Source form of the Work, excluding those notices that do not pertain to any part of the Derivative Works; and
If the Work includes a "NOTICE" text file as part of its distribution, then any Derivative Works that You distribute must include a readable copy of the attribution notices contained within such NOTICE file, excluding those notices that do not pertain to any part of the Derivative Works`
### 🙂 Expected behavior
<!-- What you expected to happen instead, and why -->
The work should contain a NOTICE text file per the convention documented here:
https://www.apache.org/licenses/LICENSE-2.0#apply
|
defect
|
missing notice file makes it harder to satisfy terms of license bug report 🔎 search terms notice file 🕗 version regression information i encountered the problem in the release i also examined the latest code in the public rep i reviewed the faq for entries about notice file 🙁 actual behavior there are copyright patent trademark and attribution notices from the source form of the work and the work does not include a notice text file as part of its distribution the license used by the project is apache public license that contains these clauses you must retain in the source form of any derivative works that you distribute all copyright patent trademark and attribution notices from the source form of the work excluding those notices that do not pertain to any part of the derivative works and if the work includes a notice text file as part of its distribution then any derivative works that you distribute must include a readable copy of the attribution notices contained within such notice file excluding those notices that do not pertain to any part of the derivative works 🙂 expected behavior the work should contain a notice text file per the convention documented here
| 1
|
57,745
| 16,025,170,288
|
IssuesEvent
|
2021-04-21 08:11:27
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Auto-switching to rooms in spaces is disorienting/broken
|
A-Spaces S-Major T-Defect Z-Rageshake
|
It's also unreliable if you are viewing a favourite room and you try to navigate to a different space
|
1.0
|
Auto-switching to rooms in spaces is disorienting/broken - It's also unreliable if you are viewing a favourite room and you try to navigate to a different space
|
defect
|
auto switching to rooms in spaces is disorienting broken it s also unreliable if you are viewing a favourite room and you try to navigate to a different space
| 1
|
55,011
| 14,121,057,162
|
IssuesEvent
|
2020-11-09 00:39:48
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
closed
|
solve_ivp fails to solve with non-zero initial time, large parameters, and methods other than LSODA
|
defect scipy.integrate
|
As the title says, `scipy.integrate.solve_ivp` fail to solve with a non-zero initial time, large parameters, and methods other than LSODA. Remove any one of those conditions and it solves.
Here is an ODE that demonstrates the problem.
```
import numpy as np
import scipy.integrate
c_0 = np.array([1.0e6])
def ode(t, y):
c = y[0]
cdot = C / c
RHS = np.array([cdot])
return(RHS)
```
Attempting to integrate from a non-zero initial time fails to solve. For example,
```
C = 1e50
t_span = (1.0, 2.0)
ivp_sol = scipy.integrate.solve_ivp(ode, t_span, c_0, method = 'Radau')
print(ivp_sol.message)
```
prints the following to the screen
```
Required step size is less than spacing between numbers.
```
Attempting to integrate from a zero initial time or with a smaller value of `C` runs fine. For example,
```
C = 1e50
t_span = (0.0, 2.0)
ivp_sol = scipy.integrate.solve_ivp(ode, t_span, c_0, method = 'Radau')
print(ivp_sol.message)
C = 1e25
t_span = (1.0, 2.0)
ivp_sol = scipy.integrate.solve_ivp(ode, t_span, c_0, method = 'Radau')
print(ivp_sol.message)
```
prints the following to the screen
```
The solver successfully reached the end of the integration interval.
The solver successfully reached the end of the integration interval.
```
This problem seems to involve all methods, except for LSODA, although LSODA still complains about small step sizes. Running
```
C = 1e50
t_span = (1.0, 2.0)
ivp_sol = scipy.integrate.solve_ivp(ode, t_span, c_0, method = 'LSODA')
print(ivp_sol.message)
```
results in several warnings like
```
lsoda-- warning..internal t (=r1) and h (=r2) are��
such that in the machine, t + h = t on the next step
(h = step size). solver will continue anyway��
in above, r1 = 0.1000000000000D+01 r2 = 0.3162277663331D-39
```
but it completes with
```
The solver successfully reached the end of the integration interval.
```
It’s not surprising that some methods might work better than others for stiff problems, but it should not matter if the initial time is zero or not.
Here are my version numbers.
```
import sys, scipy, numpy; print(scipy.__version__, numpy.__version__, sys.version_info)
```
produces
```
('1.2.1', '1.16.4', sys.version_info(major=2, minor=7, micro=16, releaselevel='final', serial=0))
```
|
1.0
|
solve_ivp fails to solve with non-zero initial time, large parameters, and methods other than LSODA - As the title says, `scipy.integrate.solve_ivp` fail to solve with a non-zero initial time, large parameters, and methods other than LSODA. Remove any one of those conditions and it solves.
Here is an ODE that demonstrates the problem.
```
import numpy as np
import scipy.integrate
c_0 = np.array([1.0e6])
def ode(t, y):
c = y[0]
cdot = C / c
RHS = np.array([cdot])
return(RHS)
```
Attempting to integrate from a non-zero initial time fails to solve. For example,
```
C = 1e50
t_span = (1.0, 2.0)
ivp_sol = scipy.integrate.solve_ivp(ode, t_span, c_0, method = 'Radau')
print(ivp_sol.message)
```
prints the following to the screen
```
Required step size is less than spacing between numbers.
```
Attempting to integrate from a zero initial time or with a smaller value of `C` runs fine. For example,
```
C = 1e50
t_span = (0.0, 2.0)
ivp_sol = scipy.integrate.solve_ivp(ode, t_span, c_0, method = 'Radau')
print(ivp_sol.message)
C = 1e25
t_span = (1.0, 2.0)
ivp_sol = scipy.integrate.solve_ivp(ode, t_span, c_0, method = 'Radau')
print(ivp_sol.message)
```
prints the following to the screen
```
The solver successfully reached the end of the integration interval.
The solver successfully reached the end of the integration interval.
```
This problem seems to involve all methods, except for LSODA, although LSODA still complains about small step sizes. Running
```
C = 1e50
t_span = (1.0, 2.0)
ivp_sol = scipy.integrate.solve_ivp(ode, t_span, c_0, method = 'LSODA')
print(ivp_sol.message)
```
results in several warnings like
```
lsoda-- warning..internal t (=r1) and h (=r2) are��
such that in the machine, t + h = t on the next step
(h = step size). solver will continue anyway��
in above, r1 = 0.1000000000000D+01 r2 = 0.3162277663331D-39
```
but it completes with
```
The solver successfully reached the end of the integration interval.
```
It’s not surprising that some methods might work better than others for stiff problems, but it should not matter if the initial time is zero or not.
Here are my version numbers.
```
import sys, scipy, numpy; print(scipy.__version__, numpy.__version__, sys.version_info)
```
produces
```
('1.2.1', '1.16.4', sys.version_info(major=2, minor=7, micro=16, releaselevel='final', serial=0))
```
|
defect
|
solve ivp fails to solve with non zero initial time large parameters and methods other than lsoda as the title says scipy integrate solve ivp fail to solve with a non zero initial time large parameters and methods other than lsoda remove any one of those conditions and it solves here is an ode that demonstrates the problem import numpy as np import scipy integrate c np array def ode t y c y cdot c c rhs np array return rhs attempting to integrate from a non zero initial time fails to solve for example c t span ivp sol scipy integrate solve ivp ode t span c method radau print ivp sol message prints the following to the screen required step size is less than spacing between numbers attempting to integrate from a zero initial time or with a smaller value of c runs fine for example c t span ivp sol scipy integrate solve ivp ode t span c method radau print ivp sol message c t span ivp sol scipy integrate solve ivp ode t span c method radau print ivp sol message prints the following to the screen the solver successfully reached the end of the integration interval the solver successfully reached the end of the integration interval this problem seems to involve all methods except for lsoda although lsoda still complains about small step sizes running c t span ivp sol scipy integrate solve ivp ode t span c method lsoda print ivp sol message results in several warnings like lsoda warning internal t and h are�� such that in the machine t h t on the next step h step size solver will continue anyway�� in above but it completes with the solver successfully reached the end of the integration interval it’s not surprising that some methods might work better than others for stiff problems but it should not matter if the initial time is zero or not here are my version numbers import sys scipy numpy print scipy version numpy version sys version info produces sys version info major minor micro releaselevel final serial
| 1
|
391,455
| 26,893,542,131
|
IssuesEvent
|
2023-02-06 10:40:26
|
Arquisoft/lomap_en1b
|
https://api.github.com/repos/Arquisoft/lomap_en1b
|
closed
|
Adding names and emails to documentation
|
documentation:page_facing_up::pen:
|
Every team member should add his/her name and email into a new section at the end of the README and merge it into develop branch via pull requests.
When the README is finished, it should be pushed into master.
|
1.0
|
Adding names and emails to documentation - Every team member should add his/her name and email into a new section at the end of the README and merge it into develop branch via pull requests.
When the README is finished, it should be pushed into master.
|
non_defect
|
adding names and emails to documentation every team member should add his her name and email into a new section at the end of the readme and merge it into develop branch via pull requests when the readme is finished it should be pushed into master
| 0
|
37,687
| 8,474,803,116
|
IssuesEvent
|
2018-10-24 17:08:25
|
brainvisa/testbidon
|
https://api.github.com/repos/brainvisa/testbidon
|
closed
|
OpenSlide plugin: ineffecient read when used with borders
|
Category: soma-io Component: Resolution Priority: Normal Status: Closed Tracker: Defect
|
---
Author Name: **Souedet, Nicolas** (Souedet, Nicolas)
Original Redmine Issue: 14573, https://bioproj.extra.cea.fr/redmine/issues/14573
Original Date: 2016-03-18
Original Assignee: Souedet, Nicolas
---
Openslide and especially CZI format is based on tiled images. To fill memory with borders and because the soma-io openslide plugin does not support strides yet, VolumeFormatReader reads line after line to fill memory. But to read a part of the line, it is necessary to read a full tile and eventually to uncompress data. Each tile is put in the openslide cache, but the size of the openslide cache is limited, and often not big enough to store all the tiles of a line.
To fix this problem, it is necessary to add support for strides in the soma-io plugin and to read data in a tiled way.
|
1.0
|
OpenSlide plugin: ineffecient read when used with borders - ---
Author Name: **Souedet, Nicolas** (Souedet, Nicolas)
Original Redmine Issue: 14573, https://bioproj.extra.cea.fr/redmine/issues/14573
Original Date: 2016-03-18
Original Assignee: Souedet, Nicolas
---
Openslide and especially CZI format is based on tiled images. To fill memory with borders and because the soma-io openslide plugin does not support strides yet, VolumeFormatReader reads line after line to fill memory. But to read a part of the line, it is necessary to read a full tile and eventually to uncompress data. Each tile is put in the openslide cache, but the size of the openslide cache is limited, and often not big enough to store all the tiles of a line.
To fix this problem, it is necessary to add support for strides in the soma-io plugin and to read data in a tiled way.
|
defect
|
openslide plugin ineffecient read when used with borders author name souedet nicolas souedet nicolas original redmine issue original date original assignee souedet nicolas openslide and especially czi format is based on tiled images to fill memory with borders and because the soma io openslide plugin does not support strides yet volumeformatreader reads line after line to fill memory but to read a part of the line it is necessary to read a full tile and eventually to uncompress data each tile is put in the openslide cache but the size of the openslide cache is limited and often not big enough to store all the tiles of a line to fix this problem it is necessary to add support for strides in the soma io plugin and to read data in a tiled way
| 1
|
249,513
| 26,945,769,816
|
IssuesEvent
|
2023-02-08 07:57:03
|
jgeraigery/pyadi-iio
|
https://api.github.com/repos/jgeraigery/pyadi-iio
|
opened
|
paramiko-3.0.0-py3-none-any.whl: 1 vulnerabilities (highest severity is: 4.8)
|
security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>paramiko-3.0.0-py3-none-any.whl</b></p></summary>
<p></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/pyadi-iio/commit/5be07ac2fc964a356b857941afa1941e3aa207ec">5be07ac2fc964a356b857941afa1941e3aa207ec</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (paramiko version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2023-23931](https://www.mend.io/vulnerability-database/CVE-2023-23931) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 4.8 | cryptography-39.0.0-cp36-abi3-manylinux_2_24_x86_64.whl | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2023-23931</summary>
### Vulnerable Library - <b>cryptography-39.0.0-cp36-abi3-manylinux_2_24_x86_64.whl</b></p>
<p>cryptography is a package which provides cryptographic recipes and primitives to Python developers.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/fc/b2/3b946e24de214fc49adeefeea6214bcbc4bce2bd745877f074d1dd13c9a2/cryptography-39.0.0-cp36-abi3-manylinux_2_24_x86_64.whl">https://files.pythonhosted.org/packages/fc/b2/3b946e24de214fc49adeefeea6214bcbc4bce2bd745877f074d1dd13c9a2/cryptography-39.0.0-cp36-abi3-manylinux_2_24_x86_64.whl</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
Dependency Hierarchy:
- paramiko-3.0.0-py3-none-any.whl (Root Library)
- :x: **cryptography-39.0.0-cp36-abi3-manylinux_2_24_x86_64.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/pyadi-iio/commit/5be07ac2fc964a356b857941afa1941e3aa207ec">5be07ac2fc964a356b857941afa1941e3aa207ec</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
cryptography is a package designed to expose cryptographic primitives and recipes to Python developers. In affected versions `Cipher.update_into` would accept Python objects which implement the buffer protocol, but provide only immutable buffers. This would allow immutable objects (such as `bytes`) to be mutated, thus violating fundamental rules of Python and resulting in corrupted output. This now correctly raises an exception. This issue has been present since `update_into` was originally introduced in cryptography 1.8.
<p>Publish Date: 2023-02-07
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-23931>CVE-2023-23931</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>4.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-23931">https://www.cve.org/CVERecord?id=CVE-2023-23931</a></p>
<p>Release Date: 2023-02-07</p>
<p>Fix Resolution: cryptography - 39.0.1</p>
</p>
<p></p>
</details>
|
True
|
paramiko-3.0.0-py3-none-any.whl: 1 vulnerabilities (highest severity is: 4.8) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>paramiko-3.0.0-py3-none-any.whl</b></p></summary>
<p></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/pyadi-iio/commit/5be07ac2fc964a356b857941afa1941e3aa207ec">5be07ac2fc964a356b857941afa1941e3aa207ec</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (paramiko version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2023-23931](https://www.mend.io/vulnerability-database/CVE-2023-23931) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 4.8 | cryptography-39.0.0-cp36-abi3-manylinux_2_24_x86_64.whl | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2023-23931</summary>
### Vulnerable Library - <b>cryptography-39.0.0-cp36-abi3-manylinux_2_24_x86_64.whl</b></p>
<p>cryptography is a package which provides cryptographic recipes and primitives to Python developers.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/fc/b2/3b946e24de214fc49adeefeea6214bcbc4bce2bd745877f074d1dd13c9a2/cryptography-39.0.0-cp36-abi3-manylinux_2_24_x86_64.whl">https://files.pythonhosted.org/packages/fc/b2/3b946e24de214fc49adeefeea6214bcbc4bce2bd745877f074d1dd13c9a2/cryptography-39.0.0-cp36-abi3-manylinux_2_24_x86_64.whl</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
Dependency Hierarchy:
- paramiko-3.0.0-py3-none-any.whl (Root Library)
- :x: **cryptography-39.0.0-cp36-abi3-manylinux_2_24_x86_64.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/pyadi-iio/commit/5be07ac2fc964a356b857941afa1941e3aa207ec">5be07ac2fc964a356b857941afa1941e3aa207ec</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
cryptography is a package designed to expose cryptographic primitives and recipes to Python developers. In affected versions `Cipher.update_into` would accept Python objects which implement the buffer protocol, but provide only immutable buffers. This would allow immutable objects (such as `bytes`) to be mutated, thus violating fundamental rules of Python and resulting in corrupted output. This now correctly raises an exception. This issue has been present since `update_into` was originally introduced in cryptography 1.8.
<p>Publish Date: 2023-02-07
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-23931>CVE-2023-23931</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>4.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-23931">https://www.cve.org/CVERecord?id=CVE-2023-23931</a></p>
<p>Release Date: 2023-02-07</p>
<p>Fix Resolution: cryptography - 39.0.1</p>
</p>
<p></p>
</details>
|
non_defect
|
paramiko none any whl vulnerabilities highest severity is vulnerable library paramiko none any whl path to dependency file requirements txt path to vulnerable library requirements txt found in head commit a href vulnerabilities cve severity cvss dependency type fixed in paramiko version remediation available medium cryptography manylinux whl transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the section details below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library cryptography manylinux whl cryptography is a package which provides cryptographic recipes and primitives to python developers library home page a href path to dependency file requirements txt path to vulnerable library requirements txt dependency hierarchy paramiko none any whl root library x cryptography manylinux whl vulnerable library found in head commit a href found in base branch master vulnerability details cryptography is a package designed to expose cryptographic primitives and recipes to python developers in affected versions cipher update into would accept python objects which implement the buffer protocol but provide only immutable buffers this would allow immutable objects such as bytes to be mutated thus violating fundamental rules of python and resulting in corrupted output this now correctly raises an exception this issue has been present since update into was originally introduced in cryptography publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution cryptography
| 0
|
57,550
| 15,855,675,157
|
IssuesEvent
|
2021-04-08 00:26:55
|
STEllAR-GROUP/hpx
|
https://api.github.com/repos/STEllAR-GROUP/hpx
|
closed
|
HPX-1.6.0 fails to build on Windows 10
|
compiler: msvc type: defect
|
## Expected Behavior
Successful build using Visual Studio 2019
## Actual Behavior
Build failed with the following errors:
Error C2182 'abstract declarator': illegal use of type 'void' 1d_stencil_channel (Examples\1D Stencil\1d_stencil_channel\1d_stencil_channel) C:\Users\phili\Documents\cpp\hpx-1.6.0\libs\core\functional\include\hpx\functional\traits\is_invocable.hpp
Note: error points to line 81 of is_invocable.hpp
## Steps to Reproduce the Problem
Follow all steps in the HPX manual regarding:
How to build HPX under Windows 10 x64 with Visual Studio 2015
## Specifications
- HPX Version: 1.6.0
- Platform (compiler, OS): Visual Studio 2019, Windows 10.
- Setting compiler to use C++14 as well as C++17 makes no difference.
|
1.0
|
HPX-1.6.0 fails to build on Windows 10 - ## Expected Behavior
Successful build using Visual Studio 2019
## Actual Behavior
Build failed with the following errors:
Error C2182 'abstract declarator': illegal use of type 'void' 1d_stencil_channel (Examples\1D Stencil\1d_stencil_channel\1d_stencil_channel) C:\Users\phili\Documents\cpp\hpx-1.6.0\libs\core\functional\include\hpx\functional\traits\is_invocable.hpp
Note: error points to line 81 of is_invocable.hpp
## Steps to Reproduce the Problem
Follow all steps in the HPX manual regarding:
How to build HPX under Windows 10 x64 with Visual Studio 2015
## Specifications
- HPX Version: 1.6.0
- Platform (compiler, OS): Visual Studio 2019, Windows 10.
- Setting compiler to use C++14 as well as C++17 makes no difference.
|
defect
|
hpx fails to build on windows expected behavior successful build using visual studio actual behavior build failed with the following errors error abstract declarator illegal use of type void stencil channel examples stencil stencil channel stencil channel c users phili documents cpp hpx libs core functional include hpx functional traits is invocable hpp note error points to line of is invocable hpp steps to reproduce the problem follow all steps in the hpx manual regarding how to build hpx under windows with visual studio specifications hpx version platform compiler os visual studio windows setting compiler to use c as well as c makes no difference
| 1
|
68,944
| 21,994,370,730
|
IssuesEvent
|
2022-05-26 03:44:26
|
SeleniumHQ/selenium
|
https://api.github.com/repos/SeleniumHQ/selenium
|
opened
|
[🐛 Bug]: Open.Selenium.InvalidStateException was occurred using IE.
|
I-defect needs-triaging
|
### What happened?
Open.Selenium.InvalidStateException was occurred using IE.
For details and reproduction methods, I wonder if you check items after 'How can we reproduce the issue?'.
This exception is occured at InternetExplorerDriver, is not occured at EdgeDriver.
I test this 👆 change to EdgeDriver class from InternetExplorerDriver class, test has all passed.
Sorry for my poor english, thank you.
### How can we reproduce the issue?
```shell
The implementation is like this:
private readonly IWebDriver _webDriver; // selenium process run class field.
...
public ctor(string webDriverDirectoryPath)
{
// get webdriver directory absolute path from ctor argument.
_webDriver = new InternetExplorerDriver(webDriverDirectoryPath);
}
public void Init()
{
// this method is navigating method for looping invoke method page.
}
// this method called by other class using foreach.
public void Invoke(TestCase testCase)
{
...
_webDriver
.FindElement(By.Id("{Input html tag id}"))
.Clear(); // error has occurred this point.
_webDriver
.FindElement(By.Id("{Input html tag id}"))
.SendKeys(testCase.Value);
...
var selectElement = new SelectElement(_webDriver
.FindElement(By.Id("{Select html tag id}"))
);
selectElement.SelectByIndex(testCase.SelectIndex); // this method not work. however using for loop and SendKey(Keys.ArrowDown) is work.
}
That👆 code runnning at Xunit testing.
```cs
var testCases = new[]
{
new TestCase(1, 1, "hoge", "001", "51000", "Yamada"),
new TestCase(2, 1, "fuga", "002", "51001", "Tanaka"),
new TestCase(3, 1, "piyo", "003", "51002", "Suzuki")
};
using var invoker = new SeleniumInvoker();
invoker.Init();
var result1 = invoker.Invoke(testCases[0]);
Assert.Equal("OK", result1.CheckResult);
Assert.Equal("Yamada", result1.CustomerName);
Assert.Empty(result1.ErrorMessages);
var result2 = invoker.GetResult(testCases[1]);
Assert.Equal("NG", result2.CheckResult);
Assert.Equal("Tanaka", result2.CustomerName);
Assert.Equal("Invalid Code", result2.ErrorMessages.Single());
var result3 = invoker.GetResult(testCases[2]);
Assert.Equal("NG", result3.CheckResult);
Assert.Equal("Suzuki", result3.CustomerName);
Assert.Equal("No Account", result3.ErrorMessages.Single());
```
```
### Relevant log output
```shell
Xunit failed message:
OpenQA.Selenium.InvalidElementStateException
A JavaScript error was encountered clearing the element. The driver assumes this is because the element is hidden, disabled or read-only, and it must not be to clear the element.
at OpenQA.Selenium.WebDriver.UnpackAndThrowOnError(Response errorResponse)
at OpenQA.Selenium.WebDriver.Execute(String driverCommandToExecute, Dictionary`2 parameters)
at OpenQA.Selenium.WebDriver.InternalExecute(String driverCommandToExecute, Dictionary`2 parameters)
at OpenQA.Selenium.WebElement.Execute(String commandToExecute, Dictionary`2 parameters)
at OpenQA.Selenium.WebElement.Clear()
```
```
### Operating System
Windows 10, Version 21H1 , OS Build 19043.1706
### Selenium version
C#, Selenium.Support 4.1.1, Selenium.WebDriver 4.1.1
### What are the browser(s) and version(s) where you see this issue?
Internet Explorer Version 21H1 (OS Build 19043.1706)
### What are the browser driver(s) and version(s) where you see this issue?
IEDriverServer.exe 4.0.0.0 (64-bit)
### Are you using Selenium Grid?
_No response_
|
1.0
|
[🐛 Bug]: Open.Selenium.InvalidStateException was occurred using IE. - ### What happened?
Open.Selenium.InvalidStateException was occurred using IE.
For details and reproduction methods, I wonder if you check items after 'How can we reproduce the issue?'.
This exception is occured at InternetExplorerDriver, is not occured at EdgeDriver.
I test this 👆 change to EdgeDriver class from InternetExplorerDriver class, test has all passed.
Sorry for my poor english, thank you.
### How can we reproduce the issue?
```shell
The implementation is like this:
private readonly IWebDriver _webDriver; // selenium process run class field.
...
public ctor(string webDriverDirectoryPath)
{
// get webdriver directory absolute path from ctor argument.
_webDriver = new InternetExplorerDriver(webDriverDirectoryPath);
}
public void Init()
{
// this method is navigating method for looping invoke method page.
}
// this method called by other class using foreach.
public void Invoke(TestCase testCase)
{
...
_webDriver
.FindElement(By.Id("{Input html tag id}"))
.Clear(); // error has occurred this point.
_webDriver
.FindElement(By.Id("{Input html tag id}"))
.SendKeys(testCase.Value);
...
var selectElement = new SelectElement(_webDriver
.FindElement(By.Id("{Select html tag id}"))
);
selectElement.SelectByIndex(testCase.SelectIndex); // this method not work. however using for loop and SendKey(Keys.ArrowDown) is work.
}
That👆 code runnning at Xunit testing.
```cs
var testCases = new[]
{
new TestCase(1, 1, "hoge", "001", "51000", "Yamada"),
new TestCase(2, 1, "fuga", "002", "51001", "Tanaka"),
new TestCase(3, 1, "piyo", "003", "51002", "Suzuki")
};
using var invoker = new SeleniumInvoker();
invoker.Init();
var result1 = invoker.Invoke(testCases[0]);
Assert.Equal("OK", result1.CheckResult);
Assert.Equal("Yamada", result1.CustomerName);
Assert.Empty(result1.ErrorMessages);
var result2 = invoker.GetResult(testCases[1]);
Assert.Equal("NG", result2.CheckResult);
Assert.Equal("Tanaka", result2.CustomerName);
Assert.Equal("Invalid Code", result2.ErrorMessages.Single());
var result3 = invoker.GetResult(testCases[2]);
Assert.Equal("NG", result3.CheckResult);
Assert.Equal("Suzuki", result3.CustomerName);
Assert.Equal("No Account", result3.ErrorMessages.Single());
```
```
### Relevant log output
```shell
Xunit failed message:
OpenQA.Selenium.InvalidElementStateException
A JavaScript error was encountered clearing the element. The driver assumes this is because the element is hidden, disabled or read-only, and it must not be to clear the element.
at OpenQA.Selenium.WebDriver.UnpackAndThrowOnError(Response errorResponse)
at OpenQA.Selenium.WebDriver.Execute(String driverCommandToExecute, Dictionary`2 parameters)
at OpenQA.Selenium.WebDriver.InternalExecute(String driverCommandToExecute, Dictionary`2 parameters)
at OpenQA.Selenium.WebElement.Execute(String commandToExecute, Dictionary`2 parameters)
at OpenQA.Selenium.WebElement.Clear()
```
```
### Operating System
Windows 10, Version 21H1 , OS Build 19043.1706
### Selenium version
C#, Selenium.Support 4.1.1, Selenium.WebDriver 4.1.1
### What are the browser(s) and version(s) where you see this issue?
Internet Explorer Version 21H1 (OS Build 19043.1706)
### What are the browser driver(s) and version(s) where you see this issue?
IEDriverServer.exe 4.0.0.0 (64-bit)
### Are you using Selenium Grid?
_No response_
|
defect
|
open selenium invalidstateexception was occurred using ie what happened open selenium invalidstateexception was occurred using ie for details and reproduction methods i wonder if you check items after how can we reproduce the issue this exception is occured at internetexplorerdriver is not occured at edgedriver i test this 👆 change to edgedriver class from internetexplorerdriver class test has all passed sorry for my poor english thank you how can we reproduce the issue shell the implementation is like this private readonly iwebdriver webdriver selenium process run class field public ctor string webdriverdirectorypath get webdriver directory absolute path from ctor argument webdriver new internetexplorerdriver webdriverdirectorypath public void init this method is navigating method for looping invoke method page this method called by other class using foreach public void invoke testcase testcase webdriver findelement by id input html tag id clear error has occurred this point webdriver findelement by id input html tag id sendkeys testcase value var selectelement new selectelement webdriver findelement by id select html tag id selectelement selectbyindex testcase selectindex this method not work however using for loop and sendkey keys arrowdown is work that👆 code runnning at xunit testing cs var testcases new new testcase hoge yamada new testcase fuga tanaka new testcase piyo suzuki using var invoker new seleniuminvoker invoker init var invoker invoke testcases assert equal ok checkresult assert equal yamada customername assert empty errormessages var invoker getresult testcases assert equal ng checkresult assert equal tanaka customername assert equal invalid code errormessages single var invoker getresult testcases assert equal ng checkresult assert equal suzuki customername assert equal no account errormessages single relevant log output shell xunit failed message openqa selenium invalidelementstateexception a javascript error was encountered clearing the element the driver assumes this is because the element is hidden disabled or read only and it must not be to clear the element at openqa selenium webdriver unpackandthrowonerror response errorresponse at openqa selenium webdriver execute string drivercommandtoexecute dictionary parameters at openqa selenium webdriver internalexecute string drivercommandtoexecute dictionary parameters at openqa selenium webelement execute string commandtoexecute dictionary parameters at openqa selenium webelement clear operating system windows version os build selenium version c selenium support selenium webdriver what are the browser s and version s where you see this issue internet explorer version os build what are the browser driver s and version s where you see this issue iedriverserver exe bit are you using selenium grid no response
| 1
|
36,852
| 8,167,589,516
|
IssuesEvent
|
2018-08-26 00:57:07
|
idaholab/raven
|
https://api.github.com/repos/idaholab/raven
|
opened
|
codecoverage is months old
|
defect devel master priority_critical
|
--------
Issue Description
--------
##### What did you expect to see happen?
The code coverage script to be functional.
The code coverage is months old since the check_coverage.sh is broken.
This is causing the master branch to be red since a while.
##### What did you see instead?
The code coverage script is currently broken
##### Do you have a suggested fix for the development team?
##### Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [ ] 1. Is it tagged with a type: defect or improvement?
- [ ] 2. Is it tagged with a priority: critical, normal or minor?
- [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [ ] 1. If the issue is a defect, is the defect fixed?
- [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [ ] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)?
- [ ] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
|
1.0
|
codecoverage is months old - --------
Issue Description
--------
##### What did you expect to see happen?
The code coverage script to be functional.
The code coverage is months old since the check_coverage.sh is broken.
This is causing the master branch to be red since a while.
##### What did you see instead?
The code coverage script is currently broken
##### Do you have a suggested fix for the development team?
##### Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [ ] 1. Is it tagged with a type: defect or improvement?
- [ ] 2. Is it tagged with a priority: critical, normal or minor?
- [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [ ] 1. If the issue is a defect, is the defect fixed?
- [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [ ] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)?
- [ ] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
|
defect
|
codecoverage is months old issue description what did you expect to see happen the code coverage script to be functional the code coverage is months old since the check coverage sh is broken this is causing the master branch to be red since a while what did you see instead the code coverage script is currently broken do you have a suggested fix for the development team please attach the input file s that generate this error the simpler the input the faster we can find the issue for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or improvement is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest stable branch if yes is there any issue tagged with stable create if needed if the issue is being closed without a merge request has an explanation of why it is being closed been provided
| 1
|
431
| 2,535,850,448
|
IssuesEvent
|
2015-01-26 08:28:05
|
localwiki/localwiki
|
https://api.github.com/repos/localwiki/localwiki
|
closed
|
django-taggit not quite working with versionutils
|
defect versioning
|
We're not using django-taggit, but it's worth noting that the following doesn't work right:
```
class Page(models.Model):
...
tags = TaggableManager()
versioning.register(Page)
versioning.register(Tag)
p_h = p.versions.most_recent()
p_h.tags.all() # <-- this will always just display an empty set.
```
Everything works right for non-historical instances. And the Tag object gets versioned correctly.
|
1.0
|
django-taggit not quite working with versionutils - We're not using django-taggit, but it's worth noting that the following doesn't work right:
```
class Page(models.Model):
...
tags = TaggableManager()
versioning.register(Page)
versioning.register(Tag)
p_h = p.versions.most_recent()
p_h.tags.all() # <-- this will always just display an empty set.
```
Everything works right for non-historical instances. And the Tag object gets versioned correctly.
|
defect
|
django taggit not quite working with versionutils we re not using django taggit but it s worth noting that the following doesn t work right class page models model tags taggablemanager versioning register page versioning register tag p h p versions most recent p h tags all this will always just display an empty set everything works right for non historical instances and the tag object gets versioned correctly
| 1
|
77,244
| 26,875,431,888
|
IssuesEvent
|
2023-02-05 00:38:46
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
closed
|
BUG: numerical issues for cdf/ppf of the betaprime distribution
|
defect scipy.stats
|
### Describe your issue.
The CDF of `stats.betaprime` loses precision if relatively small values of `b` are used.
```
from scipy import stats
from scipy import special as sc
a, b = 0.05, 0.1
x = 1e22
stats.betaprime.cdf(x, a, b) # 1.0
sc.betainc(a, b, x/(1.+x)) # 1.0, this is used as the CDF
```
The correct value is ~0.997881:
```
import mpmath as mp
mp.dps = 25
a, b = mp.mpf('0.05'), mp.mpf('0.1')
x = mp.mpf('1e22')
y = mp.fdiv(x, mp.fadd(mp.mpf('1.0'), x, dps=25), dps=25)
mp.betainc(a, b, 0.0, y, regularized=True) # mpf('0.99788115003523326')
```
To compute the correct value without increased precision, one needs to do the following (see https://dlmf.nist.gov/8.17, Eq 8.17.4):
`stats.beta.sf(1/(1+x), b, a) # 0.9978811466052919`
Similarly, the PPF added in #17562 will run into a similar issue for such values. If we continue the example for above:
```
a, b = 0.05, 0.1
x = 1e22
p = stats.beta.sf(1/(1+x), b, a)
r = sc.betaincinv(a, b, p)
r/(1-r) # 9007199254740991.0, this is the ppf value computed in 17562
r/(1-r) / x # 9.007199254740991e-07, the value is way too small
```
The fix is to solve `stats.beta.sf(1/(1+x), b, a)` for `x`:
`
1/stats.beta.isf(p, b, a) - 1 # 1.0000000000001254e+22
`
### Reproducing Code Example
```python
see above
```
### Error message
```shell
no error, just inaccurate output
```
### SciPy/NumPy/Python version information
1.9.3 1.22.4 sys.version_info(major=3, minor=9, micro=7, releaselevel='final', serial=0)
|
1.0
|
BUG: numerical issues for cdf/ppf of the betaprime distribution - ### Describe your issue.
The CDF of `stats.betaprime` loses precision if relatively small values of `b` are used.
```
from scipy import stats
from scipy import special as sc
a, b = 0.05, 0.1
x = 1e22
stats.betaprime.cdf(x, a, b) # 1.0
sc.betainc(a, b, x/(1.+x)) # 1.0, this is used as the CDF
```
The correct value is ~0.997881:
```
import mpmath as mp
mp.dps = 25
a, b = mp.mpf('0.05'), mp.mpf('0.1')
x = mp.mpf('1e22')
y = mp.fdiv(x, mp.fadd(mp.mpf('1.0'), x, dps=25), dps=25)
mp.betainc(a, b, 0.0, y, regularized=True) # mpf('0.99788115003523326')
```
To compute the correct value without increased precision, one needs to do the following (see https://dlmf.nist.gov/8.17, Eq 8.17.4):
`stats.beta.sf(1/(1+x), b, a) # 0.9978811466052919`
Similarly, the PPF added in #17562 will run into a similar issue for such values. If we continue the example for above:
```
a, b = 0.05, 0.1
x = 1e22
p = stats.beta.sf(1/(1+x), b, a)
r = sc.betaincinv(a, b, p)
r/(1-r) # 9007199254740991.0, this is the ppf value computed in 17562
r/(1-r) / x # 9.007199254740991e-07, the value is way too small
```
The fix is to solve `stats.beta.sf(1/(1+x), b, a)` for `x`:
`
1/stats.beta.isf(p, b, a) - 1 # 1.0000000000001254e+22
`
### Reproducing Code Example
```python
see above
```
### Error message
```shell
no error, just inaccurate output
```
### SciPy/NumPy/Python version information
1.9.3 1.22.4 sys.version_info(major=3, minor=9, micro=7, releaselevel='final', serial=0)
|
defect
|
bug numerical issues for cdf ppf of the betaprime distribution describe your issue the cdf of stats betaprime loses precision if relatively small values of b are used from scipy import stats from scipy import special as sc a b x stats betaprime cdf x a b sc betainc a b x x this is used as the cdf the correct value is import mpmath as mp mp dps a b mp mpf mp mpf x mp mpf y mp fdiv x mp fadd mp mpf x dps dps mp betainc a b y regularized true mpf to compute the correct value without increased precision one needs to do the following see eq stats beta sf x b a similarly the ppf added in will run into a similar issue for such values if we continue the example for above a b x p stats beta sf x b a r sc betaincinv a b p r r this is the ppf value computed in r r x the value is way too small the fix is to solve stats beta sf x b a for x stats beta isf p b a reproducing code example python see above error message shell no error just inaccurate output scipy numpy python version information sys version info major minor micro releaselevel final serial
| 1
|
29,730
| 5,845,704,384
|
IssuesEvent
|
2017-05-10 14:45:08
|
OpenMS/OpenMS
|
https://api.github.com/repos/OpenMS/OpenMS
|
opened
|
Invalid mzIdentML created from idXML with missing proteins
|
defect
|
In idXML (or internally in OpenMS) it is possible to have peptides without protein associations (e.g. from the search engine or due to filtering). However, in mzIdentML this is no longer allowed from version 1.1.0 on - which is the current version that OpenMS writes. If peptides without associated proteins are present in an idXML file, converting the file to mzid with IDFileConverter will produce an invalid result, without any warning or error. Running "FileInfo -v" on the output file will give error messages like this:
`Validation error in file '...' line ... column 34: element 'cvParam' is not allowed for content model '(PeptideEvidenceRef+,Fragmentation?,(cvParam|userParam))'`
See also this comment and the following ones: https://github.com/OpenMS/OpenMS/pull/2528#issuecomment-299894675
|
1.0
|
Invalid mzIdentML created from idXML with missing proteins - In idXML (or internally in OpenMS) it is possible to have peptides without protein associations (e.g. from the search engine or due to filtering). However, in mzIdentML this is no longer allowed from version 1.1.0 on - which is the current version that OpenMS writes. If peptides without associated proteins are present in an idXML file, converting the file to mzid with IDFileConverter will produce an invalid result, without any warning or error. Running "FileInfo -v" on the output file will give error messages like this:
`Validation error in file '...' line ... column 34: element 'cvParam' is not allowed for content model '(PeptideEvidenceRef+,Fragmentation?,(cvParam|userParam))'`
See also this comment and the following ones: https://github.com/OpenMS/OpenMS/pull/2528#issuecomment-299894675
|
defect
|
invalid mzidentml created from idxml with missing proteins in idxml or internally in openms it is possible to have peptides without protein associations e g from the search engine or due to filtering however in mzidentml this is no longer allowed from version on which is the current version that openms writes if peptides without associated proteins are present in an idxml file converting the file to mzid with idfileconverter will produce an invalid result without any warning or error running fileinfo v on the output file will give error messages like this validation error in file line column element cvparam is not allowed for content model peptideevidenceref fragmentation cvparam userparam see also this comment and the following ones
| 1
|
63,049
| 17,358,250,464
|
IssuesEvent
|
2021-07-29 16:51:29
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
When 1:1 video call is full screen, the dialpad and "three dots" buttons don't work
|
A-VoIP P2 T-Defect X-Needs-Investigation
|
### Description
When 1:1 video call is full screen, the dialpad and "three dots" buttons don't work
### Steps to reproduce
- start a video call
- go full screen (top right arrows)
- notice the controls work, but the two buttons in the bottom corners don't work
### Version information
<!-- IMPORTANT: please answer the following questions, to help us narrow down the problem -->
Element Nightly version: 2021050501
macos
|
1.0
|
When 1:1 video call is full screen, the dialpad and "three dots" buttons don't work - ### Description
When 1:1 video call is full screen, the dialpad and "three dots" buttons don't work
### Steps to reproduce
- start a video call
- go full screen (top right arrows)
- notice the controls work, but the two buttons in the bottom corners don't work
### Version information
<!-- IMPORTANT: please answer the following questions, to help us narrow down the problem -->
Element Nightly version: 2021050501
macos
|
defect
|
when video call is full screen the dialpad and three dots buttons don t work description when video call is full screen the dialpad and three dots buttons don t work steps to reproduce start a video call go full screen top right arrows notice the controls work but the two buttons in the bottom corners don t work version information element nightly version macos
| 1
|
58,456
| 16,542,689,457
|
IssuesEvent
|
2021-05-27 18:59:50
|
idaholab/moose
|
https://api.github.com/repos/idaholab/moose
|
closed
|
UserObjectInterface::hasUserObject<T> does not check type
|
C: MOOSE P: normal T: defect
|
## Bug Description
`UserObjectInterface::hasUserObject<T>` does not see if the base UO is of type T.
## Steps to Reproduce
See the method.
## Impact
Method does not perform the intended behavior.
|
1.0
|
UserObjectInterface::hasUserObject<T> does not check type - ## Bug Description
`UserObjectInterface::hasUserObject<T>` does not see if the base UO is of type T.
## Steps to Reproduce
See the method.
## Impact
Method does not perform the intended behavior.
|
defect
|
userobjectinterface hasuserobject does not check type bug description userobjectinterface hasuserobject does not see if the base uo is of type t steps to reproduce see the method impact method does not perform the intended behavior
| 1
|
180,990
| 21,630,924,595
|
IssuesEvent
|
2022-05-05 09:37:14
|
Nivaskumark/packages_apps_settings_A10_r33_CVE-2020-0188
|
https://api.github.com/repos/Nivaskumark/packages_apps_settings_A10_r33_CVE-2020-0188
|
opened
|
CVE-2020-0394 (High) detected in Settingsandroid-10.0.0_r33
|
security vulnerability
|
## CVE-2020-0394 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Settingsandroid-10.0.0_r33</b></p></summary>
<p>
<p>Library home page: <a href=https://android.googlesource.com/platform/packages/apps/Settings>https://android.googlesource.com/platform/packages/apps/Settings</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Nivaskumark/packages_apps_settings_A10_r33_CVE-2020-0188/commit/f3df08e4562c757ffb3a076c5898906fdc1afde6">f3df08e4562c757ffb3a076c5898906fdc1afde6</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/src/com/android/settings/bluetooth/BluetoothPairingDialog.java</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In onCreate of BluetoothPairingDialog.java, there is a possible tapjacking vector due to an insecure default value. This could lead to local escalation of privilege and untrusted devices accessing contact lists with no additional execution privileges needed. User interaction is needed for exploitation.Product: AndroidVersions: Android-8.0 Android-8.1 Android-9 Android-10 Android-11Android ID: A-155648639
<p>Publish Date: 2020-09-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-0394>CVE-2020-0394</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://source.android.com/security/bulletin/2020-09-01">https://source.android.com/security/bulletin/2020-09-01</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: android-8.0.0_r50,android-8.1.0_r80,android-9.0.0_r60</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-0394 (High) detected in Settingsandroid-10.0.0_r33 - ## CVE-2020-0394 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Settingsandroid-10.0.0_r33</b></p></summary>
<p>
<p>Library home page: <a href=https://android.googlesource.com/platform/packages/apps/Settings>https://android.googlesource.com/platform/packages/apps/Settings</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Nivaskumark/packages_apps_settings_A10_r33_CVE-2020-0188/commit/f3df08e4562c757ffb3a076c5898906fdc1afde6">f3df08e4562c757ffb3a076c5898906fdc1afde6</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/src/com/android/settings/bluetooth/BluetoothPairingDialog.java</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In onCreate of BluetoothPairingDialog.java, there is a possible tapjacking vector due to an insecure default value. This could lead to local escalation of privilege and untrusted devices accessing contact lists with no additional execution privileges needed. User interaction is needed for exploitation.Product: AndroidVersions: Android-8.0 Android-8.1 Android-9 Android-10 Android-11Android ID: A-155648639
<p>Publish Date: 2020-09-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-0394>CVE-2020-0394</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://source.android.com/security/bulletin/2020-09-01">https://source.android.com/security/bulletin/2020-09-01</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: android-8.0.0_r50,android-8.1.0_r80,android-9.0.0_r60</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in settingsandroid cve high severity vulnerability vulnerable library settingsandroid library home page a href found in head commit a href found in base branch master vulnerable source files src com android settings bluetooth bluetoothpairingdialog java vulnerability details in oncreate of bluetoothpairingdialog java there is a possible tapjacking vector due to an insecure default value this could lead to local escalation of privilege and untrusted devices accessing contact lists with no additional execution privileges needed user interaction is needed for exploitation product androidversions android android android android android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android android android step up your open source security game with whitesource
| 0
|
41,057
| 10,279,024,021
|
IssuesEvent
|
2019-08-25 19:18:40
|
ase379/gpprofile2017
|
https://api.github.com/repos/ase379/gpprofile2017
|
closed
|
Huge prf might crash under 64 Bit
|
defect
|
Using a huge prf file (> 2GB size) sometimes crashes. After a restart, the load works.
|
1.0
|
Huge prf might crash under 64 Bit - Using a huge prf file (> 2GB size) sometimes crashes. After a restart, the load works.
|
defect
|
huge prf might crash under bit using a huge prf file size sometimes crashes after a restart the load works
| 1
|
1,355
| 2,603,840,141
|
IssuesEvent
|
2015-02-24 18:14:24
|
chrsmith/nishazi6
|
https://api.github.com/repos/chrsmith/nishazi6
|
opened
|
沈阳包皮内长小水泡
|
auto-migrated Priority-Medium Type-Defect
|
```
沈阳包皮内长小水泡〓沈陽軍區政治部醫院性病〓TEL:024-3102
3308〓成立于1946年,68年專注于性傳播疾病的研究和治療。位�
��沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的�
��史悠久、設備精良、技術權威、專家云集,是預防、保健、
醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等��
�隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東�
��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍
后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二��
�功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:12
|
1.0
|
沈阳包皮内长小水泡 - ```
沈阳包皮内长小水泡〓沈陽軍區政治部醫院性病〓TEL:024-3102
3308〓成立于1946年,68年專注于性傳播疾病的研究和治療。位�
��沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的�
��史悠久、設備精良、技術權威、專家云集,是預防、保健、
醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等��
�隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東�
��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍
后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二��
�功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:12
|
defect
|
沈阳包皮内长小水泡 沈阳包皮内长小水泡〓沈陽軍區政治部醫院性病〓tel: 〓 , 。位� �� 。是一所與新中國同建立共輝煌的� ��史悠久、設備精良、技術權威、專家云集,是預防、保健、 醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等�� �隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東� ��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍 后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二�� �功。 original issue reported on code google com by gmail com on jun at
| 1
|
102,600
| 8,850,246,949
|
IssuesEvent
|
2019-01-08 12:43:13
|
Kademi/kademi-dev
|
https://api.github.com/repos/Kademi/kademi-dev
|
closed
|
Update Icons
|
Ready to Test - Dev
|
Some Kcom2 components are using Bootstrap Glyphicon, it should be converted to font awesome icon
https://github.com/Kademi/projects/issues/925
|
1.0
|
Update Icons - Some Kcom2 components are using Bootstrap Glyphicon, it should be converted to font awesome icon
https://github.com/Kademi/projects/issues/925
|
non_defect
|
update icons some components are using bootstrap glyphicon it should be converted to font awesome icon
| 0
|
626,006
| 19,783,643,178
|
IssuesEvent
|
2022-01-18 02:12:07
|
Baystation12/Baystation12
|
https://api.github.com/repos/Baystation12/Baystation12
|
closed
|
Shield generator not clear for user when it doesn't have power
|
Oversight :bug: Priority: Low BINGO!
|
#### Description of issue
When the shield generator does not have a cable dot-knot under it, it won't receive power. However, rather than act unpowered or rather than prompting the user, it will attempt to start, immediately fail, and show "Recovering..."
#### Difference between expected and actual behavior
Actual:
1. Shield interface can be interacted with when unpowered.
2. Shield can be turned on when unpowered.
3. Shield instantly fails when unpowered.
4. Shield will appear to be damaged to the point of failure and turn off.
Expected:
1. Shield will prompt user in redtext, show an interface error message that blocks the shield configuration, or show an "INSUFFICIENT POWER" state rather than a "RECOVERING" state in UI.
#### Steps to reproduce
Place shield generator on no wires, a straight wire, or a curved wire - anything without a wire dot in the center of the tile.
#### Specific information for locating
N/A
#### Length of time in which bug has been known to occur
Forever.
#### Client version, Server revision & Game ID
Client Version: 511
Server Revision: 326f3982b969cc893743f7dbf88f125bf1830860 - dev -
Game ID: bQ2-dege
Current map: SEV Torch
#### Issue bing
Please check whatever applies. More checkboxes checked increase your chances of the issue being looked at sooner.
- [X] Issue could be reproduced at least once
- [X] Issue could be reproduced by different players
- [X] Issue could be reproduced in multiple rounds
- [X] Issue happened in a recent (less than 7 days ago) round
- [X] [Couldn't find an existing issue about this](https://github.com/Baystation12/Baystation12/issues)
|
1.0
|
Shield generator not clear for user when it doesn't have power - #### Description of issue
When the shield generator does not have a cable dot-knot under it, it won't receive power. However, rather than act unpowered or rather than prompting the user, it will attempt to start, immediately fail, and show "Recovering..."
#### Difference between expected and actual behavior
Actual:
1. Shield interface can be interacted with when unpowered.
2. Shield can be turned on when unpowered.
3. Shield instantly fails when unpowered.
4. Shield will appear to be damaged to the point of failure and turn off.
Expected:
1. Shield will prompt user in redtext, show an interface error message that blocks the shield configuration, or show an "INSUFFICIENT POWER" state rather than a "RECOVERING" state in UI.
#### Steps to reproduce
Place shield generator on no wires, a straight wire, or a curved wire - anything without a wire dot in the center of the tile.
#### Specific information for locating
N/A
#### Length of time in which bug has been known to occur
Forever.
#### Client version, Server revision & Game ID
Client Version: 511
Server Revision: 326f3982b969cc893743f7dbf88f125bf1830860 - dev -
Game ID: bQ2-dege
Current map: SEV Torch
#### Issue bing
Please check whatever applies. More checkboxes checked increase your chances of the issue being looked at sooner.
- [X] Issue could be reproduced at least once
- [X] Issue could be reproduced by different players
- [X] Issue could be reproduced in multiple rounds
- [X] Issue happened in a recent (less than 7 days ago) round
- [X] [Couldn't find an existing issue about this](https://github.com/Baystation12/Baystation12/issues)
|
non_defect
|
shield generator not clear for user when it doesn t have power description of issue when the shield generator does not have a cable dot knot under it it won t receive power however rather than act unpowered or rather than prompting the user it will attempt to start immediately fail and show recovering difference between expected and actual behavior actual shield interface can be interacted with when unpowered shield can be turned on when unpowered shield instantly fails when unpowered shield will appear to be damaged to the point of failure and turn off expected shield will prompt user in redtext show an interface error message that blocks the shield configuration or show an insufficient power state rather than a recovering state in ui steps to reproduce place shield generator on no wires a straight wire or a curved wire anything without a wire dot in the center of the tile specific information for locating n a length of time in which bug has been known to occur forever client version server revision game id client version server revision dev game id dege current map sev torch issue bing please check whatever applies more checkboxes checked increase your chances of the issue being looked at sooner issue could be reproduced at least once issue could be reproduced by different players issue could be reproduced in multiple rounds issue happened in a recent less than days ago round
| 0
|
50,250
| 13,187,401,370
|
IssuesEvent
|
2020-08-13 03:17:59
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
closed
|
tableio and generated nugen files (Trac #386)
|
Migrated from Trac defect other
|
Hi Alex,
you asked me to write a ticket about my problem with tableio and generated nugen files.
I understand that I am not supposed to submit tickets as user "icecube", but it seems as if I do not have the needed permissions. I apologize for any inconvenience in this context.
You find a copy of the description I have sent to the dataclass mailing list below. I just have to add, that the suggestion of Jakob works. By simply filtering NO_INTERACTION events by using the neutrino_event_type of the I3MCTree I could avoid the problem.
However there is one minor inconsistency, which I do not understand. As described there have been many generated MC files, which have been processed without any problem, although containing NO_INTERACTION files.
Thanks for looking at it.
Best
Hans
===============================
Hi dataclass,
I am currently experiencing some troubles with Tableio (rootwriter) in the context of generated nugen files (5000 frames each). The problem does not occur when running on level2 processed files (i.e. IC79 datasets 6457, 6005).
The generated datasets I tried:
IC79: http://internal.icecube.wisc.edu/simulation/dataset/7747
IC80+DC6: http://internal.icecube.wisc.edu/simulation/dataset/2476
some files are processed by rootwriter smoothly, others produce the following error message:
/home/hmniederhausen/icerec/src/tableio/public/tableio/I3TableRow.h:161: FATAL: Tried to set unknown column 'EnergyLost'
/home/hmniederhausen/icerec/src/tableio/private/tableio/I3TableService.cxx:222: FATAL: Table 'I3MCTree' is still connected, which means that the output file was never properly closed. This is a BUG!
This message appears i.e. when running on file
/data/sim/IceCube/2010/generated/neutrino-generator/2476/00000-00999/nugen_nue_ic80_dc6.002476.000700.i3.gz
but not on i.e.
/data/sim/IceCube/2010/generated/neutrino-generator/2476/00000-00999/nugen_nue_ic80_dc6.002476.000800.i3.gz
Same story for dataset 7747:
/home/hmniederhausen/icerec/src/dataio/private/dataio/I3Reader.cxx:167: INFO : Opened file /data/sim/IceCube/2010/generated/neutrino-
generator/7747/00000-00999/nugen_nue_IC79.007747.000080.i3.gz
/home/hmniederhausen/icerec/src/tableio/public/tableio/I3TableRow.h:161: FATAL: Tried to set unknown column 'EnergyLost'
/home/hmniederhausen/icerec/src/tableio/private/tableio/I3TableService.cxx:222: FATAL: Table 'I3MCTree' is still connected, which means that the output file was never properly closed. This is a BUG!
whereas i.e. the files 90-99 work fine:
/home/hmniederhausen/icerec/src/dataio/private/dataio/I3Reader.cxx:167: INFO : Opened file /data/sim/IceCube/2010/generated/neutrino-
generator/7747/00000-00999/nugen_nue_IC79.007747.000099.i3.gz
I3Tray finishing...
writer: 50000 calls to physics 225.93s user 0.65s system
You can have a look at my bare tableio scripts at
/home/hmniederhausen/scripts/2476_rootwriter.py
/home/hmniederhausen/scripts/7747_generated_rootwriter.py -f insert_integer_in_range(0,10) // -f 8 gives error, -f 9 works fine
Am I doing something wrong? I am wondering what makes the difference between generated and level2 processed files when it comes down to tableio. However I need to run some root scripts on the generated 7747 to debug some problems. I would be glad if somebody has an idea.
Thanks in advance.
Cheers,
Hans
<details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/386
, reported by hans.niederhausen and owned by olivas_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-03-06T20:02:40",
"description": "Hi Alex,\n\nyou asked me to write a ticket about my problem with tableio and generated nugen files. \n\nI understand that I am not supposed to submit tickets as user \"icecube\", but it seems as if I do not have the needed permissions. I apologize for any inconvenience in this context.\n\nYou find a copy of the description I have sent to the dataclass mailing list below. I just have to add, that the suggestion of Jakob works. By simply filtering NO_INTERACTION events by using the neutrino_event_type of the I3MCTree I could avoid the problem.\n\nHowever there is one minor inconsistency, which I do not understand. As described there have been many generated MC files, which have been processed without any problem, although containing NO_INTERACTION files.\n\nThanks for looking at it.\n\nBest\nHans\n\n===============================\n\nHi dataclass,\n\nI am currently experiencing some troubles with Tableio (rootwriter) in the context of generated nugen files (5000 frames each). The problem does not occur when running on level2 processed files (i.e. IC79 datasets 6457, 6005).\n\nThe generated datasets I tried:\n IC79: http://internal.icecube.wisc.edu/simulation/dataset/7747\n IC80+DC6: http://internal.icecube.wisc.edu/simulation/dataset/2476\n\nsome files are processed by rootwriter smoothly, others produce the following error message: \n\t/home/hmniederhausen/icerec/src/tableio/public/tableio/I3TableRow.h:161: FATAL: Tried to set unknown column 'EnergyLost'\n\t/home/hmniederhausen/icerec/src/tableio/private/tableio/I3TableService.cxx:222: FATAL: Table 'I3MCTree' is still connected, which means that the output file was \tnever properly closed. This is a BUG! \n\nThis message appears i.e. when running on file \n\t/data/sim/IceCube/2010/generated/neutrino-generator/2476/00000-00999/nugen_nue_ic80_dc6.002476.000700.i3.gz \n\nbut not on i.e.\n\t/data/sim/IceCube/2010/generated/neutrino-generator/2476/00000-00999/nugen_nue_ic80_dc6.002476.000800.i3.gz \n\nSame story for dataset 7747:\n\t/home/hmniederhausen/icerec/src/dataio/private/dataio/I3Reader.cxx:167: INFO : Opened file /data/sim/IceCube/2010/generated/neutrino-\t\t\t \n\tgenerator/7747/00000-00999/nugen_nue_IC79.007747.000080.i3.gz\n\t/home/hmniederhausen/icerec/src/tableio/public/tableio/I3TableRow.h:161: FATAL: Tried to set unknown column 'EnergyLost'\n\t/home/hmniederhausen/icerec/src/tableio/private/tableio/I3TableService.cxx:222: FATAL: Table 'I3MCTree' is still connected, which means that the output file was \tnever properly closed. This is a BUG!\n\nwhereas i.e. the files 90-99 work fine:\n\t/home/hmniederhausen/icerec/src/dataio/private/dataio/I3Reader.cxx:167: INFO : Opened file /data/sim/IceCube/2010/generated/neutrino-\t\n\tgenerator/7747/00000-00999/nugen_nue_IC79.007747.000099.i3.gz\n\tI3Tray finishing...\n writer: 50000 calls to physics 225.93s user 0.65s system\n\nYou can have a look at my bare tableio scripts at \n/home/hmniederhausen/scripts/2476_rootwriter.py \n/home/hmniederhausen/scripts/7747_generated_rootwriter.py -f insert_integer_in_range(0,10) // -f 8 gives error, -f 9 works fine\n\n\nAm I doing something wrong? I am wondering what makes the difference between generated and level2 processed files when it comes down to tableio. However I need to run some root scripts on the generated 7747 to debug some problems. I would be glad if somebody has an idea.\n\nThanks in advance.\n\nCheers,\nHans",
"reporter": "hans.niederhausen",
"cc": "",
"resolution": "wontfix",
"_ts": "1425672160951089",
"component": "other",
"summary": "tableio and generated nugen files",
"priority": "minor",
"keywords": "tableio generated",
"time": "2012-03-28T02:06:12",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
tableio and generated nugen files (Trac #386) - Hi Alex,
you asked me to write a ticket about my problem with tableio and generated nugen files.
I understand that I am not supposed to submit tickets as user "icecube", but it seems as if I do not have the needed permissions. I apologize for any inconvenience in this context.
You find a copy of the description I have sent to the dataclass mailing list below. I just have to add, that the suggestion of Jakob works. By simply filtering NO_INTERACTION events by using the neutrino_event_type of the I3MCTree I could avoid the problem.
However there is one minor inconsistency, which I do not understand. As described there have been many generated MC files, which have been processed without any problem, although containing NO_INTERACTION files.
Thanks for looking at it.
Best
Hans
===============================
Hi dataclass,
I am currently experiencing some troubles with Tableio (rootwriter) in the context of generated nugen files (5000 frames each). The problem does not occur when running on level2 processed files (i.e. IC79 datasets 6457, 6005).
The generated datasets I tried:
IC79: http://internal.icecube.wisc.edu/simulation/dataset/7747
IC80+DC6: http://internal.icecube.wisc.edu/simulation/dataset/2476
some files are processed by rootwriter smoothly, others produce the following error message:
/home/hmniederhausen/icerec/src/tableio/public/tableio/I3TableRow.h:161: FATAL: Tried to set unknown column 'EnergyLost'
/home/hmniederhausen/icerec/src/tableio/private/tableio/I3TableService.cxx:222: FATAL: Table 'I3MCTree' is still connected, which means that the output file was never properly closed. This is a BUG!
This message appears i.e. when running on file
/data/sim/IceCube/2010/generated/neutrino-generator/2476/00000-00999/nugen_nue_ic80_dc6.002476.000700.i3.gz
but not on i.e.
/data/sim/IceCube/2010/generated/neutrino-generator/2476/00000-00999/nugen_nue_ic80_dc6.002476.000800.i3.gz
Same story for dataset 7747:
/home/hmniederhausen/icerec/src/dataio/private/dataio/I3Reader.cxx:167: INFO : Opened file /data/sim/IceCube/2010/generated/neutrino-
generator/7747/00000-00999/nugen_nue_IC79.007747.000080.i3.gz
/home/hmniederhausen/icerec/src/tableio/public/tableio/I3TableRow.h:161: FATAL: Tried to set unknown column 'EnergyLost'
/home/hmniederhausen/icerec/src/tableio/private/tableio/I3TableService.cxx:222: FATAL: Table 'I3MCTree' is still connected, which means that the output file was never properly closed. This is a BUG!
whereas i.e. the files 90-99 work fine:
/home/hmniederhausen/icerec/src/dataio/private/dataio/I3Reader.cxx:167: INFO : Opened file /data/sim/IceCube/2010/generated/neutrino-
generator/7747/00000-00999/nugen_nue_IC79.007747.000099.i3.gz
I3Tray finishing...
writer: 50000 calls to physics 225.93s user 0.65s system
You can have a look at my bare tableio scripts at
/home/hmniederhausen/scripts/2476_rootwriter.py
/home/hmniederhausen/scripts/7747_generated_rootwriter.py -f insert_integer_in_range(0,10) // -f 8 gives error, -f 9 works fine
Am I doing something wrong? I am wondering what makes the difference between generated and level2 processed files when it comes down to tableio. However I need to run some root scripts on the generated 7747 to debug some problems. I would be glad if somebody has an idea.
Thanks in advance.
Cheers,
Hans
<details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/386
, reported by hans.niederhausen and owned by olivas_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-03-06T20:02:40",
"description": "Hi Alex,\n\nyou asked me to write a ticket about my problem with tableio and generated nugen files. \n\nI understand that I am not supposed to submit tickets as user \"icecube\", but it seems as if I do not have the needed permissions. I apologize for any inconvenience in this context.\n\nYou find a copy of the description I have sent to the dataclass mailing list below. I just have to add, that the suggestion of Jakob works. By simply filtering NO_INTERACTION events by using the neutrino_event_type of the I3MCTree I could avoid the problem.\n\nHowever there is one minor inconsistency, which I do not understand. As described there have been many generated MC files, which have been processed without any problem, although containing NO_INTERACTION files.\n\nThanks for looking at it.\n\nBest\nHans\n\n===============================\n\nHi dataclass,\n\nI am currently experiencing some troubles with Tableio (rootwriter) in the context of generated nugen files (5000 frames each). The problem does not occur when running on level2 processed files (i.e. IC79 datasets 6457, 6005).\n\nThe generated datasets I tried:\n IC79: http://internal.icecube.wisc.edu/simulation/dataset/7747\n IC80+DC6: http://internal.icecube.wisc.edu/simulation/dataset/2476\n\nsome files are processed by rootwriter smoothly, others produce the following error message: \n\t/home/hmniederhausen/icerec/src/tableio/public/tableio/I3TableRow.h:161: FATAL: Tried to set unknown column 'EnergyLost'\n\t/home/hmniederhausen/icerec/src/tableio/private/tableio/I3TableService.cxx:222: FATAL: Table 'I3MCTree' is still connected, which means that the output file was \tnever properly closed. This is a BUG! \n\nThis message appears i.e. when running on file \n\t/data/sim/IceCube/2010/generated/neutrino-generator/2476/00000-00999/nugen_nue_ic80_dc6.002476.000700.i3.gz \n\nbut not on i.e.\n\t/data/sim/IceCube/2010/generated/neutrino-generator/2476/00000-00999/nugen_nue_ic80_dc6.002476.000800.i3.gz \n\nSame story for dataset 7747:\n\t/home/hmniederhausen/icerec/src/dataio/private/dataio/I3Reader.cxx:167: INFO : Opened file /data/sim/IceCube/2010/generated/neutrino-\t\t\t \n\tgenerator/7747/00000-00999/nugen_nue_IC79.007747.000080.i3.gz\n\t/home/hmniederhausen/icerec/src/tableio/public/tableio/I3TableRow.h:161: FATAL: Tried to set unknown column 'EnergyLost'\n\t/home/hmniederhausen/icerec/src/tableio/private/tableio/I3TableService.cxx:222: FATAL: Table 'I3MCTree' is still connected, which means that the output file was \tnever properly closed. This is a BUG!\n\nwhereas i.e. the files 90-99 work fine:\n\t/home/hmniederhausen/icerec/src/dataio/private/dataio/I3Reader.cxx:167: INFO : Opened file /data/sim/IceCube/2010/generated/neutrino-\t\n\tgenerator/7747/00000-00999/nugen_nue_IC79.007747.000099.i3.gz\n\tI3Tray finishing...\n writer: 50000 calls to physics 225.93s user 0.65s system\n\nYou can have a look at my bare tableio scripts at \n/home/hmniederhausen/scripts/2476_rootwriter.py \n/home/hmniederhausen/scripts/7747_generated_rootwriter.py -f insert_integer_in_range(0,10) // -f 8 gives error, -f 9 works fine\n\n\nAm I doing something wrong? I am wondering what makes the difference between generated and level2 processed files when it comes down to tableio. However I need to run some root scripts on the generated 7747 to debug some problems. I would be glad if somebody has an idea.\n\nThanks in advance.\n\nCheers,\nHans",
"reporter": "hans.niederhausen",
"cc": "",
"resolution": "wontfix",
"_ts": "1425672160951089",
"component": "other",
"summary": "tableio and generated nugen files",
"priority": "minor",
"keywords": "tableio generated",
"time": "2012-03-28T02:06:12",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
|
defect
|
tableio and generated nugen files trac hi alex you asked me to write a ticket about my problem with tableio and generated nugen files i understand that i am not supposed to submit tickets as user icecube but it seems as if i do not have the needed permissions i apologize for any inconvenience in this context you find a copy of the description i have sent to the dataclass mailing list below i just have to add that the suggestion of jakob works by simply filtering no interaction events by using the neutrino event type of the i could avoid the problem however there is one minor inconsistency which i do not understand as described there have been many generated mc files which have been processed without any problem although containing no interaction files thanks for looking at it best hans hi dataclass i am currently experiencing some troubles with tableio rootwriter in the context of generated nugen files frames each the problem does not occur when running on processed files i e datasets the generated datasets i tried some files are processed by rootwriter smoothly others produce the following error message home hmniederhausen icerec src tableio public tableio h fatal tried to set unknown column energylost home hmniederhausen icerec src tableio private tableio cxx fatal table is still connected which means that the output file was never properly closed this is a bug this message appears i e when running on file data sim icecube generated neutrino generator nugen nue gz but not on i e data sim icecube generated neutrino generator nugen nue gz same story for dataset home hmniederhausen icerec src dataio private dataio cxx info opened file data sim icecube generated neutrino generator nugen nue gz home hmniederhausen icerec src tableio public tableio h fatal tried to set unknown column energylost home hmniederhausen icerec src tableio private tableio cxx fatal table is still connected which means that the output file was never properly closed this is a bug whereas i e the files work fine home hmniederhausen icerec src dataio private dataio cxx info opened file data sim icecube generated neutrino generator nugen nue gz finishing writer calls to physics user system you can have a look at my bare tableio scripts at home hmniederhausen scripts rootwriter py home hmniederhausen scripts generated rootwriter py f insert integer in range f gives error f works fine am i doing something wrong i am wondering what makes the difference between generated and processed files when it comes down to tableio however i need to run some root scripts on the generated to debug some problems i would be glad if somebody has an idea thanks in advance cheers hans migrated from reported by hans niederhausen and owned by olivas json status closed changetime description hi alex n nyou asked me to write a ticket about my problem with tableio and generated nugen files n ni understand that i am not supposed to submit tickets as user icecube but it seems as if i do not have the needed permissions i apologize for any inconvenience in this context n nyou find a copy of the description i have sent to the dataclass mailing list below i just have to add that the suggestion of jakob works by simply filtering no interaction events by using the neutrino event type of the i could avoid the problem n nhowever there is one minor inconsistency which i do not understand as described there have been many generated mc files which have been processed without any problem although containing no interaction files n nthanks for looking at it n nbest nhans n n n nhi dataclass n ni am currently experiencing some troubles with tableio rootwriter in the context of generated nugen files frames each the problem does not occur when running on processed files i e datasets n nthe generated datasets i tried n files are processed by rootwriter smoothly others produce the following error message n t home hmniederhausen icerec src tableio public tableio h fatal tried to set unknown column energylost n t home hmniederhausen icerec src tableio private tableio cxx fatal table is still connected which means that the output file was tnever properly closed this is a bug n nthis message appears i e when running on file n t data sim icecube generated neutrino generator nugen nue gz n nbut not on i e n t data sim icecube generated neutrino generator nugen nue gz n nsame story for dataset n t home hmniederhausen icerec src dataio private dataio cxx info opened file data sim icecube generated neutrino t t t n tgenerator nugen nue gz n t home hmniederhausen icerec src tableio public tableio h fatal tried to set unknown column energylost n t home hmniederhausen icerec src tableio private tableio cxx fatal table is still connected which means that the output file was tnever properly closed this is a bug n nwhereas i e the files work fine n t home hmniederhausen icerec src dataio private dataio cxx info opened file data sim icecube generated neutrino t n tgenerator nugen nue gz n finishing n writer calls to physics user system n nyou can have a look at my bare tableio scripts at n home hmniederhausen scripts rootwriter py n home hmniederhausen scripts generated rootwriter py f insert integer in range f gives error f works fine n n nam i doing something wrong i am wondering what makes the difference between generated and processed files when it comes down to tableio however i need to run some root scripts on the generated to debug some problems i would be glad if somebody has an idea n nthanks in advance n ncheers nhans reporter hans niederhausen cc resolution wontfix ts component other summary tableio and generated nugen files priority minor keywords tableio generated time milestone owner olivas type defect
| 1
|
40,570
| 10,054,793,349
|
IssuesEvent
|
2019-07-22 03:19:26
|
line/armeria
|
https://api.github.com/repos/line/armeria
|
closed
|
`RequestLogAvailabilityException: REQUEST_FIRST_BYTES_TRANSFERRED` when `RequestLogAvailability` is `COMPLETE`
|
defect
|
Reported by @adriancole:
```
WARN 88577 --- [orageComponent}] c.l.a.c.l.RequestLogListenerInvoker : onRequestLog() failed with an exception:com.linecorp.armeria.common.logging.RequestLogAvailabilityException: REQUEST_FIRST_BYTES_TRANSFERRED
at com.linecorp.armeria.common.logging.RequestLog.ensureAvailability(RequestLog.java:120) ~[armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.common.logging.DefaultRequestLog.requestFirstBytesTransferredTimeNanos(DefaultRequestLog.java:446) ~[armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.client.brave.BraveClient.lambda$execute$2(BraveClient.java:126) ~[armeria-brave-0.88.0.jar!/:?]
at com.linecorp.armeria.common.logging.RequestLogListenerInvoker.invokeOnRequestLog(RequestLogListenerInvoker.java:38) [armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.common.logging.DefaultRequestLog.notifyListeners(DefaultRequestLog.java:1036) [armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.common.logging.DefaultRequestLog.updateAvailability(DefaultRequestLog.java:993) [armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.common.logging.DefaultRequestLog.endResponse0(DefaultRequestLog.java:970) [armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.common.logging.DefaultRequestLog.endResponse0(DefaultRequestLog.java:949) [armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.common.logging.DefaultRequestLog.endResponse(DefaultRequestLog.java:935) [armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.internal.ClientUtil.executeWithFallback(ClientUtil.java:50) [armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.client.UserClient.execute(UserClient.java:159) [armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.client.DefaultHttpClient.execute(DefaultHttpClient.java:61) [armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.client.DefaultHttpClient.execute(DefaultHttpClient.java:78) [armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.client.DefaultHttpClient.execute(DefaultHttpClient.java:74) [armeria-0.88.0.jar!/:?]
at zipkin2.elasticsearch.internal.client.HttpCall.sendRequest(HttpCall.java:145) [zipkin-storage-elasticsearch-2.15.1-SNAPSHOT.jar!/:?]
at zipkin2.elasticsearch.internal.client.HttpCall.doExecute(HttpCall.java:101) [zipkin-storage-elasticsearch-2.15.1-SNAPSHOT.jar!/:?]
at zipkin2.Call$Base.execute(Call.java:379) [zipkin-2.15.1-SNAPSHOT.jar!/:?]
at zipkin2.reporter.AsyncReporter$BoundedAsyncReporter.flush(AsyncReporter.java:286) [zipkin-reporter-2.8.4.jar!/:?]
at zipkin2.reporter.AsyncReporter$Builder$1.run(AsyncReporter.java:190) [zipkin-reporter-2.8.4.jar!/:?]
```
In BraveClient.java:
```java
ctx.log().addListener(log -> {
SpanTags.logWireSend(span, log.requestFirstBytesTransferredTimeNanos(), log);
// If the client timed-out the request, we will have never received any response data at all.
if (log.isAvailable(RequestLogAvailability.RESPONSE_FIRST_BYTES_TRANSFERRED)) {
SpanTags.logWireReceive(span, log.responseFirstBytesTransferredTimeNanos(), log);
}
finishSpan(span, log);
}, RequestLogAvailability.COMPLETE);
```
/cc @anuraaga
|
1.0
|
`RequestLogAvailabilityException: REQUEST_FIRST_BYTES_TRANSFERRED` when `RequestLogAvailability` is `COMPLETE` - Reported by @adriancole:
```
WARN 88577 --- [orageComponent}] c.l.a.c.l.RequestLogListenerInvoker : onRequestLog() failed with an exception:com.linecorp.armeria.common.logging.RequestLogAvailabilityException: REQUEST_FIRST_BYTES_TRANSFERRED
at com.linecorp.armeria.common.logging.RequestLog.ensureAvailability(RequestLog.java:120) ~[armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.common.logging.DefaultRequestLog.requestFirstBytesTransferredTimeNanos(DefaultRequestLog.java:446) ~[armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.client.brave.BraveClient.lambda$execute$2(BraveClient.java:126) ~[armeria-brave-0.88.0.jar!/:?]
at com.linecorp.armeria.common.logging.RequestLogListenerInvoker.invokeOnRequestLog(RequestLogListenerInvoker.java:38) [armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.common.logging.DefaultRequestLog.notifyListeners(DefaultRequestLog.java:1036) [armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.common.logging.DefaultRequestLog.updateAvailability(DefaultRequestLog.java:993) [armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.common.logging.DefaultRequestLog.endResponse0(DefaultRequestLog.java:970) [armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.common.logging.DefaultRequestLog.endResponse0(DefaultRequestLog.java:949) [armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.common.logging.DefaultRequestLog.endResponse(DefaultRequestLog.java:935) [armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.internal.ClientUtil.executeWithFallback(ClientUtil.java:50) [armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.client.UserClient.execute(UserClient.java:159) [armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.client.DefaultHttpClient.execute(DefaultHttpClient.java:61) [armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.client.DefaultHttpClient.execute(DefaultHttpClient.java:78) [armeria-0.88.0.jar!/:?]
at com.linecorp.armeria.client.DefaultHttpClient.execute(DefaultHttpClient.java:74) [armeria-0.88.0.jar!/:?]
at zipkin2.elasticsearch.internal.client.HttpCall.sendRequest(HttpCall.java:145) [zipkin-storage-elasticsearch-2.15.1-SNAPSHOT.jar!/:?]
at zipkin2.elasticsearch.internal.client.HttpCall.doExecute(HttpCall.java:101) [zipkin-storage-elasticsearch-2.15.1-SNAPSHOT.jar!/:?]
at zipkin2.Call$Base.execute(Call.java:379) [zipkin-2.15.1-SNAPSHOT.jar!/:?]
at zipkin2.reporter.AsyncReporter$BoundedAsyncReporter.flush(AsyncReporter.java:286) [zipkin-reporter-2.8.4.jar!/:?]
at zipkin2.reporter.AsyncReporter$Builder$1.run(AsyncReporter.java:190) [zipkin-reporter-2.8.4.jar!/:?]
```
In BraveClient.java:
```java
ctx.log().addListener(log -> {
SpanTags.logWireSend(span, log.requestFirstBytesTransferredTimeNanos(), log);
// If the client timed-out the request, we will have never received any response data at all.
if (log.isAvailable(RequestLogAvailability.RESPONSE_FIRST_BYTES_TRANSFERRED)) {
SpanTags.logWireReceive(span, log.responseFirstBytesTransferredTimeNanos(), log);
}
finishSpan(span, log);
}, RequestLogAvailability.COMPLETE);
```
/cc @anuraaga
|
defect
|
requestlogavailabilityexception request first bytes transferred when requestlogavailability is complete reported by adriancole warn c l a c l requestloglistenerinvoker onrequestlog failed with an exception com linecorp armeria common logging requestlogavailabilityexception request first bytes transferred at com linecorp armeria common logging requestlog ensureavailability requestlog java at com linecorp armeria common logging defaultrequestlog requestfirstbytestransferredtimenanos defaultrequestlog java at com linecorp armeria client brave braveclient lambda execute braveclient java at com linecorp armeria common logging requestloglistenerinvoker invokeonrequestlog requestloglistenerinvoker java at com linecorp armeria common logging defaultrequestlog notifylisteners defaultrequestlog java at com linecorp armeria common logging defaultrequestlog updateavailability defaultrequestlog java at com linecorp armeria common logging defaultrequestlog defaultrequestlog java at com linecorp armeria common logging defaultrequestlog defaultrequestlog java at com linecorp armeria common logging defaultrequestlog endresponse defaultrequestlog java at com linecorp armeria internal clientutil executewithfallback clientutil java at com linecorp armeria client userclient execute userclient java at com linecorp armeria client defaulthttpclient execute defaulthttpclient java at com linecorp armeria client defaulthttpclient execute defaulthttpclient java at com linecorp armeria client defaulthttpclient execute defaulthttpclient java at elasticsearch internal client httpcall sendrequest httpcall java at elasticsearch internal client httpcall doexecute httpcall java at call base execute call java at reporter asyncreporter boundedasyncreporter flush asyncreporter java at reporter asyncreporter builder run asyncreporter java in braveclient java java ctx log addlistener log spantags logwiresend span log requestfirstbytestransferredtimenanos log if the client timed out the request we will have never received any response data at all if log isavailable requestlogavailability response first bytes transferred spantags logwirereceive span log responsefirstbytestransferredtimenanos log finishspan span log requestlogavailability complete cc anuraaga
| 1
|
70,967
| 23,391,993,353
|
IssuesEvent
|
2022-08-11 18:47:54
|
vector-im/element-android
|
https://api.github.com/repos/vector-im/element-android
|
opened
|
App Layout: Bottom Sheet
|
T-Defect X-Needs-Design Z-AppLayout
|
### Steps to reproduce
We had several bugs and issues raised during internal testing regarding the Spaces bottom sheet.
### Outcome
#### What did you expect?
#### What happened instead?
### Your phone model
_No response_
### Operating system version
_No response_
### Application version and app store
App Layout APK
### Homeserver
_No response_
### Will you send logs?
No
### Are you willing to provide a PR?
No
|
1.0
|
App Layout: Bottom Sheet - ### Steps to reproduce
We had several bugs and issues raised during internal testing regarding the Spaces bottom sheet.
### Outcome
#### What did you expect?
#### What happened instead?
### Your phone model
_No response_
### Operating system version
_No response_
### Application version and app store
App Layout APK
### Homeserver
_No response_
### Will you send logs?
No
### Are you willing to provide a PR?
No
|
defect
|
app layout bottom sheet steps to reproduce we had several bugs and issues raised during internal testing regarding the spaces bottom sheet outcome what did you expect what happened instead your phone model no response operating system version no response application version and app store app layout apk homeserver no response will you send logs no are you willing to provide a pr no
| 1
|
65,423
| 7,878,740,311
|
IssuesEvent
|
2018-06-26 11:14:45
|
openebs/openebs
|
https://api.github.com/repos/openebs/openebs
|
closed
|
Document the steps to run GRPC application
|
contributor-docs kind/design summerhack
|
We need to create a document on how we're going to place the GRPC related stubs in `maya` project. It will be helpful for community.
|
1.0
|
Document the steps to run GRPC application - We need to create a document on how we're going to place the GRPC related stubs in `maya` project. It will be helpful for community.
|
non_defect
|
document the steps to run grpc application we need to create a document on how we re going to place the grpc related stubs in maya project it will be helpful for community
| 0
|
52,472
| 13,224,768,742
|
IssuesEvent
|
2020-08-17 19:48:33
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
opened
|
dataio - more ncurses issues blocking "dataio-shovel" (Trac #2268)
|
Incomplete Migration Migrated from Trac analysis defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2268">https://code.icecube.wisc.edu/projects/icecube/ticket/2268</a>, reported by negaand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-09-04T12:46:55",
"_ts": "1567601215767531",
"description": "After updating `ncurses` detection, building `dataio-shovel` dies with:\n\n{{{\n[ 62%] Building CXX object dataio/CMakeFiles/dataio-shovel.dir/private/shovel/View.cxx.o\n/build/buildslave/chernobog/SuSE/source/dataio/private/shovel/View.cxx: In function \u2018void resize(int)\u2019:\n/build/buildslave/chernobog/SuSE/source/dataio/private/shovel/View.cxx:99:25: error: lvalue required as left operand of assignment\n LINES = win.ws_row;\n ^~~~~~\n/build/buildslave/chernobog/SuSE/source/dataio/private/shovel/View.cxx:102:24: error: lvalue required as left operand of assignment\n COLS = win.ws_col;\n ^~~~~~\nmake[2]: *** [dataio/CMakeFiles/dataio-shovel.dir/build.make:111: dataio/CMakeFiles/dataio-shovel.dir/private/shovel/View.cxx.o] Error 1\nmake[1]: *** [CMakeFiles/Makefile2:5679: dataio/CMakeFiles/dataio-shovel.dir/all] Error 2\nmake: *** [Makefile:141: all] Error 2\n}}}\n\nMaybe a `TIOCGSIZE`/`TIOCGWINSZ` issue. \n\nSUSE is such a super minor OS for us, kicking it down the road. ",
"reporter": "nega",
"cc": "",
"resolution": "worksforme",
"time": "2019-04-03T18:55:54",
"component": "analysis",
"summary": "dataio - more ncurses issues blocking \"dataio-shovel\"",
"priority": "normal",
"keywords": "dataio dataio-shovel curses ncurses janitors",
"milestone": "Autumnal Equinox 2019",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
dataio - more ncurses issues blocking "dataio-shovel" (Trac #2268) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2268">https://code.icecube.wisc.edu/projects/icecube/ticket/2268</a>, reported by negaand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-09-04T12:46:55",
"_ts": "1567601215767531",
"description": "After updating `ncurses` detection, building `dataio-shovel` dies with:\n\n{{{\n[ 62%] Building CXX object dataio/CMakeFiles/dataio-shovel.dir/private/shovel/View.cxx.o\n/build/buildslave/chernobog/SuSE/source/dataio/private/shovel/View.cxx: In function \u2018void resize(int)\u2019:\n/build/buildslave/chernobog/SuSE/source/dataio/private/shovel/View.cxx:99:25: error: lvalue required as left operand of assignment\n LINES = win.ws_row;\n ^~~~~~\n/build/buildslave/chernobog/SuSE/source/dataio/private/shovel/View.cxx:102:24: error: lvalue required as left operand of assignment\n COLS = win.ws_col;\n ^~~~~~\nmake[2]: *** [dataio/CMakeFiles/dataio-shovel.dir/build.make:111: dataio/CMakeFiles/dataio-shovel.dir/private/shovel/View.cxx.o] Error 1\nmake[1]: *** [CMakeFiles/Makefile2:5679: dataio/CMakeFiles/dataio-shovel.dir/all] Error 2\nmake: *** [Makefile:141: all] Error 2\n}}}\n\nMaybe a `TIOCGSIZE`/`TIOCGWINSZ` issue. \n\nSUSE is such a super minor OS for us, kicking it down the road. ",
"reporter": "nega",
"cc": "",
"resolution": "worksforme",
"time": "2019-04-03T18:55:54",
"component": "analysis",
"summary": "dataio - more ncurses issues blocking \"dataio-shovel\"",
"priority": "normal",
"keywords": "dataio dataio-shovel curses ncurses janitors",
"milestone": "Autumnal Equinox 2019",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
defect
|
dataio more ncurses issues blocking dataio shovel trac migrated from json status closed changetime ts description after updating ncurses detection building dataio shovel dies with n n n building cxx object dataio cmakefiles dataio shovel dir private shovel view cxx o n build buildslave chernobog suse source dataio private shovel view cxx in function resize int n build buildslave chernobog suse source dataio private shovel view cxx error lvalue required as left operand of assignment n lines win ws row n n build buildslave chernobog suse source dataio private shovel view cxx error lvalue required as left operand of assignment n cols win ws col n nmake error nmake error nmake error n n nmaybe a tiocgsize tiocgwinsz issue n nsuse is such a super minor os for us kicking it down the road reporter nega cc resolution worksforme time component analysis summary dataio more ncurses issues blocking dataio shovel priority normal keywords dataio dataio shovel curses ncurses janitors milestone autumnal equinox owner nega type defect
| 1
|
42,736
| 11,224,112,035
|
IssuesEvent
|
2020-01-08 01:03:21
|
codepotent/Update-Manager
|
https://api.github.com/repos/codepotent/Update-Manager
|
closed
|
Footer version number is inconsistent with current version
|
Status: Completed Type: Defect
|
# Description
The version noted in the footer of the Update Manager plugin is '1.0.0' while the actual version is '1.0.0-rc2'. Not critical, but, not accurate.
|
1.0
|
Footer version number is inconsistent with current version - # Description
The version noted in the footer of the Update Manager plugin is '1.0.0' while the actual version is '1.0.0-rc2'. Not critical, but, not accurate.
|
defect
|
footer version number is inconsistent with current version description the version noted in the footer of the update manager plugin is while the actual version is not critical but not accurate
| 1
|
48,653
| 13,184,711,378
|
IssuesEvent
|
2020-08-12 19:57:17
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
opened
|
dartboard on icecode (Trac #10)
|
Incomplete Migration Migrated from Trac defect infrastructure
|
<details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/10
, reported by troy and owned by cgils_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-07-06T14:14:06",
"description": "needs startup scripts, bring up autobuild boxes.\n\n",
"reporter": "troy",
"cc": "",
"resolution": "fixed",
"_ts": "1183731246000000",
"component": "infrastructure",
"summary": "dartboard on icecode",
"priority": "normal",
"keywords": "",
"time": "2007-06-03T16:29:58",
"milestone": "",
"owner": "cgils",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
dartboard on icecode (Trac #10) - <details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/10
, reported by troy and owned by cgils_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-07-06T14:14:06",
"description": "needs startup scripts, bring up autobuild boxes.\n\n",
"reporter": "troy",
"cc": "",
"resolution": "fixed",
"_ts": "1183731246000000",
"component": "infrastructure",
"summary": "dartboard on icecode",
"priority": "normal",
"keywords": "",
"time": "2007-06-03T16:29:58",
"milestone": "",
"owner": "cgils",
"type": "defect"
}
```
</p>
</details>
|
defect
|
dartboard on icecode trac migrated from reported by troy and owned by cgils json status closed changetime description needs startup scripts bring up autobuild boxes n n reporter troy cc resolution fixed ts component infrastructure summary dartboard on icecode priority normal keywords time milestone owner cgils type defect
| 1
|
72,110
| 23,947,553,230
|
IssuesEvent
|
2022-09-12 08:43:13
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
I did a thread and it went wrong for Travis (maybe other folks too?)
|
T-Defect S-Minor O-Uncommon A-Threads
|
### Steps to reproduce
1. Where are you starting? What can you see?
I made a thread and typed lots of messages in it
### Outcome
The thread looked good for me, but for Travis it looked like a giant pile of replies.
### Operating system
Windows 10
### Application version
Element Nightly version: 0.0.1-nightly.2022090801 Olm version: 3.2.12
### How did you install the app?
website
### Homeserver
lant.uk
### Will you send logs?
Yes
|
1.0
|
I did a thread and it went wrong for Travis (maybe other folks too?) - ### Steps to reproduce
1. Where are you starting? What can you see?
I made a thread and typed lots of messages in it
### Outcome
The thread looked good for me, but for Travis it looked like a giant pile of replies.
### Operating system
Windows 10
### Application version
Element Nightly version: 0.0.1-nightly.2022090801 Olm version: 3.2.12
### How did you install the app?
website
### Homeserver
lant.uk
### Will you send logs?
Yes
|
defect
|
i did a thread and it went wrong for travis maybe other folks too steps to reproduce where are you starting what can you see i made a thread and typed lots of messages in it outcome the thread looked good for me but for travis it looked like a giant pile of replies operating system windows application version element nightly version nightly olm version how did you install the app website homeserver lant uk will you send logs yes
| 1
|
36,085
| 4,713,602,782
|
IssuesEvent
|
2016-10-14 20:40:07
|
WordPress/twentyseventeen
|
https://api.github.com/repos/WordPress/twentyseventeen
|
closed
|
Explore viewport units for responsive typography
|
design enhancement question
|
Placeholder issue for now. I'll come back later and explain what I'm thinking.
|
1.0
|
Explore viewport units for responsive typography - Placeholder issue for now. I'll come back later and explain what I'm thinking.
|
non_defect
|
explore viewport units for responsive typography placeholder issue for now i ll come back later and explain what i m thinking
| 0
|
8,539
| 2,611,516,986
|
IssuesEvent
|
2015-02-27 05:51:50
|
chrsmith/hedgewars
|
https://api.github.com/repos/chrsmith/hedgewars
|
closed
|
hang while connecting to custom server via hwplay scheme if you have empty nickname
|
auto-migrated Priority-Low Type-Defect
|
```
What steps will reproduce the problem?
1. Associate Hedgewars file extension on Linux
2. Click on (eg) hwplay://m8y.org
3. Wait indefinitely
What is the expected output? What do you see instead?
Instead of hanging in the 'connecting' page it should connect to server! Same
code, only happens on Linux (why?).
It's also possible to cause this by calling $(hedgewars hwplay://m8y.org) from
the command line
```
Original issue reported on code.google.com by `vittorio...@gmail.com` on 25 Mar 2013 at 11:21
* Blocking: #580
|
1.0
|
hang while connecting to custom server via hwplay scheme if you have empty nickname - ```
What steps will reproduce the problem?
1. Associate Hedgewars file extension on Linux
2. Click on (eg) hwplay://m8y.org
3. Wait indefinitely
What is the expected output? What do you see instead?
Instead of hanging in the 'connecting' page it should connect to server! Same
code, only happens on Linux (why?).
It's also possible to cause this by calling $(hedgewars hwplay://m8y.org) from
the command line
```
Original issue reported on code.google.com by `vittorio...@gmail.com` on 25 Mar 2013 at 11:21
* Blocking: #580
|
defect
|
hang while connecting to custom server via hwplay scheme if you have empty nickname what steps will reproduce the problem associate hedgewars file extension on linux click on eg hwplay org wait indefinitely what is the expected output what do you see instead instead of hanging in the connecting page it should connect to server same code only happens on linux why it s also possible to cause this by calling hedgewars hwplay org from the command line original issue reported on code google com by vittorio gmail com on mar at blocking
| 1
|
51,217
| 13,207,396,112
|
IssuesEvent
|
2020-08-14 22:56:44
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
opened
|
I3DOMCalibration class needs a suck-ectomy (Trac #54)
|
Incomplete Migration Migrated from Trac defect offline-software
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/54">https://code.icecube.wisc.edu/projects/icecube/ticket/54</a>, reported by blaufussand owned by blaufuss</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-11-11T03:51:18",
"_ts": "1194753078000000",
"description": "There's gotta be some performance to be gained by redoing the internals of\nthis class.\n\n typedef map<unsigned, map<unsigned, LinearFit> > ATWDBinParam_t;\n\nCan't be the optimal way.",
"reporter": "blaufuss",
"cc": "",
"resolution": "duplicate",
"time": "2007-06-08T17:16:45",
"component": "offline-software",
"summary": "I3DOMCalibration class needs a suck-ectomy",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "blaufuss",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
I3DOMCalibration class needs a suck-ectomy (Trac #54) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/54">https://code.icecube.wisc.edu/projects/icecube/ticket/54</a>, reported by blaufussand owned by blaufuss</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-11-11T03:51:18",
"_ts": "1194753078000000",
"description": "There's gotta be some performance to be gained by redoing the internals of\nthis class.\n\n typedef map<unsigned, map<unsigned, LinearFit> > ATWDBinParam_t;\n\nCan't be the optimal way.",
"reporter": "blaufuss",
"cc": "",
"resolution": "duplicate",
"time": "2007-06-08T17:16:45",
"component": "offline-software",
"summary": "I3DOMCalibration class needs a suck-ectomy",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "blaufuss",
"type": "defect"
}
```
</p>
</details>
|
defect
|
class needs a suck ectomy trac migrated from json status closed changetime ts description there s gotta be some performance to be gained by redoing the internals of nthis class n n typedef map atwdbinparam t n ncan t be the optimal way reporter blaufuss cc resolution duplicate time component offline software summary class needs a suck ectomy priority normal keywords milestone owner blaufuss type defect
| 1
|
525,516
| 15,255,262,986
|
IssuesEvent
|
2021-02-20 15:24:58
|
google/oboe
|
https://api.github.com/repos/google/oboe
|
closed
|
LiveEffect sample will crash when earphone plug is connected after plug-out earphone
|
bug high priority
|
Android version(s): 8.0.0
Android device(s): Garaxy S8
Oboe version: 1.5.1 (1.5.0 )
App name used for testing: LiveEffect sample
**Short description**
Try to disconnect and reconnect earphone plug, on Garaxy S8, app will crash after reconnect earphone plug.
**Steps to reproduce**
Start Engine
→ Plug-out earphone
→ wait few seconds
→ plug-in earphone without tapping stop engine button
→Crash happens
**Expected behavior**
not crash
**Actual behavior**
crash error has occured
```
2021-02-17 00:09:27.246 23407-23419/com.google.oboe.samples.liveeffect A/libc: Fatal signal 11 (SIGSEGV), code 1, fault addr 0xffffffffffffffe8 in tid 23419 (Binder:23407_1)
2021-02-17 00:09:27.387 23534-23534/? A/DEBUG: *** *** *** *** *** *** *** *** *** *** *** *** *** *** *** ***
2021-02-17 00:09:27.387 23534-23534/? A/DEBUG: Build fingerprint: 'samsung/dreamltexx/dreamlte:8.0.0/R16NW/G950FXXU4CRK1:user/release-keys'
2021-02-17 00:09:27.387 23534-23534/? A/DEBUG: Revision: '10'
2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: ABI: 'arm64'
2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: pid: 23407, tid: 23419, name: Binder:23407_1 >>> com.google.oboe.samples.liveeffect <<<
2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: signal 11 (SIGSEGV), code 1 (SEGV_MAPERR), fault addr 0xffffffffffffffe8
2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: x0 00000076778891c0 x1 00000076778cf580 x2 0000000000000001 x3 0000000000000000
2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: x4 0000000000000000 x5 0000000000000000 x6 0000000000000000 x7 000000765c24ee38
2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: x8 00000076590c0508 x9 0000000000000000 x10 0000000000000008 x11 0000000000000000
2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: x12 0000000000000000 x13 0000000000000004 x14 000000000000001e x15 0000000000000008
2021-02-17 00:09:27.389 23534-23534/? A/DEBUG: x16 00000076780776e0 x17 000000767bc9fa70 x18 0000000000000020 x19 00000076778cf6c0
2021-02-17 00:09:27.389 23534-23534/? A/DEBUG: x20 0000000000000001 x21 0000000000000000 x22 0000000000000000 x23 000000767bc9c3c0
2021-02-17 00:09:27.389 23534-23534/? A/DEBUG: x24 000000765c24f570 x25 000000765c153000 x26 00000076650770c0 x27 0000000000000000
2021-02-17 00:09:27.389 23534-23534/? A/DEBUG: x28 0000000000000000 x29 000000765c24f340 x30 000000767802c204
2021-02-17 00:09:27.389 23534-23534/? A/DEBUG: sp 000000765c24f320 pc 000000767802c220 pstate 0000000060000000
2021-02-17 00:09:27.431 23534-23534/? A/DEBUG: backtrace:
2021-02-17 00:09:27.431 23534-23534/? A/DEBUG: #00 pc 0000000000063220 /system/lib64/libbinder.so (_ZN7android14IPCThreadState20processPendingDerefsEv+128)
2021-02-17 00:09:27.431 23534-23534/? A/DEBUG: #01 pc 0000000000063294 /system/lib64/libbinder.so (_ZN7android14IPCThreadState14joinThreadPoolEb+52)
2021-02-17 00:09:27.431 23534-23534/? A/DEBUG: #02 pc 0000000000085408 /system/lib64/libbinder.so (_ZN7android10PoolThread10threadLoopEv+24)
2021-02-17 00:09:27.432 23534-23534/? A/DEBUG: #03 pc 0000000000011be8 /system/lib64/libutils.so (_ZN7android6Thread11_threadLoopEPv+280)
2021-02-17 00:09:27.432 23534-23534/? A/DEBUG: #04 pc 00000000000fd6c8 /system/lib64/libandroid_runtime.so (_ZN7android14AndroidRuntime15javaThreadShellEPv+136)
2021-02-17 00:09:27.432 23534-23534/? A/DEBUG: #05 pc 000000000006de3c /system/lib64/libc.so (_ZL15__pthread_startPv+220)
2021-02-17 00:09:27.432 23534-23534/? A/DEBUG: #06 pc 000000000001f9c4 /system/lib64/libc.so (__start_thread+68)
```
**Device**
**Any additional context**
On Pixel3, crash doesn't happen, but we can't hear audio playback until restart oboe engine.
( we need to stop , and restart engine to enable audio playback after plug-out earphone when engine running).
I don't have any idea to prevent this crash on c++ code...
|
1.0
|
LiveEffect sample will crash when earphone plug is connected after plug-out earphone - Android version(s): 8.0.0
Android device(s): Garaxy S8
Oboe version: 1.5.1 (1.5.0 )
App name used for testing: LiveEffect sample
**Short description**
Try to disconnect and reconnect earphone plug, on Garaxy S8, app will crash after reconnect earphone plug.
**Steps to reproduce**
Start Engine
→ Plug-out earphone
→ wait few seconds
→ plug-in earphone without tapping stop engine button
→Crash happens
**Expected behavior**
not crash
**Actual behavior**
crash error has occured
```
2021-02-17 00:09:27.246 23407-23419/com.google.oboe.samples.liveeffect A/libc: Fatal signal 11 (SIGSEGV), code 1, fault addr 0xffffffffffffffe8 in tid 23419 (Binder:23407_1)
2021-02-17 00:09:27.387 23534-23534/? A/DEBUG: *** *** *** *** *** *** *** *** *** *** *** *** *** *** *** ***
2021-02-17 00:09:27.387 23534-23534/? A/DEBUG: Build fingerprint: 'samsung/dreamltexx/dreamlte:8.0.0/R16NW/G950FXXU4CRK1:user/release-keys'
2021-02-17 00:09:27.387 23534-23534/? A/DEBUG: Revision: '10'
2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: ABI: 'arm64'
2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: pid: 23407, tid: 23419, name: Binder:23407_1 >>> com.google.oboe.samples.liveeffect <<<
2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: signal 11 (SIGSEGV), code 1 (SEGV_MAPERR), fault addr 0xffffffffffffffe8
2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: x0 00000076778891c0 x1 00000076778cf580 x2 0000000000000001 x3 0000000000000000
2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: x4 0000000000000000 x5 0000000000000000 x6 0000000000000000 x7 000000765c24ee38
2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: x8 00000076590c0508 x9 0000000000000000 x10 0000000000000008 x11 0000000000000000
2021-02-17 00:09:27.388 23534-23534/? A/DEBUG: x12 0000000000000000 x13 0000000000000004 x14 000000000000001e x15 0000000000000008
2021-02-17 00:09:27.389 23534-23534/? A/DEBUG: x16 00000076780776e0 x17 000000767bc9fa70 x18 0000000000000020 x19 00000076778cf6c0
2021-02-17 00:09:27.389 23534-23534/? A/DEBUG: x20 0000000000000001 x21 0000000000000000 x22 0000000000000000 x23 000000767bc9c3c0
2021-02-17 00:09:27.389 23534-23534/? A/DEBUG: x24 000000765c24f570 x25 000000765c153000 x26 00000076650770c0 x27 0000000000000000
2021-02-17 00:09:27.389 23534-23534/? A/DEBUG: x28 0000000000000000 x29 000000765c24f340 x30 000000767802c204
2021-02-17 00:09:27.389 23534-23534/? A/DEBUG: sp 000000765c24f320 pc 000000767802c220 pstate 0000000060000000
2021-02-17 00:09:27.431 23534-23534/? A/DEBUG: backtrace:
2021-02-17 00:09:27.431 23534-23534/? A/DEBUG: #00 pc 0000000000063220 /system/lib64/libbinder.so (_ZN7android14IPCThreadState20processPendingDerefsEv+128)
2021-02-17 00:09:27.431 23534-23534/? A/DEBUG: #01 pc 0000000000063294 /system/lib64/libbinder.so (_ZN7android14IPCThreadState14joinThreadPoolEb+52)
2021-02-17 00:09:27.431 23534-23534/? A/DEBUG: #02 pc 0000000000085408 /system/lib64/libbinder.so (_ZN7android10PoolThread10threadLoopEv+24)
2021-02-17 00:09:27.432 23534-23534/? A/DEBUG: #03 pc 0000000000011be8 /system/lib64/libutils.so (_ZN7android6Thread11_threadLoopEPv+280)
2021-02-17 00:09:27.432 23534-23534/? A/DEBUG: #04 pc 00000000000fd6c8 /system/lib64/libandroid_runtime.so (_ZN7android14AndroidRuntime15javaThreadShellEPv+136)
2021-02-17 00:09:27.432 23534-23534/? A/DEBUG: #05 pc 000000000006de3c /system/lib64/libc.so (_ZL15__pthread_startPv+220)
2021-02-17 00:09:27.432 23534-23534/? A/DEBUG: #06 pc 000000000001f9c4 /system/lib64/libc.so (__start_thread+68)
```
**Device**
**Any additional context**
On Pixel3, crash doesn't happen, but we can't hear audio playback until restart oboe engine.
( we need to stop , and restart engine to enable audio playback after plug-out earphone when engine running).
I don't have any idea to prevent this crash on c++ code...
|
non_defect
|
liveeffect sample will crash when earphone plug is connected after plug out earphone android version s android device s garaxy oboe version app name used for testing liveeffect sample short description try to disconnect and reconnect earphone plug on garaxy app will crash after reconnect earphone plug steps to reproduce start engine → plug out earphone → wait few seconds → plug in earphone without tapping stop engine button →crash happens expected behavior not crash actual behavior crash error has occured com google oboe samples liveeffect a libc fatal signal sigsegv code fault addr in tid binder a debug a debug build fingerprint samsung dreamltexx dreamlte user release keys a debug revision a debug abi a debug pid tid name binder com google oboe samples liveeffect a debug signal sigsegv code segv maperr fault addr a debug a debug a debug a debug a debug a debug a debug a debug a debug sp pc pstate a debug backtrace a debug pc system libbinder so a debug pc system libbinder so a debug pc system libbinder so a debug pc system libutils so threadloopepv a debug pc system libandroid runtime so a debug pc system libc so pthread startpv a debug pc system libc so start thread device any additional context on crash doesn t happen but we can t hear audio playback until restart oboe engine we need to stop and restart engine to enable audio playback after plug out earphone when engine running i don t have any idea to prevent this crash on c code
| 0
|
47,006
| 13,056,013,581
|
IssuesEvent
|
2020-07-30 03:23:32
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
opened
|
[icetray] I3TestModuleFactory doesn't work in python3 (Trac #2256)
|
Incomplete Migration Migrated from Trac combo core defect
|
Migrated from https://code.icecube.wisc.edu/ticket/2256
```json
{
"status": "closed",
"changetime": "2020-06-24T12:31:42",
"description": "It looks like the `unittest` works differently in python 2 and 3. In 2 you can keep running the same unit test over and over again, but in 3 after you run it once it gets deleted and becomes a `None`\n{{{\n$ python CoincSuite/resources/test/AfterpulseTest.py \nNOTICE (I3Tray): I3Tray finishing... (I3Tray.cxx:497 in void I3Tray::Execute(unsigned int))\nNOTICE (TesterModule): \nRecombination Decisions by AfterpulseTester:\n YES : 1\n NO : 0\n UNDECIDED : 0\n (TesterModule.cxx:45 in virtual void TesterModule::Finish())\nCoincSuite/resources/test/AfterpulseTest.py:142: DeprecationWarning: Please use assertTrue instead.\n self.assert_(self.frame.Has(SplitName+\"RecombAttempts\"))\nCoincSuite/resources/test/AfterpulseTest.py:144: DeprecationWarning: Please use assertTrue instead.\n self.assert_(ra[0] == \"AfterpulseTester\")\n.\n----------------------------------------------------------------------\nRan 1 test in 0.001s\n\nOK\nCoincSuite/resources/test/AfterpulseTest.py:156: DeprecationWarning: Please use assertTrue instead.\n self.assert_(eh.sub_event_stream==\"split\" and eh.sub_event_id==0)\n.\n----------------------------------------------------------------------\nRan 1 test in 0.000s\n\nOK\nERROR (I3Module): TestPhysicsSequence: Exception thrown (I3Module.cxx:127 in void I3Module::Do(void (I3Module::*)()))\nTraceback (most recent call last):\n File \"CoincSuite/resources/test/AfterpulseTest.py\", line 194, in <module>\n tray.Execute()\n File \"/Users/kmeagher/icecube/combo/release/lib/I3Tray.py\", line 256, in Execute\n super(I3Tray, self).Execute()\n File \"/Users/kmeagher/icecube/combo/release/lib/icecube/icetray/test_module.py\", line 83, in RunTests\n test_result = self.test_runner.run(self.test_suites)\n File \"/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/runner.py\", line 176, in run\n test(result)\n File \"/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/suite.py\", line 84, in __call__\n return self.run(*args, **kwds)\n File \"/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/suite.py\", line 122, in run\n test(result)\nTypeError: 'NoneType' object is not callable\n\n}}}\n",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1593001902142004",
"component": "combo core",
"summary": "[icetray] I3TestModuleFactory doesn't work in python3",
"priority": "blocker",
"keywords": "",
"time": "2019-03-18T19:19:29",
"milestone": "Autumnal Equinox 2020",
"owner": "olivas",
"type": "defect"
}
```
|
1.0
|
[icetray] I3TestModuleFactory doesn't work in python3 (Trac #2256) - Migrated from https://code.icecube.wisc.edu/ticket/2256
```json
{
"status": "closed",
"changetime": "2020-06-24T12:31:42",
"description": "It looks like the `unittest` works differently in python 2 and 3. In 2 you can keep running the same unit test over and over again, but in 3 after you run it once it gets deleted and becomes a `None`\n{{{\n$ python CoincSuite/resources/test/AfterpulseTest.py \nNOTICE (I3Tray): I3Tray finishing... (I3Tray.cxx:497 in void I3Tray::Execute(unsigned int))\nNOTICE (TesterModule): \nRecombination Decisions by AfterpulseTester:\n YES : 1\n NO : 0\n UNDECIDED : 0\n (TesterModule.cxx:45 in virtual void TesterModule::Finish())\nCoincSuite/resources/test/AfterpulseTest.py:142: DeprecationWarning: Please use assertTrue instead.\n self.assert_(self.frame.Has(SplitName+\"RecombAttempts\"))\nCoincSuite/resources/test/AfterpulseTest.py:144: DeprecationWarning: Please use assertTrue instead.\n self.assert_(ra[0] == \"AfterpulseTester\")\n.\n----------------------------------------------------------------------\nRan 1 test in 0.001s\n\nOK\nCoincSuite/resources/test/AfterpulseTest.py:156: DeprecationWarning: Please use assertTrue instead.\n self.assert_(eh.sub_event_stream==\"split\" and eh.sub_event_id==0)\n.\n----------------------------------------------------------------------\nRan 1 test in 0.000s\n\nOK\nERROR (I3Module): TestPhysicsSequence: Exception thrown (I3Module.cxx:127 in void I3Module::Do(void (I3Module::*)()))\nTraceback (most recent call last):\n File \"CoincSuite/resources/test/AfterpulseTest.py\", line 194, in <module>\n tray.Execute()\n File \"/Users/kmeagher/icecube/combo/release/lib/I3Tray.py\", line 256, in Execute\n super(I3Tray, self).Execute()\n File \"/Users/kmeagher/icecube/combo/release/lib/icecube/icetray/test_module.py\", line 83, in RunTests\n test_result = self.test_runner.run(self.test_suites)\n File \"/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/runner.py\", line 176, in run\n test(result)\n File \"/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/suite.py\", line 84, in __call__\n return self.run(*args, **kwds)\n File \"/usr/local/Cellar/python/3.7.2_2/Frameworks/Python.framework/Versions/3.7/lib/python3.7/unittest/suite.py\", line 122, in run\n test(result)\nTypeError: 'NoneType' object is not callable\n\n}}}\n",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1593001902142004",
"component": "combo core",
"summary": "[icetray] I3TestModuleFactory doesn't work in python3",
"priority": "blocker",
"keywords": "",
"time": "2019-03-18T19:19:29",
"milestone": "Autumnal Equinox 2020",
"owner": "olivas",
"type": "defect"
}
```
|
defect
|
doesn t work in trac migrated from json status closed changetime description it looks like the unittest works differently in python and in you can keep running the same unit test over and over again but in after you run it once it gets deleted and becomes a none n n python coincsuite resources test afterpulsetest py nnotice finishing cxx in void execute unsigned int nnotice testermodule nrecombination decisions by afterpulsetester n yes n no n undecided n testermodule cxx in virtual void testermodule finish ncoincsuite resources test afterpulsetest py deprecationwarning please use asserttrue instead n self assert self frame has splitname recombattempts ncoincsuite resources test afterpulsetest py deprecationwarning please use asserttrue instead n self assert ra afterpulsetester n n nran test in n nok ncoincsuite resources test afterpulsetest py deprecationwarning please use asserttrue instead n self assert eh sub event stream split and eh sub event id n n nran test in n nok nerror testphysicssequence exception thrown cxx in void do void ntraceback most recent call last n file coincsuite resources test afterpulsetest py line in n tray execute n file users kmeagher icecube combo release lib py line in execute n super self execute n file users kmeagher icecube combo release lib icecube icetray test module py line in runtests n test result self test runner run self test suites n file usr local cellar python frameworks python framework versions lib unittest runner py line in run n test result n file usr local cellar python frameworks python framework versions lib unittest suite py line in call n return self run args kwds n file usr local cellar python frameworks python framework versions lib unittest suite py line in run n test result ntypeerror nonetype object is not callable n n n reporter kjmeagher cc resolution fixed ts component combo core summary doesn t work in priority blocker keywords time milestone autumnal equinox owner olivas type defect
| 1
|
28,843
| 5,390,079,462
|
IssuesEvent
|
2017-02-25 10:10:57
|
patmun/pynetdicom
|
https://api.github.com/repos/patmun/pynetdicom
|
closed
|
Missing self
|
auto-migrated Priority-Medium Type-Defect
|
```
https://code.google.com/p/pynetdicom/source/browse/source/netdicom/SOPclass.py?s
pec=svn2ebb51548127bf4cc6de867e550f44102e7c4cbf&r=2ebb51548127bf4cc6de867e550f44
102e7c4cbf#351
should be self.CannotUnderstand
```
Original issue reported on code.google.com by `agrothberg` on 22 Apr 2015 at 8:06
|
1.0
|
Missing self - ```
https://code.google.com/p/pynetdicom/source/browse/source/netdicom/SOPclass.py?s
pec=svn2ebb51548127bf4cc6de867e550f44102e7c4cbf&r=2ebb51548127bf4cc6de867e550f44
102e7c4cbf#351
should be self.CannotUnderstand
```
Original issue reported on code.google.com by `agrothberg` on 22 Apr 2015 at 8:06
|
defect
|
missing self pec r should be self cannotunderstand original issue reported on code google com by agrothberg on apr at
| 1
|
11,797
| 13,911,122,977
|
IssuesEvent
|
2020-10-20 16:57:37
|
elementor/elementor
|
https://api.github.com/repos/elementor/elementor
|
closed
|
Advanced Editor Tools (previously TinyMCE Advanced) fails to load in Elementor
|
compatibility status/merged
|
https://github.com/elementor/elementor/blob/bb96e7a2cde78e1000997319accb2bc193c48fa4/core/editor/editor.php#L584
When the _Advanced Editor Tools (previously TinyMCE Advanced)_ plugin is activated and a text widget is selected in Elementor, the Elementor panel is empty.
The author of the plugin has tracked down the issue but says that it is something that needs to be fixed in Elementor. He provides the solution here:
https://wordpress.org/support/topic/editor-fails-to-load-in-elementor/
A temporary solution is to uncheck the "Other instances of the Classic (TinyMCE) editor in wp-admin" option in the plugin's settings
(Note: I don't know anything about GitHub so I hope I've done this right!)
<img width="1084" alt="elementor-panel" src="https://user-images.githubusercontent.com/26119379/95342674-98282980-08af-11eb-9d13-7e7dc04ef21c.png">
|
True
|
Advanced Editor Tools (previously TinyMCE Advanced) fails to load in Elementor - https://github.com/elementor/elementor/blob/bb96e7a2cde78e1000997319accb2bc193c48fa4/core/editor/editor.php#L584
When the _Advanced Editor Tools (previously TinyMCE Advanced)_ plugin is activated and a text widget is selected in Elementor, the Elementor panel is empty.
The author of the plugin has tracked down the issue but says that it is something that needs to be fixed in Elementor. He provides the solution here:
https://wordpress.org/support/topic/editor-fails-to-load-in-elementor/
A temporary solution is to uncheck the "Other instances of the Classic (TinyMCE) editor in wp-admin" option in the plugin's settings
(Note: I don't know anything about GitHub so I hope I've done this right!)
<img width="1084" alt="elementor-panel" src="https://user-images.githubusercontent.com/26119379/95342674-98282980-08af-11eb-9d13-7e7dc04ef21c.png">
|
non_defect
|
advanced editor tools previously tinymce advanced fails to load in elementor when the advanced editor tools previously tinymce advanced plugin is activated and a text widget is selected in elementor the elementor panel is empty the author of the plugin has tracked down the issue but says that it is something that needs to be fixed in elementor he provides the solution here a temporary solution is to uncheck the other instances of the classic tinymce editor in wp admin option in the plugin s settings note i don t know anything about github so i hope i ve done this right img width alt elementor panel src
| 0
|
77,196
| 26,834,736,287
|
IssuesEvent
|
2023-02-02 18:33:38
|
Software-Hardware-Codesign/jme-alloc
|
https://api.github.com/repos/Software-Hardware-Codesign/jme-alloc
|
closed
|
[Core-Bug] Writing on a buffer after destruction leads to a jvm monitor crash
|
invalid core defect
|
Writing on a buffer after destroying it will lead to a jvm crash with the following logs, despite the ability to print the buffer capacity, limit and the current position.
The jvm crash log:
```
--------------- P R O C E S S ---------------
VM state: at safepoint (shutting down)
VM Mutex/Monitor currently owned by a thread: ([mutex/lock_event])
[0x00007ff5180117d0] Threads_lock - owner thread: 0x00007ff518125c60
Heap address: 0x0000000746e00000, size: 2962 MB, Compressed Oops mode: Zero based, Oop shift amount: 3
CDS archive(s) mapped at: [0x0000000800000000-0x0000000800be3000-0x0000000800be3000), size 12464128, SharedBaseAddress: 0x0000000800000000, ArchiveRelocationMode: 0.
Compressed class space mapped at: 0x0000000800c00000-0x0000000840c00000, reserved size: 1073741824
Narrow klass base: 0x0000000800000000, Narrow klass shift: 0, Narrow klass range: 0x100000000
GC Precious Log:
<Skipped>
Heap:
garbage-first heap total 194560K, used 5065K [0x0000000746e00000, 0x0000000800000000)
region size 2048K, 2 young (4096K), 0 survivors (0K)
Metaspace used 841K, committed 1024K, reserved 1056768K
class space used 78K, committed 192K, reserved 1048576K
```
A workaround is to nullify the `java.nio.ByteBuffer` reference internally from jni via the invocation api after destroying its memory; because this memory is now owned by another allocated object, so accessing it using this reference and writing on it might be dangerous if it is not thread protected.
|
1.0
|
[Core-Bug] Writing on a buffer after destruction leads to a jvm monitor crash - Writing on a buffer after destroying it will lead to a jvm crash with the following logs, despite the ability to print the buffer capacity, limit and the current position.
The jvm crash log:
```
--------------- P R O C E S S ---------------
VM state: at safepoint (shutting down)
VM Mutex/Monitor currently owned by a thread: ([mutex/lock_event])
[0x00007ff5180117d0] Threads_lock - owner thread: 0x00007ff518125c60
Heap address: 0x0000000746e00000, size: 2962 MB, Compressed Oops mode: Zero based, Oop shift amount: 3
CDS archive(s) mapped at: [0x0000000800000000-0x0000000800be3000-0x0000000800be3000), size 12464128, SharedBaseAddress: 0x0000000800000000, ArchiveRelocationMode: 0.
Compressed class space mapped at: 0x0000000800c00000-0x0000000840c00000, reserved size: 1073741824
Narrow klass base: 0x0000000800000000, Narrow klass shift: 0, Narrow klass range: 0x100000000
GC Precious Log:
<Skipped>
Heap:
garbage-first heap total 194560K, used 5065K [0x0000000746e00000, 0x0000000800000000)
region size 2048K, 2 young (4096K), 0 survivors (0K)
Metaspace used 841K, committed 1024K, reserved 1056768K
class space used 78K, committed 192K, reserved 1048576K
```
A workaround is to nullify the `java.nio.ByteBuffer` reference internally from jni via the invocation api after destroying its memory; because this memory is now owned by another allocated object, so accessing it using this reference and writing on it might be dangerous if it is not thread protected.
|
defect
|
writing on a buffer after destruction leads to a jvm monitor crash writing on a buffer after destroying it will lead to a jvm crash with the following logs despite the ability to print the buffer capacity limit and the current position the jvm crash log p r o c e s s vm state at safepoint shutting down vm mutex monitor currently owned by a thread threads lock owner thread heap address size mb compressed oops mode zero based oop shift amount cds archive s mapped at size sharedbaseaddress archiverelocationmode compressed class space mapped at reserved size narrow klass base narrow klass shift narrow klass range gc precious log heap garbage first heap total used region size young survivors metaspace used committed reserved class space used committed reserved a workaround is to nullify the java nio bytebuffer reference internally from jni via the invocation api after destroying its memory because this memory is now owned by another allocated object so accessing it using this reference and writing on it might be dangerous if it is not thread protected
| 1
|
1,851
| 2,603,972,471
|
IssuesEvent
|
2015-02-24 19:00:36
|
chrsmith/nishazi6
|
https://api.github.com/repos/chrsmith/nishazi6
|
opened
|
沈阳病毒疣是否传染
|
auto-migrated Priority-Medium Type-Defect
|
```
沈阳病毒疣是否传染〓沈陽軍區政治部醫院性病〓TEL:024-3102
3308〓成立于1946年,68年專注于性傳播疾病的研究和治療。位�
��沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的�
��史悠久、設備精良、技術權威、專家云集,是預防、保健、
醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等��
�隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東�
��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍
后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二��
�功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:00
|
1.0
|
沈阳病毒疣是否传染 - ```
沈阳病毒疣是否传染〓沈陽軍區政治部醫院性病〓TEL:024-3102
3308〓成立于1946年,68年專注于性傳播疾病的研究和治療。位�
��沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的�
��史悠久、設備精良、技術權威、專家云集,是預防、保健、
醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等��
�隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東�
��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍
后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二��
�功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:00
|
defect
|
沈阳病毒疣是否传染 沈阳病毒疣是否传染〓沈陽軍區政治部醫院性病〓tel: 〓 , 。位� �� 。是一所與新中國同建立共輝煌的� ��史悠久、設備精良、技術權威、專家云集,是預防、保健、 醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等�� �隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東� ��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍 后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二�� �功。 original issue reported on code google com by gmail com on jun at
| 1
|
409,343
| 27,734,355,434
|
IssuesEvent
|
2023-03-15 10:12:27
|
pires/go-proxyproto
|
https://api.github.com/repos/pires/go-proxyproto
|
closed
|
Set-up automated releases
|
help wanted documentation good first issue
|
Rely on Github Actions, such as https://github.com/softprops/action-gh-release, to automatically create releases based on tags.
|
1.0
|
Set-up automated releases - Rely on Github Actions, such as https://github.com/softprops/action-gh-release, to automatically create releases based on tags.
|
non_defect
|
set up automated releases rely on github actions such as to automatically create releases based on tags
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.